diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-043437fb-fde6-4180-b927-bb5cfebf6b0e1759058799627-2025_09_28-13.27.24.771/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-043437fb-fde6-4180-b927-bb5cfebf6b0e1759058799627-2025_09_28-13.27.24.771/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..5743c1e260911f0b4ec021cd7d7a7850e641c671 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-043437fb-fde6-4180-b927-bb5cfebf6b0e1759058799627-2025_09_28-13.27.24.771/source.csv @@ -0,0 +1,6192 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +2,334,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"1:27:24 PM [info] Activating crowd-code\n1:27:24 PM [info] Recording started\n1:27:24 PM [info] Initializing git provider using file system watchers...\n",Log,tab +3,412,"extension-output-pdoom-org.crowd-code-#1-crowd-code",150,0,"1:27:25 PM [info] Git repository found\n1:27:25 PM [info] Git provider initialized successfully\n1:27:25 PM [info] Initial git state: [object Object]\n",Log,content +4,152506,"TERMINAL",0,0,"git branch",,terminal_command +5,152573,"TERMINAL",0,0,"]633;C[?1h=\r",,terminal_output +6,153100,"TERMINAL",0,0," action-mapper\r\n add-noise-to-combat-exposure-bias\r\n add-wandb-name-and-tags\r\n before-nnx\r\n causal-mem-reduce\r\n causal-spatiotemporal-kv-cache\r\n causal-st-transformer\r\n causal-transformer-dynamics-model\r\n causal-transformer-nnx-no-kv-cache\r\n coinrun-gt-actions\r\n convert-to-jax-array-in-iter\r\n correct-batched-sampling\r\n dev\r\n dont-let-tf-see-gpu\r\n feat/darkness-filter\r\n feat/explicit-image-dims\r\n fix-action-padding-lam-future-information-access\r\n fix-sampling\r\n fix-transformer-forwardpass\r\n fix/dyn-restore-after-nnx-upgrade\r\n fix/spatiotemporal-pe-once-in-STTransformer\r\n generate-minatar-breakout-dataset\r\n grad-norm-log-and-clip\r\n grain-dataloader\r\n gt-actions\r\n hotfix/eval-full-frame-fix\r\n hotfix/fix-val-loss-maskgit-masking\r\n hotfix/full-frame-eval-only-calculate-last-frame-metrics\r\n hotfix/sampling-shapes-error\r\n input_pipeline/add-npy2array_record\r\n logging-variants\r\n lr-schedules\r\n main\r\n maskgit-different-maskprob-per-sample\r\n maskgit-sampling-iterative-unmasking-fix\r\n:",,terminal_output +7,154537,"TERMINAL",0,0,"\r metrics-logging-for-dynamics-model\r\n:",,terminal_output +8,155111,"TERMINAL",0,0,"\r monkey-patch\r\n:\r new-arch-sampling\r\n:\r preprocess_video\r\n:\r refactor-full-frame-val-loss\r\n:\r refactor-tmp\r\n:",,terminal_output +9,155173,"TERMINAL",0,0,"\r remove-restore-branching\r\n:",,terminal_output +10,155526,"TERMINAL",0,0,"\r revised-dataloader\r\n:",,terminal_output +11,155707,"TERMINAL",0,0,"\r runner\r\n:",,terminal_output +12,155849,"TERMINAL",0,0,"\r runner-grain\r\n:",,terminal_output +13,156095,"TERMINAL",0,0,"\r sample-ali-branch\r\n:",,terminal_output +14,156190,"TERMINAL",0,0,"\r sample-from-different-topologies\r\n:",,terminal_output +15,156402,"TERMINAL",0,0,"\r sampling-script-add-metrics\r\n:",,terminal_output +16,156588,"TERMINAL",0,0,"\r sampling-startframe-indexing-fix\r\n:",,terminal_output +17,157549,"TERMINAL",0,0,"\rM dont-let-tf-see-gpu\r\n\r:",,terminal_output +18,158763,"TERMINAL",0,0,"\rM dev\r\n\r:\rM correct-batched-sampling\r\n\r:\rM convert-to-jax-array-in-iter\r\n\r:\rM coinrun-gt-actions\r\n\r:\rM causal-transformer-nnx-no-kv-cache\r\n\r:\rM causal-transformer-dynamics-model\r\n\r:\rM causal-st-transformer\r\n\r:\rM causal-spatiotemporal-kv-cache\r\n\r:\rM causal-mem-reduce\r\n\r:\rM before-nnx\r\n\r:\rM add-wandb-name-and-tags\r\n\r:\rM add-noise-to-combat-exposure-bias\r\n\r:\rM action-mapper\r\n\r:\r\r:\r\r:\r\r:\r\r:\r\r:\r\r:\r\r:\r\r:\r\r:\r\r:\r\r:\r\r:\r\r:",,terminal_output +19,168386,"TERMINAL",0,0,"\r\r:",,terminal_output +20,169178,"TERMINAL",0,0,"\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +21,172190,"TERMINAL",0,0,"git diff",,terminal_command +22,172248,"TERMINAL",0,0,"]633;C[?1h=\r",,terminal_output +23,172338,"TERMINAL",0,0,"\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +24,174928,"TERMINAL",0,0,"git checkout add-noise-to-combat-exposure-bias",,terminal_command +25,174968,"TERMINAL",0,0,"]633;C",,terminal_output +26,175274,"TERMINAL",0,0,"Switched to branch 'add-noise-to-combat-exposure-bias'\r\nYour branch is ahead of 'origin/add-noise-to-combat-exposure-bias' by 1 commit.\r\n (use ""git push"" to publish your local commits)\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +27,175299,"",0,0,"Switched from branch 'zloss-runs' to 'add-noise-to-combat-exposure-bias'",,git_branch_checkout +28,180632,"TERMINAL",0,0,"git log",,terminal_command +29,180686,"TERMINAL",0,0,"]633;C",,terminal_output +30,180817,"TERMINAL",0,0,"[?1h=\rcommit 28c431527970770c7a4c95d76f8919a8f4041418 (HEAD -> add-noise-to-combat-exposure-bias)\r\nAuthor: Mihir Mahajan \r\nDate: Sat Sep 27 19:13:23 2025 +0200\r\n\r\n max noise to 1 in sampling logic\r\n\r\ncommit fa9afacf62a0974e93a32bfeb9b120f4fce42993 (origin/add-noise-to-combat-exposure-bias)\r\nAuthor: Mihir Mahajan \r\nDate: Thu Sep 25 16:25:36 2025 +0200\r\n\r\n fix noise augmentation logic\r\n\r\ncommit 69d3b80a13d11d464da81a745f04da3cbb2c5585\r\nAuthor: Mihir Mahajan \r\nDate: Thu Sep 25 11:08:55 2025 +0200\r\n\r\n added noise level to sample.py\r\n\r\ncommit 82cb201ccccac9d9e90c160b948196128d5f6ec8\r\nMerge: 3d8ba61 8993b1d\r\nAuthor: Mihir Mahajan \r\nDate: Thu Sep 25 11:02:45 2025 +0200\r\n\r\n Merge branch 'main' into add-noise-to-combat-exposure-bias\r\n\r\ncommit 3d8ba61c33be90eab3955303e9e4472d6c44b78b\r\nAuthor: Mihir Mahajan \r\nDate: Wed Sep 24 18:39:12 2025 +0200\r\n\r\n first attempt to implementing noise level feature\r\n\r\ncommit 8993b1d39267b826933c22e1911d50f2702dd634 (main)\r\nAuthor: Franz Srambical <79149449+emergenz@users.noreply.github.com>\r\nDate: Wed Sep 24 17:04:15 2025 +0200\r\n\r\n:",,terminal_output +31,182188,"TERMINAL",0,0,"\r\r:",,terminal_output +32,182488,"TERMINAL",0,0,"\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +33,184138,"TERMINAL",0,0,"git push",,terminal_command +34,184201,"TERMINAL",0,0,"]633;C",,terminal_output +35,185629,"TERMINAL",0,0,"Enumerating objects: 7, done.\r\nCounting objects: 14% (1/7)\rCounting objects: 28% (2/7)\rCounting objects: 42% (3/7)\rCounting objects: 57% (4/7)\rCounting objects: 71% (5/7)\rCounting objects: 85% (6/7)\rCounting objects: 100% (7/7)\rCounting objects: 100% (7/7), done.\r\nDelta compression using up to 152 threads\r\nCompressing objects: 25% (1/4)\rCompressing objects: 50% (2/4)\rCompressing objects: 75% (3/4)\rCompressing objects: 100% (4/4)\rCompressing objects: 100% (4/4), done.\r\nWriting objects: 25% (1/4)\rWriting objects: 50% (2/4)\rWriting objects: 75% (3/4)\rWriting objects: 100% (4/4)\rWriting objects: 100% (4/4), 374 bytes | 374.00 KiB/s, done.\r\nTotal 4 (delta 3), reused 0 (delta 0), pack-reused 0\r\n",,terminal_output +36,185736,"TERMINAL",0,0,"remote: Resolving deltas: 0% (0/3)\rremote: Resolving deltas: 33% (1/3)\rremote: Resolving deltas: 66% (2/3)\rremote: Resolving deltas: 100% (3/3)\rremote: Resolving deltas: 100% (3/3), completed with 3 local objects.\r\n",,terminal_output +37,186264,"TERMINAL",0,0,"To github.com:p-doom/jasmine.git\r\n fa9afac..28c4315 add-noise-to-combat-exposure-bias -> add-noise-to-combat-exposure-bias\r\n",,terminal_output +38,186279,"TERMINAL",0,0,"]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +39,191827,"TERMINAL",0,0,"--time=01:00:00 --partition=accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +40,191850,"TERMINAL",0,0,"]633;C]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +41,202738,"jasmine/models/dynamics.py",0,0,"from typing import Dict\n\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\n\nfrom utils.nn import STTransformer, Transformer\n\n\nclass DynamicsMaskGIT(nnx.Module):\n """"""\n MaskGIT dynamics model\n\n Dimension keys:\n B: batch size\n T: sequence length\n N: number of patches per frame\n L: latent dimension\n V: vocabulary size (number of latents)\n """"""\n\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_latents: int,\n latent_action_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n max_noise_level: float,\n noise_buckets: int,\n mask_limit: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_latents = num_latents\n self.latent_action_dim = latent_action_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.max_noise_level = max_noise_level\n self.noise_buckets = noise_buckets\n self.mask_limit = mask_limit\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.transformer = STTransformer(\n self.model_dim,\n self.model_dim,\n self.ffn_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.patch_embed = nnx.Embed(self.num_latents, self.model_dim, rngs=rngs)\n self.mask_token = nnx.Param(\n nnx.initializers.lecun_uniform()(rngs.params(), (1, 1, 1, self.model_dim))\n )\n self.action_up = nnx.Linear(\n self.latent_action_dim,\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.noise_level_embed = nnx.Embed(\n self.noise_buckets, self.model_dim, rngs=rngs\n )\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n ) -> tuple[jax.Array, jax.Array]:\n # --- Mask videos ---\n video_tokens_BTN = batch[""video_tokens""]\n latent_actions_BTm11L = batch[""latent_actions""]\n vid_embed_BTNM = self.patch_embed(video_tokens_BTN)\n\n B, T, N, M = vid_embed_BTNM.shape\n rng, _rng_prob, *_rngs_mask = jax.random.split(batch[""mask_rng""], B + 2)\n mask_prob = jax.random.uniform(_rng_prob, shape=(B,), minval=self.mask_limit)\n per_sample_shape = vid_embed_BTNM.shape[1:-1]\n mask = jax.vmap(\n lambda rng, prob: jax.random.bernoulli(rng, prob, per_sample_shape),\n in_axes=(0, 0),\n )(jnp.asarray(_rngs_mask), mask_prob)\n mask = mask.at[:, 0].set(False)\n vid_embed_BTNM = jnp.where(\n jnp.expand_dims(mask, -1), self.mask_token.value, vid_embed_BTNM\n )\n\n # --- Sample noise ---\n rng, _rng_noise = jax.random.split(rng)\n noise_level_B = jax.random.uniform(\n _rng_noise, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B / self.max_noise_level) * self.noise_buckets\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n vid_embed_BTNM = (\n jnp.sqrt(1 - noise_level_B111) * vid_embed_BTNM\n + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n\n # --- Predict transition ---\n act_embed_BTm11M = self.action_up(latent_actions_BTm11L)\n padded_act_embed_BT1M = jnp.pad(\n act_embed_BTm11M, ((0, 0), (1, 0), (0, 0), (0, 0))\n )\n vid_embed_BTNp2M = jnp.concatenate(\n [padded_act_embed_BT1M, noise_level_embed_BT1M, vid_embed_BTNM], axis=2\n )\n logits_BTNp2V = self.transformer(vid_embed_BTNp2M)\n logits_BTNV = logits_BTNp2V[:, :, 2:]\n return logits_BTNV, mask\n\n\nclass DynamicsCausal(nnx.Module):\n """"""Causal dynamics model""""""\n\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_latents: int,\n latent_action_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_latents = num_latents\n self.latent_action_dim = latent_action_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.decode = decode\n\n self.transformer = Transformer(\n self.model_dim,\n self.model_dim,\n self.ffn_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=self.decode,\n rngs=rngs,\n )\n self.patch_embed = nnx.Embed(self.num_latents, self.model_dim, rngs=rngs)\n self.action_up = nnx.Linear(\n self.latent_action_dim,\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n ) -> tuple[jax.Array, jax.Array]:\n video_tokens_BTN = batch[""video_tokens""]\n latent_actions_BTm11L = batch[""latent_actions""]\n vid_embed_BTNM = self.patch_embed(video_tokens_BTN)\n act_embed_BTm11M = self.action_up(latent_actions_BTm11L)\n padded_act_embed_BT1M = jnp.pad(\n act_embed_BTm11M, ((0, 0), (1, 0), (0, 0), (0, 0))\n )\n vid_embed_BTNp1M = jnp.concatenate(\n [padded_act_embed_BT1M, vid_embed_BTNM], axis=2\n )\n logits_BTNp1V = self.transformer(vid_embed_BTNp1M)\n logits_BTNV = logits_BTNp1V[:, :, :-1]\n return logits_BTNV, jnp.ones_like(video_tokens_BTN)\n",python,tab +42,205211,"jasmine/models/dynamics.py",0,0,"",python,tab +43,211182,"jasmine/models/dynamics.py",2273,0,"",python,selection_mouse +44,211184,"jasmine/models/dynamics.py",2272,0,"",python,selection_command +45,211969,"jasmine/models/dynamics.py",0,0,"",python,selection_command +46,212233,"jasmine/models/dynamics.py",0,23,"from typing import Dict",python,selection_command +47,212544,"jasmine/models/dynamics.py",0,6768,"from typing import Dict\n\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\n\nfrom utils.nn import STTransformer, Transformer\n\n\nclass DynamicsMaskGIT(nnx.Module):\n """"""\n MaskGIT dynamics model\n\n Dimension keys:\n B: batch size\n T: sequence length\n N: number of patches per frame\n L: latent dimension\n V: vocabulary size (number of latents)\n """"""\n\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_latents: int,\n latent_action_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n max_noise_level: float,\n noise_buckets: int,\n mask_limit: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_latents = num_latents\n self.latent_action_dim = latent_action_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.max_noise_level = max_noise_level\n self.noise_buckets = noise_buckets\n self.mask_limit = mask_limit\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.transformer = STTransformer(\n self.model_dim,\n self.model_dim,\n self.ffn_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.patch_embed = nnx.Embed(self.num_latents, self.model_dim, rngs=rngs)\n self.mask_token = nnx.Param(\n nnx.initializers.lecun_uniform()(rngs.params(), (1, 1, 1, self.model_dim))\n )\n self.action_up = nnx.Linear(\n self.latent_action_dim,\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.noise_level_embed = nnx.Embed(\n self.noise_buckets, self.model_dim, rngs=rngs\n )\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n ) -> tuple[jax.Array, jax.Array]:\n # --- Mask videos ---\n video_tokens_BTN = batch[""video_tokens""]\n latent_actions_BTm11L = batch[""latent_actions""]\n vid_embed_BTNM = self.patch_embed(video_tokens_BTN)\n\n B, T, N, M = vid_embed_BTNM.shape\n rng, _rng_prob, *_rngs_mask = jax.random.split(batch[""mask_rng""], B + 2)\n mask_prob = jax.random.uniform(_rng_prob, shape=(B,), minval=self.mask_limit)\n per_sample_shape = vid_embed_BTNM.shape[1:-1]\n mask = jax.vmap(\n lambda rng, prob: jax.random.bernoulli(rng, prob, per_sample_shape),\n in_axes=(0, 0),\n )(jnp.asarray(_rngs_mask), mask_prob)\n mask = mask.at[:, 0].set(False)\n vid_embed_BTNM = jnp.where(\n jnp.expand_dims(mask, -1), self.mask_token.value, vid_embed_BTNM\n )\n\n # --- Sample noise ---\n rng, _rng_noise = jax.random.split(rng)\n noise_level_B = jax.random.uniform(\n _rng_noise, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B / self.max_noise_level) * self.noise_buckets\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n vid_embed_BTNM = (\n jnp.sqrt(1 - noise_level_B111) * vid_embed_BTNM\n + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n\n # --- Predict transition ---\n act_embed_BTm11M = self.action_up(latent_actions_BTm11L)\n padded_act_embed_BT1M = jnp.pad(\n act_embed_BTm11M, ((0, 0), (1, 0), (0, 0), (0, 0))\n )\n vid_embed_BTNp2M = jnp.concatenate(\n [padded_act_embed_BT1M, noise_level_embed_BT1M, vid_embed_BTNM], axis=2\n )\n logits_BTNp2V = self.transformer(vid_embed_BTNp2M)\n logits_BTNV = logits_BTNp2V[:, :, 2:]\n return logits_BTNV, mask\n\n\nclass DynamicsCausal(nnx.Module):\n """"""Causal dynamics model""""""\n\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_latents: int,\n latent_action_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_latents = num_latents\n self.latent_action_dim = latent_action_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.decode = decode\n\n self.transformer = Transformer(\n self.model_dim,\n self.model_dim,\n self.ffn_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=self.decode,\n rngs=rngs,\n )\n self.patch_embed = nnx.Embed(self.num_latents, self.model_dim, rngs=rngs)\n self.action_up = nnx.Linear(\n self.latent_action_dim,\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n ) -> tuple[jax.Array, jax.Array]:\n video_tokens_BTN = batch[""video_tokens""]\n latent_actions_BTm11L = batch[""latent_actions""]\n vid_embed_BTNM = self.patch_embed(video_tokens_BTN)\n act_embed_BTm11M = self.action_up(latent_actions_BTm11L)\n padded_act_embed_BT1M = jnp.pad(\n act_embed_BTm11M, ((0, 0), (1, 0), (0, 0), (0, 0))\n )\n vid_embed_BTNp1M = jnp.concatenate(\n [padded_act_embed_BT1M, vid_embed_BTNM], axis=2\n )\n logits_BTNp1V = self.transformer(vid_embed_BTNp1M)\n logits_BTNV = logits_BTNp1V[:, :, :-1]\n return logits_BTNV, jnp.ones_like(video_tokens_BTN)\n",python,selection_command +48,547761,"jasmine/models/dynamics.py",6707,0,"",python,selection_mouse +49,547795,"jasmine/models/dynamics.py",6706,0,"",python,selection_command +50,551560,"jasmine/models/dynamics.py",1324,0,"",python,selection_mouse +51,569588,"jasmine/models/dynamics.py",2604,0,"",python,selection_mouse +52,569729,"jasmine/models/dynamics.py",2600,9,"_rng_prob",python,selection_mouse +53,571211,"jasmine/models/dynamics.py",2617,0,"",python,selection_mouse +54,571348,"jasmine/models/dynamics.py",2612,10,"_rngs_mask",python,selection_mouse +55,572053,"jasmine/models/dynamics.py",2606,0,"",python,selection_mouse +56,572220,"jasmine/models/dynamics.py",2600,9,"_rng_prob",python,selection_mouse +57,575985,"jasmine/models/dynamics.py",2615,0,"",python,selection_mouse +58,576160,"jasmine/models/dynamics.py",2612,10,"_rngs_mask",python,selection_mouse +59,578631,"jasmine/models/dynamics.py",2596,0,"",python,selection_mouse +60,578763,"jasmine/models/dynamics.py",2595,3,"rng",python,selection_mouse +61,580342,"jasmine/models/dynamics.py",3291,0,"",python,selection_mouse +62,580532,"jasmine/models/dynamics.py",3287,10,"_rng_noise",python,selection_mouse +63,590668,"jasmine/models/dynamics.py",3229,0,"",python,selection_mouse +64,591994,"jasmine/models/dynamics.py",3231,0,"",python,selection_command +65,594216,"jasmine/models/dynamics.py",3229,0,"",python,selection_mouse +66,595452,"jasmine/models/dynamics.py",3229,0,",",python,content +67,595456,"jasmine/models/dynamics.py",3230,0,"",python,selection_keyboard +68,595519,"jasmine/models/dynamics.py",3230,0," ",python,content +69,595520,"jasmine/models/dynamics.py",3231,0,"",python,selection_keyboard +70,596122,"jasmine/models/dynamics.py",3231,0,"3",python,content +71,596122,"jasmine/models/dynamics.py",3232,0,"",python,selection_keyboard +72,597536,"jasmine/models/dynamics.py",3206,0,"",python,selection_mouse +73,598170,"jasmine/models/dynamics.py",3206,0,",",python,content +74,598172,"jasmine/models/dynamics.py",3207,0,"",python,selection_keyboard +75,598258,"jasmine/models/dynamics.py",3207,0," ",python,content +76,598259,"jasmine/models/dynamics.py",3208,0,"",python,selection_keyboard +77,598829,"jasmine/models/dynamics.py",3208,0,"_",python,content +78,598830,"jasmine/models/dynamics.py",3209,0,"",python,selection_keyboard +79,599291,"jasmine/models/dynamics.py",3209,0,"r",python,content +80,599293,"jasmine/models/dynamics.py",3210,0,"",python,selection_keyboard +81,599483,"jasmine/models/dynamics.py",3210,0,"n",python,content +82,599486,"jasmine/models/dynamics.py",3211,0,"",python,selection_keyboard +83,599838,"jasmine/models/dynamics.py",3211,0,"g",python,content +84,599840,"jasmine/models/dynamics.py",3212,0,"",python,selection_keyboard +85,601643,"jasmine/models/dynamics.py",3212,0,"_",python,content +86,601645,"jasmine/models/dynamics.py",3213,0,"",python,selection_keyboard +87,602010,"jasmine/models/dynamics.py",3213,0,"n",python,content +88,602012,"jasmine/models/dynamics.py",3214,0,"",python,selection_keyboard +89,602206,"jasmine/models/dynamics.py",3214,0,"o",python,content +90,602207,"jasmine/models/dynamics.py",3215,0,"",python,selection_keyboard +91,602265,"jasmine/models/dynamics.py",3215,0,"i",python,content +92,602267,"jasmine/models/dynamics.py",3216,0,"",python,selection_keyboard +93,602310,"jasmine/models/dynamics.py",3216,0,"s",python,content +94,602311,"jasmine/models/dynamics.py",3217,0,"",python,selection_keyboard +95,602464,"jasmine/models/dynamics.py",3217,0,"e",python,content +96,602465,"jasmine/models/dynamics.py",3218,0,"",python,selection_keyboard +97,603489,"jasmine/models/dynamics.py",3207,0,"",python,selection_command +98,603627,"jasmine/models/dynamics.py",3206,0,"",python,selection_command +99,603894,"jasmine/models/dynamics.py",3205,1,"",python,content +100,604031,"jasmine/models/dynamics.py",3204,1,"",python,content +101,604166,"jasmine/models/dynamics.py",3203,1,"",python,content +102,604304,"jasmine/models/dynamics.py",3202,1,"",python,content +103,604445,"jasmine/models/dynamics.py",3201,1,"",python,content +104,606948,"jasmine/models/dynamics.py",3201,0,"n",python,content +105,606949,"jasmine/models/dynamics.py",3202,0,"",python,selection_keyboard +106,607156,"jasmine/models/dynamics.py",3202,0,"o",python,content +107,607157,"jasmine/models/dynamics.py",3203,0,"",python,selection_keyboard +108,607258,"jasmine/models/dynamics.py",3203,0,"i",python,content +109,607260,"jasmine/models/dynamics.py",3204,0,"",python,selection_keyboard +110,607270,"jasmine/models/dynamics.py",3204,0,"s",python,content +111,607271,"jasmine/models/dynamics.py",3205,0,"",python,selection_keyboard +112,607458,"jasmine/models/dynamics.py",3205,0,"e",python,content +113,607459,"jasmine/models/dynamics.py",3206,0,"",python,selection_keyboard +114,607612,"jasmine/models/dynamics.py",3206,0,"_",python,content +115,607614,"jasmine/models/dynamics.py",3207,0,"",python,selection_keyboard +116,607834,"jasmine/models/dynamics.py",3207,0,"l",python,content +117,607835,"jasmine/models/dynamics.py",3208,0,"",python,selection_keyboard +118,607928,"jasmine/models/dynamics.py",3208,0,"e",python,content +119,607930,"jasmine/models/dynamics.py",3209,0,"",python,selection_keyboard +120,608084,"jasmine/models/dynamics.py",3209,0,"v",python,content +121,608085,"jasmine/models/dynamics.py",3210,0,"",python,selection_keyboard +122,608496,"jasmine/models/dynamics.py",3209,1,"",python,content +123,608574,"jasmine/models/dynamics.py",3208,1,"",python,content +124,608934,"jasmine/models/dynamics.py",3208,0,"v",python,content +125,608935,"jasmine/models/dynamics.py",3209,0,"",python,selection_keyboard +126,608983,"jasmine/models/dynamics.py",3209,0,"l",python,content +127,608984,"jasmine/models/dynamics.py",3210,0,"",python,selection_keyboard +128,609564,"jasmine/models/dynamics.py",3209,0,"",python,selection_command +129,609652,"jasmine/models/dynamics.py",3276,0,"",python,selection_command +130,610004,"jasmine/models/dynamics.py",3275,0,"",python,selection_command +131,610153,"jasmine/models/dynamics.py",3274,0,"",python,selection_command +132,610304,"jasmine/models/dynamics.py",3273,0,"",python,selection_command +133,610450,"jasmine/models/dynamics.py",3272,0,"",python,selection_command +134,610783,"jasmine/models/dynamics.py",3316,0,"",python,selection_command +135,611006,"jasmine/models/dynamics.py",3315,0,"",python,selection_command +136,611176,"jasmine/models/dynamics.py",3314,0,"",python,selection_command +137,611328,"jasmine/models/dynamics.py",3313,0,"",python,selection_command +138,611512,"jasmine/models/dynamics.py",3312,0,"",python,selection_command +139,612004,"jasmine/models/dynamics.py",3313,0,"",python,selection_command +140,612174,"jasmine/models/dynamics.py",3314,0,"",python,selection_command +141,612311,"jasmine/models/dynamics.py",3315,0,"",python,selection_command +142,612454,"jasmine/models/dynamics.py",3316,0,"",python,selection_command +143,612969,"jasmine/models/dynamics.py",3316,0,"_",python,content +144,612970,"jasmine/models/dynamics.py",3317,0,"",python,selection_keyboard +145,613248,"jasmine/models/dynamics.py",3317,0,"l",python,content +146,613249,"jasmine/models/dynamics.py",3318,0,"",python,selection_keyboard +147,613419,"jasmine/models/dynamics.py",3318,0,"v",python,content +148,613421,"jasmine/models/dynamics.py",3319,0,"",python,selection_keyboard +149,613504,"jasmine/models/dynamics.py",3319,0,"l",python,content +150,613506,"jasmine/models/dynamics.py",3320,0,"",python,selection_keyboard +151,614047,"jasmine/models/dynamics.py",3319,0,"",python,selection_command +152,615218,"jasmine/models/dynamics.py",3430,0,"",python,selection_mouse +153,615388,"jasmine/models/dynamics.py",3423,10,"_rng_noise",python,selection_mouse +154,616436,"jasmine/models/dynamics.py",3401,0,"",python,selection_mouse +155,616567,"jasmine/models/dynamics.py",3392,10,"noise_BTNM",python,selection_mouse +156,617195,"jasmine/models/dynamics.py",3383,0,"",python,selection_mouse +157,617210,"jasmine/models/dynamics.py",3382,0,"",python,selection_command +158,617733,"jasmine/models/dynamics.py",3317,0,"",python,selection_mouse +159,617915,"jasmine/models/dynamics.py",3306,14,"_rng_noise_lvl",python,selection_mouse +160,618894,"jasmine/models/dynamics.py",3427,0,"",python,selection_mouse +161,619039,"jasmine/models/dynamics.py",3423,10,"_rng_noise",python,selection_mouse +162,649655,"jasmine/models/dynamics.py",3409,0,"",python,selection_mouse +163,649797,"jasmine/models/dynamics.py",3409,6,"random",python,selection_mouse +164,650085,"jasmine/models/dynamics.py",3409,7,"random.",python,selection_mouse +165,650165,"jasmine/models/dynamics.py",3409,13,"random.normal",python,selection_mouse +166,650527,"jasmine/models/dynamics.py",3417,0,"",python,selection_mouse +167,650528,"jasmine/models/dynamics.py",3416,6,"normal",python,selection_mouse +168,650689,"jasmine/models/dynamics.py",3409,13,"random.normal",python,selection_mouse +169,650773,"jasmine/models/dynamics.py",3408,14,".random.normal",python,selection_mouse +170,650831,"jasmine/models/dynamics.py",3405,17,"jax.random.normal",python,selection_mouse +171,651332,"jasmine/models/dynamics.py",3405,0,"",python,selection_mouse +172,651333,"jasmine/models/dynamics.py",3405,3,"jax",python,selection_mouse +173,651549,"jasmine/models/dynamics.py",3405,4,"jax.",python,selection_mouse +174,651570,"jasmine/models/dynamics.py",3405,10,"jax.random",python,selection_mouse +175,651646,"jasmine/models/dynamics.py",3405,11,"jax.random.",python,selection_mouse +176,651711,"jasmine/models/dynamics.py",3405,17,"jax.random.normal",python,selection_mouse +177,652256,"jasmine/models/dynamics.py",3418,0,"",python,selection_mouse +178,652257,"jasmine/models/dynamics.py",3416,6,"normal",python,selection_mouse +179,652472,"jasmine/models/dynamics.py",3409,13,"random.normal",python,selection_mouse +180,652674,"jasmine/models/dynamics.py",3408,14,".random.normal",python,selection_mouse +181,652737,"jasmine/models/dynamics.py",3405,17,"jax.random.normal",python,selection_mouse +182,653349,"jasmine/models/dynamics.py",3406,0,"",python,selection_mouse +183,653834,"jasmine/models/dynamics.py",3399,0,"",python,selection_mouse +184,653964,"jasmine/models/dynamics.py",3392,10,"noise_BTNM",python,selection_mouse +185,657974,"jasmine/models/dynamics.py",3945,0,"",python,selection_mouse +186,658104,"jasmine/models/dynamics.py",3933,16,"noise_level_B111",python,selection_mouse +187,667789,"jasmine/models/dynamics.py",3907,0,"",python,selection_mouse +188,667808,"jasmine/models/dynamics.py",3906,0,"",python,selection_command +189,668795,"jasmine/models/dynamics.py",3779,0,"",python,selection_mouse +190,669352,"jasmine/models/dynamics.py",3858,0,"",python,selection_mouse +191,669957,"jasmine/models/dynamics.py",3867,0,"",python,selection_mouse +192,670094,"jasmine/models/dynamics.py",3861,7,"reshape",python,selection_mouse +193,670722,"jasmine/models/dynamics.py",3852,0,"",python,selection_mouse +194,670862,"jasmine/models/dynamics.py",3847,13,"noise_level_B",python,selection_mouse +195,671406,"jasmine/models/dynamics.py",3865,0,"",python,selection_mouse +196,671554,"jasmine/models/dynamics.py",3861,7,"reshape",python,selection_mouse +197,672292,"jasmine/models/dynamics.py",3853,0,"",python,selection_mouse +198,672438,"jasmine/models/dynamics.py",3847,13,"noise_level_B",python,selection_mouse +199,724610,"jasmine/models/dynamics.py",3483,0,"",python,selection_mouse +200,725166,"jasmine/models/dynamics.py",3521,0,"",python,selection_mouse +201,726359,"jasmine/models/dynamics.py",3558,0,"",python,selection_mouse +202,727329,"jasmine/models/dynamics.py",3594,0,"",python,selection_mouse +203,727341,"jasmine/models/dynamics.py",3593,0,"",python,selection_command +204,727987,"jasmine/models/dynamics.py",3526,0,"",python,selection_mouse +205,728087,"jasmine/models/dynamics.py",3524,4,"self",python,selection_mouse +206,728702,"jasmine/models/dynamics.py",3536,0,"",python,selection_mouse +207,728857,"jasmine/models/dynamics.py",3529,15,"max_noise_level",python,selection_mouse +208,790658,"jasmine/models/dynamics.py",3619,0,"",python,selection_mouse +209,790793,"jasmine/models/dynamics.py",3603,20,"noise_bucket_idx_B11",python,selection_mouse +210,793354,"jasmine/models/dynamics.py",3479,0,"",python,selection_mouse +211,793559,"jasmine/models/dynamics.py",3463,18,"noise_bucket_idx_B",python,selection_mouse +212,801566,"jasmine/models/dynamics.py",3523,0,"",python,selection_mouse +213,803530,"jasmine/models/dynamics.py",3524,0,"",python,selection_command +214,804829,"jasmine/models/dynamics.py",3484,0,"",python,selection_command +215,805429,"jasmine/models/dynamics.py",3455,39," noise_bucket_idx_B = jnp.floor(",python,selection_command +216,805612,"jasmine/models/dynamics.py",3455,111," noise_bucket_idx_B = jnp.floor(\n (noise_level_B / self.max_noise_level) * self.noise_buckets",python,selection_command +217,805747,"jasmine/models/dynamics.py",3455,139," noise_bucket_idx_B = jnp.floor(\n (noise_level_B / self.max_noise_level) * self.noise_buckets\n ).astype(jnp.int32)",python,selection_command +218,805893,"jasmine/models/dynamics.py",3455,0,"",python,selection_command +219,806024,"jasmine/models/dynamics.py",3384,0,"",python,selection_command +220,806203,"jasmine/models/dynamics.py",3374,0,"",python,selection_command +221,806513,"jasmine/models/dynamics.py",3384,0,"",python,selection_command +222,806672,"jasmine/models/dynamics.py",3454,0,"\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B / self.max_noise_level) * self.noise_buckets\n ).astype(jnp.int32)",python,content +223,806713,"jasmine/models/dynamics.py",3463,0,"",python,selection_command +224,807713,"jasmine/models/dynamics.py",3464,0,"",python,selection_command +225,808212,"jasmine/models/dynamics.py",3465,0,"",python,selection_command +226,808299,"jasmine/models/dynamics.py",3466,0,"",python,selection_command +227,808361,"jasmine/models/dynamics.py",3467,0,"",python,selection_command +228,808362,"jasmine/models/dynamics.py",3468,0,"",python,selection_command +229,808362,"jasmine/models/dynamics.py",3469,0,"",python,selection_command +230,808379,"jasmine/models/dynamics.py",3470,0,"",python,selection_command +231,808424,"jasmine/models/dynamics.py",3471,0,"",python,selection_command +232,808425,"jasmine/models/dynamics.py",3472,0,"",python,selection_command +233,808480,"jasmine/models/dynamics.py",3473,0,"",python,selection_command +234,808481,"jasmine/models/dynamics.py",3474,0,"",python,selection_command +235,808536,"jasmine/models/dynamics.py",3475,0,"",python,selection_command +236,808545,"jasmine/models/dynamics.py",3476,0,"",python,selection_command +237,808605,"jasmine/models/dynamics.py",3477,0,"",python,selection_command +238,808606,"jasmine/models/dynamics.py",3478,0,"",python,selection_command +239,808639,"jasmine/models/dynamics.py",3479,0,"",python,selection_command +240,808693,"jasmine/models/dynamics.py",3480,0,"",python,selection_command +241,808750,"jasmine/models/dynamics.py",3520,0,"",python,selection_command +242,808935,"jasmine/models/dynamics.py",3592,0,"",python,selection_command +243,809137,"jasmine/models/dynamics.py",3593,0,"",python,selection_command +244,809525,"jasmine/models/dynamics.py",3521,0,"",python,selection_command +245,809749,"jasmine/models/dynamics.py",3522,0,"",python,selection_command +246,811581,"jasmine/models/dynamics.py",3522,1,"",python,content +247,812653,"jasmine/models/dynamics.py",3522,0,"*",python,content +248,812654,"jasmine/models/dynamics.py",3523,0,"",python,selection_keyboard +249,813685,"jasmine/models/dynamics.py",3523,0," ",python,content +250,813686,"jasmine/models/dynamics.py",3524,0,"",python,selection_keyboard +251,813872,"jasmine/models/dynamics.py",3524,0,"s",python,content +252,813874,"jasmine/models/dynamics.py",3525,0,"",python,selection_keyboard +253,814029,"jasmine/models/dynamics.py",3525,0,"e",python,content +254,814031,"jasmine/models/dynamics.py",3526,0,"",python,selection_keyboard +255,814113,"jasmine/models/dynamics.py",3526,0,"l",python,content +256,814115,"jasmine/models/dynamics.py",3527,0,"",python,selection_keyboard +257,814243,"jasmine/models/dynamics.py",3527,0,"f",python,content +258,814245,"jasmine/models/dynamics.py",3528,0,"",python,selection_keyboard +259,814311,"jasmine/models/dynamics.py",3528,0,".",python,content +260,814313,"jasmine/models/dynamics.py",3529,0,"",python,selection_keyboard +261,814728,"jasmine/models/dynamics.py",3529,0,"n",python,content +262,814730,"jasmine/models/dynamics.py",3530,0,"",python,selection_keyboard +263,814896,"jasmine/models/dynamics.py",3530,0,"o",python,content +264,814898,"jasmine/models/dynamics.py",3531,0,"",python,selection_keyboard +265,814961,"jasmine/models/dynamics.py",3531,0,"i",python,content +266,814962,"jasmine/models/dynamics.py",3532,0,"",python,selection_keyboard +267,815084,"jasmine/models/dynamics.py",3532,0,"s",python,content +268,815085,"jasmine/models/dynamics.py",3533,0,"",python,selection_keyboard +269,815224,"jasmine/models/dynamics.py",3533,0,"e",python,content +270,815226,"jasmine/models/dynamics.py",3534,0,"",python,selection_keyboard +271,815373,"jasmine/models/dynamics.py",3534,0,"_",python,content +272,815375,"jasmine/models/dynamics.py",3535,0,"",python,selection_keyboard +273,815680,"jasmine/models/dynamics.py",3535,0,"b",python,content +274,815682,"jasmine/models/dynamics.py",3536,0,"",python,selection_keyboard +275,815806,"jasmine/models/dynamics.py",3536,0,"u",python,content +276,815808,"jasmine/models/dynamics.py",3537,0,"",python,selection_keyboard +277,815927,"jasmine/models/dynamics.py",3537,0,"c",python,content +278,815929,"jasmine/models/dynamics.py",3538,0,"",python,selection_keyboard +279,816090,"jasmine/models/dynamics.py",3538,0,"k",python,content +280,816092,"jasmine/models/dynamics.py",3539,0,"",python,selection_keyboard +281,816177,"jasmine/models/dynamics.py",3539,0,"e",python,content +282,816179,"jasmine/models/dynamics.py",3540,0,"",python,selection_keyboard +283,816356,"jasmine/models/dynamics.py",3540,0,"t",python,content +284,816358,"jasmine/models/dynamics.py",3541,0,"",python,selection_keyboard +285,816543,"jasmine/models/dynamics.py",3541,0,"s",python,content +286,816545,"jasmine/models/dynamics.py",3542,0,"",python,selection_keyboard +287,817087,"jasmine/models/dynamics.py",3543,0,"",python,selection_command +288,817596,"jasmine/models/dynamics.py",3543,0,"/",python,content +289,817597,"jasmine/models/dynamics.py",3544,0,"",python,selection_keyboard +290,818097,"jasmine/models/dynamics.py",3543,0,"",python,selection_command +291,818280,"jasmine/models/dynamics.py",3542,0,"",python,selection_command +292,818825,"jasmine/models/dynamics.py",3542,0,")",python,content +293,818826,"jasmine/models/dynamics.py",3543,0,"",python,selection_keyboard +294,818946,"jasmine/models/dynamics.py",3543,0," ",python,content +295,818947,"jasmine/models/dynamics.py",3544,0,"",python,selection_keyboard +296,820000,"jasmine/models/dynamics.py",3543,1,"",python,content +297,821574,"jasmine/models/dynamics.py",3545,0," ",python,content +298,821576,"jasmine/models/dynamics.py",3546,0,"",python,selection_keyboard +299,823536,"jasmine/models/dynamics.py",3566,1,"",python,content +300,823802,"jasmine/models/dynamics.py",3566,2,"",python,content +301,824127,"jasmine/models/dynamics.py",3566,5,"",python,content +302,824381,"jasmine/models/dynamics.py",3566,1,"",python,content +303,824699,"jasmine/models/dynamics.py",3566,13,"",python,content +304,835313,"jasmine/models/dynamics.py",3565,0,"",python,selection_command +305,836235,"jasmine/models/dynamics.py",3593,0,"",python,selection_command +306,836424,"jasmine/models/dynamics.py",3633,0,"",python,selection_command +307,837092,"jasmine/models/dynamics.py",3595,112,"",python,content +308,837136,"jasmine/models/dynamics.py",3603,0,"",python,selection_command +309,837574,"jasmine/models/dynamics.py",3595,28,"",python,content +310,837615,"jasmine/models/dynamics.py",3603,0,"",python,selection_command +311,838549,"jasmine/models/dynamics.py",3603,0,").astype(jnp.int32)\n ",python,content +312,838554,"jasmine/models/dynamics.py",3603,0,"",python,selection_command +313,838848,"jasmine/models/dynamics.py",3603,0,"noise_bucket_idx_B = jnp.floor(\n (noise_level_B / self.max_noise_level) * self.noise_buckets\n ",python,content +314,838876,"jasmine/models/dynamics.py",3633,0,"",python,selection_command +315,865140,"jasmine/models/dynamics.py",3593,0,"",python,selection_command +316,865448,"jasmine/models/dynamics.py",3533,0,"",python,selection_command +317,865629,"jasmine/models/dynamics.py",3493,0,"",python,selection_command +318,865754,"jasmine/models/dynamics.py",3422,0,"",python,selection_command +319,865988,"jasmine/models/dynamics.py",3454,0,"\n ",python,content +320,867008,"jasmine/models/dynamics.py",3463,0,"#",python,content +321,867009,"jasmine/models/dynamics.py",3464,0,"",python,selection_keyboard +322,867102,"jasmine/models/dynamics.py",3464,0," ",python,content +323,867103,"jasmine/models/dynamics.py",3465,0,"",python,selection_keyboard +324,867713,"jasmine/models/dynamics.py",3465,0,"W",python,content +325,867714,"jasmine/models/dynamics.py",3466,0,"",python,selection_keyboard +326,867906,"jasmine/models/dynamics.py",3466,0,"e",python,content +327,867907,"jasmine/models/dynamics.py",3467,0,"",python,selection_keyboard +328,867977,"jasmine/models/dynamics.py",3467,0," ",python,content +329,867978,"jasmine/models/dynamics.py",3468,0,"",python,selection_keyboard +330,868642,"jasmine/models/dynamics.py",3468,0,"c",python,content +331,868643,"jasmine/models/dynamics.py",3469,0,"",python,selection_keyboard +332,868768,"jasmine/models/dynamics.py",3469,0,"a",python,content +333,868769,"jasmine/models/dynamics.py",3470,0,"",python,selection_keyboard +334,868833,"jasmine/models/dynamics.py",3470,0,"l",python,content +335,868834,"jasmine/models/dynamics.py",3471,0,"",python,selection_keyboard +336,868975,"jasmine/models/dynamics.py",3471,0,"c",python,content +337,868976,"jasmine/models/dynamics.py",3472,0,"",python,selection_keyboard +338,869068,"jasmine/models/dynamics.py",3472,0,"u",python,content +339,869069,"jasmine/models/dynamics.py",3473,0,"",python,selection_keyboard +340,869243,"jasmine/models/dynamics.py",3473,0,"l",python,content +341,869244,"jasmine/models/dynamics.py",3474,0,"",python,selection_keyboard +342,869426,"jasmine/models/dynamics.py",3474,0,"a",python,content +343,869427,"jasmine/models/dynamics.py",3475,0,"",python,selection_keyboard +344,869607,"jasmine/models/dynamics.py",3475,0,"t",python,content +345,869608,"jasmine/models/dynamics.py",3476,0,"",python,selection_keyboard +346,869686,"jasmine/models/dynamics.py",3476,0,"e",python,content +347,869687,"jasmine/models/dynamics.py",3477,0,"",python,selection_keyboard +348,869758,"jasmine/models/dynamics.py",3477,0," ",python,content +349,869759,"jasmine/models/dynamics.py",3478,0,"",python,selection_keyboard +350,871423,"jasmine/models/dynamics.py",3478,0,"n",python,content +351,871425,"jasmine/models/dynamics.py",3479,0,"",python,selection_keyboard +352,871718,"jasmine/models/dynamics.py",3479,0,"i",python,content +353,871719,"jasmine/models/dynamics.py",3480,0,"",python,selection_keyboard +354,872090,"jasmine/models/dynamics.py",3479,1,"",python,content +355,872260,"jasmine/models/dynamics.py",3479,0,"o",python,content +356,872261,"jasmine/models/dynamics.py",3480,0,"",python,selection_keyboard +357,872344,"jasmine/models/dynamics.py",3480,0,"i",python,content +358,872345,"jasmine/models/dynamics.py",3481,0,"",python,selection_keyboard +359,872373,"jasmine/models/dynamics.py",3481,0,"s",python,content +360,872374,"jasmine/models/dynamics.py",3482,0,"",python,selection_keyboard +361,872551,"jasmine/models/dynamics.py",3482,0,"e",python,content +362,872552,"jasmine/models/dynamics.py",3483,0,"",python,selection_keyboard +363,872783,"jasmine/models/dynamics.py",3483,0,"_",python,content +364,872784,"jasmine/models/dynamics.py",3484,0,"",python,selection_keyboard +365,873437,"jasmine/models/dynamics.py",3484,0,"v",python,content +366,873438,"jasmine/models/dynamics.py",3485,0,"",python,selection_keyboard +367,873520,"jasmine/models/dynamics.py",3485,0,"l",python,content +368,873523,"jasmine/models/dynamics.py",3486,0,"",python,selection_keyboard +369,873632,"jasmine/models/dynamics.py",3486,0,"v",python,content +370,873633,"jasmine/models/dynamics.py",3487,0,"",python,selection_keyboard +371,873963,"jasmine/models/dynamics.py",3486,1,"",python,content +372,874073,"jasmine/models/dynamics.py",3485,1,"",python,content +373,874182,"jasmine/models/dynamics.py",3484,1,"",python,content +374,874395,"jasmine/models/dynamics.py",3484,0,"l",python,content +375,874396,"jasmine/models/dynamics.py",3485,0,"",python,selection_keyboard +376,874737,"jasmine/models/dynamics.py",3485,0,"e",python,content +377,874738,"jasmine/models/dynamics.py",3486,0,"",python,selection_keyboard +378,874877,"jasmine/models/dynamics.py",3486,0,"v",python,content +379,874877,"jasmine/models/dynamics.py",3487,0,"",python,selection_keyboard +380,875001,"jasmine/models/dynamics.py",3487,0,"e",python,content +381,875002,"jasmine/models/dynamics.py",3488,0,"",python,selection_keyboard +382,875115,"jasmine/models/dynamics.py",3488,0,"l",python,content +383,875117,"jasmine/models/dynamics.py",3489,0,"",python,selection_keyboard +384,875694,"jasmine/models/dynamics.py",3489,0," ",python,content +385,875695,"jasmine/models/dynamics.py",3490,0,"",python,selection_keyboard +386,875972,"jasmine/models/dynamics.py",3490,0,"*",python,content +387,875973,"jasmine/models/dynamics.py",3491,0,"",python,selection_keyboard +388,876509,"jasmine/models/dynamics.py",3491,0," ",python,content +389,876510,"jasmine/models/dynamics.py",3492,0,"",python,selection_keyboard +390,876896,"jasmine/models/dynamics.py",3492,0,"n",python,content +391,876897,"jasmine/models/dynamics.py",3493,0,"",python,selection_keyboard +392,877096,"jasmine/models/dynamics.py",3493,0,"o",python,content +393,877098,"jasmine/models/dynamics.py",3494,0,"",python,selection_keyboard +394,877557,"jasmine/models/dynamics.py",3493,1,"",python,content +395,877676,"jasmine/models/dynamics.py",3492,1,"",python,content +396,878569,"jasmine/models/dynamics.py",3492,0,"n",python,content +397,878570,"jasmine/models/dynamics.py",3493,0,"",python,selection_keyboard +398,878805,"jasmine/models/dynamics.py",3493,0,"o",python,content +399,878806,"jasmine/models/dynamics.py",3494,0,"",python,selection_keyboard +400,878880,"jasmine/models/dynamics.py",3494,0,"i",python,content +401,878881,"jasmine/models/dynamics.py",3495,0,"",python,selection_keyboard +402,878973,"jasmine/models/dynamics.py",3495,0,"s",python,content +403,878974,"jasmine/models/dynamics.py",3496,0,"",python,selection_keyboard +404,879134,"jasmine/models/dynamics.py",3496,0,"e",python,content +405,879135,"jasmine/models/dynamics.py",3497,0,"",python,selection_keyboard +406,879316,"jasmine/models/dynamics.py",3497,0,"_",python,content +407,879317,"jasmine/models/dynamics.py",3498,0,"",python,selection_keyboard +408,880009,"jasmine/models/dynamics.py",3498,0,"b",python,content +409,880011,"jasmine/models/dynamics.py",3499,0,"",python,selection_keyboard +410,880143,"jasmine/models/dynamics.py",3499,0,"u",python,content +411,880144,"jasmine/models/dynamics.py",3500,0,"",python,selection_keyboard +412,880280,"jasmine/models/dynamics.py",3500,0,"c",python,content +413,880281,"jasmine/models/dynamics.py",3501,0,"",python,selection_keyboard +414,880399,"jasmine/models/dynamics.py",3501,0,"k",python,content +415,880401,"jasmine/models/dynamics.py",3502,0,"",python,selection_keyboard +416,880514,"jasmine/models/dynamics.py",3502,0,"e",python,content +417,880515,"jasmine/models/dynamics.py",3503,0,"",python,selection_keyboard +418,880728,"jasmine/models/dynamics.py",3503,0,"t",python,content +419,880729,"jasmine/models/dynamics.py",3504,0,"",python,selection_keyboard +420,880896,"jasmine/models/dynamics.py",3504,0,"s",python,content +421,880897,"jasmine/models/dynamics.py",3505,0,"",python,selection_keyboard +422,881093,"jasmine/models/dynamics.py",3505,0," ",python,content +423,881094,"jasmine/models/dynamics.py",3506,0,"",python,selection_keyboard +424,881752,"jasmine/models/dynamics.py",3506,0,"/",python,content +425,881753,"jasmine/models/dynamics.py",3507,0,"",python,selection_keyboard +426,882543,"jasmine/models/dynamics.py",3507,0," ",python,content +427,882544,"jasmine/models/dynamics.py",3508,0,"",python,selection_keyboard +428,883793,"jasmine/models/dynamics.py",3508,0,"m",python,content +429,883794,"jasmine/models/dynamics.py",3509,0,"",python,selection_keyboard +430,883883,"jasmine/models/dynamics.py",3509,0,"a",python,content +431,883885,"jasmine/models/dynamics.py",3510,0,"",python,selection_keyboard +432,884102,"jasmine/models/dynamics.py",3510,0,"x",python,content +433,884103,"jasmine/models/dynamics.py",3511,0,"",python,selection_keyboard +434,884346,"jasmine/models/dynamics.py",3511,0,"_",python,content +435,884347,"jasmine/models/dynamics.py",3512,0,"",python,selection_keyboard +436,884664,"jasmine/models/dynamics.py",3512,0,"n",python,content +437,884665,"jasmine/models/dynamics.py",3513,0,"",python,selection_keyboard +438,884843,"jasmine/models/dynamics.py",3513,0,"o",python,content +439,884844,"jasmine/models/dynamics.py",3514,0,"",python,selection_keyboard +440,884925,"jasmine/models/dynamics.py",3514,0,"i",python,content +441,884926,"jasmine/models/dynamics.py",3515,0,"",python,selection_keyboard +442,885012,"jasmine/models/dynamics.py",3515,0,"s",python,content +443,885014,"jasmine/models/dynamics.py",3516,0,"",python,selection_keyboard +444,885139,"jasmine/models/dynamics.py",3516,0,"e",python,content +445,885140,"jasmine/models/dynamics.py",3517,0,"",python,selection_keyboard +446,885829,"jasmine/models/dynamics.py",3517,0," ",python,content +447,885830,"jasmine/models/dynamics.py",3518,0,"",python,selection_keyboard +448,886181,"jasmine/models/dynamics.py",3517,1,"",python,content +449,886459,"jasmine/models/dynamics.py",3517,0,"_",python,content +450,886460,"jasmine/models/dynamics.py",3518,0,"",python,selection_keyboard +451,886717,"jasmine/models/dynamics.py",3518,0,"l",python,content +452,886718,"jasmine/models/dynamics.py",3519,0,"",python,selection_keyboard +453,886799,"jasmine/models/dynamics.py",3519,0,"e",python,content +454,886800,"jasmine/models/dynamics.py",3520,0,"",python,selection_keyboard +455,887016,"jasmine/models/dynamics.py",3520,0,"v",python,content +456,887017,"jasmine/models/dynamics.py",3521,0,"",python,selection_keyboard +457,887163,"jasmine/models/dynamics.py",3521,0,"e",python,content +458,887164,"jasmine/models/dynamics.py",3522,0,"",python,selection_keyboard +459,887264,"jasmine/models/dynamics.py",3522,0,"l",python,content +460,887264,"jasmine/models/dynamics.py",3523,0,"",python,selection_keyboard +461,887354,"jasmine/models/dynamics.py",3523,0," ",python,content +462,887355,"jasmine/models/dynamics.py",3524,0,"",python,selection_keyboard +463,887635,"jasmine/models/dynamics.py",3524,0,"i",python,content +464,887636,"jasmine/models/dynamics.py",3525,0,"",python,selection_keyboard +465,887684,"jasmine/models/dynamics.py",3525,0,"n",python,content +466,887686,"jasmine/models/dynamics.py",3526,0,"",python,selection_keyboard +467,887833,"jasmine/models/dynamics.py",3526,0,"s",python,content +468,887834,"jasmine/models/dynamics.py",3527,0,"",python,selection_keyboard +469,888047,"jasmine/models/dynamics.py",3527,0,"t",python,content +470,888048,"jasmine/models/dynamics.py",3528,0,"",python,selection_keyboard +471,888058,"jasmine/models/dynamics.py",3528,0,"e",python,content +472,888059,"jasmine/models/dynamics.py",3529,0,"",python,selection_keyboard +473,888213,"jasmine/models/dynamics.py",3529,0,"a",python,content +474,888214,"jasmine/models/dynamics.py",3530,0,"",python,selection_keyboard +475,888290,"jasmine/models/dynamics.py",3530,0,"d",python,content +476,888291,"jasmine/models/dynamics.py",3531,0,"",python,selection_keyboard +477,888341,"jasmine/models/dynamics.py",3531,0," ",python,content +478,888342,"jasmine/models/dynamics.py",3532,0,"",python,selection_keyboard +479,888513,"jasmine/models/dynamics.py",3532,0,"o",python,content +480,888514,"jasmine/models/dynamics.py",3533,0,"",python,selection_keyboard +481,888563,"jasmine/models/dynamics.py",3533,0,"f",python,content +482,888564,"jasmine/models/dynamics.py",3534,0,"",python,selection_keyboard +483,888643,"jasmine/models/dynamics.py",3534,0," ",python,content +484,888644,"jasmine/models/dynamics.py",3535,0,"",python,selection_keyboard +485,893670,"jasmine/models/dynamics.py",3478,0,"",python,selection_mouse +486,894987,"jasmine/models/dynamics.py",3478,0,"(",python,content +487,894988,"jasmine/models/dynamics.py",3479,0,"",python,selection_keyboard +488,895876,"jasmine/models/dynamics.py",3506,0,"",python,selection_mouse +489,896510,"jasmine/models/dynamics.py",3506,0,")",python,content +490,896512,"jasmine/models/dynamics.py",3507,0,"",python,selection_keyboard +491,897421,"jasmine/models/dynamics.py",3537,0,"",python,selection_mouse +492,910260,"jasmine/models/dynamics.py",3729,0,"",python,selection_mouse +493,910355,"jasmine/models/dynamics.py",3729,1," ",python,selection_mouse +494,910379,"jasmine/models/dynamics.py",3729,2," (",python,selection_mouse +495,910403,"jasmine/models/dynamics.py",3729,4," (no",python,selection_mouse +496,910408,"jasmine/models/dynamics.py",3729,7," (noise",python,selection_mouse +497,910449,"jasmine/models/dynamics.py",3729,9," (noise_l",python,selection_mouse +498,910487,"jasmine/models/dynamics.py",3729,12," (noise_leve",python,selection_mouse +499,910487,"jasmine/models/dynamics.py",3729,14," (noise_level_",python,selection_mouse +500,910488,"jasmine/models/dynamics.py",3729,16," (noise_level_B ",python,selection_mouse +501,910496,"jasmine/models/dynamics.py",3729,20," (noise_level_B / se",python,selection_mouse +502,910547,"jasmine/models/dynamics.py",3729,24," (noise_level_B / self.m",python,selection_mouse +503,910547,"jasmine/models/dynamics.py",3729,25," (noise_level_B / self.ma",python,selection_mouse +504,910548,"jasmine/models/dynamics.py",3729,26," (noise_level_B / self.max",python,selection_mouse +505,910574,"jasmine/models/dynamics.py",3729,27," (noise_level_B / self.max_",python,selection_mouse +506,910646,"jasmine/models/dynamics.py",3729,30," (noise_level_B / self.max_noi",python,selection_mouse +507,910647,"jasmine/models/dynamics.py",3729,32," (noise_level_B / self.max_noise",python,selection_mouse +508,910647,"jasmine/models/dynamics.py",3729,88," (noise_level_B / self.max_noise_level) * self.noise_buckets\n ).astype(jnp.int32)",python,selection_mouse +509,910972,"jasmine/models/dynamics.py",3729,57," (noise_level_B / self.max_noise_level) * self.noise_buck",python,selection_mouse +510,911003,"jasmine/models/dynamics.py",3729,58," (noise_level_B / self.max_noise_level) * self.noise_bucke",python,selection_mouse +511,911045,"jasmine/models/dynamics.py",3729,59," (noise_level_B / self.max_noise_level) * self.noise_bucket",python,selection_mouse +512,911114,"jasmine/models/dynamics.py",3729,60," (noise_level_B / self.max_noise_level) * self.noise_buckets",python,selection_mouse +513,915744,"jasmine/models/dynamics.py",3789,0,"",python,selection_mouse +514,916704,"jasmine/models/dynamics.py",3783,6,"uckets",python,selection_mouse +515,916719,"jasmine/models/dynamics.py",3773,16,"lf.noise_buckets",python,selection_mouse +516,916758,"jasmine/models/dynamics.py",3762,27,"level) * self.noise_buckets",python,selection_mouse +517,916759,"jasmine/models/dynamics.py",3755,34,"_noise_level) * self.noise_buckets",python,selection_mouse +518,916771,"jasmine/models/dynamics.py",3753,36,"ax_noise_level) * self.noise_buckets",python,selection_mouse +519,916782,"jasmine/models/dynamics.py",3747,42,"self.max_noise_level) * self.noise_buckets",python,selection_mouse +520,916839,"jasmine/models/dynamics.py",3742,47,"_B / self.max_noise_level) * self.noise_buckets",python,selection_mouse +521,916840,"jasmine/models/dynamics.py",3732,57,"oise_level_B / self.max_noise_level) * self.noise_buckets",python,selection_mouse +522,916840,"jasmine/models/dynamics.py",3731,58,"noise_level_B / self.max_noise_level) * self.noise_buckets",python,selection_mouse +523,916851,"jasmine/models/dynamics.py",3730,59,"(noise_level_B / self.max_noise_level) * self.noise_buckets",python,selection_mouse +524,916862,"jasmine/models/dynamics.py",3729,60," (noise_level_B / self.max_noise_level) * self.noise_buckets",python,selection_mouse +525,916929,"jasmine/models/dynamics.py",3728,61," (noise_level_B / self.max_noise_level) * self.noise_buckets",python,selection_mouse +526,917278,"jasmine/models/dynamics.py",3729,60," (noise_level_B / self.max_noise_level) * self.noise_buckets",python,selection_mouse +527,917370,"jasmine/models/dynamics.py",3730,59,"(noise_level_B / self.max_noise_level) * self.noise_buckets",python,selection_mouse +528,918292,"jasmine/models/dynamics.py",3730,0,"",python,selection_mouse +529,919174,"jasmine/models/dynamics.py",3730,2,"(n",python,selection_mouse +530,919175,"jasmine/models/dynamics.py",3730,5,"(nois",python,selection_mouse +531,919175,"jasmine/models/dynamics.py",3730,10,"(noise_lev",python,selection_mouse +532,919176,"jasmine/models/dynamics.py",3730,14,"(noise_level_B",python,selection_mouse +533,919188,"jasmine/models/dynamics.py",3730,87,"(noise_level_B / self.max_noise_level) * self.noise_buckets\n ).astype(jnp.int32)",python,selection_mouse +534,919580,"jasmine/models/dynamics.py",3730,51,"(noise_level_B / self.max_noise_level) * self.noise",python,selection_mouse +535,920009,"jasmine/models/dynamics.py",3730,49,"(noise_level_B / self.max_noise_level) * self.noi",python,selection_mouse +536,920030,"jasmine/models/dynamics.py",3730,48,"(noise_level_B / self.max_noise_level) * self.no",python,selection_mouse +537,920045,"jasmine/models/dynamics.py",3730,47,"(noise_level_B / self.max_noise_level) * self.n",python,selection_mouse +538,920101,"jasmine/models/dynamics.py",3730,44,"(noise_level_B / self.max_noise_level) * sel",python,selection_mouse +539,920102,"jasmine/models/dynamics.py",3730,87,"(noise_level_B / self.max_noise_level) * self.noise_buckets\n ).astype(jnp.int32)",python,selection_mouse +540,920409,"jasmine/models/dynamics.py",3730,58,"(noise_level_B / self.max_noise_level) * self.noise_bucket",python,selection_mouse +541,920425,"jasmine/models/dynamics.py",3730,59,"(noise_level_B / self.max_noise_level) * self.noise_buckets",python,selection_mouse +542,920733,"jasmine/models/dynamics.py",3730,58,"(noise_level_B / self.max_noise_level) * self.noise_bucket",python,selection_mouse +543,921083,"jasmine/models/dynamics.py",3730,59,"(noise_level_B / self.max_noise_level) * self.noise_buckets",python,selection_mouse +544,923997,"jasmine/models/dynamics.py",3537,0,"",python,selection_mouse +545,924679,"jasmine/models/dynamics.py",3537,0,"(noise_level_B / self.max_noise_level) * self.noise_buckets",python,content +546,926669,"jasmine/models/dynamics.py",3583,0,"",python,selection_mouse +547,927069,"jasmine/models/dynamics.py",3582,1,"",python,content +548,927190,"jasmine/models/dynamics.py",3581,1,"",python,content +549,927324,"jasmine/models/dynamics.py",3580,1,"",python,content +550,927458,"jasmine/models/dynamics.py",3579,1,"",python,content +551,927574,"jasmine/models/dynamics.py",3578,1,"",python,content +552,928917,"jasmine/models/dynamics.py",3559,0,"",python,selection_command +553,929298,"jasmine/models/dynamics.py",3558,1,"",python,content +554,929775,"jasmine/models/dynamics.py",3557,1,"",python,content +555,930051,"jasmine/models/dynamics.py",3554,3,"",python,content +556,931604,"jasmine/models/dynamics.py",3586,0," ",python,content +557,931606,"jasmine/models/dynamics.py",3587,0,"",python,selection_keyboard +558,931704,"jasmine/models/dynamics.py",3587,0,"f",python,content +559,931705,"jasmine/models/dynamics.py",3588,0,"",python,selection_keyboard +560,931868,"jasmine/models/dynamics.py",3588,0,"r",python,content +561,931869,"jasmine/models/dynamics.py",3589,0,"",python,selection_keyboard +562,931907,"jasmine/models/dynamics.py",3589,0," ",python,content +563,931908,"jasmine/models/dynamics.py",3590,0,"",python,selection_keyboard +564,932184,"jasmine/models/dynamics.py",3589,1,"",python,content +565,932312,"jasmine/models/dynamics.py",3588,1,"",python,content +566,932475,"jasmine/models/dynamics.py",3588,0,"o",python,content +567,932476,"jasmine/models/dynamics.py",3589,0,"",python,selection_keyboard +568,932538,"jasmine/models/dynamics.py",3589,0,"r",python,content +569,932539,"jasmine/models/dynamics.py",3590,0,"",python,selection_keyboard +570,932582,"jasmine/models/dynamics.py",3590,0," ",python,content +571,932584,"jasmine/models/dynamics.py",3591,0,"",python,selection_keyboard +572,932744,"jasmine/models/dynamics.py",3591,0,"n",python,content +573,932745,"jasmine/models/dynamics.py",3592,0,"",python,selection_keyboard +574,932932,"jasmine/models/dynamics.py",3592,0,"u",python,content +575,932933,"jasmine/models/dynamics.py",3593,0,"",python,selection_keyboard +576,933092,"jasmine/models/dynamics.py",3593,0,"m",python,content +577,933093,"jasmine/models/dynamics.py",3594,0,"",python,selection_keyboard +578,933254,"jasmine/models/dynamics.py",3594,0,"e",python,content +579,933255,"jasmine/models/dynamics.py",3595,0,"",python,selection_keyboard +580,933381,"jasmine/models/dynamics.py",3595,0,"r",python,content +581,933382,"jasmine/models/dynamics.py",3596,0,"",python,selection_keyboard +582,933499,"jasmine/models/dynamics.py",3596,0,"i",python,content +583,933499,"jasmine/models/dynamics.py",3597,0,"",python,selection_keyboard +584,933610,"jasmine/models/dynamics.py",3597,0,"c",python,content +585,933611,"jasmine/models/dynamics.py",3598,0,"",python,selection_keyboard +586,933741,"jasmine/models/dynamics.py",3598,0,"a",python,content +587,933742,"jasmine/models/dynamics.py",3599,0,"",python,selection_keyboard +588,933864,"jasmine/models/dynamics.py",3599,0,"l",python,content +589,933865,"jasmine/models/dynamics.py",3600,0,"",python,selection_keyboard +590,933941,"jasmine/models/dynamics.py",3600,0," ",python,content +591,933942,"jasmine/models/dynamics.py",3601,0,"",python,selection_keyboard +592,934064,"jasmine/models/dynamics.py",3601,0,"s",python,content +593,934065,"jasmine/models/dynamics.py",3602,0,"",python,selection_keyboard +594,934286,"jasmine/models/dynamics.py",3602,0,"t",python,content +595,934287,"jasmine/models/dynamics.py",3603,0,"",python,selection_keyboard +596,934389,"jasmine/models/dynamics.py",3603,0,"a",python,content +597,934390,"jasmine/models/dynamics.py",3604,0,"",python,selection_keyboard +598,934456,"jasmine/models/dynamics.py",3604,0,"b",python,content +599,934457,"jasmine/models/dynamics.py",3605,0,"",python,selection_keyboard +600,934712,"jasmine/models/dynamics.py",3605,0,"i",python,content +601,934713,"jasmine/models/dynamics.py",3606,0,"",python,selection_keyboard +602,934985,"jasmine/models/dynamics.py",3606,0,"l",python,content +603,934987,"jasmine/models/dynamics.py",3607,0,"",python,selection_keyboard +604,935150,"jasmine/models/dynamics.py",3607,0,"i",python,content +605,935151,"jasmine/models/dynamics.py",3608,0,"",python,selection_keyboard +606,935829,"jasmine/models/dynamics.py",3608,0,"t",python,content +607,935830,"jasmine/models/dynamics.py",3609,0,"",python,selection_keyboard +608,936711,"jasmine/models/dynamics.py",3609,0,"y",python,content +609,936713,"jasmine/models/dynamics.py",3610,0,"",python,selection_keyboard +610,936775,"jasmine/models/dynamics.py",3610,0,".",python,content +611,936776,"jasmine/models/dynamics.py",3611,0,"",python,selection_keyboard +612,939558,"jasmine/models/dynamics.py",3537,0,"",python,selection_mouse +613,939973,"jasmine/models/dynamics.py",3537,0,"\n # ",python,content +614,943431,"jasmine/models/dynamics.py",3478,0,"",python,selection_mouse +615,944246,"jasmine/models/dynamics.py",3478,0,"`",python,content +616,944247,"jasmine/models/dynamics.py",3479,0,"",python,selection_keyboard +617,944938,"jasmine/models/dynamics.py",3478,1,"`",python,content +618,944953,"jasmine/models/dynamics.py",3478,1,"`",python,content +619,944961,"jasmine/models/dynamics.py",3478,1,"",python,content +620,944962,"jasmine/models/dynamics.py",3478,0,"",python,selection_keyboard +621,944978,"jasmine/models/dynamics.py",3478,0,"`",python,content +622,944979,"jasmine/models/dynamics.py",3479,0,"",python,selection_keyboard +623,946081,"jasmine/models/dynamics.py",3526,0,"",python,selection_mouse +624,946837,"jasmine/models/dynamics.py",3526,0,"``",python,content +625,946838,"jasmine/models/dynamics.py",3527,0,"",python,selection_keyboard +626,947380,"jasmine/models/dynamics.py",3526,1,"`",python,content +627,947396,"jasmine/models/dynamics.py",3526,1,"`",python,content +628,947418,"jasmine/models/dynamics.py",3526,1,"",python,content +629,947418,"jasmine/models/dynamics.py",3526,0,"",python,selection_keyboard +630,947429,"jasmine/models/dynamics.py",3526,0,"`",python,content +631,947430,"jasmine/models/dynamics.py",3527,0,"",python,selection_keyboard +632,948972,"jasmine/models/dynamics.py",3527,1,"",python,content +633,950576,"jasmine/models/dynamics.py",3550,0,"",python,selection_mouse +634,951236,"jasmine/models/dynamics.py",3550,0,"`",python,content +635,951237,"jasmine/models/dynamics.py",3551,0,"",python,selection_keyboard +636,951723,"jasmine/models/dynamics.py",3550,1,"`",python,content +637,951737,"jasmine/models/dynamics.py",3550,1,"`",python,content +638,951748,"jasmine/models/dynamics.py",3550,1,"",python,content +639,951749,"jasmine/models/dynamics.py",3550,0,"",python,selection_keyboard +640,951761,"jasmine/models/dynamics.py",3550,0,"`",python,content +641,951762,"jasmine/models/dynamics.py",3551,0,"",python,selection_keyboard +642,953282,"jasmine/models/dynamics.py",3600,0,"",python,selection_mouse +643,953860,"jasmine/models/dynamics.py",3600,0,"``",python,content +644,953862,"jasmine/models/dynamics.py",3601,0,"",python,selection_keyboard +645,954231,"jasmine/models/dynamics.py",3600,1,"`",python,content +646,954243,"jasmine/models/dynamics.py",3600,1,"`",python,content +647,954257,"jasmine/models/dynamics.py",3600,1,"",python,content +648,954258,"jasmine/models/dynamics.py",3600,0,"",python,selection_keyboard +649,954270,"jasmine/models/dynamics.py",3600,0,"`",python,content +650,954270,"jasmine/models/dynamics.py",3601,0,"",python,selection_keyboard +651,954953,"jasmine/models/dynamics.py",3601,1,"",python,content +652,955888,"jasmine/models/dynamics.py",3539,0,"",python,selection_mouse +653,956391,"jasmine/models/dynamics.py",3666,0,"",python,selection_mouse +654,957644,"jasmine/models/dynamics.py",3539,0,"",python,selection_mouse +655,958217,"jasmine/models/dynamics.py",3766,0,"",python,selection_mouse +656,958728,"jasmine/models/dynamics.py",3765,0,"",python,selection_command +657,960338,"jasmine/models/dynamics.py",3793,0,"",python,selection_command +658,1014411,"jasmine/models/dynamics.py",3766,0,"",python,selection_mouse +659,1014423,"jasmine/models/dynamics.py",3765,0,"",python,selection_command +660,1016200,"jasmine/models/dynamics.py",3793,0,"",python,selection_command +661,1018733,"jasmine/models/dynamics.py",3767,40,"",python,content +662,1018754,"jasmine/models/dynamics.py",3779,0,"",python,selection_command +663,1019102,"jasmine/models/dynamics.py",3767,72,"",python,content +664,1019151,"jasmine/models/dynamics.py",3775,0,"",python,selection_command +665,1019584,"jasmine/models/dynamics.py",3767,28,"",python,content +666,1019601,"jasmine/models/dynamics.py",3775,0,"",python,selection_command +667,1021500,"jasmine/models/dynamics.py",3765,0,"",python,selection_mouse +668,1021989,"jasmine/models/dynamics.py",3766,0,"",python,selection_mouse +669,1021990,"jasmine/models/dynamics.py",3765,0,"",python,selection_command +670,1024746,"jasmine/models/dynamics.py",3605,0,"",python,selection_mouse +671,1025581,"jasmine/models/dynamics.py",3606,0,"",python,selection_command +672,1026597,"jasmine/models/dynamics.py",3605,0,"",python,selection_command +673,1031411,"jasmine/models/dynamics.py",3857,0,"",python,selection_mouse +674,1031988,"jasmine/models/dynamics.py",3791,0,"",python,selection_mouse +675,1032617,"jasmine/models/dynamics.py",4015,0,"",python,selection_mouse +676,1033228,"jasmine/models/dynamics.py",3936,0,"",python,selection_mouse +677,1033945,"jasmine/models/dynamics.py",4075,0,"",python,selection_mouse +678,1038445,"jasmine/models/dynamics.py",4274,0,"",python,selection_mouse +679,1039175,"jasmine/models/dynamics.py",4203,0,"",python,selection_mouse +680,1039185,"jasmine/models/dynamics.py",4202,0,"",python,selection_command +681,1040799,"jasmine/models/dynamics.py",4031,0,"",python,selection_mouse +682,1041429,"jasmine/models/dynamics.py",4052,0,"\n ",python,content +683,1042216,"jasmine/models/dynamics.py",4061,0,"#",python,content +684,1042217,"jasmine/models/dynamics.py",4062,0,"",python,selection_keyboard +685,1042638,"jasmine/models/dynamics.py",4062,0," ",python,content +686,1042639,"jasmine/models/dynamics.py",4063,0,"",python,selection_keyboard +687,1043191,"jasmine/models/dynamics.py",4063,0,"e",python,content +688,1043192,"jasmine/models/dynamics.py",4064,0,"",python,selection_keyboard +689,1043461,"jasmine/models/dynamics.py",4064,0,"x",python,content +690,1043462,"jasmine/models/dynamics.py",4065,0,"",python,selection_keyboard +691,1043571,"jasmine/models/dynamics.py",4065,0,"p",python,content +692,1043573,"jasmine/models/dynamics.py",4066,0,"",python,selection_keyboard +693,1043643,"jasmine/models/dynamics.py",4066,0,"l",python,content +694,1043644,"jasmine/models/dynamics.py",4067,0,"",python,selection_keyboard +695,1043851,"jasmine/models/dynamics.py",4067,0,"i",python,content +696,1043852,"jasmine/models/dynamics.py",4068,0,"",python,selection_keyboard +697,1045144,"jasmine/models/dynamics.py",4063,5,"",python,content +698,1045898,"jasmine/models/dynamics.py",4063,0,"c",python,content +699,1045899,"jasmine/models/dynamics.py",4064,0,"",python,selection_keyboard +700,1046036,"jasmine/models/dynamics.py",4064,0,"a",python,content +701,1046037,"jasmine/models/dynamics.py",4065,0,"",python,selection_keyboard +702,1046149,"jasmine/models/dynamics.py",4065,0,"l",python,content +703,1046150,"jasmine/models/dynamics.py",4066,0,"",python,selection_keyboard +704,1046228,"jasmine/models/dynamics.py",4066,0,"c",python,content +705,1046229,"jasmine/models/dynamics.py",4067,0,"",python,selection_keyboard +706,1046351,"jasmine/models/dynamics.py",4067,0,"u",python,content +707,1046352,"jasmine/models/dynamics.py",4068,0,"",python,selection_keyboard +708,1046519,"jasmine/models/dynamics.py",4068,0,"l",python,content +709,1046520,"jasmine/models/dynamics.py",4069,0,"",python,selection_keyboard +710,1046746,"jasmine/models/dynamics.py",4069,0,"a",python,content +711,1046747,"jasmine/models/dynamics.py",4070,0,"",python,selection_keyboard +712,1046858,"jasmine/models/dynamics.py",4070,0,"t",python,content +713,1046859,"jasmine/models/dynamics.py",4071,0,"",python,selection_keyboard +714,1046951,"jasmine/models/dynamics.py",4071,0,"e",python,content +715,1046952,"jasmine/models/dynamics.py",4072,0,"",python,selection_keyboard +716,1047036,"jasmine/models/dynamics.py",4072,0," ",python,content +717,1047037,"jasmine/models/dynamics.py",4073,0,"",python,selection_keyboard +718,1048149,"jasmine/models/dynamics.py",4073,0,"s",python,content +719,1048150,"jasmine/models/dynamics.py",4074,0,"",python,selection_keyboard +720,1048347,"jasmine/models/dynamics.py",4074,0,"q",python,content +721,1048348,"jasmine/models/dynamics.py",4075,0,"",python,selection_keyboard +722,1048586,"jasmine/models/dynamics.py",4075,0,"r",python,content +723,1048587,"jasmine/models/dynamics.py",4076,0,"",python,selection_keyboard +724,1048745,"jasmine/models/dynamics.py",4076,0,"t",python,content +725,1048746,"jasmine/models/dynamics.py",4077,0,"",python,selection_keyboard +726,1048819,"jasmine/models/dynamics.py",4077,0," ",python,content +727,1048820,"jasmine/models/dynamics.py",4078,0,"",python,selection_keyboard +728,1051896,"jasmine/models/dynamics.py",4078,0,"s",python,content +729,1051898,"jasmine/models/dynamics.py",4079,0,"",python,selection_keyboard +730,1052095,"jasmine/models/dynamics.py",4079,0,"a",python,content +731,1052096,"jasmine/models/dynamics.py",4080,0,"",python,selection_keyboard +732,1052208,"jasmine/models/dynamics.py",4080,0,"f",python,content +733,1052210,"jasmine/models/dynamics.py",4081,0,"",python,selection_keyboard +734,1052485,"jasmine/models/dynamics.py",4081,0,"l",python,content +735,1052486,"jasmine/models/dynamics.py",4082,0,"",python,selection_keyboard +736,1052736,"jasmine/models/dynamics.py",4081,1,"",python,content +737,1052824,"jasmine/models/dynamics.py",4081,0,"e",python,content +738,1052825,"jasmine/models/dynamics.py",4082,0,"",python,selection_keyboard +739,1052949,"jasmine/models/dynamics.py",4082,0,"l",python,content +740,1052950,"jasmine/models/dynamics.py",4083,0,"",python,selection_keyboard +741,1053064,"jasmine/models/dynamics.py",4083,0,"y",python,content +742,1053065,"jasmine/models/dynamics.py",4084,0,"",python,selection_keyboard +743,1053228,"jasmine/models/dynamics.py",4084,0,"\n ",python,content +744,1113217,"jasmine/models/dynamics.py",4245,0,"",python,selection_mouse +745,1113432,"jasmine/models/dynamics.py",4244,1,"\n",python,selection_mouse +746,1113503,"jasmine/models/dynamics.py",4189,56," + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n",python,selection_mouse +747,1113580,"jasmine/models/dynamics.py",4188,57," + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n",python,selection_mouse +748,1113608,"jasmine/models/dynamics.py",4187,58," + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n",python,selection_mouse +749,1113626,"jasmine/models/dynamics.py",4126,119," jnp.sqrt(1 - noise_level_B111) * vid_embed_BTNM\n + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n",python,selection_mouse +750,1113649,"jasmine/models/dynamics.py",4125,120," jnp.sqrt(1 - noise_level_B111) * vid_embed_BTNM\n + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n",python,selection_mouse +751,1113702,"jasmine/models/dynamics.py",4124,121," jnp.sqrt(1 - noise_level_B111) * vid_embed_BTNM\n + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n",python,selection_mouse +752,1113766,"jasmine/models/dynamics.py",4123,122," jnp.sqrt(1 - noise_level_B111) * vid_embed_BTNM\n + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n",python,selection_mouse +753,1113767,"jasmine/models/dynamics.py",4096,149," vid_embed_BTNM = (\n jnp.sqrt(1 - noise_level_B111) * vid_embed_BTNM\n + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n",python,selection_mouse +754,1113783,"jasmine/models/dynamics.py",4095,150," vid_embed_BTNM = (\n jnp.sqrt(1 - noise_level_B111) * vid_embed_BTNM\n + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n",python,selection_mouse +755,1113853,"jasmine/models/dynamics.py",4094,151," vid_embed_BTNM = (\n jnp.sqrt(1 - noise_level_B111) * vid_embed_BTNM\n + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n",python,selection_mouse +756,1113897,"jasmine/models/dynamics.py",4085,160," \n vid_embed_BTNM = (\n jnp.sqrt(1 - noise_level_B111) * vid_embed_BTNM\n + jnp.sqrt(noise_level_B111) * noise_BTNM\n )\n",python,selection_mouse +757,1115457,"jasmine/models/dynamics.py",4085,160,"",python,content +758,1115799,"jasmine/models/dynamics.py",4085,0,"# safe sqrt: clip argument to >= 0\none_minus_noise = jnp.clip(1.0 - noise_level_B.reshape(B, 1, 1, 1), a_min=0.0)\nsqrt_one_minus = jnp.sqrt(one_minus_noise)\nsqrt_noise = jnp.sqrt(jnp.clip(noise_level_B.reshape(B, 1, 1, 1), a_min=0.0))\n\nvid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,content +759,1117012,"jasmine/models/dynamics.py",4394,0,"",python,selection_command +760,1118191,"jasmine/models/dynamics.py",4321,74,"vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +761,1118357,"jasmine/models/dynamics.py",4320,75,"\nvid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +762,1118520,"jasmine/models/dynamics.py",4242,153,"sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B.reshape(B, 1, 1, 1), a_min=0.0))\n\nvid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +763,1118663,"jasmine/models/dynamics.py",4199,196,"sqrt_one_minus = jnp.sqrt(one_minus_noise)\nsqrt_noise = jnp.sqrt(jnp.clip(noise_level_B.reshape(B, 1, 1, 1), a_min=0.0))\n\nvid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +764,1118764,"jasmine/models/dynamics.py",4120,275,"one_minus_noise = jnp.clip(1.0 - noise_level_B.reshape(B, 1, 1, 1), a_min=0.0)\nsqrt_one_minus = jnp.sqrt(one_minus_noise)\nsqrt_noise = jnp.sqrt(jnp.clip(noise_level_B.reshape(B, 1, 1, 1), a_min=0.0))\n\nvid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +765,1119143,"jasmine/models/dynamics.py",4085,310,"# safe sqrt: clip argument to >= 0\none_minus_noise = jnp.clip(1.0 - noise_level_B.reshape(B, 1, 1, 1), a_min=0.0)\nsqrt_one_minus = jnp.sqrt(one_minus_noise)\nsqrt_noise = jnp.sqrt(jnp.clip(noise_level_B.reshape(B, 1, 1, 1), a_min=0.0))\n\nvid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +766,1119396,"jasmine/models/dynamics.py",4085,0,"",python,selection_command +767,1119555,"jasmine/models/dynamics.py",4321,0," ",python,content +768,1119555,"jasmine/models/dynamics.py",4242,0," ",python,content +769,1119555,"jasmine/models/dynamics.py",4199,0," ",python,content +770,1119555,"jasmine/models/dynamics.py",4120,0," ",python,content +771,1119555,"jasmine/models/dynamics.py",4085,0," ",python,content +772,1119939,"jasmine/models/dynamics.py",4341,0," ",python,content +773,1119939,"jasmine/models/dynamics.py",4258,0," ",python,content +774,1119939,"jasmine/models/dynamics.py",4211,0," ",python,content +775,1119939,"jasmine/models/dynamics.py",4128,0," ",python,content +776,1119940,"jasmine/models/dynamics.py",4089,0," ",python,content +777,1120254,"jasmine/models/dynamics.py",4092,0,"",python,selection_command +778,1120407,"jasmine/models/dynamics.py",4060,0,"",python,selection_command +779,1120805,"jasmine/models/dynamics.py",4053,32,"",python,content +780,1120857,"jasmine/models/dynamics.py",4061,0,"",python,selection_command +781,1122493,"jasmine/models/dynamics.py",4067,0,"",python,selection_mouse +782,1122652,"jasmine/models/dynamics.py",4063,4,"safe",python,selection_mouse +783,1123176,"jasmine/models/dynamics.py",4011,0,"",python,selection_mouse +784,1123298,"jasmine/models/dynamics.py",4000,16,"noise_level_B111",python,selection_mouse +785,1125098,"jasmine/models/dynamics.py",4033,0,"",python,selection_mouse +786,1125684,"jasmine/models/dynamics.py",4027,0,"",python,selection_mouse +787,1127539,"jasmine/models/dynamics.py",4009,0,"",python,selection_mouse +788,1129149,"jasmine/models/dynamics.py",4137,0,"",python,selection_mouse +789,1129354,"jasmine/models/dynamics.py",4137,1,"n",python,selection_mouse +790,1129355,"jasmine/models/dynamics.py",4137,3,"noi",python,selection_mouse +791,1129433,"jasmine/models/dynamics.py",4137,4,"nois",python,selection_mouse +792,1129434,"jasmine/models/dynamics.py",4137,6,"noise_",python,selection_mouse +793,1129434,"jasmine/models/dynamics.py",4137,7,"noise_l",python,selection_mouse +794,1129435,"jasmine/models/dynamics.py",4137,9,"noise_lev",python,selection_mouse +795,1129435,"jasmine/models/dynamics.py",4137,10,"noise_leve",python,selection_mouse +796,1129455,"jasmine/models/dynamics.py",4137,11,"noise_level",python,selection_mouse +797,1129474,"jasmine/models/dynamics.py",4137,12,"noise_level_",python,selection_mouse +798,1129492,"jasmine/models/dynamics.py",4137,13,"noise_level_B",python,selection_mouse +799,1129504,"jasmine/models/dynamics.py",4137,15,"noise_level_B.r",python,selection_mouse +800,1129516,"jasmine/models/dynamics.py",4137,16,"noise_level_B.re",python,selection_mouse +801,1129573,"jasmine/models/dynamics.py",4137,96,"noise_level_B.reshape(B, 1, 1, 1), a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)",python,selection_mouse +802,1129860,"jasmine/models/dynamics.py",4137,28,"noise_level_B.reshape(B, 1, ",python,selection_mouse +803,1129954,"jasmine/models/dynamics.py",4137,29,"noise_level_B.reshape(B, 1, 1",python,selection_mouse +804,1129955,"jasmine/models/dynamics.py",4137,30,"noise_level_B.reshape(B, 1, 1,",python,selection_mouse +805,1130032,"jasmine/models/dynamics.py",4137,31,"noise_level_B.reshape(B, 1, 1, ",python,selection_mouse +806,1130125,"jasmine/models/dynamics.py",4137,32,"noise_level_B.reshape(B, 1, 1, 1",python,selection_mouse +807,1132005,"jasmine/models/dynamics.py",4137,33,"noise_level_B.reshape(B, 1, 1, 1)",python,selection_command +808,1133272,"jasmine/models/dynamics.py",4137,33,"",python,content +809,1134063,"jasmine/models/dynamics.py",4137,0,"n",python,content +810,1134065,"jasmine/models/dynamics.py",4138,0,"",python,selection_keyboard +811,1134266,"jasmine/models/dynamics.py",4138,0,"o",python,content +812,1134268,"jasmine/models/dynamics.py",4139,0,"",python,selection_keyboard +813,1134346,"jasmine/models/dynamics.py",4139,0,"i",python,content +814,1134347,"jasmine/models/dynamics.py",4140,0,"",python,selection_keyboard +815,1134436,"jasmine/models/dynamics.py",4140,0,"s",python,content +816,1134437,"jasmine/models/dynamics.py",4141,0,"",python,selection_keyboard +817,1134600,"jasmine/models/dynamics.py",4141,0,"e",python,content +818,1134601,"jasmine/models/dynamics.py",4142,0,"",python,selection_keyboard +819,1135274,"jasmine/models/dynamics.py",4142,0,"_",python,content +820,1135276,"jasmine/models/dynamics.py",4143,0,"",python,selection_keyboard +821,1135509,"jasmine/models/dynamics.py",4143,0,"l",python,content +822,1135511,"jasmine/models/dynamics.py",4144,0,"",python,selection_keyboard +823,1135658,"jasmine/models/dynamics.py",4144,0,"e",python,content +824,1135660,"jasmine/models/dynamics.py",4145,0,"",python,selection_keyboard +825,1135794,"jasmine/models/dynamics.py",4145,0,"v",python,content +826,1135796,"jasmine/models/dynamics.py",4146,0,"",python,selection_keyboard +827,1135913,"jasmine/models/dynamics.py",4146,0,"e",python,content +828,1135914,"jasmine/models/dynamics.py",4147,0,"",python,selection_keyboard +829,1135985,"jasmine/models/dynamics.py",4147,0,"l",python,content +830,1135986,"jasmine/models/dynamics.py",4148,0,"",python,selection_keyboard +831,1136325,"jasmine/models/dynamics.py",4148,0,"_",python,content +832,1136327,"jasmine/models/dynamics.py",4149,0,"",python,selection_keyboard +833,1137116,"jasmine/models/dynamics.py",4149,0,"B",python,content +834,1137118,"jasmine/models/dynamics.py",4150,0,"",python,selection_keyboard +835,1137775,"jasmine/models/dynamics.py",4150,0,"1",python,content +836,1137776,"jasmine/models/dynamics.py",4151,0,"",python,selection_keyboard +837,1137956,"jasmine/models/dynamics.py",4151,0,"1",python,content +838,1137957,"jasmine/models/dynamics.py",4152,0,"",python,selection_keyboard +839,1138049,"jasmine/models/dynamics.py",4152,0,"1",python,content +840,1138051,"jasmine/models/dynamics.py",4153,0,"",python,selection_keyboard +841,1138557,"jasmine/models/dynamics.py",4137,16,"noise_level_B111",python,content +842,1139079,"jasmine/models/dynamics.py",4152,0,"",python,selection_command +843,1139196,"jasmine/models/dynamics.py",4215,0,"",python,selection_command +844,1139816,"jasmine/models/dynamics.py",4273,0,"",python,selection_command +845,1140015,"jasmine/models/dynamics.py",4272,0,"",python,selection_command +846,1140515,"jasmine/models/dynamics.py",4271,0,"",python,selection_command +847,1140579,"jasmine/models/dynamics.py",4270,0,"",python,selection_command +848,1140580,"jasmine/models/dynamics.py",4269,0,"",python,selection_command +849,1140669,"jasmine/models/dynamics.py",4268,0,"",python,selection_command +850,1140670,"jasmine/models/dynamics.py",4267,0,"",python,selection_command +851,1140671,"jasmine/models/dynamics.py",4266,0,"",python,selection_command +852,1140692,"jasmine/models/dynamics.py",4265,0,"",python,selection_command +853,1140720,"jasmine/models/dynamics.py",4264,0,"",python,selection_command +854,1140780,"jasmine/models/dynamics.py",4263,0,"",python,selection_command +855,1140780,"jasmine/models/dynamics.py",4262,0,"",python,selection_command +856,1140833,"jasmine/models/dynamics.py",4261,0,"",python,selection_command +857,1140839,"jasmine/models/dynamics.py",4260,0,"",python,selection_command +858,1140922,"jasmine/models/dynamics.py",4259,0,"",python,selection_command +859,1141109,"jasmine/models/dynamics.py",4258,0,"",python,selection_command +860,1141213,"jasmine/models/dynamics.py",4257,0,"",python,selection_command +861,1141338,"jasmine/models/dynamics.py",4256,0,"",python,selection_command +862,1141609,"jasmine/models/dynamics.py",4256,13,"",python,content +863,1141906,"jasmine/models/dynamics.py",4256,1,"",python,content +864,1142203,"jasmine/models/dynamics.py",4256,7,"",python,content +865,1142476,"jasmine/models/dynamics.py",4256,1,"",python,content +866,1142784,"jasmine/models/dynamics.py",4256,1,"",python,content +867,1143030,"jasmine/models/dynamics.py",4256,2,"",python,content +868,1143291,"jasmine/models/dynamics.py",4256,1,"",python,content +869,1143550,"jasmine/models/dynamics.py",4256,2,"",python,content +870,1143851,"jasmine/models/dynamics.py",4256,1,"",python,content +871,1144153,"jasmine/models/dynamics.py",4256,2,"",python,content +872,1144418,"jasmine/models/dynamics.py",4256,1,"",python,content +873,1144861,"jasmine/models/dynamics.py",4256,1,"",python,content +874,1145394,"jasmine/models/dynamics.py",4256,0,"n",python,content +875,1145395,"jasmine/models/dynamics.py",4257,0,"",python,selection_keyboard +876,1145590,"jasmine/models/dynamics.py",4257,0,"o",python,content +877,1145592,"jasmine/models/dynamics.py",4258,0,"",python,selection_keyboard +878,1145676,"jasmine/models/dynamics.py",4258,0,"i",python,content +879,1145678,"jasmine/models/dynamics.py",4259,0,"",python,selection_keyboard +880,1145695,"jasmine/models/dynamics.py",4259,0,"s",python,content +881,1145696,"jasmine/models/dynamics.py",4260,0,"",python,selection_keyboard +882,1145869,"jasmine/models/dynamics.py",4260,0,"e",python,content +883,1145871,"jasmine/models/dynamics.py",4261,0,"",python,selection_keyboard +884,1146091,"jasmine/models/dynamics.py",4261,0,"_",python,content +885,1146092,"jasmine/models/dynamics.py",4262,0,"",python,selection_keyboard +886,1146310,"jasmine/models/dynamics.py",4262,0,"l",python,content +887,1146312,"jasmine/models/dynamics.py",4263,0,"",python,selection_keyboard +888,1146399,"jasmine/models/dynamics.py",4263,0,"e",python,content +889,1146400,"jasmine/models/dynamics.py",4264,0,"",python,selection_keyboard +890,1146673,"jasmine/models/dynamics.py",4264,0,"e",python,content +891,1146675,"jasmine/models/dynamics.py",4265,0,"",python,selection_keyboard +892,1147012,"jasmine/models/dynamics.py",4264,1,"",python,content +893,1147066,"jasmine/models/dynamics.py",4264,0,"v",python,content +894,1147067,"jasmine/models/dynamics.py",4265,0,"",python,selection_keyboard +895,1147199,"jasmine/models/dynamics.py",4265,0,"e",python,content +896,1147200,"jasmine/models/dynamics.py",4266,0,"",python,selection_keyboard +897,1147302,"jasmine/models/dynamics.py",4266,0,"l",python,content +898,1147303,"jasmine/models/dynamics.py",4267,0,"",python,selection_keyboard +899,1147633,"jasmine/models/dynamics.py",4267,0,"_",python,content +900,1147634,"jasmine/models/dynamics.py",4268,0,"",python,selection_keyboard +901,1148384,"jasmine/models/dynamics.py",4268,0,"B",python,content +902,1148385,"jasmine/models/dynamics.py",4269,0,"",python,selection_keyboard +903,1148719,"jasmine/models/dynamics.py",4269,0,"1",python,content +904,1148720,"jasmine/models/dynamics.py",4270,0,"",python,selection_keyboard +905,1148983,"jasmine/models/dynamics.py",4270,0,"1",python,content +906,1148984,"jasmine/models/dynamics.py",4271,0,"",python,selection_keyboard +907,1149162,"jasmine/models/dynamics.py",4271,0,"1",python,content +908,1149163,"jasmine/models/dynamics.py",4272,0,"",python,selection_keyboard +909,1149722,"jasmine/models/dynamics.py",4256,16,"noise_level_B111",python,content +910,1150902,"jasmine/models/dynamics.py",4333,0,"",python,selection_mouse +911,1151429,"jasmine/models/dynamics.py",4286,0,"",python,selection_mouse +912,1152596,"jasmine/models/dynamics.py",4316,0,"",python,selection_mouse +913,1152968,"jasmine/models/dynamics.py",4312,14,"sqrt_one_minus",python,selection_mouse +914,1153797,"jasmine/models/dynamics.py",4338,0,"",python,selection_mouse +915,1153955,"jasmine/models/dynamics.py",4329,14,"vid_embed_BTNM",python,selection_mouse +916,1155654,"jasmine/models/dynamics.py",4301,0,"",python,selection_mouse +917,1155810,"jasmine/models/dynamics.py",4295,14,"vid_embed_BTNM",python,selection_mouse +918,1158486,"jasmine/models/dynamics.py",4406,0,"",python,selection_mouse +919,1159132,"jasmine/models/dynamics.py",4354,0,"",python,selection_mouse +920,1159285,"jasmine/models/dynamics.py",4346,10,"sqrt_noise",python,selection_mouse +921,1159792,"jasmine/models/dynamics.py",4364,0,"",python,selection_mouse +922,1159945,"jasmine/models/dynamics.py",4359,10,"noise_BTNM",python,selection_mouse +923,1169227,"jasmine/models/dynamics.py",4052,0,"",python,selection_mouse +924,1170168,"jasmine/models/dynamics.py",4052,0,"\n ",python,content +925,1180135,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_command +926,1180172,"TERMINAL",0,0,"]633;C]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +927,1189861,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --job-name=train_dyn_default_breakout_longer\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n# source .venv/bin/activate\n\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\n# lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\n# tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\n\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3518963\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\n\nenv | grep SLURM\n\nexport PYTHONUNBUFFERED=1\n\nsrun python jasmine/train_dynamics.py \\n --save_ckpt \\n --image_height=10 \\n --image_width=10 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=120 \\n --init_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=250 \\n --log_checkpoint_interval=250 \\n --dyna_type=maskgit \\n --log \\n --name=breakout-dyn-noise-lvl-default-$slurm_job_id \\n --tags dyn breakout noise-lvl default \\n --entity instant-uv \\n --project jafar \\n --patch_size 4 \\n --lam_patch_size 4 \\n --warmup_steps 100 \\n --wsd_decay_steps 1000 \\n --num_steps 5000 \\n --data_dir $array_records_dir_train \\n --val_data_dir $array_records_dir_val \\n --tokenizer_checkpoint $tokenizer_checkpoint \\n --lam_checkpoint $lam_checkpoint \\n --val_interval 500 \\n --eval_full_frame \\n",shellscript,tab +928,1194171,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2426,0,"",shellscript,selection_command +929,1205100,"TERMINAL",0,0,"sync-runner",,terminal_command +930,1205148,"TERMINAL",0,0,"]633;Csending incremental file list\r\n",,terminal_output +931,1211361,"TERMINAL",0,0,"jasmine/\r\njasmine/sample.py\r\njasmine/train_dynamics.py\r\njasmine/models/dynamics.py\r\n",,terminal_output +932,1211516,"TERMINAL",0,0,"\r\nsent 79,543 bytes received 277 bytes 10,642.67 bytes/sec\r\ntotal size is 28,028,497 speedup is 351.15\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +933,1217624,"TERMINAL",0,0,"salloc --time=03:00:00 --partition=accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +934,1217713,"TERMINAL",0,0,"]633;Csalloc: Granted job allocation 3527669\r\n",,terminal_output +935,1217842,"TERMINAL",0,0,"salloc: Waiting for resource configuration\r\n",,terminal_output +936,1244878,"TERMINAL",0,0,"salloc: Nodes hkn0702 are ready for job\r\n",,terminal_output +937,1245893,"TERMINAL",0,0,"]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h[tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +938,1249143,"TERMINAL",0,0,"s",,terminal_output +939,1249206,"TERMINAL",0,0,"o",,terminal_output +940,1249268,"TERMINAL",0,0,"u",,terminal_output +941,1249331,"TERMINAL",0,0,"r",,terminal_output +942,1249555,"TERMINAL",0,0,"c",,terminal_output +943,1249641,"TERMINAL",0,0," e",,terminal_output +944,1250256,"TERMINAL",0,0,"",,terminal_output +945,1250390,"TERMINAL",0,0,"",,terminal_output +946,1250500,"TERMINAL",0,0,"e",,terminal_output +947,1250569,"TERMINAL",0,0," ",,terminal_output +948,1250675,"TERMINAL",0,0,".",,terminal_output +949,1251192,"TERMINAL",0,0,"v",,terminal_output +950,1251257,"TERMINAL",0,0,"e",,terminal_output +951,1251443,"TERMINAL",0,0,"nv/",,terminal_output +952,1251572,"TERMINAL",0,0,"b",,terminal_output +953,1251742,"TERMINAL",0,0,"in/",,terminal_output +954,1252019,"TERMINAL",0,0,"a",,terminal_output +955,1252081,"TERMINAL",0,0,"c",,terminal_output +956,1252264,"TERMINAL",0,0,"tivate",,terminal_output +957,1252567,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +958,1257984,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +959,1259458,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +960,1259473,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +961,1263253,"TERMINAL",0,0,"s",,terminal_output +962,1263476,"TERMINAL",0,0,"h",,terminal_output +963,1263662,"TERMINAL",0,0," ",,terminal_output +964,1263860,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",,terminal_output +965,1264062,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=48:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\n# source .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\n# lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\r\n# tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\r\n\r\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3518963\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=250 \\r\n --log_checkpoint_interval=250 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-noise-lvl-default-$slurm_job_id \\r\n --tags dyn breakout noise-lvl default \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 100 \\r\n --wsd_decay_steps 1000 \\r\n --num_steps 5000 \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --lam_checkpoint $lam_checkpoint \\r\n --val_interval 500 \\r\n --eval_full_frame \\r\n",,terminal_output +966,1264341,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=3602395\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0702\r\nSLURM_JOB_START_TIME=1759060062\r\nSLURM_STEP_NODELIST=hkn0702\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759070862\r\nSLURM_PMI2_SRUN_PORT=38045\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3527669\r\nSLURM_PTY_PORT=34957\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=36\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e8.hkn0702\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=156\r\nSLURM_NODELIST=hkn0702\r\nSLURM_SRUN_COMM_PORT=35753\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3527669\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0702\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=35753\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0702\r\n",,terminal_output +967,1264646,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +968,1266633,"jasmine/models/dynamics.py",0,0,"",python,tab +969,1277673,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +970,1283809,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +971,1284206,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +972,1285249,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20250928_134848-3q3izg97\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run breakout-dyn-noise-lvl-default-3527669\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/3q3izg97\r\n",,terminal_output +973,1285642,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26561024, 'lam': 16900976, 'tokenizer': 33750256, 'total': 77212256}\r\n",,terminal_output +974,1289920,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +975,1353197,"jasmine/models/dynamics.py",0,0,"",python,tab +976,1363676,"TERMINAL",0,0,"Total memory size: 2.7 GB, Output size: 0.9 GB, Temp size: 1.9 GB, Argument size: 0.9 GB, Host temp size: 0.0 GB.\r\nFLOPs: 4.797e+11, Bytes: 5.190e+10 (48.3 GB), Intensity: 9.2 FLOPs/byte\r\nStarting training from step 0...\r\n",,terminal_output +977,1364179,"TERMINAL",0,0,"\r\nMemstats: After params initialized:\r\n\tUsing (GB) 1.1 / 38.7 (2.842377%) on cuda:0\r\n",,terminal_output +978,1364361,"jasmine/models/dynamics.py",4172,0,"",python,selection_mouse +979,1364513,"jasmine/models/dynamics.py",4172,1,"0",python,selection_mouse +980,1364688,"jasmine/models/dynamics.py",4170,3,"0.0",python,selection_mouse +981,1364734,"jasmine/models/dynamics.py",4169,4,"=0.0",python,selection_mouse +982,1364790,"jasmine/models/dynamics.py",4164,9,"a_min=0.0",python,selection_mouse +983,1365419,"jasmine/models/dynamics.py",4165,0,"",python,selection_mouse +984,1365550,"jasmine/models/dynamics.py",4164,5,"a_min",python,selection_mouse +985,1365798,"jasmine/models/dynamics.py",4164,6,"a_min=",python,selection_mouse +986,1365863,"jasmine/models/dynamics.py",4164,7,"a_min=0",python,selection_mouse +987,1365980,"jasmine/models/dynamics.py",4164,8,"a_min=0.",python,selection_mouse +988,1366495,"jasmine/models/dynamics.py",4172,0,"",python,selection_mouse +989,1366496,"jasmine/models/dynamics.py",4172,1,"0",python,selection_mouse +990,1366691,"jasmine/models/dynamics.py",4171,2,".0",python,selection_mouse +991,1366713,"jasmine/models/dynamics.py",4170,3,"0.0",python,selection_mouse +992,1366734,"jasmine/models/dynamics.py",4169,4,"=0.0",python,selection_mouse +993,1366772,"jasmine/models/dynamics.py",4164,9,"a_min=0.0",python,selection_mouse +994,1367158,"jasmine/models/dynamics.py",4166,0,"",python,selection_mouse +995,1372687,"jasmine/models/dynamics.py",4154,0,"",python,selection_mouse +996,1372869,"jasmine/models/dynamics.py",4146,16,"noise_level_B111",python,selection_mouse +997,1373745,"jasmine/models/dynamics.py",4276,0,"",python,selection_mouse +998,1373890,"jasmine/models/dynamics.py",4265,16,"noise_level_B111",python,selection_mouse +999,1384307,"jasmine/models/dynamics.py",3648,0,"",python,selection_mouse +1000,1384415,"jasmine/models/dynamics.py",3635,18,"noise_bucket_idx_B",python,selection_mouse +1001,1430982,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=06:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --job-name=train_dyn_default_breakout_longer\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n# source .venv/bin/activate\n\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\n# lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\n# tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\n\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3518963\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\n\nenv | grep SLURM\n\nsrun python jasmine/train_dynamics.py \\n --save_ckpt \\n --image_height=10 \\n --image_width=10 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=120 \\n --init_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --dyna_type=maskgit \\n --log \\n --name=breakout-dyn-50ksteps-3e-5-noise-lvl-$slurm_job_id \\n --tags dyn breakout 50ksteps 3e-5 noise-lvl \\n --entity instant-uv \\n --project jafar \\n --patch_size 4 \\n --lam_patch_size 4 \\n --warmup_steps 1000 \\n --wsd_decay_steps 10000 \\n --num_steps 50000 \\n --data_dir $array_records_dir_train \\n --val_data_dir $array_records_dir_val \\n --tokenizer_checkpoint $tokenizer_checkpoint \\n --lam_checkpoint $lam_checkpoint \\n --val_interval 1000 \\n --eval_full_frame \\n",shellscript,tab +1002,1432505,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=06:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --job-name=train_dyn_default_gt_actions_breakout_longer\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n# source .venv/bin/activate\n\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect_big/train\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect_big/val\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn-gt-actions/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\n\nenv | grep SLURM\n\nsrun python jasmine/train_dynamics.py \\n --save_ckpt \\n --image_height=10 \\n --image_width=10 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=120 \\n --init_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=250 \\n --log_checkpoint_interval=250 \\n --dyna_type=maskgit \\n --log \\n --name=breakout-dyn-50ksteps-noise-lvl-gt-actions-lr-3e-5-$slurm_job_id \\n --tags dyn breakout 50ksteps noise-lvl 3e-5 \\n --entity instant-uv \\n --project jafar \\n --patch_size 4 \\n --lam_patch_size 4 \\n --warmup_steps 100 \\n --wsd_decay_steps 1000 \\n --num_steps 50000 \\n --use_gt_actions \\n --data_dir $array_records_dir_train \\n --val_data_dir $array_records_dir_val \\n --tokenizer_checkpoint $tokenizer_checkpoint \\n --val_interval 750 \\n --eval_full_frame \\n",shellscript,tab +1003,1433550,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=06:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --job-name=train_dyn_default_gt_actions_breakout_longer_smaller_lr\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n# source .venv/bin/activate\n\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect_big/train\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect_big/val\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn-gt-actions/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\n\nenv | grep SLURM\n\nsrun python jasmine/train_dynamics.py \\n --save_ckpt \\n --image_height=10 \\n --image_width=10 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=120 \\n --init_lr=0 \\n --max_lr=1e-5 \\n --log_image_interval=250 \\n --log_checkpoint_interval=250 \\n --dyna_type=maskgit \\n --log \\n --name=breakout-dyn-50ksteps-noise-lvl-gt-actions-lr-1e-5-$slurm_job_id \\n --tags dyn breakout 50ksteps noise-lvl 1e-5 \\n --entity instant-uv \\n --project jafar \\n --patch_size 4 \\n --lam_patch_size 4 \\n --warmup_steps 100 \\n --wsd_decay_steps 1000 \\n --num_steps 50000 \\n --use_gt_actions \\n --data_dir $array_records_dir_train \\n --val_data_dir $array_records_dir_val \\n --tokenizer_checkpoint $tokenizer_checkpoint \\n --val_interval 750 \\n --eval_full_frame \\n",shellscript,tab +1004,1434461,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --job-name=train_dyn_default_breakout_longer\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n# source .venv/bin/activate\n\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect_big/train\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect_big/val\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn-gt-actions/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\n\nenv | grep SLURM\n\nexport PYTHONUNBUFFERED=1\n\nsrun python jasmine/train_dynamics.py \\n --save_ckpt \\n --image_height=10 \\n --image_width=10 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=120 \\n --init_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --dyna_type=maskgit \\n --log \\n --name=breakout-dyn-default-noise-lvl-gt-actions-$slurm_job_id \\n --tags dyn breakout default noise-lvl \\n --entity instant-uv \\n --project jafar \\n --patch_size 4 \\n --lam_patch_size 4 \\n --warmup_steps 100 \\n --wsd_decay_steps 1000 \\n --num_steps 5000 \\n --use_gt_actions \\n --data_dir $array_records_dir_train \\n --val_data_dir $array_records_dir_val \\n --tokenizer_checkpoint $tokenizer_checkpoint \\n --val_interval 1000 \\n --eval_full_frame \\n",shellscript,tab +1005,1436849,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",85,0,"",shellscript,selection_mouse +1006,1437426,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",84,0,"",shellscript,selection_mouse +1007,1438415,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",83,1,"",shellscript,content +1008,1438517,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",82,1,"",shellscript,content +1009,1439260,"TERMINAL",0,0,"WARNING:absl:[process=0][thread=MainThread][operation_id=1] _SignalingThread.join() waiting for signals ([, ]) blocking the main thread will slow down blocking save times. This is likely due to main thread calling result() on a CommitFuture.\r\nWARNING:absl:[process=0][thread=MainThread][operation_id=1] _SignalingThread.join() waiting for signals ([, ]) blocking the main thread will slow down blocking save times. This is likely due to main thread calling result() on a CommitFuture.\r\n",,terminal_output +1010,1439873,"TERMINAL",0,0,"Saved checkpoint at step 250\r\n",,terminal_output +1011,1441373,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",82,0,"0",shellscript,content +1012,1441374,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",83,0,"",shellscript,selection_keyboard +1013,1441433,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",83,0,"6",shellscript,content +1014,1441434,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",84,0,"",shellscript,selection_keyboard +1015,1443650,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3527669.0 task 0: running\r\n",,terminal_output +1016,1443883,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3527669.0\r\nsrun: forcing job termination\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-7:\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3527669.0 ON hkn0702 CANCELLED AT 2025-09-28T13:51:28 ***\r\n",,terminal_output +1017,1444073,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3527669.0\r\nsrun: job abort in progress\r\n",,terminal_output +1018,1444239,"TERMINAL",0,0,"]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1019,1460767,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",0,0,"",shellscript,tab +1020,1462285,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",0,0,"",shellscript,tab +1021,1463229,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",0,0,"",shellscript,tab +1022,1463770,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",0,0,"",shellscript,tab +1023,1465087,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +1024,1466594,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",84,0,"",shellscript,selection_mouse +1025,1467503,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",83,1,"",shellscript,content +1026,1467594,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",82,1,"",shellscript,content +1027,1468168,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",82,0,"0",shellscript,content +1028,1468169,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",83,0,"",shellscript,selection_keyboard +1029,1468543,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",83,0,"6",shellscript,content +1030,1468544,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",84,0,"",shellscript,selection_keyboard +1031,1470697,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_smaller_lr_50k.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=06:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --job-name=train_dyn_default_breakout_longer_smaller_lr\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n# source .venv/bin/activate\n\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\n# lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\n# tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\n\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3518963\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\n\nenv | grep SLURM\n\nsrun python jasmine/train_dynamics.py \\n --save_ckpt \\n --image_height=10 \\n --image_width=10 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=120 \\n --init_lr=0 \\n --max_lr=1e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --dyna_type=maskgit \\n --log \\n --name=breakout-dyn-50ksteps-noise-lvl-$slurm_job_id \\n --tags dyn breakout 50ksteps noise-lvl 1e-5 \\n --entity instant-uv \\n --project jafar \\n --patch_size 4 \\n --lam_patch_size 4 \\n --warmup_steps 1000 \\n --wsd_decay_steps 10000 \\n --num_steps 50000 \\n --data_dir $array_records_dir_train \\n --val_data_dir $array_records_dir_val \\n --tokenizer_checkpoint $tokenizer_checkpoint \\n --lam_checkpoint $lam_checkpoint \\n --val_interval 1000 \\n --eval_full_frame \\n",shellscript,tab +1032,1474613,"TERMINAL",0,0,"s",,terminal_output +1033,1474790,"TERMINAL",0,0,"y",,terminal_output +1034,1474843,"TERMINAL",0,0,"n",,terminal_output +1035,1475035,"TERMINAL",0,0,"c",,terminal_output +1036,1475098,"TERMINAL",0,0,"-",,terminal_output +1037,1475502,"TERMINAL",0,0,"r",,terminal_output +1038,1475636,"TERMINAL",0,0,"u",,terminal_output +1039,1475773,"TERMINAL",0,0,"n",,terminal_output +1040,1475946,"TERMINAL",0,0,"e",,terminal_output +1041,1476076,"TERMINAL",0,0,"n",,terminal_output +1042,1476411,"TERMINAL",0,0,"",,terminal_output +1043,1476517,"TERMINAL",0,0,"",,terminal_output +1044,1477042,"TERMINAL",0,0,"n",,terminal_output +1045,1477206,"TERMINAL",0,0,"er",,terminal_output +1046,1477337,"TERMINAL",0,0,"\r\n[?2004l\rsending incremental file list\r\n",,terminal_output +1047,1483170,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh\r\n",,terminal_output +1048,1483246,"TERMINAL",0,0,"\r\nsent 37,713 bytes received 253 bytes 5,840.92 bytes/sec\r\ntotal size is 28,028,497 speedup is 738.25\r\n]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1049,1484867,"TERMINAL",0,0,"s",,terminal_output +1050,1484971,"TERMINAL",0,0,"b",,terminal_output +1051,1485093,"TERMINAL",0,0,"a",,terminal_output +1052,1485435,"TERMINAL",0,0,"",,terminal_output +1053,1486475,"TERMINAL",0,0,"a",,terminal_output +1054,1486537,"TERMINAL",0,0,"t",,terminal_output +1055,1486676,"TERMINAL",0,0,"c",,terminal_output +1056,1486739,"TERMINAL",0,0,"h",,terminal_output +1057,1487139,"TERMINAL",0,0,"_",,terminal_output +1058,1487335,"TERMINAL",0,0,"d",,terminal_output +1059,1487456,"TERMINAL",0,0,"i",,terminal_output +1060,1487518,"TERMINAL",0,0,"r",,terminal_output +1061,1487661,"TERMINAL",0,0," ",,terminal_output +1062,1488323,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_smaller_lr_50k.sh",,terminal_output +1063,1488980,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_smaller_lr_50k.s",,terminal_output +1064,1489216,"TERMINAL",0,0,"",,terminal_output +1065,1489426,"TERMINAL",0,0,"",,terminal_output +1066,1490315,"TERMINAL",0,0,"",,terminal_output +1067,1490522,"TERMINAL",0,0,"\r\r\n\r",,terminal_output +1068,1490581,"TERMINAL",0,0,"",,terminal_output +1069,1490699,"TERMINAL",0,0,"",,terminal_output +1070,1490849,"TERMINAL",0,0,"",,terminal_output +1071,1491527,"TERMINAL",0,0,"",,terminal_output +1072,1491669,"TERMINAL",0,0,"",,terminal_output +1073,1491833,"TERMINAL",0,0,"",,terminal_output +1074,1491969,"TERMINAL",0,0,"",,terminal_output +1075,1492376,"TERMINAL",0,0,"\r\n[?2004l\rsbatching all files in /home/hk-project-p0023960/tum_cte0515/Projects/jasmine/slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/\r\nSubmitted batch job 3527671\r\nSubmitted batch job 3527672\r\n",,terminal_output +1076,1492439,"TERMINAL",0,0,"Submitted batch job 3527673\r\nSubmitted batch job 3527674\r\nSubmitted batch job 3527675\r\n",,terminal_output +1077,1492507,"TERMINAL",0,0,"Submitted batch job 3527676\r\n]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1078,1493953,"TERMINAL",0,0,"q",,terminal_output +1079,1494014,"TERMINAL",0,0,"u",,terminal_output +1080,1494119,"TERMINAL",0,0,"e",,terminal_output +1081,1494181,"TERMINAL",0,0,"u",,terminal_output +1082,1494243,"TERMINAL",0,0,"e",,terminal_output +1083,1494386,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: squeu... hkn0702.localdomain: Sun Sep 28 13:52:19 2025JOBID PARTITION NAME USER ST\tTIME NODESNODELIST(REASON)3527671 accelerat train_dy tum_cte0 R\t0:01\t 1hkn07023527672 accelerat train_dy tum_cte0 R\t0:01\t 1hkn07023527673 accelerat train_dy tum_cte0 R\t0:01\t 1hkn07023527674 accelerat train_dy tum_cte0 R\t0:01\t 1hkn05343527675 accelerat train_dy tum_cte0 R\t0:01\t 1hkn08073527676 accelerat train_dy tum_cte0 R\t0:01\t 1hkn08073527669 accelerat interact tum_cte0 R\t4:37\t 1hkn0702",,terminal_output +1084,1495389,"TERMINAL",0,0,"202222228",,terminal_output +1085,1496481,"TERMINAL",0,0,"13333339",,terminal_output +1086,1497408,"TERMINAL",0,0,"244444440",,terminal_output +1087,1497779,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1088,1505553,"TERMINAL",0,0,"s",,terminal_output +1089,1505749,"TERMINAL",0,0,"c",,terminal_output +1090,1505866,"TERMINAL",0,0,"a",,terminal_output +1091,1506026,"TERMINAL",0,0,"c",,terminal_output +1092,1506372,"TERMINAL",0,0,"",,terminal_output +1093,1506578,"TERMINAL",0,0,"n",,terminal_output +1094,1506683,"TERMINAL",0,0,"c",,terminal_output +1095,1506745,"TERMINAL",0,0,"e",,terminal_output +1096,1506808,"TERMINAL",0,0,"l",,terminal_output +1097,1506976,"TERMINAL",0,0," ",,terminal_output +1098,1507232,"TERMINAL",0,0,"3527671",,terminal_output +1099,1507850,"TERMINAL",0,0,"3527671 ",,terminal_output +1100,1509345,"TERMINAL",0,0,"3527672",,terminal_output +1101,1509859,"TERMINAL",0,0,"3527672 ",,terminal_output +1102,1512153,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1103,1521804,"TERMINAL",0,0,"q",,terminal_output +1104,1521950,"TERMINAL",0,0,"u",,terminal_output +1105,1522006,"TERMINAL",0,0,"e",,terminal_output +1106,1522068,"TERMINAL",0,0,"u",,terminal_output +1107,1522129,"TERMINAL",0,0,"e",,terminal_output +1108,1522319,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: squeu... hkn0702.localdomain: Sun Sep 28 13:52:46 2025JOBID PARTITION NAME USER ST\tTIME NODESNODELIST(REASON)3527671 accelerat train_dy tum_cte0 R\t0:28\t 1hkn07023527672 accelerat train_dy tum_cte0 R\t0:28\t 1hkn07023527673 accelerat train_dy tum_cte0 R\t0:28\t 1hkn07023527674 accelerat train_dy tum_cte0 R\t0:28\t 1hkn05343527675 accelerat train_dy tum_cte0 R\t0:28\t 1hkn08073527676 accelerat train_dy tum_cte0 R\t0:28\t 1hkn08073527669 accelerat interact tum_cte0 R\t5:04\t 1hkn0702",,terminal_output +1109,1523465,"TERMINAL",0,0,"73030303030306",,terminal_output +1110,1523543,"TERMINAL",0,0,"Every 1.0s: squeu... hkn0702.localdomain: Sun Sep 28 13:52:48 2025JOBID PARTITION NAME USER ST\tTIME NODESNODELIST(REASON)3527671 accelerat train_dy tum_cte0 R\t0:30\t 1hkn07023527672 accelerat train_dy tum_cte0 R\t0:30\t 1hkn07023527673 accelerat train_dy tum_cte0 R\t0:30\t 1hkn07023527674 accelerat train_dy tum_cte0 R\t0:30\t 1hkn05343527675 accelerat train_dy tum_cte0 R\t0:30\t 1hkn08073527676 accelerat train_dy tum_cte0 R\t0:30\t 1hkn08073527669 accelerat interact tum_cte0 R\t5:06\t 1hkn0702Every 1.0s: squeue --mehkn0702.localdomain: Sun Sep 28 13:52:48 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3527671 accelerat train_dy tum_cte0 R\t0:30\t 1 hkn07023527672 accelerat train_dy tum_cte0 R\t0:30\t 1 hkn07023527673 accelerat train_dy tum_cte0 R\t0:30\t 1 hkn07023527674 accelerat train_dy tum_cte0 R\t0:30\t 1 hkn05343527675 accelerat train_dy tum_cte0 R\t0:30\t 1 hkn08073527676 accelerat train_dy tum_cte0 R\t0:30\t 1 hkn08073527669 accelerat interact tum_cte0 R\t5:06\t 1 hkn0702",,terminal_output +1111,1524542,"TERMINAL",0,0,"91111117",,terminal_output +1112,1525313,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1113,1528583,"TERMINAL",0,0,"gi",,terminal_output +1114,1528703,"TERMINAL",0,0,"t",,terminal_output +1115,1528767,"TERMINAL",0,0," ",,terminal_output +1116,1528983,"TERMINAL",0,0,"st",,terminal_output +1117,1529428,"TERMINAL",0,0,"a",,terminal_output +1118,1529481,"TERMINAL",0,0,"t",,terminal_output +1119,1529616,"TERMINAL",0,0,"u",,terminal_output +1120,1529735,"TERMINAL",0,0,"s",,terminal_output +1121,1529879,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +1122,1530047,"TERMINAL",0,0,"On branch add-noise-to-combat-exposure-bias\r\nYour branch is up to date with 'origin/add-noise-to-combat-exposure-bias'.\r\n\r\n",,terminal_output +1123,1530170,"TERMINAL",0,0,"Last commands done (2 commands done):\r\n pick ba37453 feat: generate coinrun dataset with val split\r\n pick faadd10 feat: implemented validation loss for all three models\r\nNext commands to do (26 remaining commands):\r\n pick 9a17dbb fix: pass val data path to dataloader\r\n pick 6e69cdb fix typo in image logging\r\n (use ""git rebase --edit-todo"" to view and edit)\r\nYou are currently editing a commit while rebasing branch 'gt-actions' on 'c7522f2'.\r\n (use ""git commit --amend"" to amend the current commit)\r\n (use ""git rebase --continue"" once you are satisfied with your changes)\r\n\r\nChanges not staged for commit:\r\n (use ""git add ..."" to update what will be committed)\r\n (use ""git restore ..."" to discard changes in working directory)\r\n\tmodified: jasmine/models/dynamics.py\r\n\r\nUntracked files:\r\n (use ""git add ..."" to include in what will be committed)\r\n\tdiff.diff\r\n\tdiff2.diff\r\n\tinput_pipeline/\r\n\tkiller.sh\r\n\tkiller_partition.sh\r\n\tlog.log\r\n\toverfit_dir.zip\r\n\trequirements-franz.txt\r\n\tsamples/\r\n\tscripts_cremers/\r\n\tslurm/\r\n\ttest.py\r\n\tutils/\r\n\tuv.lock\r\n\r\nno changes added to commit (use ""git add"" and/or ""git commit -a"")\r\n]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1124,1531156,"TERMINAL",0,0,"gi",,terminal_output +1125,1531286,"TERMINAL",0,0,"t",,terminal_output +1126,1531349,"TERMINAL",0,0," ",,terminal_output +1127,1531552,"TERMINAL",0,0,"a",,terminal_output +1128,1531686,"TERMINAL",0,0,"d",,terminal_output +1129,1531873,"TERMINAL",0,0,"d",,terminal_output +1130,1531934,"TERMINAL",0,0," ",,terminal_output +1131,1532005,"TERMINAL",0,0,"j",,terminal_output +1132,1532125,"TERMINAL",0,0,"a",,terminal_output +1133,1532336,"TERMINAL",0,0,"s",,terminal_output +1134,1533242,"TERMINAL",0,0,"mine/",,terminal_output +1135,1533302,"TERMINAL",0,0,"m",,terminal_output +1136,1533492,"TERMINAL",0,0,"odels/",,terminal_output +1137,1534008,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +1138,1534124,"TERMINAL",0,0,"]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1139,1534761,"TERMINAL",0,0,"g",,terminal_output +1140,1534813,"TERMINAL",0,0,"i",,terminal_output +1141,1534880,"TERMINAL",0,0,"t",,terminal_output +1142,1534986,"TERMINAL",0,0," ",,terminal_output +1143,1535120,"TERMINAL",0,0,"c",,terminal_output +1144,1535175,"TERMINAL",0,0,"o",,terminal_output +1145,1535350,"TERMINAL",0,0,"m",,terminal_output +1146,1535497,"TERMINAL",0,0,"m",,terminal_output +1147,1535567,"TERMINAL",0,0,"i",,terminal_output +1148,1535679,"TERMINAL",0,0,"t",,terminal_output +1149,1535743,"TERMINAL",0,0," ",,terminal_output +1150,1536159,"TERMINAL",0,0,"-",,terminal_output +1151,1536320,"TERMINAL",0,0,"a",,terminal_output +1152,1536381,"TERMINAL",0,0,"m",,terminal_output +1153,1536447,"TERMINAL",0,0," ",,terminal_output +1154,1536670,"TERMINAL",0,0,"""",,terminal_output +1155,1536959,"TERMINAL",0,0,"m",,terminal_output +1156,1537031,"TERMINAL",0,0,"a",,terminal_output +1157,1537211,"TERMINAL",0,0,"d",,terminal_output +1158,1537274,"TERMINAL",0,0,"e",,terminal_output +1159,1537415,"TERMINAL",0,0," ",,terminal_output +1160,1538092,"TERMINAL",0,0,"n",,terminal_output +1161,1538245,"TERMINAL",0,0,"o",,terminal_output +1162,1538312,"TERMINAL",0,0,"i",,terminal_output +1163,1538415,"TERMINAL",0,0,"s",,terminal_output +1164,1538548,"TERMINAL",0,0,"e",,terminal_output +1165,1538611,"TERMINAL",0,0," ",,terminal_output +1166,1538778,"TERMINAL",0,0,"a",,terminal_output +1167,1538890,"TERMINAL",0,0,"u",,terminal_output +1168,1539230,"TERMINAL",0,0,"g",,terminal_output +1169,1539423,"TERMINAL",0,0,"m",,terminal_output +1170,1539502,"TERMINAL",0,0,"e",,terminal_output +1171,1539605,"TERMINAL",0,0,"n",,terminal_output +1172,1539666,"TERMINAL",0,0,"t",,terminal_output +1173,1540199,"TERMINAL",0,0,"a",,terminal_output +1174,1540409,"TERMINAL",0,0,"t",,terminal_output +1175,1540474,"TERMINAL",0,0,"i",,terminal_output +1176,1540536,"TERMINAL",0,0,"o",,terminal_output +1177,1540699,"TERMINAL",0,0,"n",,terminal_output +1178,1540771,"TERMINAL",0,0," ",,terminal_output +1179,1540915,"TERMINAL",0,0,"p",,terminal_output +1180,1540975,"TERMINAL",0,0,"r",,terminal_output +1181,1541104,"TERMINAL",0,0,"e",,terminal_output +1182,1541540,"TERMINAL",0,0,"",,terminal_output +1183,1541693,"TERMINAL",0,0,"",,terminal_output +1184,1541814,"TERMINAL",0,0,"",,terminal_output +1185,1541887,"TERMINAL",0,0,"s",,terminal_output +1186,1542100,"TERMINAL",0,0,"a",,terminal_output +1187,1542235,"TERMINAL",0,0,"f",,terminal_output +1188,1542356,"TERMINAL",0,0,"e",,terminal_output +1189,1542419,"TERMINAL",0,0,"r \r",,terminal_output +1190,1545860,"TERMINAL",0,0,"""",,terminal_output +1191,1547202,"TERMINAL",0,0,"\r",,terminal_output +1192,1547485,"TERMINAL",0,0," """,,terminal_output +1193,1547622,"TERMINAL",0,0,"f""",,terminal_output +1194,1547684,"TERMINAL",0,0,"o""",,terminal_output +1195,1547846,"TERMINAL",0,0,"r""",,terminal_output +1196,1548259,"TERMINAL",0,0," """,,terminal_output +1197,1548489,"TERMINAL",0,0,"m""",,terminal_output +1198,1548693,"TERMINAL",0,0,"i""",,terminal_output +1199,1548903,"TERMINAL",0,0,"x""",,terminal_output +1200,1550026,"TERMINAL",0,0,"""",,terminal_output +1201,1550291,"TERMINAL",0,0,"\r""",,terminal_output +1202,1550984,"TERMINAL",0,0,"\r""\r",,terminal_output +1203,1551869,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +1204,1553417,"TERMINAL",0,0,"black....................................................................",,terminal_output +1205,1554656,"TERMINAL",0,0,"Failed\r\n- hook id: black\r\n- files were modified by this hook\r\n\r\nreformatted jasmine/models/dynamics.py\r\n\r\nAll done! ✨ 🍰 ✨\r\n1 file reformatted.\r\n\r\n]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1206,1555542,"TERMINAL",0,0,"git commit -am ""made noise augmentation safer""",,terminal_output +1207,1555762,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +1208,1556489,"TERMINAL",0,0,"black....................................................................",,terminal_output +1209,1556602,"TERMINAL",0,0,"Passed\r\n",,terminal_output +1210,1557010,"TERMINAL",0,0,"[add-noise-to-combat-exposure-bias 57dee33] made noise augmentation safer\r\n 1 file changed, 11 insertions(+), 7 deletions(-)\r\n]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1211,1559277,"jasmine/models/dynamics.py",0,0,"",python,tab +1212,1559394,"jasmine/models/dynamics.py",3455,607," # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n",python,content +1213,1563699,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1214,1565249,"jasmine/models/dynamics.py",0,0,"",python,tab +1215,1580171,"jasmine/models/dynamics.py",4679,0,"",python,selection_mouse +1216,1580279,"jasmine/models/dynamics.py",4666,22,"noise_level_embed_BT1M",python,selection_mouse +1217,1619264,"jasmine/models/dynamics.py",3765,0,"",python,selection_command +1218,1625454,"jasmine/models/dynamics.py",3347,0,"",python,selection_mouse +1219,1626134,"jasmine/models/dynamics.py",3383,0,"",python,selection_mouse +1220,1626147,"jasmine/models/dynamics.py",3382,0,"",python,selection_command +1221,1635128,"jasmine/models/dynamics.py",3641,0,"",python,selection_mouse +1222,1762835,"jasmine/models/dynamics.py",3796,0,"",python,selection_mouse +1223,1763465,"jasmine/models/dynamics.py",3761,0,"",python,selection_mouse +1224,1764178,"jasmine/models/dynamics.py",3641,0,"",python,selection_mouse +1225,1764946,"jasmine/models/dynamics.py",3754,0,"",python,selection_mouse +1226,1769955,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1227,1770293,"jasmine/models/dynamics.py",3856,0,"",python,selection_mouse +1228,1770925,"jasmine/models/dynamics.py",3765,0,"",python,selection_mouse +1229,1770938,"jasmine/models/dynamics.py",3764,0,"",python,selection_command +1230,1772750,"jasmine/models/dynamics.py",3361,0,"",python,selection_mouse +1231,1876478,"jasmine/models/dynamics.py",3765,0,"",python,selection_mouse +1232,1876483,"jasmine/models/dynamics.py",3764,0,"",python,selection_command +1233,1877313,"jasmine/models/dynamics.py",3765,0,"\n ",python,content +1234,1878417,"jasmine/models/dynamics.py",3770,4,"",python,content +1235,1878611,"jasmine/models/dynamics.py",3766,4,"",python,content +1236,1878961,"jasmine/models/dynamics.py",3765,1,"",python,content +1237,1908981,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",0,0,"",shellscript,tab +1238,1910889,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",0,0,"",shellscript,tab +1239,1913716,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",0,0,"",shellscript,tab +1240,1917112,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2047,0,"",shellscript,selection_mouse +1241,1917853,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2048,0,"",shellscript,selection_command +1242,1918617,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2047,1,"",shellscript,content +1243,1918718,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2046,1,"",shellscript,content +1244,1918949,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2045,1,"",shellscript,content +1245,1919059,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2045,0,"1",shellscript,content +1246,1919060,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2046,0,"",shellscript,selection_keyboard +1247,1919169,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2046,0,"0",shellscript,content +1248,1919170,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2047,0,"",shellscript,selection_keyboard +1249,1919295,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2047,0,"0",shellscript,content +1250,1919296,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2048,0,"",shellscript,selection_keyboard +1251,1919391,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2048,0,"0",shellscript,content +1252,1919392,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",2049,0,"",shellscript,selection_keyboard +1253,1921125,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",0,0,"",shellscript,tab +1254,1922649,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",2059,0,"",shellscript,selection_mouse +1255,1923515,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",2058,1,"",shellscript,content +1256,1923619,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",2057,1,"",shellscript,content +1257,1923766,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",2056,1,"",shellscript,content +1258,1923902,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",2056,0,"1",shellscript,content +1259,1923903,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",2057,0,"",shellscript,selection_keyboard +1260,1923995,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",2057,0,"0",shellscript,content +1261,1923996,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",2058,0,"",shellscript,selection_keyboard +1262,1924128,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",2058,0,"0",shellscript,content +1263,1924129,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",2059,0,"",shellscript,selection_keyboard +1264,1924230,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",2059,0,"0",shellscript,content +1265,1924231,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",2060,0,"",shellscript,selection_keyboard +1266,1926430,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_smaller_lr_50k.sh",0,0,"",shellscript,tab +1267,1927164,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",0,0,"",shellscript,tab +1268,1929475,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_smaller_lr_50k.sh",0,0,"",shellscript,tab +1269,1931692,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +1270,1933640,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2399,0,"",shellscript,selection_mouse +1271,1934510,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2398,1,"",shellscript,content +1272,1934652,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2397,1,"",shellscript,content +1273,1934783,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2396,1,"",shellscript,content +1274,1935005,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2396,0,"1",shellscript,content +1275,1935006,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2397,0,"",shellscript,selection_keyboard +1276,1935200,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2397,0,"0",shellscript,content +1277,1935201,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2398,0,"",shellscript,selection_keyboard +1278,1935344,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2398,0,"0",shellscript,content +1279,1935345,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2399,0,"",shellscript,selection_keyboard +1280,1935428,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2399,0,"0",shellscript,content +1281,1935429,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2400,0,"",shellscript,selection_keyboard +1282,1936786,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",2427,0,"",shellscript,selection_mouse +1283,1984911,"jasmine/models/dynamics.py",0,0,"",python,tab +1284,2005896,"jasmine/models/dynamics.py",3372,0,"",python,selection_mouse +1285,2006915,"jasmine/models/dynamics.py",3373,0,"",python,selection_command +1286,3449504,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1287,3550065,"TERMINAL",0,0,"c",,terminal_output +1288,3550693,"TERMINAL",0,0,"",,terminal_output +1289,3550930,"TERMINAL",0,0,"q",,terminal_output +1290,3550991,"TERMINAL",0,0,"u",,terminal_output +1291,3551103,"TERMINAL",0,0,"e",,terminal_output +1292,3551203,"TERMINAL",0,0,"u",,terminal_output +1293,3551265,"TERMINAL",0,0,"e",,terminal_output +1294,3551547,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: squeue --mehkn0702.localdomain: Sun Sep 28 14:26:36 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3527671 accelerat train_dy tum_cte0 R34:18\t 1 hkn07023527672 accelerat train_dy tum_cte0 R34:18\t 1 hkn07023527674 accelerat train_dy tum_cte0 R34:18\t 1 hkn05343527676 accelerat train_dy tum_cte0 R34:18\t 1 hkn08073527669 accelerat interact tum_cte0 R38:54\t 1 hkn0702",,terminal_output +1295,3552586,"TERMINAL",0,0,"799995",,terminal_output +1296,3553611,"TERMINAL",0,0,"8202020206",,terminal_output +1297,3553835,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1298,3554437,"TERMINAL",0,0,"c",,terminal_output +1299,3554619,"TERMINAL",0,0,"le",,terminal_output +1300,3554822,"TERMINAL",0,0,"a",,terminal_output +1301,3555059,"TERMINAL",0,0,"r\r\n[?2004l\r]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +1302,3711861,"jasmine/genie.py",0,0,"from typing import Dict\n\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\nimport orbax.checkpoint as ocp\n\nfrom models.dynamics import DynamicsMaskGIT, DynamicsCausal\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\n\nclass Genie(nnx.Module):\n """"""Genie model""""""\n\n def __init__(\n self,\n in_dim: int,\n tokenizer_dim: int,\n tokenizer_ffn_dim: int,\n latent_patch_dim: int,\n num_patch_latents: int,\n patch_size: int,\n tokenizer_num_blocks: int,\n tokenizer_num_heads: int,\n lam_dim: int,\n lam_ffn_dim: int,\n latent_action_dim: int,\n num_actions: int,\n lam_patch_size: int,\n lam_num_blocks: int,\n lam_num_heads: int,\n lam_co_train: bool,\n use_gt_actions: bool,\n dyna_type: str,\n dyna_dim: int,\n dyna_ffn_dim: int,\n dyna_num_blocks: int,\n dyna_num_heads: int,\n max_noise_level: float,\n noise_buckets: int,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n dropout: float = 0.0,\n mask_limit: float = 0.0,\n ):\n # --- Tokenizer ---\n self.in_dim = in_dim\n self.tokenizer_dim = tokenizer_dim\n self.tokenizer_ffn_dim = tokenizer_ffn_dim\n self.latent_patch_dim = latent_patch_dim\n self.num_patch_latents = num_patch_latents\n self.patch_size = patch_size\n self.tokenizer_num_blocks = tokenizer_num_blocks\n self.tokenizer_num_heads = tokenizer_num_heads\n # --- LAM ---\n self.lam_dim = lam_dim\n self.lam_ffn_dim = lam_ffn_dim\n self.latent_action_dim = latent_action_dim\n self.num_actions = num_actions\n self.lam_patch_size = lam_patch_size\n self.lam_num_blocks = lam_num_blocks\n self.lam_num_heads = lam_num_heads\n self.lam_co_train = lam_co_train\n self.use_gt_actions = use_gt_actions\n # --- Dynamics ---\n self.dyna_type = dyna_type\n self.dyna_dim = dyna_dim\n self.dyna_ffn_dim = dyna_ffn_dim\n self.dyna_num_blocks = dyna_num_blocks\n self.dyna_num_heads = dyna_num_heads\n self.max_noise_level = max_noise_level\n self.noise_buckets = noise_buckets\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.dropout = dropout\n self.mask_limit = mask_limit\n self.decode = decode\n\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n ffn_dim=self.tokenizer_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n if self.use_gt_actions:\n self.action_embed = nnx.Embed(\n self.num_actions, self.latent_action_dim, rngs=rngs\n )\n self.lam = None\n else:\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n ffn_dim=self.lam_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.action_embed = None\n if self.dyna_type == ""maskgit"":\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n max_noise_level=self.max_noise_level,\n noise_buckets=self.noise_buckets,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n elif self.dyna_type == ""causal"":\n self.dynamics = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n max_noise_level=self.max_noise_level,\n noise_buckets=self.noise_buckets,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n rngs=rngs,\n )\n else:\n raise ValueError(f""Invalid dynamics type: {self.dyna_type}"")\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n ) -> Dict[str, jax.Array]:\n videos_BTHWC = batch[""videos""]\n tokenizer_outputs = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_indices_BTN = tokenizer_outputs[""indices""]\n latent_actions_BTm11L = None\n action_embeddings_BTm11L = None\n if self.use_gt_actions:\n assert self.action_embed is not None\n action_indices_E = None\n action_embeddings_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n action_embeddings_BTm11L = action_embeddings_BT1L[:, :-1]\n else:\n assert self.lam is not None\n lam_outputs = self.lam.vq_encode(videos_BTHWC, training=False)\n z_q_BTm11L = lam_outputs[""z_q""]\n action_indices_E = lam_outputs[""indices""]\n latent_actions_BTm11L = jax.lax.cond(\n self.lam_co_train,\n lambda: z_q_BTm11L,\n lambda: jax.lax.stop_gradient(z_q_BTm11L),\n )\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(token_indices_BTN),\n latent_actions=(\n action_embeddings_BTm11L\n if self.use_gt_actions\n else latent_actions_BTm11L\n ),\n )\n outputs[""mask_rng""] = batch[""rng""]\n dyna_logits_BTNV, dyna_mask = self.dynamics(outputs)\n outputs[""token_logits""] = dyna_logits_BTNV\n outputs[""mask""] = dyna_mask\n mle_indices_BTN = jnp.argmax(outputs[""token_logits""], axis=-1)\n H, W = batch[""videos""].shape[2:4]\n outputs[""recon""] = self.tokenizer.decode(mle_indices_BTN, (H, W))\n if action_indices_E is not None:\n outputs[""lam_indices""] = action_indices_E\n return outputs\n\n def sample(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n maskgit_steps: int = 25,\n ) -> tuple[jax.Array, jax.Array]:\n if self.dyna_type == ""maskgit"":\n return self.sample_maskgit(\n batch, seq_len, 0.0, maskgit_steps, temperature, sample_argmax\n )\n elif self.dyna_type == ""causal"":\n return self.sample_causal(batch, seq_len, temperature, sample_argmax)\n else:\n raise ValueError(f""Dynamics model type unknown: {self.dyna_type}"")\n\n def sample_maskgit(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n noise_level: float,\n steps: int = 25,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n P: S * N\n """"""\n assert isinstance(self.dynamics, DynamicsMaskGIT)\n assert (\n noise_level < self.max_noise_level\n ), ""Noise level must me smaller than max_noise_level.""\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n init_logits_BSNV = jnp.zeros(\n shape=(*token_idxs_BSN.shape, self.num_patch_latents)\n )\n noise_level = jnp.array(noise_level)\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, :-1]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n # --- Extract submodule state ---\n dynamics_state = nnx.state(self.dynamics)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def maskgit_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, mask_BSN, action_tokens_EL = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_maskgit = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n max_noise_level=self.max_noise_level,\n noise_buckets=self.noise_buckets,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_maskgit, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_maskgit.patch_embed(token_idxs_BSN)\n mask_token_111M = dynamics_maskgit.mask_token.value\n mask_expanded_BSN1 = mask_BSN[..., None]\n vid_embed_BSNM = jnp.where(\n mask_expanded_BSN1, mask_token_111M, vid_embed_BSNM\n )\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_maskgit.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n # TODO mihir\n\n rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 / self.max_noise_level) * self.noise_buckets\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n vid_embed_BSNM += jnp.expand_dims(noise_level_B11, -1)\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (steps * 2))\n step_temp = temperature * (1.0 - unmasked_ratio)\n final_logits_BSNp2V = (\n dynamics_maskgit.transformer(vid_embed_BSNp2M) / step_temp\n )\n final_logits_BSNV = final_logits_BSNp2V[:, :, 2:]\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_BSN = jnp.argmax(final_logits_BSNV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_BSN = jax.random.categorical(_rng, final_logits_BSNV)\n gather_fn = jax.vmap(jax.vmap(jax.vmap(lambda x, y: x[y])))\n final_token_probs_BSN = gather_fn(\n jax.nn.softmax(final_logits_BSNV), sampled_token_idxs_BSN\n )\n final_token_probs_BSN += ~mask_BSN\n # Update masked tokens and logits only\n token_idxs_BSN = jnp.where(mask_BSN, sampled_token_idxs_BSN, token_idxs_BSN)\n logits_BSNV = jnp.where(\n jnp.expand_dims(mask_BSN, -1), final_logits_BSNV, logits_BSNV\n )\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n final_token_probs_flat_BP = einops.rearrange(\n final_token_probs_BSN, ""b s n -> b (s n)""\n )\n idx_mask_P = (\n jnp.arange(final_token_probs_flat_BP.shape[-1])\n <= N - num_unmasked_tokens\n )\n sorted_idxs_BP = jnp.argsort(final_token_probs_flat_BP, axis=-1)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask_P))\n mask_flat_BP = einops.rearrange(mask_BSN, ""b s n -> b (s n)"")\n new_mask_flat_BP = mask_update_fn(mask_flat_BP, sorted_idxs_BP)\n new_mask_BSN = einops.rearrange(new_mask_flat_BP, ""b (s n) -> b s n"", n=N)\n\n new_carry = (\n rng,\n token_idxs_BSN,\n logits_BSNV,\n new_mask_BSN,\n action_tokens_EL,\n )\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # Mask current frame (i.e., t == step_t)\n mask_S = jnp.arange(seq_len) == step_t\n mask_BSN = jnp.broadcast_to(mask_S[None, :, None], (B, seq_len, N)).astype(\n bool\n )\n masked_token_idxs_BSN = current_token_idxs_BSN * ~mask_BSN\n masked_logits_BSNV = current_logits_BSNV * jnp.expand_dims(~mask_BSN, -1)\n\n # --- Initialize and run MaskGIT loop ---\n init_carry_maskgit = (\n step_rng,\n masked_token_idxs_BSN,\n masked_logits_BSNV,\n mask_BSN,\n action_tokens_EL,\n )\n final_carry_maskgit = maskgit_step_fn(init_carry_maskgit, jnp.arange(steps))\n updated_token_idxs_BSN = final_carry_maskgit[1]\n updated_logits_BSNV = final_carry_maskgit[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, init_logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def sample_causal(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n assert isinstance(self.dynamics, DynamicsCausal)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n logits_BSNV = jnp.zeros((*token_idxs_BSN.shape, self.num_patch_latents))\n dynamics_state = nnx.state(self.dynamics)\n\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, :-1]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def causal_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step_n: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_causal = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=self.decode,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_causal, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_causal.patch_embed(token_idxs_BSN)\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_causal.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n vid_embed_BSNp1M = jnp.concatenate([act_embed_BS1M, vid_embed_BSNM], axis=2)\n final_logits_BTNp1V = (\n dynamics_causal.transformer(vid_embed_BSNp1M, (step_t, step_n))\n / temperature\n )\n final_logits_BV = final_logits_BTNp1V[:, step_t, step_n, :]\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_B = jnp.argmax(final_logits_BV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_B = jax.random.categorical(_rng, final_logits_BV)\n # Update next tokens only\n token_idxs_BSN = token_idxs_BSN.at[:, step_t, step_n].set(\n sampled_token_idxs_B\n )\n logits_BSNV = logits_BSNV.at[:, step_t, step_n].set(final_logits_BV)\n\n new_carry = (rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t)\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # --- Initialize and run causal loop ---\n init_carry_causal = (\n step_rng,\n current_token_idxs_BSN,\n current_logits_BSNV,\n action_tokens_EL,\n step_t,\n )\n final_carry_causal = causal_step_fn(init_carry_causal, jnp.arange(N))\n updated_token_idxs_BSN = final_carry_causal[1]\n updated_logits_BSNV = final_carry_causal[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def vq_encode(self, batch: Dict[str, jax.Array], training: bool) -> jax.Array:\n # --- Preprocess videos ---\n assert self.lam is not None\n video_BTHWC = batch[""videos""]\n lam_output: Dict[str, jax.Array] = self.lam.vq_encode(\n video_BTHWC, training=training\n )\n lam_indices_E = lam_output[""indices""]\n return lam_indices_E\n\n\n# FIXME (f.srambical): add conversion script for old checkpoints\ndef restore_genie_components(\n optimizer: nnx.ModelAndOptimizer,\n sharding: jax.sharding.NamedSharding,\n rng: jax.Array,\n args,\n) -> nnx.ModelAndOptimizer:\n """"""Restore pre-trained Genie components""""""\n rng_tokenizer, rng_lam = jax.random.split(rng)\n rngs_tokenizer = nnx.Rngs(rng_tokenizer)\n rngs_lam = nnx.Rngs(rng_lam)\n\n tx = optimizer.tx\n model = optimizer.model\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n tokenizer_checkpoint_manager = ocp.CheckpointManager(\n directory=args.tokenizer_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n ffn_dim=args.tokenizer_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_tokenizer,\n )\n dummy_tokenizer_optimizer = nnx.ModelAndOptimizer(dummy_tokenizer, tx)\n dummy_tokenizer_optimizer_state = nnx.state(dummy_tokenizer_optimizer)\n abstract_sharded_tokenizer_optimizer_state = _create_abstract_sharded_pytree(\n dummy_tokenizer_optimizer_state, sharding\n )\n restored_tokenizer = tokenizer_checkpoint_manager.restore(\n step=tokenizer_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_tokenizer_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_tokenizer_optimizer.model, restored_tokenizer.model)\n model.tokenizer = dummy_tokenizer_optimizer.model\n tokenizer_checkpoint_manager.close()\n\n if args.lam_checkpoint:\n lam_checkpoint_manager = ocp.CheckpointManager(\n directory=args.lam_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n ffn_dim=args.lam_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_lam,\n )\n dummy_lam_optimizer = nnx.ModelAndOptimizer(dummy_lam, tx)\n dummy_lam_optimizer_state = nnx.state(dummy_lam_optimizer)\n abstract_sharded_lam_optimizer_state = _create_abstract_sharded_pytree(\n dummy_lam_optimizer_state, sharding\n )\n restored_lam_optimizer = lam_checkpoint_manager.restore(\n step=lam_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_lam_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_lam_optimizer.model, restored_lam_optimizer.model)\n model.lam = dummy_lam_optimizer.model\n # Remove the LAM decoder to save memory and avoid unnecessary computation.\n del model.lam.decoder\n lam_checkpoint_manager.close()\n\n # Reinitialize the optimizer states\n optimizer = nnx.ModelAndOptimizer(model, tx)\n return optimizer\n\n\ndef _create_abstract_sharded_pytree(\n pytree_template: nnx.GraphState, sharding_spec: jax.sharding.NamedSharding\n) -> jax.Array:\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)\n",python,tab +1303,3716643,"jasmine/genie.py",9621,0,"",python,selection_mouse +1304,3716644,"jasmine/genie.py",9620,0,"",python,selection_command +1305,3821596,"jasmine/genie.py",9125,0,"",python,selection_mouse +1306,3822263,"jasmine/genie.py",9302,0,"",python,selection_mouse +1307,3822272,"jasmine/genie.py",9301,0,"",python,selection_command +1308,3822838,"jasmine/genie.py",9436,0,"",python,selection_mouse +1309,3822839,"jasmine/genie.py",9435,0,"",python,selection_command +1310,3827755,"jasmine/genie.py",0,0,"",python,tab +1311,3855214,"jasmine/genie.py",7628,0,"",python,selection_mouse +1312,3855228,"jasmine/genie.py",7627,0,"",python,selection_command +1313,3871471,"jasmine/train_dynamics.py",0,0,"import os\n\n\nos.environ.setdefault(""XLA_PYTHON_CLIENT_MEM_FRACTION"", ""0.98"")\n\nfrom dataclasses import dataclass, field\nimport itertools\nfrom typing import cast, Optional\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\nfrom utils.train_utils import (\n get_lr_schedule,\n count_parameters_by_component,\n print_mem_stats,\n print_compiled_memory_stats,\n print_compiled_cost_analysis,\n)\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n # Tokenizer\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_type: str = ""maskgit"" # supported options: maskgit, causal\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n max_noise_level: float = 0.7\n noise_buckets: int = 10\n dropout: float = 0.0\n mask_limit: float = 0.5\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n use_gt_actions: bool = False\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n eval_full_frame: bool = False\n val_maskgit_steps: int = 25\n val_temperature: float = 1\n val_sample_argmax: bool = False\n wandb_id: str = """"\n\n\ndef build_model(args: Args, rng: jax.Array) -> tuple[Genie, jax.Array]:\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_actions=args.num_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n use_gt_actions=args.use_gt_actions,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n max_noise_level=args.max_noise_level,\n noise_buckets=args.noise_buckets,\n mask_limit=args.mask_limit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n decode=False,\n rngs=rngs,\n )\n if args.use_gt_actions:\n assert (\n not args.lam_checkpoint\n ), ""Cannot use LAM when using ground-truth actions.""\n else:\n assert genie.lam is not None\n del genie.lam.decoder\n return genie, rng\n\n\ndef build_optimizer(genie: Genie, args: Args) -> nnx.ModelAndOptimizer:\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.param_dtype, # moments in full precision\n )\n optimizer = nnx.ModelAndOptimizer(genie, tx)\n return optimizer\n\n\ndef build_mesh_and_sharding(\n num_devices: int,\n) -> tuple[Mesh, NamedSharding, NamedSharding, NamedSharding]:\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n actions_sharding = NamedSharding(mesh, PartitionSpec(""data"", None))\n return mesh, replicated_sharding, videos_sharding, actions_sharding\n\n\ndef shard_optimizer_states(\n optimizer: nnx.ModelAndOptimizer, replicated_sharding: NamedSharding\n) -> None:\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n\ndef build_dataloader(args: Args, data_dir: str) -> grain.DataLoaderIterator:\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(data_dir, x)\n for x in os.listdir(data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n return grain_iterator\n\n\ndef build_checkpoint_manager(args: Args) -> Optional[ocp.CheckpointManager]:\n if args.restore_ckpt or args.save_ckpt:\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(\n ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler\n ),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(\n ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler\n ),\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n return checkpoint_manager\n else:\n return None\n\n\ndef restore_or_initialize_components(\n args: Args,\n checkpoint_manager: Optional[ocp.CheckpointManager],\n optimizer: nnx.ModelAndOptimizer,\n train_iterator: grain.DataLoaderIterator,\n rng: jax.Array,\n replicated_sharding: NamedSharding,\n val_iterator: Optional[grain.DataLoaderIterator],\n restore_step: Optional[int] = None,\n) -> tuple[\n int,\n nnx.ModelAndOptimizer,\n grain.DataLoaderIterator,\n grain.DataLoaderIterator,\n jax.Array,\n]:\n step = 0\n if checkpoint_manager and restore_step is None:\n restore_step = checkpoint_manager.latest_step()\n if args.restore_ckpt:\n assert checkpoint_manager is not None\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n if val_iterator:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n val_dataloader_state=grain.checkpoint.CheckpointRestore(val_iterator), # type: ignore\n )\n else:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n )\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(), args=restore_args\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n train_iterator = restored[""train_dataloader_state""]\n if val_iterator:\n val_iterator = restored[""val_dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n rng, _rng = jax.random.split(rng)\n optimizer = restore_genie_components(optimizer, replicated_sharding, _rng, args)\n return step, optimizer, train_iterator, val_iterator, rng\n\n\ndef _calculate_step_metrics(\n outputs: dict[str, jax.Array],\n gt: jax.Array,\n num_actions: int,\n num_patch_latents: int,\n) -> tuple[jax.Array, dict]:\n mask = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n gt_val = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt_val, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt_val, recon)).mean()\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]),\n size=num_patch_latents,\n fill_value=0,\n )\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n psnr=psnr,\n ssim=ssim,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n if ""lam_indices"" in outputs.keys():\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]),\n size=num_actions,\n fill_value=0,\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n metrics[""codebook_usage_lam""] = codebook_usage_lam\n return ce_loss, metrics\n\n\ndef main(args: Args) -> None:\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n genie, rng = build_model(args, rng)\n _, params, _ = nnx.split(genie, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n optimizer = build_optimizer(genie, args)\n del genie\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n _, replicated_sharding, videos_sharding, actions_sharding = build_mesh_and_sharding(\n num_devices\n )\n\n shard_optimizer_states(optimizer, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n checkpoint_manager = build_checkpoint_manager(args)\n\n # --- Create DataLoaderIterator from dataloader ---\n train_iterator = build_dataloader(args, args.data_dir)\n val_iterator = None\n if args.val_data_dir:\n val_iterator = build_dataloader(args, args.val_data_dir)\n\n # --- Restore checkpoint ---\n step, optimizer, train_iterator, val_iterator, rng = (\n restore_or_initialize_components(\n args,\n checkpoint_manager,\n optimizer,\n train_iterator,\n rng,\n replicated_sharding,\n val_iterator,\n )\n )\n\n # --- Define loss and train step (close over args) ---\n def dynamics_loss_fn(\n model: Genie,\n inputs: dict,\n ) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs)\n ce_loss, metrics = _calculate_step_metrics(\n outputs, gt, args.num_actions, args.num_patch_latents\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n @nnx.jit(donate_argnums=0)\n def train_step(\n optimizer: nnx.ModelAndOptimizer, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(model: Genie) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n model.train()\n return dynamics_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n optimizer.model\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return loss, recon, metrics\n\n @nnx.jit\n def val_step(genie: Genie, inputs: dict) -> dict:\n """"""Evaluate model and compute metrics""""""\n genie.eval()\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n (loss, (recon, metrics)) = dynamics_loss_fn(genie, inputs)\n val_output = {""loss"": loss, ""recon"": recon, ""metrics"": metrics}\n\n # --- Evaluate full frame prediction (sampling) ---\n if args.eval_full_frame:\n inputs[""videos""] = gt.astype(args.dtype)\n tokenizer_outputs = genie.tokenizer.vq_encode(\n inputs[""videos""], training=False\n )\n tokens_full_frame = tokenizer_outputs[""indices""]\n lam_indices_E = None\n if not args.use_gt_actions:\n lam_indices_E = genie.vq_encode(inputs, training=False)\n inputs[""latent_actions""] = lam_indices_E\n inputs[""videos""] = inputs[""videos""][\n :, :-1\n ] # remove last frame for generation\n recon_full_frame, logits_full_frame = genie.sample(\n inputs,\n args.seq_len,\n args.val_temperature,\n args.val_sample_argmax,\n args.val_maskgit_steps,\n )\n # Calculate metrics for the last frame only\n step_outputs = {\n ""recon"": recon_full_frame[:, -1],\n ""token_logits"": logits_full_frame[:, -1],\n ""video_tokens"": tokens_full_frame[:, -1],\n ""mask"": jnp.ones_like(tokens_full_frame[:, -1]),\n }\n if lam_indices_E is not None:\n lam_indices_B = lam_indices_E.reshape((-1, args.seq_len - 1))[:, -1]\n step_outputs[""lam_indices""] = lam_indices_B\n\n loss_full_frame, metrics_full_frame = _calculate_step_metrics(\n step_outputs, gt[:, -1], args.num_actions, args.num_patch_latents\n )\n val_output.update(\n {\n ""loss_full_frame"": loss_full_frame,\n ""recon_full_frame"": recon_full_frame,\n ""metrics_full_frame"": metrics_full_frame,\n }\n )\n return val_output\n\n def calculate_validation_metrics(val_dataloader, genie, rng):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n loss_full_frame_per_step = []\n metrics_full_frame_per_step = []\n batch = None\n recon = None\n recon_full_frame = None\n for batch in val_dataloader:\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n val_outputs = val_step(genie, batch)\n loss_per_step.append(val_outputs[""loss""])\n metrics_per_step.append(val_outputs[""metrics""])\n recon = val_outputs[""recon""]\n if args.eval_full_frame:\n loss_full_frame_per_step.append(val_outputs[""loss_full_frame""])\n metrics_full_frame_per_step.append(val_outputs[""metrics_full_frame""])\n recon_full_frame = val_outputs[""recon_full_frame""]\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(\n f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}""\n )\n\n val_metrics = {\n f""val_{key}"": np.mean([float(m[key]) for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n val_metrics[""val_loss""] = np.mean(loss_per_step)\n if args.eval_full_frame:\n val_metrics_full_frame = {\n f""val_full_frame_{key}"": np.mean(\n [float(m[key]) for m in metrics_full_frame_per_step]\n )\n for key in metrics_full_frame_per_step[0].keys()\n }\n val_metrics.update(val_metrics_full_frame)\n val_metrics[""val_full_frame_loss""] = np.mean(loss_full_frame_per_step)\n return val_metrics, batch, recon, recon_full_frame\n\n # --- TRAIN LOOP ---\n dataloader_train = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in train_iterator\n )\n dataloader_val = None\n if val_iterator:\n dataloader_val = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in val_iterator\n )\n if jax.process_index() == 0:\n first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader_train = itertools.chain([first_batch], dataloader_train)\n print(f""Starting training from step {step}..."")\n first_step = step\n while step < args.num_steps:\n for batch in dataloader_train:\n # --- Train step ---\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n loss, recon, metrics = train_step(optimizer, batch)\n if step == first_step:\n print_mem_stats(""After params initialized"")\n step += 1\n\n # --- Validation loss ---\n val_results = {}\n if dataloader_val and step % args.val_interval == 0:\n rng, _rng_mask_val = jax.random.split(rng, 2)\n print(""Calculating validation metrics..."")\n val_metrics, val_gt_batch, val_recon, val_recon_full_frame = (\n calculate_validation_metrics(\n dataloader_val, optimizer.model, _rng_mask_val\n )\n )\n print(f""Step {step}, validation loss: {val_metrics['val_loss']}"")\n val_results = {\n ""metrics"": val_metrics,\n ""gt_batch"": val_gt_batch,\n ""recon"": val_recon,\n ""full_frame"": val_recon_full_frame,\n }\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n log_dict = {""loss"": loss, ""step"": step, **metrics}\n if val_results:\n log_dict.update(val_results[""metrics""])\n wandb.log(log_dict)\n if step % args.log_image_interval == 0:\n gt_seq = batch[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if val_results:\n val_results[""gt_seq_val""] = (\n val_results[""gt_batch""][""videos""][0].astype(jnp.float32)\n / 255.0\n )\n val_results[""recon_seq_val""] = val_results[""recon""][0].clip(\n 0, 1\n )\n val_comparison_seq = jnp.concatenate(\n (val_results[""gt_seq_val""], val_results[""recon_seq_val""]),\n axis=1,\n )\n val_results[""val_comparison_seq""] = einops.rearrange(\n val_comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if args.eval_full_frame:\n val_results[""full_frame_seq_val""] = val_results[\n ""full_frame""\n ][0].clip(0, 1)\n val_results[""val_full_frame_comparison_seq""] = (\n jnp.concatenate(\n (\n val_results[""gt_seq_val""],\n val_results[""full_frame_seq_val""],\n ),\n axis=1,\n )\n )\n val_results[""val_full_frame_comparison_seq""] = (\n einops.rearrange(\n val_results[""val_full_frame_comparison_seq""] * 255,\n ""t h w c -> h (t w) c"",\n )\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n if val_results:\n log_images.update(\n dict(\n val_image=wandb.Image(\n np.asarray(\n val_results[""gt_seq_val""][args.seq_len - 1]\n )\n ),\n val_recon=wandb.Image(\n np.asarray(\n val_results[""recon_seq_val""][\n args.seq_len - 1\n ]\n )\n ),\n val_true_vs_recon=wandb.Image(\n np.asarray(\n val_results[""val_comparison_seq""].astype(\n np.uint8\n )\n )\n ),\n )\n )\n if args.eval_full_frame:\n log_images.update(\n dict(\n val_full_frame=wandb.Image(\n np.asarray(\n val_results[""full_frame_seq_val""][\n args.seq_len - 1\n ]\n )\n ),\n val_true_vs_full_frame=wandb.Image(\n np.asarray(\n val_results[\n ""val_full_frame_comparison_seq""\n ].astype(np.uint8)\n )\n ),\n )\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n assert checkpoint_manager is not None\n optimizer_state = nnx.state(optimizer)\n if val_iterator:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n val_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n val_iterator # type: ignore\n ),\n )\n else:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n )\n checkpoint_manager.save(step, args=ckpt_manager_args)\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n if checkpoint_manager:\n checkpoint_manager.close()\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n main(args)\n",python,tab +1314,3873419,"jasmine/train_dynamics.py",0,0,"",python,tab +1315,3875537,"jasmine/train_dynamics.py",11,0,"",python,selection_mouse +1316,3954520,"jasmine/train_dynamics.py",17617,0,"",python,selection_mouse +1317,3954535,"jasmine/train_dynamics.py",17616,0,"",python,selection_command +1318,3955480,"jasmine/train_dynamics.py",17647,0,"",python,selection_mouse +1319,3955485,"jasmine/train_dynamics.py",17646,0,"",python,selection_command +1320,3958547,"jasmine/genie.py",0,0,"",python,tab +1321,3965002,"jasmine/genie.py",7659,0,"",python,selection_mouse +1322,3966189,"jasmine/genie.py",7628,0,"",python,selection_mouse +1323,3966191,"jasmine/genie.py",7627,0,"",python,selection_command +1324,3967422,"jasmine/genie.py",7628,0,"\n ",python,content +1325,3967840,"jasmine/genie.py",7637,0,"n",python,content +1326,3967841,"jasmine/genie.py",7638,0,"",python,selection_keyboard +1327,3968079,"jasmine/genie.py",7638,0,"o",python,content +1328,3968081,"jasmine/genie.py",7639,0,"",python,selection_keyboard +1329,3969303,"jasmine/genie.py",7639,0,"i",python,content +1330,3969306,"jasmine/genie.py",7640,0,"",python,selection_keyboard +1331,3969375,"jasmine/genie.py",7640,0,"s",python,content +1332,3969377,"jasmine/genie.py",7641,0,"",python,selection_keyboard +1333,3969523,"jasmine/genie.py",7641,0,"e",python,content +1334,3969525,"jasmine/genie.py",7642,0,"",python,selection_keyboard +1335,3969754,"jasmine/genie.py",7642,0,"_",python,content +1336,3969756,"jasmine/genie.py",7643,0,"",python,selection_keyboard +1337,3970131,"jasmine/genie.py",7643,0,"l",python,content +1338,3970133,"jasmine/genie.py",7644,0,"",python,selection_keyboard +1339,3970268,"jasmine/genie.py",7644,0,"e",python,content +1340,3970270,"jasmine/genie.py",7645,0,"",python,selection_keyboard +1341,3970449,"jasmine/genie.py",7645,0,"v",python,content +1342,3970450,"jasmine/genie.py",7646,0,"",python,selection_keyboard +1343,3970581,"jasmine/genie.py",7646,0,"e",python,content +1344,3970582,"jasmine/genie.py",7647,0,"",python,selection_keyboard +1345,3970688,"jasmine/genie.py",7647,0,"l",python,content +1346,3970690,"jasmine/genie.py",7648,0,"",python,selection_keyboard +1347,3971403,"jasmine/genie.py",7648,0,":",python,content +1348,3971404,"jasmine/genie.py",7649,0,"",python,selection_keyboard +1349,3971538,"jasmine/genie.py",7649,0," ",python,content +1350,3971539,"jasmine/genie.py",7650,0,"",python,selection_keyboard +1351,3973034,"jasmine/genie.py",7650,0,"f",python,content +1352,3973036,"jasmine/genie.py",7651,0,"",python,selection_keyboard +1353,3973201,"jasmine/genie.py",7651,0,"l",python,content +1354,3973202,"jasmine/genie.py",7652,0,"",python,selection_keyboard +1355,3973403,"jasmine/genie.py",7652,0,"o",python,content +1356,3973404,"jasmine/genie.py",7653,0,"",python,selection_keyboard +1357,3973450,"jasmine/genie.py",7653,0,"a",python,content +1358,3973452,"jasmine/genie.py",7654,0,"",python,selection_keyboard +1359,3974023,"jasmine/genie.py",7654,0,"t",python,content +1360,3974024,"jasmine/genie.py",7655,0,"",python,selection_keyboard +1361,3974340,"jasmine/genie.py",7655,0," ",python,content +1362,3974341,"jasmine/genie.py",7656,0,"",python,selection_keyboard +1363,3974585,"jasmine/genie.py",7656,0,"=",python,content +1364,3974587,"jasmine/genie.py",7657,0,"",python,selection_keyboard +1365,3974733,"jasmine/genie.py",7657,0," ",python,content +1366,3974733,"jasmine/genie.py",7658,0,"",python,selection_keyboard +1367,3975398,"jasmine/genie.py",7658,0,"0",python,content +1368,3975399,"jasmine/genie.py",7659,0,"",python,selection_keyboard +1369,3975865,"jasmine/genie.py",7659,0,".",python,content +1370,3975866,"jasmine/genie.py",7660,0,"",python,selection_keyboard +1371,3977465,"jasmine/genie.py",7917,0,"",python,selection_mouse +1372,3977994,"jasmine/genie.py",7916,0,"",python,selection_command +1373,3978273,"jasmine/genie.py",7915,1,"",python,content +1374,3978405,"jasmine/genie.py",7914,1,"",python,content +1375,3978548,"jasmine/genie.py",7913,1,"",python,content +1376,3978826,"jasmine/genie.py",7913,0,"n",python,content +1377,3978827,"jasmine/genie.py",7914,0,"",python,selection_keyboard +1378,3979494,"jasmine/genie.py",7914,0,"o",python,content +1379,3979495,"jasmine/genie.py",7915,0,"",python,selection_keyboard +1380,3979553,"jasmine/genie.py",7915,0,"i",python,content +1381,3979555,"jasmine/genie.py",7916,0,"",python,selection_keyboard +1382,3979719,"jasmine/genie.py",7916,0,"s",python,content +1383,3979720,"jasmine/genie.py",7917,0,"",python,selection_keyboard +1384,3979900,"jasmine/genie.py",7917,0,"e",python,content +1385,3979901,"jasmine/genie.py",7918,0,"",python,selection_keyboard +1386,3980086,"jasmine/genie.py",7918,0,"_",python,content +1387,3980087,"jasmine/genie.py",7919,0,"",python,selection_keyboard +1388,3980311,"jasmine/genie.py",7919,0,"l",python,content +1389,3980313,"jasmine/genie.py",7920,0,"",python,selection_keyboard +1390,3980444,"jasmine/genie.py",7920,0,"e",python,content +1391,3980446,"jasmine/genie.py",7921,0,"",python,selection_keyboard +1392,3980544,"jasmine/genie.py",7921,0,"v",python,content +1393,3980546,"jasmine/genie.py",7922,0,"",python,selection_keyboard +1394,3980641,"jasmine/genie.py",7922,0,"e",python,content +1395,3980643,"jasmine/genie.py",7923,0,"",python,selection_keyboard +1396,3980708,"jasmine/genie.py",7923,0,"l",python,content +1397,3980710,"jasmine/genie.py",7924,0,"",python,selection_keyboard +1398,3981539,"jasmine/genie.py",8058,0,"",python,selection_mouse +1399,3982632,"jasmine/genie.py",7981,0,"",python,selection_mouse +1400,3983884,"jasmine/genie.py",7880,0,"",python,selection_mouse +1401,3984673,"jasmine/genie.py",7921,0,"",python,selection_mouse +1402,3985796,"jasmine/genie.py",7660,0,"",python,selection_mouse +1403,3986392,"jasmine/genie.py",7660,0,",",python,content +1404,3986393,"jasmine/genie.py",7661,0,"",python,selection_keyboard +1405,3988816,"jasmine/genie.py",8078,0,"",python,selection_mouse +1406,3991377,"jasmine/genie.py",8078,0,"n",python,content +1407,3991379,"jasmine/genie.py",8079,0,"",python,selection_keyboard +1408,3991588,"jasmine/genie.py",8079,0,"o",python,content +1409,3991589,"jasmine/genie.py",8080,0,"",python,selection_keyboard +1410,3991664,"jasmine/genie.py",8080,0,"i",python,content +1411,3991665,"jasmine/genie.py",8081,0,"",python,selection_keyboard +1412,3991768,"jasmine/genie.py",8081,0,"s",python,content +1413,3991769,"jasmine/genie.py",8082,0,"",python,selection_keyboard +1414,3991948,"jasmine/genie.py",8082,0,"e",python,content +1415,3991950,"jasmine/genie.py",8083,0,"",python,selection_keyboard +1416,3992830,"jasmine/genie.py",8083,0,"_",python,content +1417,3992831,"jasmine/genie.py",8084,0,"",python,selection_keyboard +1418,3993562,"jasmine/genie.py",8084,0,"l",python,content +1419,3993564,"jasmine/genie.py",8085,0,"",python,selection_keyboard +1420,3993600,"jasmine/genie.py",8085,0,"e",python,content +1421,3993601,"jasmine/genie.py",8086,0,"",python,selection_keyboard +1422,3994099,"jasmine/genie.py",8086,0,"v",python,content +1423,3994101,"jasmine/genie.py",8087,0,"",python,selection_keyboard +1424,3994195,"jasmine/genie.py",8087,0,"e",python,content +1425,3994197,"jasmine/genie.py",8088,0,"",python,selection_keyboard +1426,3994262,"jasmine/genie.py",8088,0,"l",python,content +1427,3994264,"jasmine/genie.py",8089,0,"",python,selection_keyboard +1428,3994646,"jasmine/genie.py",8089,0,",",python,content +1429,3994648,"jasmine/genie.py",8090,0,"",python,selection_keyboard +1430,3994807,"jasmine/genie.py",8090,0," ",python,content +1431,3994808,"jasmine/genie.py",8091,0,"",python,selection_keyboard +1432,3997925,"jasmine/genie.py",7693,0,"",python,selection_mouse +1433,3999810,"jasmine/genie.py",7628,0,"",python,selection_mouse +1434,4000435,"jasmine/genie.py",7660,0,"",python,selection_mouse +1435,4006378,"jasmine/genie.py",8055,0,"",python,selection_mouse +1436,4006907,"jasmine/genie.py",17920,0,"",python,selection_command +1437,4008036,"jasmine/genie.py",18007,0,"",python,selection_mouse +1438,4008652,"jasmine/genie.py",17985,0,"",python,selection_mouse +1439,4009073,"jasmine/genie.py",17985,0,"\n ",python,content +1440,4009446,"jasmine/genie.py",17990,4,"",python,content +1441,4009636,"jasmine/genie.py",17986,4,"",python,content +1442,4009865,"jasmine/genie.py",17985,1,"",python,content +1443,4010297,"jasmine/genie.py",18007,0,"",python,selection_command +1444,4010553,"jasmine/genie.py",18007,0,"\n ",python,content +1445,4010963,"jasmine/genie.py",18016,0,"n",python,content +1446,4010964,"jasmine/genie.py",18017,0,"",python,selection_keyboard +1447,4011147,"jasmine/genie.py",18017,0,"o",python,content +1448,4011150,"jasmine/genie.py",18018,0,"",python,selection_keyboard +1449,4011219,"jasmine/genie.py",18018,0,"i",python,content +1450,4011220,"jasmine/genie.py",18019,0,"",python,selection_keyboard +1451,4011403,"jasmine/genie.py",18019,0,"e",python,content +1452,4011404,"jasmine/genie.py",18020,0,"",python,selection_keyboard +1453,4011913,"jasmine/genie.py",18019,1,"",python,content +1454,4011974,"jasmine/genie.py",18019,0,"s",python,content +1455,4011975,"jasmine/genie.py",18020,0,"",python,selection_keyboard +1456,4012099,"jasmine/genie.py",18020,0,"e",python,content +1457,4012101,"jasmine/genie.py",18021,0,"",python,selection_keyboard +1458,4012249,"jasmine/genie.py",18021,0,"_",python,content +1459,4012251,"jasmine/genie.py",18022,0,"",python,selection_keyboard +1460,4012512,"jasmine/genie.py",18022,0,"l",python,content +1461,4012514,"jasmine/genie.py",18023,0,"",python,selection_keyboard +1462,4012710,"jasmine/genie.py",18023,0,"e",python,content +1463,4012712,"jasmine/genie.py",18024,0,"",python,selection_keyboard +1464,4013003,"jasmine/genie.py",18023,1,"",python,content +1465,4013639,"jasmine/genie.py",18023,0,"e",python,content +1466,4013641,"jasmine/genie.py",18024,0,"",python,selection_keyboard +1467,4014090,"jasmine/genie.py",18024,0,"v",python,content +1468,4014091,"jasmine/genie.py",18025,0,"",python,selection_keyboard +1469,4014152,"jasmine/genie.py",18025,0,"e",python,content +1470,4014153,"jasmine/genie.py",18026,0,"",python,selection_keyboard +1471,4014302,"jasmine/genie.py",18026,0,"l",python,content +1472,4014303,"jasmine/genie.py",18027,0,"",python,selection_keyboard +1473,4015162,"jasmine/genie.py",18027,0,":",python,content +1474,4015163,"jasmine/genie.py",18028,0,"",python,selection_keyboard +1475,4015396,"jasmine/genie.py",18028,0," ",python,content +1476,4015397,"jasmine/genie.py",18029,0,"",python,selection_keyboard +1477,4015582,"jasmine/genie.py",18029,0,"f",python,content +1478,4015583,"jasmine/genie.py",18030,0,"",python,selection_keyboard +1479,4015662,"jasmine/genie.py",18030,0,"l",python,content +1480,4015663,"jasmine/genie.py",18031,0,"",python,selection_keyboard +1481,4015836,"jasmine/genie.py",18031,0,"o",python,content +1482,4015837,"jasmine/genie.py",18032,0,"",python,selection_keyboard +1483,4015903,"jasmine/genie.py",18032,0,"a",python,content +1484,4015904,"jasmine/genie.py",18033,0,"",python,selection_keyboard +1485,4016139,"jasmine/genie.py",18033,0,"t",python,content +1486,4016140,"jasmine/genie.py",18034,0,"",python,selection_keyboard +1487,4016834,"jasmine/genie.py",18034,0,",",python,content +1488,4016835,"jasmine/genie.py",18035,0,"",python,selection_keyboard +1489,4028274,"jasmine/genie.py",9655,0,"",python,selection_mouse +1490,4029144,"jasmine/genie.py",9654,0,"",python,selection_command +1491,4029744,"jasmine/genie.py",9613,62," ), ""Noise level must me smaller than max_noise_level.""",python,selection_command +1492,4029987,"jasmine/genie.py",9566,109," noise_level < self.max_noise_level\n ), ""Noise level must me smaller than max_noise_level.""",python,selection_command +1493,4030090,"jasmine/genie.py",9549,126," assert (\n noise_level < self.max_noise_level\n ), ""Noise level must me smaller than max_noise_level.""",python,selection_command +1494,4030720,"jasmine/genie.py",9549,127,"",python,content +1495,4030768,"jasmine/genie.py",9557,0,"",python,selection_command +1496,4033967,"jasmine/genie.py",7801,0,"",python,selection_mouse +1497,4033978,"jasmine/genie.py",7800,0,"",python,selection_command +1498,4036243,"jasmine/genie.py",7801,0,"\n assert (\n noise_level < self.max_noise_level\n ), ""Noise level must me smaller than max_noise_level.""",python,content +1499,4036283,"jasmine/genie.py",7810,0,"",python,selection_command +1500,4037318,"jasmine/genie.py",7827,0,"",python,selection_command +1501,4037542,"jasmine/genie.py",7874,0,"",python,selection_command +1502,4102142,"jasmine/genie.py",11911,0,"",python,selection_mouse +1503,4102176,"jasmine/genie.py",11910,0,"",python,selection_command +1504,4182529,"jasmine/genie.py",5265,0,"",python,selection_mouse +1505,4183842,"jasmine/genie.py",5280,0,"",python,selection_mouse +1506,4183992,"jasmine/genie.py",5275,15,"max_noise_level",python,selection_mouse +1507,4184690,"jasmine/genie.py",5267,0,"",python,selection_mouse +1508,4184856,"jasmine/genie.py",5254,15,"max_noise_level",python,selection_mouse +1509,4185467,"jasmine/genie.py",5367,0,"",python,selection_mouse +1510,4185624,"jasmine/genie.py",5358,11,"param_dtype",python,selection_mouse +1511,4186163,"jasmine/genie.py",5318,0,"",python,selection_mouse +1512,4186302,"jasmine/genie.py",5308,13,"noise_buckets",python,selection_mouse +1513,4194854,"jasmine/models/dynamics.py",0,0,"",python,tab +1514,4198629,"jasmine/models/dynamics.py",0,0,"",python,tab +1515,4390762,"jasmine/models/dynamics.py",3364,0,"",python,selection_mouse +1516,4390870,"jasmine/models/dynamics.py",3358,15,"max_noise_level",python,selection_mouse +1517,4396696,"jasmine/genie.py",0,0,"",python,tab +1518,4410672,"jasmine/genie.py",3437,0,"",python,selection_mouse +1519,4410680,"jasmine/genie.py",3436,0,"",python,selection_command +1520,4411416,"jasmine/genie.py",3381,0,"",python,selection_mouse +1521,4411425,"jasmine/genie.py",3380,0,"",python,selection_command +1522,4457875,"jasmine/train_dynamics.py",0,0,"",python,tab +1523,4459724,"jasmine/train_dynamics.py",17576,0,"",python,selection_mouse +1524,4460456,"jasmine/train_dynamics.py",17647,0,"",python,selection_mouse +1525,4460460,"jasmine/train_dynamics.py",17646,0,"",python,selection_command +1526,4462405,"jasmine/train_dynamics.py",17647,0,"\n ",python,content +1527,4481023,"jasmine/train_dynamics.py",17610,0,"",python,selection_mouse +1528,4481813,"jasmine/train_dynamics.py",17648,16,"",python,content +1529,4481814,"jasmine/train_dynamics.py",17610,0,"b",python,content +1530,4481815,"jasmine/train_dynamics.py",17611,0,"",python,selection_keyboard +1531,4481851,"jasmine/train_dynamics.py",17611,0,"a",python,content +1532,4481852,"jasmine/train_dynamics.py",17612,0,"",python,selection_keyboard +1533,4481995,"jasmine/train_dynamics.py",17612,0,"t",python,content +1534,4481996,"jasmine/train_dynamics.py",17613,0,"",python,selection_keyboard +1535,4482201,"jasmine/train_dynamics.py",17613,0,"h",python,content +1536,4482203,"jasmine/train_dynamics.py",17614,0,"",python,selection_keyboard +1537,4482595,"jasmine/train_dynamics.py",17613,1,"",python,content +1538,4482769,"jasmine/train_dynamics.py",17613,0,"c",python,content +1539,4482771,"jasmine/train_dynamics.py",17614,0,"",python,selection_keyboard +1540,4482849,"jasmine/train_dynamics.py",17614,0,"h",python,content +1541,4482850,"jasmine/train_dynamics.py",17615,0,"",python,selection_keyboard +1542,4483553,"jasmine/train_dynamics.py",17615,0,"?",python,content +1543,4483557,"jasmine/train_dynamics.py",17616,0,"",python,selection_keyboard +1544,4484555,"jasmine/train_dynamics.py",17615,1,"",python,content +1545,4484775,"jasmine/train_dynamics.py",17615,0,"=",python,content +1546,4484776,"jasmine/train_dynamics.py",17616,0,"",python,selection_keyboard +1547,4485229,"jasmine/train_dynamics.py",17646,0,"",python,selection_command +1548,4487045,"jasmine/train_dynamics.py",17640,0,"s",python,content +1549,4487047,"jasmine/train_dynamics.py",17641,0,"",python,selection_keyboard +1550,4487256,"jasmine/train_dynamics.py",17641,0,"e",python,content +1551,4487258,"jasmine/train_dynamics.py",17642,0,"",python,selection_keyboard +1552,4487389,"jasmine/train_dynamics.py",17642,0,"q",python,content +1553,4487390,"jasmine/train_dynamics.py",17643,0,"",python,selection_keyboard +1554,4487555,"jasmine/train_dynamics.py",17643,0,"_",python,content +1555,4487557,"jasmine/train_dynamics.py",17644,0,"",python,selection_keyboard +1556,4487913,"jasmine/train_dynamics.py",17644,0,"e",python,content +1557,4487915,"jasmine/train_dynamics.py",17645,0,"",python,selection_keyboard +1558,4488250,"jasmine/train_dynamics.py",17644,1,"",python,content +1559,4488419,"jasmine/train_dynamics.py",17644,0,"l",python,content +1560,4488420,"jasmine/train_dynamics.py",17645,0,"",python,selection_keyboard +1561,4488501,"jasmine/train_dynamics.py",17645,0,"e",python,content +1562,4488503,"jasmine/train_dynamics.py",17646,0,"",python,selection_keyboard +1563,4488768,"jasmine/train_dynamics.py",17646,0,"n",python,content +1564,4488770,"jasmine/train_dynamics.py",17647,0,"",python,selection_keyboard +1565,4489512,"jasmine/train_dynamics.py",17647,0,"=",python,content +1566,4489514,"jasmine/train_dynamics.py",17648,0,"",python,selection_keyboard +1567,4489980,"jasmine/train_dynamics.py",17662,0,"",python,selection_command +1568,4490698,"jasmine/train_dynamics.py",17662,0," ",python,content +1569,4490903,"jasmine/train_dynamics.py",17666,0," ",python,content +1570,4491085,"jasmine/train_dynamics.py",17670,0," ",python,content +1571,4491273,"jasmine/train_dynamics.py",17674,0," ",python,content +1572,4497640,"jasmine/train_dynamics.py",17695,0,"",python,selection_command +1573,4498351,"jasmine/train_dynamics.py",17695,0,"t",python,content +1574,4498351,"jasmine/train_dynamics.py",17662,16,"",python,content +1575,4498353,"jasmine/train_dynamics.py",17680,0,"",python,selection_keyboard +1576,4498529,"jasmine/train_dynamics.py",17680,0,"e",python,content +1577,4498530,"jasmine/train_dynamics.py",17681,0,"",python,selection_keyboard +1578,4498606,"jasmine/train_dynamics.py",17681,0,"m",python,content +1579,4498607,"jasmine/train_dynamics.py",17682,0,"",python,selection_keyboard +1580,4498804,"jasmine/train_dynamics.py",17682,0,"p",python,content +1581,4498806,"jasmine/train_dynamics.py",17683,0,"",python,selection_keyboard +1582,4498922,"jasmine/train_dynamics.py",17683,0,"e",python,content +1583,4498924,"jasmine/train_dynamics.py",17684,0,"",python,selection_keyboard +1584,4499094,"jasmine/train_dynamics.py",17684,0,"r",python,content +1585,4499095,"jasmine/train_dynamics.py",17685,0,"",python,selection_keyboard +1586,4499287,"jasmine/train_dynamics.py",17685,0,"a",python,content +1587,4499288,"jasmine/train_dynamics.py",17686,0,"",python,selection_keyboard +1588,4499516,"jasmine/train_dynamics.py",17686,0,"t",python,content +1589,4499517,"jasmine/train_dynamics.py",17687,0,"",python,selection_keyboard +1590,4499624,"jasmine/train_dynamics.py",17687,0,"u",python,content +1591,4499625,"jasmine/train_dynamics.py",17688,0,"",python,selection_keyboard +1592,4499733,"jasmine/train_dynamics.py",17688,0,"r",python,content +1593,4499735,"jasmine/train_dynamics.py",17689,0,"",python,selection_keyboard +1594,4499894,"jasmine/train_dynamics.py",17689,0,"e",python,content +1595,4499896,"jasmine/train_dynamics.py",17690,0,"",python,selection_keyboard +1596,4500173,"jasmine/train_dynamics.py",17690,0,"=",python,content +1597,4500174,"jasmine/train_dynamics.py",17691,0,"",python,selection_keyboard +1598,4500869,"jasmine/train_dynamics.py",17741,0,"",python,selection_command +1599,4503414,"jasmine/train_dynamics.py",17729,0,"s",python,content +1600,4503416,"jasmine/train_dynamics.py",17730,0,"",python,selection_keyboard +1601,4503581,"jasmine/train_dynamics.py",17730,0,"a",python,content +1602,4503581,"jasmine/train_dynamics.py",17731,0,"",python,selection_keyboard +1603,4503598,"jasmine/train_dynamics.py",17731,0,"m",python,content +1604,4503599,"jasmine/train_dynamics.py",17732,0,"",python,selection_keyboard +1605,4503793,"jasmine/train_dynamics.py",17732,0,"p",python,content +1606,4503794,"jasmine/train_dynamics.py",17733,0,"",python,selection_keyboard +1607,4503853,"jasmine/train_dynamics.py",17733,0,"l",python,content +1608,4503854,"jasmine/train_dynamics.py",17734,0,"",python,selection_keyboard +1609,4503955,"jasmine/train_dynamics.py",17734,0,"e",python,content +1610,4503956,"jasmine/train_dynamics.py",17735,0,"",python,selection_keyboard +1611,4504185,"jasmine/train_dynamics.py",17735,0,"_",python,content +1612,4504187,"jasmine/train_dynamics.py",17736,0,"",python,selection_keyboard +1613,4504366,"jasmine/train_dynamics.py",17736,0,"a",python,content +1614,4504368,"jasmine/train_dynamics.py",17737,0,"",python,selection_keyboard +1615,4504626,"jasmine/train_dynamics.py",17737,0,"r",python,content +1616,4504627,"jasmine/train_dynamics.py",17738,0,"",python,selection_keyboard +1617,4505139,"jasmine/train_dynamics.py",17738,0,"g",python,content +1618,4505140,"jasmine/train_dynamics.py",17739,0,"",python,selection_keyboard +1619,4505368,"jasmine/train_dynamics.py",17739,0,"m",python,content +1620,4505370,"jasmine/train_dynamics.py",17740,0,"",python,selection_keyboard +1621,4505501,"jasmine/train_dynamics.py",17740,0,"a",python,content +1622,4505502,"jasmine/train_dynamics.py",17741,0,"",python,selection_keyboard +1623,4505721,"jasmine/train_dynamics.py",17741,0,"x",python,content +1624,4505722,"jasmine/train_dynamics.py",17742,0,"",python,selection_keyboard +1625,4506084,"jasmine/train_dynamics.py",17742,0,"=",python,content +1626,4506085,"jasmine/train_dynamics.py",17743,0,"",python,selection_keyboard +1627,4506725,"jasmine/train_dynamics.py",17742,0,"",python,selection_command +1628,4506865,"jasmine/train_dynamics.py",17796,0,"",python,selection_command +1629,4507237,"jasmine/train_dynamics.py",17783,0,"",python,selection_command +1630,4508366,"jasmine/train_dynamics.py",17783,0,"m",python,content +1631,4508367,"jasmine/train_dynamics.py",17784,0,"",python,selection_keyboard +1632,4508455,"jasmine/train_dynamics.py",17784,0,"a",python,content +1633,4508457,"jasmine/train_dynamics.py",17785,0,"",python,selection_keyboard +1634,4508525,"jasmine/train_dynamics.py",17785,0,"s",python,content +1635,4508526,"jasmine/train_dynamics.py",17786,0,"",python,selection_keyboard +1636,4508652,"jasmine/train_dynamics.py",17786,0,"k",python,content +1637,4508653,"jasmine/train_dynamics.py",17787,0,"",python,selection_keyboard +1638,4508877,"jasmine/train_dynamics.py",17787,0,"g",python,content +1639,4508878,"jasmine/train_dynamics.py",17788,0,"",python,selection_keyboard +1640,4508971,"jasmine/train_dynamics.py",17788,0,"i",python,content +1641,4508972,"jasmine/train_dynamics.py",17789,0,"",python,selection_keyboard +1642,4509081,"jasmine/train_dynamics.py",17789,0,"t",python,content +1643,4509082,"jasmine/train_dynamics.py",17790,0,"",python,selection_keyboard +1644,4509302,"jasmine/train_dynamics.py",17790,0,"_",python,content +1645,4509304,"jasmine/train_dynamics.py",17791,0,"",python,selection_keyboard +1646,4509536,"jasmine/train_dynamics.py",17791,0,"s",python,content +1647,4509537,"jasmine/train_dynamics.py",17792,0,"",python,selection_keyboard +1648,4509721,"jasmine/train_dynamics.py",17792,0,"t",python,content +1649,4509722,"jasmine/train_dynamics.py",17793,0,"",python,selection_keyboard +1650,4509930,"jasmine/train_dynamics.py",17793,0,"e",python,content +1651,4509931,"jasmine/train_dynamics.py",17794,0,"",python,selection_keyboard +1652,4510056,"jasmine/train_dynamics.py",17794,0,"p",python,content +1653,4510058,"jasmine/train_dynamics.py",17795,0,"",python,selection_keyboard +1654,4510145,"jasmine/train_dynamics.py",17795,0,"s",python,content +1655,4510147,"jasmine/train_dynamics.py",17796,0,"",python,selection_keyboard +1656,4510306,"jasmine/train_dynamics.py",17796,0,"=",python,content +1657,4510307,"jasmine/train_dynamics.py",17797,0,"",python,selection_keyboard +1658,4511058,"jasmine/train_dynamics.py",17796,0,"",python,selection_command +1659,4511159,"jasmine/train_dynamics.py",17742,0,"",python,selection_command +1660,4511303,"jasmine/train_dynamics.py",17692,0,"",python,selection_command +1661,4511575,"jasmine/train_dynamics.py",17662,0,"",python,selection_command +1662,4512639,"jasmine/train_dynamics.py",17662,0," ",python,content +1663,4512840,"jasmine/train_dynamics.py",17666,0," ",python,content +1664,4513145,"jasmine/train_dynamics.py",17670,0," ",python,content +1665,4513410,"jasmine/train_dynamics.py",17674,0," ",python,content +1666,4521380,"jasmine/genie.py",0,0,"",python,tab +1667,4526888,"jasmine/train_dynamics.py",0,0,"",python,tab +1668,4527291,"jasmine/train_dynamics.py",17678,0,"n",python,content +1669,4527292,"jasmine/train_dynamics.py",17679,0,"",python,selection_keyboard +1670,4527503,"jasmine/train_dynamics.py",17679,0,"o",python,content +1671,4527504,"jasmine/train_dynamics.py",17680,0,"",python,selection_keyboard +1672,4527597,"jasmine/train_dynamics.py",17680,0,"i",python,content +1673,4527599,"jasmine/train_dynamics.py",17681,0,"",python,selection_keyboard +1674,4527722,"jasmine/train_dynamics.py",17681,0,"s",python,content +1675,4527723,"jasmine/train_dynamics.py",17682,0,"",python,selection_keyboard +1676,4527867,"jasmine/train_dynamics.py",17682,0,"e",python,content +1677,4527869,"jasmine/train_dynamics.py",17683,0,"",python,selection_keyboard +1678,4528116,"jasmine/train_dynamics.py",17683,0,"_",python,content +1679,4528118,"jasmine/train_dynamics.py",17684,0,"",python,selection_keyboard +1680,4528379,"jasmine/train_dynamics.py",17684,0,"l",python,content +1681,4528381,"jasmine/train_dynamics.py",17685,0,"",python,selection_keyboard +1682,4528611,"jasmine/train_dynamics.py",17685,0,"e",python,content +1683,4528612,"jasmine/train_dynamics.py",17686,0,"",python,selection_keyboard +1684,4528758,"jasmine/train_dynamics.py",17686,0,"v",python,content +1685,4528760,"jasmine/train_dynamics.py",17687,0,"",python,selection_keyboard +1686,4528899,"jasmine/train_dynamics.py",17687,0,"e",python,content +1687,4528901,"jasmine/train_dynamics.py",17688,0,"",python,selection_keyboard +1688,4528965,"jasmine/train_dynamics.py",17688,0,"l",python,content +1689,4528967,"jasmine/train_dynamics.py",17689,0,"",python,selection_keyboard +1690,4529815,"jasmine/train_dynamics.py",17689,0,"=",python,content +1691,4529817,"jasmine/train_dynamics.py",17690,0,"",python,selection_keyboard +1692,4530153,"jasmine/train_dynamics.py",17690,0,"0",python,content +1693,4530154,"jasmine/train_dynamics.py",17691,0,"",python,selection_keyboard +1694,4530256,"jasmine/train_dynamics.py",17691,0,".",python,content +1695,4530258,"jasmine/train_dynamics.py",17692,0,"",python,selection_keyboard +1696,4533360,"jasmine/train_dynamics.py",17692,0,",",python,content +1697,4533361,"jasmine/train_dynamics.py",17693,0,"",python,selection_keyboard +1698,4536122,"jasmine/train_dynamics.py",17848,0,"",python,selection_mouse +1699,4536647,"jasmine/train_dynamics.py",17797,0,"",python,selection_mouse +1700,4537331,"jasmine/train_dynamics.py",17851,0,"",python,selection_mouse +1701,4609343,"jasmine/models/dynamics.py",0,0,"",python,tab +1702,4612320,"jasmine/models/dynamics.py",0,0,"",python,tab +1703,4617540,"jasmine/models/dynamics.py",4017,0,"",python,selection_mouse +1704,4654611,"jasmine/models/dynamics.py",2241,0,"",python,selection_mouse +1705,4655336,"jasmine/models/dynamics.py",2240,0,"",python,selection_mouse +1706,4655338,"jasmine/models/dynamics.py",2239,0,"",python,selection_command +1707,4657786,"jasmine/models/dynamics.py",2240,0,"\n ",python,content +1708,4658073,"jasmine/models/dynamics.py",2249,0,"\n ",python,content +1709,4658073,"jasmine/models/dynamics.py",2241,8,"",python,content +1710,4658694,"jasmine/models/dynamics.py",2246,4,"",python,content +1711,4658886,"jasmine/models/dynamics.py",2246,0,"d",python,content +1712,4658887,"jasmine/models/dynamics.py",2247,0,"",python,selection_keyboard +1713,4659015,"jasmine/models/dynamics.py",2247,0,"e",python,content +1714,4659016,"jasmine/models/dynamics.py",2248,0,"",python,selection_keyboard +1715,4659113,"jasmine/models/dynamics.py",2248,0,"f",python,content +1716,4659115,"jasmine/models/dynamics.py",2249,0,"",python,selection_keyboard +1717,4659560,"jasmine/models/dynamics.py",2249,0," ",python,content +1718,4659562,"jasmine/models/dynamics.py",2250,0,"",python,selection_keyboard +1719,4659753,"jasmine/models/dynamics.py",2250,0,"_",python,content +1720,4659754,"jasmine/models/dynamics.py",2251,0,"",python,selection_keyboard +1721,4660015,"jasmine/models/dynamics.py",2251,0,"a",python,content +1722,4660017,"jasmine/models/dynamics.py",2252,0,"",python,selection_keyboard +1723,4660134,"jasmine/models/dynamics.py",2252,0,"p",python,content +1724,4660136,"jasmine/models/dynamics.py",2253,0,"",python,selection_keyboard +1725,4660276,"jasmine/models/dynamics.py",2253,0,"p",python,content +1726,4660277,"jasmine/models/dynamics.py",2254,0,"",python,selection_keyboard +1727,4660363,"jasmine/models/dynamics.py",2254,0,"l",python,content +1728,4660364,"jasmine/models/dynamics.py",2255,0,"",python,selection_keyboard +1729,4660533,"jasmine/models/dynamics.py",2255,0,"y",python,content +1730,4660534,"jasmine/models/dynamics.py",2256,0,"",python,selection_keyboard +1731,4660788,"jasmine/models/dynamics.py",2256,0,"_",python,content +1732,4660789,"jasmine/models/dynamics.py",2257,0,"",python,selection_keyboard +1733,4661441,"jasmine/models/dynamics.py",2257,0,"n",python,content +1734,4661442,"jasmine/models/dynamics.py",2258,0,"",python,selection_keyboard +1735,4661591,"jasmine/models/dynamics.py",2258,0,"o",python,content +1736,4661592,"jasmine/models/dynamics.py",2259,0,"",python,selection_keyboard +1737,4661675,"jasmine/models/dynamics.py",2259,0,"i",python,content +1738,4661676,"jasmine/models/dynamics.py",2260,0,"",python,selection_keyboard +1739,4661746,"jasmine/models/dynamics.py",2260,0,"s",python,content +1740,4661747,"jasmine/models/dynamics.py",2261,0,"",python,selection_keyboard +1741,4661921,"jasmine/models/dynamics.py",2261,0,"e",python,content +1742,4661921,"jasmine/models/dynamics.py",2262,0,"",python,selection_keyboard +1743,4662106,"jasmine/models/dynamics.py",2262,0,"_",python,content +1744,4662107,"jasmine/models/dynamics.py",2263,0,"",python,selection_keyboard +1745,4662311,"jasmine/models/dynamics.py",2263,0,"a",python,content +1746,4662312,"jasmine/models/dynamics.py",2264,0,"",python,selection_keyboard +1747,4662454,"jasmine/models/dynamics.py",2264,0,"u",python,content +1748,4662455,"jasmine/models/dynamics.py",2265,0,"",python,selection_keyboard +1749,4662621,"jasmine/models/dynamics.py",2265,0,"g",python,content +1750,4662622,"jasmine/models/dynamics.py",2266,0,"",python,selection_keyboard +1751,4662857,"jasmine/models/dynamics.py",2266,0,"m",python,content +1752,4662858,"jasmine/models/dynamics.py",2267,0,"",python,selection_keyboard +1753,4662970,"jasmine/models/dynamics.py",2267,0,"e",python,content +1754,4662971,"jasmine/models/dynamics.py",2268,0,"",python,selection_keyboard +1755,4663096,"jasmine/models/dynamics.py",2268,0,"n",python,content +1756,4663097,"jasmine/models/dynamics.py",2269,0,"",python,selection_keyboard +1757,4663254,"jasmine/models/dynamics.py",2269,0,"t",python,content +1758,4663255,"jasmine/models/dynamics.py",2270,0,"",python,selection_keyboard +1759,4663522,"jasmine/models/dynamics.py",2270,0,"a",python,content +1760,4663523,"jasmine/models/dynamics.py",2271,0,"",python,selection_keyboard +1761,4663708,"jasmine/models/dynamics.py",2271,0,"t",python,content +1762,4663709,"jasmine/models/dynamics.py",2272,0,"",python,selection_keyboard +1763,4663793,"jasmine/models/dynamics.py",2272,0,"i",python,content +1764,4663794,"jasmine/models/dynamics.py",2273,0,"",python,selection_keyboard +1765,4663879,"jasmine/models/dynamics.py",2273,0,"o",python,content +1766,4663879,"jasmine/models/dynamics.py",2274,0,"",python,selection_keyboard +1767,4664017,"jasmine/models/dynamics.py",2274,0,"n",python,content +1768,4664018,"jasmine/models/dynamics.py",2275,0,"",python,selection_keyboard +1769,4665735,"jasmine/models/dynamics.py",2275,0,"()",python,content +1770,4665737,"jasmine/models/dynamics.py",2276,0,"",python,selection_keyboard +1771,4666398,"jasmine/models/dynamics.py",2277,0,"",python,selection_command +1772,4666800,"jasmine/models/dynamics.py",2276,0,"",python,selection_command +1773,4667009,"jasmine/models/dynamics.py",2276,0,"s",python,content +1774,4667010,"jasmine/models/dynamics.py",2277,0,"",python,selection_keyboard +1775,4667209,"jasmine/models/dynamics.py",2277,0,"e",python,content +1776,4667210,"jasmine/models/dynamics.py",2278,0,"",python,selection_keyboard +1777,4667338,"jasmine/models/dynamics.py",2278,0,"l",python,content +1778,4667339,"jasmine/models/dynamics.py",2279,0,"",python,selection_keyboard +1779,4667542,"jasmine/models/dynamics.py",2279,0,"f",python,content +1780,4667543,"jasmine/models/dynamics.py",2280,0,"",python,selection_keyboard +1781,4667655,"jasmine/models/dynamics.py",2280,0,",",python,content +1782,4667657,"jasmine/models/dynamics.py",2281,0,"",python,selection_keyboard +1783,4667997,"jasmine/models/dynamics.py",2280,1,"",python,content +1784,4668239,"jasmine/models/dynamics.py",2281,0,"",python,selection_command +1785,4668662,"jasmine/models/dynamics.py",2281,0,":",python,content +1786,4668663,"jasmine/models/dynamics.py",2282,0,"",python,selection_keyboard +1787,4668856,"jasmine/models/dynamics.py",2282,0,"\n ",python,content +1788,4669354,"jasmine/models/dynamics.py",2291,0,"p",python,content +1789,4669355,"jasmine/models/dynamics.py",2292,0,"",python,selection_keyboard +1790,4669433,"jasmine/models/dynamics.py",2292,0,"a",python,content +1791,4669435,"jasmine/models/dynamics.py",2293,0,"",python,selection_keyboard +1792,4669548,"jasmine/models/dynamics.py",2293,0,"s",python,content +1793,4669549,"jasmine/models/dynamics.py",2294,0,"",python,selection_keyboard +1794,4669699,"jasmine/models/dynamics.py",2294,0,"s",python,content +1795,4669700,"jasmine/models/dynamics.py",2295,0,"",python,selection_keyboard +1796,4669982,"jasmine/models/dynamics.py",2294,0,"",python,selection_command +1797,4674092,"jasmine/models/dynamics.py",4075,0,"",python,selection_mouse +1798,4676199,"jasmine/models/dynamics.py",4341,0,"",python,selection_mouse +1799,4676827,"jasmine/models/dynamics.py",4456,0,"",python,selection_mouse +1800,4677482,"jasmine/models/dynamics.py",4372,0,"",python,selection_mouse +1801,4681540,"jasmine/models/dynamics.py",4359,0,"",python,selection_mouse +1802,4681683,"jasmine/models/dynamics.py",4350,14,"vid_embed_BTNM",python,selection_mouse +1803,4685629,"jasmine/models/dynamics.py",4731,0,"",python,selection_mouse +1804,4685756,"jasmine/models/dynamics.py",4721,22,"noise_level_embed_BT1M",python,selection_mouse +1805,4688000,"jasmine/models/dynamics.py",4360,0,"",python,selection_mouse +1806,4688153,"jasmine/models/dynamics.py",4350,14,"vid_embed_BTNM",python,selection_mouse +1807,4693463,"jasmine/models/dynamics.py",4360,0,"",python,selection_mouse +1808,4697376,"jasmine/models/dynamics.py",4342,82," vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1809,4697594,"jasmine/models/dynamics.py",4341,83,"\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1810,4698129,"jasmine/models/dynamics.py",4272,152," sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1811,4698131,"jasmine/models/dynamics.py",4221,203," sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1812,4698154,"jasmine/models/dynamics.py",4151,273," one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1813,4698180,"jasmine/models/dynamics.py",4108,316," # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1814,4698242,"jasmine/models/dynamics.py",4107,317,"\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1815,4698243,"jasmine/models/dynamics.py",4046,378," noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1816,4698281,"jasmine/models/dynamics.py",3966,458," noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1817,4698305,"jasmine/models/dynamics.py",3888,536," noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1818,4698328,"jasmine/models/dynamics.py",3821,603," noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1819,4698389,"jasmine/models/dynamics.py",3793,631," ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1820,4698390,"jasmine/models/dynamics.py",3721,703," (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1821,4698418,"jasmine/models/dynamics.py",3681,743," noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1822,4698463,"jasmine/models/dynamics.py",3594,830," # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1823,4698567,"jasmine/models/dynamics.py",3510,914," # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1824,4698567,"jasmine/models/dynamics.py",3439,985," noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1825,4698640,"jasmine/models/dynamics.py",3429,995," )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1826,4698771,"jasmine/models/dynamics.py",3349,1075," _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1827,4698914,"jasmine/models/dynamics.py",3305,1119," noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1828,4699057,"jasmine/models/dynamics.py",3238,1186," rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1829,4699226,"jasmine/models/dynamics.py",3207,1217," # --- Sample noise ---\n rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1830,4699584,"jasmine/models/dynamics.py",3238,1186," rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_command +1831,4703119,"jasmine/models/dynamics.py",3238,0,"",python,selection_command +1832,4704766,"jasmine/models/dynamics.py",2259,0,"",python,selection_mouse +1833,4705412,"jasmine/models/dynamics.py",2282,0,"\n rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,content +1834,4705445,"jasmine/models/dynamics.py",2291,0,"",python,selection_command +1835,4706022,"jasmine/models/dynamics.py",2358,0,"",python,selection_command +1836,4706515,"jasmine/models/dynamics.py",2402,0,"",python,selection_command +1837,4706567,"jasmine/models/dynamics.py",2482,0,"",python,selection_command +1838,4706578,"jasmine/models/dynamics.py",2492,0,"",python,selection_command +1839,4706598,"jasmine/models/dynamics.py",2563,0,"",python,selection_command +1840,4706632,"jasmine/models/dynamics.py",2647,0,"",python,selection_command +1841,4706696,"jasmine/models/dynamics.py",2734,0,"",python,selection_command +1842,4706697,"jasmine/models/dynamics.py",2774,0,"",python,selection_command +1843,4706708,"jasmine/models/dynamics.py",2846,0,"",python,selection_command +1844,4706796,"jasmine/models/dynamics.py",2874,0,"",python,selection_command +1845,4706797,"jasmine/models/dynamics.py",2941,0,"",python,selection_command +1846,4706815,"jasmine/models/dynamics.py",3019,0,"",python,selection_command +1847,4706857,"jasmine/models/dynamics.py",3099,0,"",python,selection_command +1848,4706867,"jasmine/models/dynamics.py",3152,0,"",python,selection_command +1849,4706889,"jasmine/models/dynamics.py",3161,0,"",python,selection_command +1850,4706957,"jasmine/models/dynamics.py",3204,0,"",python,selection_command +1851,4706966,"jasmine/models/dynamics.py",3274,0,"",python,selection_command +1852,4707006,"jasmine/models/dynamics.py",3325,0,"",python,selection_command +1853,4707029,"jasmine/models/dynamics.py",3386,0,"",python,selection_command +1854,4707072,"jasmine/models/dynamics.py",3395,0,"",python,selection_command +1855,4707080,"jasmine/models/dynamics.py",3478,0,"",python,selection_command +1856,4707139,"jasmine/models/dynamics.py",3483,0,"",python,selection_command +1857,4707140,"jasmine/models/dynamics.py",3492,0,"",python,selection_command +1858,4707169,"jasmine/models/dynamics.py",3510,0,"",python,selection_command +1859,4707193,"jasmine/models/dynamics.py",3524,0,"",python,selection_command +1860,4707260,"jasmine/models/dynamics.py",3561,0,"",python,selection_command +1861,4707261,"jasmine/models/dynamics.py",3599,0,"",python,selection_command +1862,4707290,"jasmine/models/dynamics.py",3629,0,"",python,selection_command +1863,4707313,"jasmine/models/dynamics.py",3678,0,"",python,selection_command +1864,4707348,"jasmine/models/dynamics.py",3734,0,"",python,selection_command +1865,4707576,"jasmine/models/dynamics.py",3678,0,"",python,selection_command +1866,4707729,"jasmine/models/dynamics.py",3629,0,"",python,selection_command +1867,4707891,"jasmine/models/dynamics.py",3599,0,"",python,selection_command +1868,4708030,"jasmine/models/dynamics.py",3561,0,"",python,selection_command +1869,4708177,"jasmine/models/dynamics.py",3524,0,"",python,selection_command +1870,4708347,"jasmine/models/dynamics.py",3510,0,"",python,selection_command +1871,4708482,"jasmine/models/dynamics.py",3492,0,"",python,selection_command +1872,4708630,"jasmine/models/dynamics.py",3483,0,"",python,selection_command +1873,4708779,"jasmine/models/dynamics.py",3478,0,"",python,selection_command +1874,4709142,"jasmine/models/dynamics.py",3470,13,"",python,content +1875,4709344,"jasmine/models/dynamics.py",3387,0,"",python,selection_command +1876,4709846,"jasmine/models/dynamics.py",3386,0,"",python,selection_command +1877,4709866,"jasmine/models/dynamics.py",3317,0,"",python,selection_command +1878,4709949,"jasmine/models/dynamics.py",3266,0,"",python,selection_command +1879,4709950,"jasmine/models/dynamics.py",3196,0,"",python,selection_command +1880,4709957,"jasmine/models/dynamics.py",3153,0,"",python,selection_command +1881,4710022,"jasmine/models/dynamics.py",3152,0,"",python,selection_command +1882,4710023,"jasmine/models/dynamics.py",3091,0,"",python,selection_command +1883,4710062,"jasmine/models/dynamics.py",3011,0,"",python,selection_command +1884,4710089,"jasmine/models/dynamics.py",2933,0,"",python,selection_command +1885,4710154,"jasmine/models/dynamics.py",2866,0,"",python,selection_command +1886,4710156,"jasmine/models/dynamics.py",2838,0,"",python,selection_command +1887,4710166,"jasmine/models/dynamics.py",2766,0,"",python,selection_command +1888,4710246,"jasmine/models/dynamics.py",2726,0,"",python,selection_command +1889,4710249,"jasmine/models/dynamics.py",2639,0,"",python,selection_command +1890,4710359,"jasmine/models/dynamics.py",2726,0,"",python,selection_command +1891,4710845,"jasmine/models/dynamics.py",2766,0,"",python,selection_command +1892,4710873,"jasmine/models/dynamics.py",2838,0,"",python,selection_command +1893,4710911,"jasmine/models/dynamics.py",2866,0,"",python,selection_command +1894,4710977,"jasmine/models/dynamics.py",2933,0,"",python,selection_command +1895,4710978,"jasmine/models/dynamics.py",3011,0,"",python,selection_command +1896,4710999,"jasmine/models/dynamics.py",3091,0,"",python,selection_command +1897,4711063,"jasmine/models/dynamics.py",3152,0,"",python,selection_command +1898,4711063,"jasmine/models/dynamics.py",3153,0,"",python,selection_command +1899,4711085,"jasmine/models/dynamics.py",3196,0,"",python,selection_command +1900,4711109,"jasmine/models/dynamics.py",3266,0,"",python,selection_command +1901,4711274,"jasmine/models/dynamics.py",3317,0,"",python,selection_command +1902,4711452,"jasmine/models/dynamics.py",3386,0,"",python,selection_command +1903,4711623,"jasmine/models/dynamics.py",3387,0,"",python,selection_command +1904,4711821,"jasmine/models/dynamics.py",3469,0,"\n ",python,content +1905,4712336,"jasmine/models/dynamics.py",3478,0,"r",python,content +1906,4712337,"jasmine/models/dynamics.py",3479,0,"",python,selection_keyboard +1907,4712509,"jasmine/models/dynamics.py",3479,0,"e",python,content +1908,4712510,"jasmine/models/dynamics.py",3480,0,"",python,selection_keyboard +1909,4712576,"jasmine/models/dynamics.py",3480,0,"t",python,content +1910,4712578,"jasmine/models/dynamics.py",3481,0,"",python,selection_keyboard +1911,4712741,"jasmine/models/dynamics.py",3481,0,"u",python,content +1912,4712743,"jasmine/models/dynamics.py",3482,0,"",python,selection_keyboard +1913,4712957,"jasmine/models/dynamics.py",3482,0,"n",python,content +1914,4712959,"jasmine/models/dynamics.py",3483,0,"",python,selection_keyboard +1915,4713319,"jasmine/models/dynamics.py",3482,1,"",python,content +1916,4713430,"jasmine/models/dynamics.py",3482,0,"r",python,content +1917,4713431,"jasmine/models/dynamics.py",3483,0,"",python,selection_keyboard +1918,4713506,"jasmine/models/dynamics.py",3483,0,"n",python,content +1919,4713508,"jasmine/models/dynamics.py",3484,0,"",python,selection_keyboard +1920,4713604,"jasmine/models/dynamics.py",3484,0," ",python,content +1921,4713606,"jasmine/models/dynamics.py",3485,0,"",python,selection_keyboard +1922,4713871,"jasmine/models/dynamics.py",3485,0,"v",python,content +1923,4713872,"jasmine/models/dynamics.py",3486,0,"",python,selection_keyboard +1924,4713956,"jasmine/models/dynamics.py",3486,0,"i",python,content +1925,4713958,"jasmine/models/dynamics.py",3487,0,"",python,selection_keyboard +1926,4714272,"jasmine/models/dynamics.py",3487,0,"d",python,content +1927,4714273,"jasmine/models/dynamics.py",3488,0,"",python,selection_keyboard +1928,4714447,"jasmine/models/dynamics.py",3488,0,"_",python,content +1929,4714449,"jasmine/models/dynamics.py",3489,0,"",python,selection_keyboard +1930,4714702,"jasmine/models/dynamics.py",3489,0,"e",python,content +1931,4714703,"jasmine/models/dynamics.py",3490,0,"",python,selection_keyboard +1932,4715702,"jasmine/models/dynamics.py",3485,5,"vid_embed_BTNM",python,content +1933,4716051,"jasmine/models/dynamics.py",3498,0,"",python,selection_command +1934,4716169,"jasmine/models/dynamics.py",3415,0,"",python,selection_command +1935,4716653,"jasmine/models/dynamics.py",3386,0,"",python,selection_command +1936,4716714,"jasmine/models/dynamics.py",3345,0,"",python,selection_command +1937,4716715,"jasmine/models/dynamics.py",3294,0,"",python,selection_command +1938,4716754,"jasmine/models/dynamics.py",3224,0,"",python,selection_command +1939,4716783,"jasmine/models/dynamics.py",3181,0,"",python,selection_command +1940,4716804,"jasmine/models/dynamics.py",3152,0,"",python,selection_command +1941,4716841,"jasmine/models/dynamics.py",3119,0,"",python,selection_command +1942,4716868,"jasmine/models/dynamics.py",3039,0,"",python,selection_command +1943,4716909,"jasmine/models/dynamics.py",2961,0,"",python,selection_command +1944,4716957,"jasmine/models/dynamics.py",2894,0,"",python,selection_command +1945,4717000,"jasmine/models/dynamics.py",2864,0,"",python,selection_command +1946,4717048,"jasmine/models/dynamics.py",2794,0,"",python,selection_command +1947,4717066,"jasmine/models/dynamics.py",2754,0,"",python,selection_command +1948,4717066,"jasmine/models/dynamics.py",2667,0,"",python,selection_command +1949,4717096,"jasmine/models/dynamics.py",2583,0,"",python,selection_command +1950,4717123,"jasmine/models/dynamics.py",2512,0,"",python,selection_command +1951,4717136,"jasmine/models/dynamics.py",2482,0,"",python,selection_command +1952,4717184,"jasmine/models/dynamics.py",2422,0,"",python,selection_command +1953,4717218,"jasmine/models/dynamics.py",2378,0,"",python,selection_command +1954,4717236,"jasmine/models/dynamics.py",2311,0,"",python,selection_command +1955,4717299,"jasmine/models/dynamics.py",2270,0,"",python,selection_command +1956,4717376,"jasmine/models/dynamics.py",2241,0,"",python,selection_command +1957,4717793,"jasmine/models/dynamics.py",2270,0,"",python,selection_command +1958,4718126,"jasmine/models/dynamics.py",2271,0,"",python,selection_command +1959,4718480,"jasmine/models/dynamics.py",2282,0,"",python,selection_command +1960,4718796,"jasmine/models/dynamics.py",2281,0,"",python,selection_command +1961,4718981,"jasmine/models/dynamics.py",2280,0,"",python,selection_command +1962,4719298,"jasmine/models/dynamics.py",2280,0,",",python,content +1963,4719300,"jasmine/models/dynamics.py",2281,0,"",python,selection_keyboard +1964,4719399,"jasmine/models/dynamics.py",2281,0," ",python,content +1965,4719399,"jasmine/models/dynamics.py",2282,0,"",python,selection_keyboard +1966,4721062,"jasmine/models/dynamics.py",2282,0,"v",python,content +1967,4721063,"jasmine/models/dynamics.py",2283,0,"",python,selection_keyboard +1968,4721130,"jasmine/models/dynamics.py",2283,0,"i",python,content +1969,4721131,"jasmine/models/dynamics.py",2284,0,"",python,selection_keyboard +1970,4721416,"jasmine/models/dynamics.py",2284,0,"d",python,content +1971,4721417,"jasmine/models/dynamics.py",2285,0,"",python,selection_keyboard +1972,4721773,"jasmine/models/dynamics.py",2285,0,"_",python,content +1973,4721774,"jasmine/models/dynamics.py",2286,0,"",python,selection_keyboard +1974,4722453,"jasmine/models/dynamics.py",2286,0,"e",python,content +1975,4722454,"jasmine/models/dynamics.py",2287,0,"",python,selection_keyboard +1976,4722564,"jasmine/models/dynamics.py",2287,0,"m",python,content +1977,4722565,"jasmine/models/dynamics.py",2288,0,"",python,selection_keyboard +1978,4722805,"jasmine/models/dynamics.py",2288,0,"b",python,content +1979,4722806,"jasmine/models/dynamics.py",2289,0,"",python,selection_keyboard +1980,4722902,"jasmine/models/dynamics.py",2289,0,"e",python,content +1981,4722903,"jasmine/models/dynamics.py",2290,0,"",python,selection_keyboard +1982,4723003,"jasmine/models/dynamics.py",2290,0,"d",python,content +1983,4723004,"jasmine/models/dynamics.py",2291,0,"",python,selection_keyboard +1984,4723382,"jasmine/models/dynamics.py",2291,0,"-",python,content +1985,4723383,"jasmine/models/dynamics.py",2292,0,"",python,selection_keyboard +1986,4724462,"jasmine/models/dynamics.py",2291,1,"",python,content +1987,4724836,"jasmine/models/dynamics.py",2291,0,"_",python,content +1988,4724837,"jasmine/models/dynamics.py",2292,0,"",python,selection_keyboard +1989,4725175,"jasmine/models/dynamics.py",2292,0,"B",python,content +1990,4725175,"jasmine/models/dynamics.py",2293,0,"",python,selection_keyboard +1991,4725408,"jasmine/models/dynamics.py",2293,0,"T",python,content +1992,4725409,"jasmine/models/dynamics.py",2294,0,"",python,selection_keyboard +1993,4726329,"jasmine/models/dynamics.py",2294,0,"N",python,content +1994,4726330,"jasmine/models/dynamics.py",2295,0,"",python,selection_keyboard +1995,4726475,"jasmine/models/dynamics.py",2295,0,"M",python,content +1996,4726476,"jasmine/models/dynamics.py",2296,0,"",python,selection_keyboard +1997,4726912,"jasmine/models/dynamics.py",2296,0,"^",python,content +1998,4726913,"jasmine/models/dynamics.py",2297,0,"",python,selection_keyboard +1999,4727475,"jasmine/models/dynamics.py",2296,1,"",python,content +2000,4727476,"jasmine/models/dynamics.py",2296,0,"",python,selection_keyboard +2001,4727568,"jasmine/models/dynamics.py",2295,0,"",python,selection_command +2002,4727978,"jasmine/models/dynamics.py",2298,0,"\n ",python,content +2003,4729399,"jasmine/models/dynamics.py",2307,0,"B",python,content +2004,4729400,"jasmine/models/dynamics.py",2308,0,"",python,selection_keyboard +2005,4730325,"jasmine/models/dynamics.py",2308,0,",",python,content +2006,4730326,"jasmine/models/dynamics.py",2309,0,"",python,selection_keyboard +2007,4730551,"jasmine/models/dynamics.py",2309,0," ",python,content +2008,4730552,"jasmine/models/dynamics.py",2310,0,"",python,selection_keyboard +2009,4730832,"jasmine/models/dynamics.py",2310,0,"T",python,content +2010,4730833,"jasmine/models/dynamics.py",2311,0,"",python,selection_keyboard +2011,4731093,"jasmine/models/dynamics.py",2311,0,",",python,content +2012,4731093,"jasmine/models/dynamics.py",2312,0,"",python,selection_keyboard +2013,4731162,"jasmine/models/dynamics.py",2312,0," ",python,content +2014,4731163,"jasmine/models/dynamics.py",2313,0,"",python,selection_keyboard +2015,4731443,"jasmine/models/dynamics.py",2313,0,"N",python,content +2016,4731443,"jasmine/models/dynamics.py",2314,0,"",python,selection_keyboard +2017,4731791,"jasmine/models/dynamics.py",2314,0,",",python,content +2018,4731792,"jasmine/models/dynamics.py",2315,0,"",python,selection_keyboard +2019,4731898,"jasmine/models/dynamics.py",2315,0," ",python,content +2020,4731899,"jasmine/models/dynamics.py",2316,0,"",python,selection_keyboard +2021,4732102,"jasmine/models/dynamics.py",2316,0,"M",python,content +2022,4732103,"jasmine/models/dynamics.py",2317,0,"",python,selection_keyboard +2023,4732528,"jasmine/models/dynamics.py",2317,0," ",python,content +2024,4732529,"jasmine/models/dynamics.py",2318,0,"",python,selection_keyboard +2025,4732790,"jasmine/models/dynamics.py",2318,0,"=",python,content +2026,4732791,"jasmine/models/dynamics.py",2319,0,"",python,selection_keyboard +2027,4732906,"jasmine/models/dynamics.py",2319,0," ",python,content +2028,4732907,"jasmine/models/dynamics.py",2320,0,"",python,selection_keyboard +2029,4733397,"jasmine/models/dynamics.py",2320,0,"v",python,content +2030,4733398,"jasmine/models/dynamics.py",2321,0,"",python,selection_keyboard +2031,4733513,"jasmine/models/dynamics.py",2321,0,"i",python,content +2032,4733514,"jasmine/models/dynamics.py",2322,0,"",python,selection_keyboard +2033,4733645,"jasmine/models/dynamics.py",2322,0,"d",python,content +2034,4733646,"jasmine/models/dynamics.py",2323,0,"",python,selection_keyboard +2035,4733973,"jasmine/models/dynamics.py",2323,0,"_",python,content +2036,4733973,"jasmine/models/dynamics.py",2324,0,"",python,selection_keyboard +2037,4734448,"jasmine/models/dynamics.py",2320,4,"vid_embed_BTNM",python,content +2038,4734801,"jasmine/models/dynamics.py",2334,0,".",python,content +2039,4734802,"jasmine/models/dynamics.py",2335,0,"",python,selection_keyboard +2040,4735169,"jasmine/models/dynamics.py",2335,0,"s",python,content +2041,4735170,"jasmine/models/dynamics.py",2336,0,"",python,selection_keyboard +2042,4735281,"jasmine/models/dynamics.py",2336,0,"h",python,content +2043,4735282,"jasmine/models/dynamics.py",2337,0,"",python,selection_keyboard +2044,4735416,"jasmine/models/dynamics.py",2337,0,"a",python,content +2045,4735418,"jasmine/models/dynamics.py",2338,0,"",python,selection_keyboard +2046,4735498,"jasmine/models/dynamics.py",2338,0,"p",python,content +2047,4735500,"jasmine/models/dynamics.py",2339,0,"",python,selection_keyboard +2048,4735667,"jasmine/models/dynamics.py",2339,0,"e",python,content +2049,4735668,"jasmine/models/dynamics.py",2340,0,"",python,selection_keyboard +2050,4736016,"jasmine/models/dynamics.py",2339,0,"",python,selection_command +2051,4736315,"jasmine/models/dynamics.py",2381,0,"",python,selection_command +2052,4736479,"jasmine/models/dynamics.py",2448,0,"",python,selection_command +2053,4736644,"jasmine/models/dynamics.py",2492,0,"",python,selection_command +2054,4736999,"jasmine/models/dynamics.py",2540,0,"",python,selection_command +2055,4737150,"jasmine/models/dynamics.py",2582,0,"",python,selection_command +2056,4737728,"jasmine/models/dynamics.py",2583,0,"",python,selection_command +2057,4737913,"jasmine/models/dynamics.py",2540,0,"",python,selection_command +2058,4738362,"jasmine/models/dynamics.py",2493,0,"",python,selection_command +2059,4738531,"jasmine/models/dynamics.py",2449,0,"",python,selection_command +2060,4739801,"jasmine/models/dynamics.py",2382,0,"",python,selection_command +2061,4739947,"jasmine/models/dynamics.py",2339,0,"",python,selection_command +2062,4740311,"jasmine/models/dynamics.py",2283,0,"",python,selection_command +2063,4740639,"jasmine/models/dynamics.py",2298,0,"",python,selection_command +2064,4740747,"jasmine/models/dynamics.py",2297,0,"",python,selection_command +2065,4741536,"jasmine/models/dynamics.py",2296,0,"",python,selection_command +2066,4741738,"jasmine/models/dynamics.py",2296,0,",",python,content +2067,4741740,"jasmine/models/dynamics.py",2297,0,"",python,selection_keyboard +2068,4741850,"jasmine/models/dynamics.py",2297,0," ",python,content +2069,4741851,"jasmine/models/dynamics.py",2298,0,"",python,selection_keyboard +2070,4742022,"jasmine/models/dynamics.py",2298,0,"r",python,content +2071,4742023,"jasmine/models/dynamics.py",2299,0,"",python,selection_keyboard +2072,4742109,"jasmine/models/dynamics.py",2299,0,"n",python,content +2073,4742110,"jasmine/models/dynamics.py",2300,0,"",python,selection_keyboard +2074,4742195,"jasmine/models/dynamics.py",2300,0,"g",python,content +2075,4742195,"jasmine/models/dynamics.py",2301,0,"",python,selection_keyboard +2076,4742530,"jasmine/models/dynamics.py",2300,0,"",python,selection_command +2077,4742759,"jasmine/models/dynamics.py",2344,0,"",python,selection_command +2078,4742942,"jasmine/models/dynamics.py",2404,0,"",python,selection_command +2079,4743091,"jasmine/models/dynamics.py",2455,0,"",python,selection_command +2080,4744495,"jasmine/models/dynamics.py",2337,0,"",python,selection_mouse +2081,4745041,"jasmine/models/dynamics.py",2379,0,"",python,selection_mouse +2082,4745209,"jasmine/models/dynamics.py",2375,10,"_rng_noise",python,selection_mouse +2083,4748012,"jasmine/models/dynamics.py",3101,0,"",python,selection_mouse +2084,4748164,"jasmine/models/dynamics.py",3082,22,"noise_level_embed_BT1M",python,selection_mouse +2085,4749415,"jasmine/models/dynamics.py",3562,0,"",python,selection_mouse +2086,4749427,"jasmine/models/dynamics.py",3561,0,"",python,selection_command +2087,4750343,"jasmine/models/dynamics.py",3562,0,"",python,selection_command +2088,4750519,"jasmine/models/dynamics.py",3562,0,",",python,content +2089,4750521,"jasmine/models/dynamics.py",3563,0,"",python,selection_keyboard +2090,4750597,"jasmine/models/dynamics.py",3563,0," ",python,content +2091,4750598,"jasmine/models/dynamics.py",3564,0,"",python,selection_keyboard +2092,4750906,"jasmine/models/dynamics.py",3564,0,"n",python,content +2093,4750907,"jasmine/models/dynamics.py",3565,0,"",python,selection_keyboard +2094,4751102,"jasmine/models/dynamics.py",3565,0,"o",python,content +2095,4751103,"jasmine/models/dynamics.py",3566,0,"",python,selection_keyboard +2096,4751187,"jasmine/models/dynamics.py",3566,0,"i",python,content +2097,4751188,"jasmine/models/dynamics.py",3567,0,"",python,selection_keyboard +2098,4751252,"jasmine/models/dynamics.py",3567,0,"s",python,content +2099,4751254,"jasmine/models/dynamics.py",3568,0,"",python,selection_keyboard +2100,4751384,"jasmine/models/dynamics.py",3568,0,"e",python,content +2101,4751385,"jasmine/models/dynamics.py",3569,0,"",python,selection_keyboard +2102,4751602,"jasmine/models/dynamics.py",3569,0,"_",python,content +2103,4751604,"jasmine/models/dynamics.py",3570,0,"",python,selection_keyboard +2104,4751899,"jasmine/models/dynamics.py",3570,0,"l",python,content +2105,4751900,"jasmine/models/dynamics.py",3571,0,"",python,selection_keyboard +2106,4752048,"jasmine/models/dynamics.py",3571,0,"e",python,content +2107,4752049,"jasmine/models/dynamics.py",3572,0,"",python,selection_keyboard +2108,4752216,"jasmine/models/dynamics.py",3572,0,"v",python,content +2109,4752217,"jasmine/models/dynamics.py",3573,0,"",python,selection_keyboard +2110,4756005,"jasmine/models/dynamics.py",3564,9,"noise_level_embed_BT1M",python,content +2111,4757177,"jasmine/models/dynamics.py",3018,0,"",python,selection_mouse +2112,4757314,"jasmine/models/dynamics.py",3004,22,"noise_level_embed_B11M",python,selection_mouse +2113,4757878,"jasmine/models/dynamics.py",3097,0,"",python,selection_mouse +2114,4758020,"jasmine/models/dynamics.py",3082,22,"noise_level_embed_BT1M",python,selection_mouse +2115,4759999,"jasmine/models/dynamics.py",3556,0,"",python,selection_mouse +2116,4760147,"jasmine/models/dynamics.py",3548,14,"vid_embed_BTNM",python,selection_mouse +2117,4763459,"jasmine/models/dynamics.py",3472,0,"",python,selection_mouse +2118,4764970,"jasmine/models/dynamics.py",3468,0,"",python,selection_mouse +2119,4766304,"jasmine/models/dynamics.py",3458,0,"",python,selection_mouse +2120,4767694,"jasmine/models/dynamics.py",3458,0,"n",python,content +2121,4767695,"jasmine/models/dynamics.py",3459,0,"",python,selection_keyboard +2122,4767846,"jasmine/models/dynamics.py",3459,0,"o",python,content +2123,4767847,"jasmine/models/dynamics.py",3460,0,"",python,selection_keyboard +2124,4767909,"jasmine/models/dynamics.py",3460,0,"i",python,content +2125,4767911,"jasmine/models/dynamics.py",3461,0,"",python,selection_keyboard +2126,4768007,"jasmine/models/dynamics.py",3461,0,"s",python,content +2127,4768008,"jasmine/models/dynamics.py",3462,0,"",python,selection_keyboard +2128,4768177,"jasmine/models/dynamics.py",3462,0,"e",python,content +2129,4768179,"jasmine/models/dynamics.py",3463,0,"",python,selection_keyboard +2130,4768377,"jasmine/models/dynamics.py",3463,0,"_",python,content +2131,4768379,"jasmine/models/dynamics.py",3464,0,"",python,selection_keyboard +2132,4768573,"jasmine/models/dynamics.py",3464,0,"a",python,content +2133,4768574,"jasmine/models/dynamics.py",3465,0,"",python,selection_keyboard +2134,4768740,"jasmine/models/dynamics.py",3465,0,"u",python,content +2135,4768741,"jasmine/models/dynamics.py",3466,0,"",python,selection_keyboard +2136,4768903,"jasmine/models/dynamics.py",3466,0,"g",python,content +2137,4768904,"jasmine/models/dynamics.py",3467,0,"",python,selection_keyboard +2138,4769114,"jasmine/models/dynamics.py",3467,0,"m",python,content +2139,4769115,"jasmine/models/dynamics.py",3468,0,"",python,selection_keyboard +2140,4769236,"jasmine/models/dynamics.py",3468,0,"e",python,content +2141,4769237,"jasmine/models/dynamics.py",3469,0,"",python,selection_keyboard +2142,4769319,"jasmine/models/dynamics.py",3469,0,"n",python,content +2143,4769320,"jasmine/models/dynamics.py",3470,0,"",python,selection_keyboard +2144,4769488,"jasmine/models/dynamics.py",3470,0,"t",python,content +2145,4769489,"jasmine/models/dynamics.py",3471,0,"",python,selection_keyboard +2146,4769714,"jasmine/models/dynamics.py",3471,0,"e",python,content +2147,4769714,"jasmine/models/dynamics.py",3472,0,"",python,selection_keyboard +2148,4769782,"jasmine/models/dynamics.py",3472,0,"d",python,content +2149,4769783,"jasmine/models/dynamics.py",3473,0,"",python,selection_keyboard +2150,4769973,"jasmine/models/dynamics.py",3473,0,"_",python,content +2151,4769973,"jasmine/models/dynamics.py",3474,0,"",python,selection_keyboard +2152,4771118,"jasmine/models/dynamics.py",3565,0,"",python,selection_mouse +2153,4772604,"jasmine/models/dynamics.py",3564,0,"",python,selection_mouse +2154,4774003,"jasmine/models/dynamics.py",3564,0,"n",python,content +2155,4774004,"jasmine/models/dynamics.py",3565,0,"",python,selection_keyboard +2156,4774219,"jasmine/models/dynamics.py",3565,0,"o",python,content +2157,4774220,"jasmine/models/dynamics.py",3566,0,"",python,selection_keyboard +2158,4774236,"jasmine/models/dynamics.py",3566,0,"i",python,content +2159,4774236,"jasmine/models/dynamics.py",3567,0,"",python,selection_keyboard +2160,4774313,"jasmine/models/dynamics.py",3567,0,"s",python,content +2161,4774314,"jasmine/models/dynamics.py",3568,0,"",python,selection_keyboard +2162,4774484,"jasmine/models/dynamics.py",3568,0,"e",python,content +2163,4774485,"jasmine/models/dynamics.py",3569,0,"",python,selection_keyboard +2164,4774657,"jasmine/models/dynamics.py",3569,0,"_",python,content +2165,4774657,"jasmine/models/dynamics.py",3570,0,"",python,selection_keyboard +2166,4774930,"jasmine/models/dynamics.py",3570,0,"a",python,content +2167,4774931,"jasmine/models/dynamics.py",3571,0,"",python,selection_keyboard +2168,4775369,"jasmine/models/dynamics.py",3571,0,"u",python,content +2169,4775370,"jasmine/models/dynamics.py",3572,0,"",python,selection_keyboard +2170,4775494,"jasmine/models/dynamics.py",3572,0,"g",python,content +2171,4775496,"jasmine/models/dynamics.py",3573,0,"",python,selection_keyboard +2172,4775740,"jasmine/models/dynamics.py",3573,0,"m",python,content +2173,4775741,"jasmine/models/dynamics.py",3574,0,"",python,selection_keyboard +2174,4775852,"jasmine/models/dynamics.py",3574,0,"e",python,content +2175,4775853,"jasmine/models/dynamics.py",3575,0,"",python,selection_keyboard +2176,4775913,"jasmine/models/dynamics.py",3575,0,"n",python,content +2177,4775914,"jasmine/models/dynamics.py",3576,0,"",python,selection_keyboard +2178,4776096,"jasmine/models/dynamics.py",3576,0,"t",python,content +2179,4776097,"jasmine/models/dynamics.py",3577,0,"",python,selection_keyboard +2180,4776249,"jasmine/models/dynamics.py",3577,0,"e",python,content +2181,4776250,"jasmine/models/dynamics.py",3578,0,"",python,selection_keyboard +2182,4776329,"jasmine/models/dynamics.py",3578,0,"d",python,content +2183,4776330,"jasmine/models/dynamics.py",3579,0,"",python,selection_keyboard +2184,4776768,"jasmine/models/dynamics.py",3579,0,"_",python,content +2185,4776770,"jasmine/models/dynamics.py",3580,0,"",python,selection_keyboard +2186,4777616,"jasmine/models/dynamics.py",3677,0,"",python,selection_mouse +2187,4778420,"jasmine/models/dynamics.py",3602,0,"",python,selection_mouse +2188,4778559,"jasmine/models/dynamics.py",3596,22,"noise_level_embed_BT1M",python,selection_mouse +2189,4784234,"jasmine/models/dynamics.py",5691,0,"",python,selection_mouse +2190,4784413,"jasmine/models/dynamics.py",5691,3,"qrt",python,selection_mouse +2191,4784429,"jasmine/models/dynamics.py",5691,7,"qrt_one",python,selection_mouse +2192,4784487,"jasmine/models/dynamics.py",5691,10,"qrt_one_mi",python,selection_mouse +2193,4784488,"jasmine/models/dynamics.py",5691,13,"qrt_one_minus",python,selection_mouse +2194,4784488,"jasmine/models/dynamics.py",5691,15,"qrt_one_minus *",python,selection_mouse +2195,4784496,"jasmine/models/dynamics.py",5691,18,"qrt_one_minus * vi",python,selection_mouse +2196,4784512,"jasmine/models/dynamics.py",5691,23,"qrt_one_minus * vid_emb",python,selection_mouse +2197,4784569,"jasmine/models/dynamics.py",5691,25,"qrt_one_minus * vid_embed",python,selection_mouse +2198,4784570,"jasmine/models/dynamics.py",5691,27,"qrt_one_minus * vid_embed_B",python,selection_mouse +2199,4784570,"jasmine/models/dynamics.py",5691,29,"qrt_one_minus * vid_embed_BTN",python,selection_mouse +2200,4784579,"jasmine/models/dynamics.py",5691,31,"qrt_one_minus * vid_embed_BTNM ",python,selection_mouse +2201,4784598,"jasmine/models/dynamics.py",5691,33,"qrt_one_minus * vid_embed_BTNM + ",python,selection_mouse +2202,4784652,"jasmine/models/dynamics.py",5691,35,"qrt_one_minus * vid_embed_BTNM + sq",python,selection_mouse +2203,4784652,"jasmine/models/dynamics.py",5691,37,"qrt_one_minus * vid_embed_BTNM + sqrt",python,selection_mouse +2204,4784653,"jasmine/models/dynamics.py",5691,38,"qrt_one_minus * vid_embed_BTNM + sqrt_",python,selection_mouse +2205,4784666,"jasmine/models/dynamics.py",5691,40,"qrt_one_minus * vid_embed_BTNM + sqrt_no",python,selection_mouse +2206,4784679,"jasmine/models/dynamics.py",5691,42,"qrt_one_minus * vid_embed_BTNM + sqrt_nois",python,selection_mouse +2207,4784740,"jasmine/models/dynamics.py",5691,44,"qrt_one_minus * vid_embed_BTNM + sqrt_noise ",python,selection_mouse +2208,4784740,"jasmine/models/dynamics.py",5691,46,"qrt_one_minus * vid_embed_BTNM + sqrt_noise * ",python,selection_mouse +2209,4784741,"jasmine/models/dynamics.py",5691,48,"qrt_one_minus * vid_embed_BTNM + sqrt_noise * no",python,selection_mouse +2210,4784745,"jasmine/models/dynamics.py",5691,50,"qrt_one_minus * vid_embed_BTNM + sqrt_noise * nois",python,selection_mouse +2211,4784762,"jasmine/models/dynamics.py",5691,51,"qrt_one_minus * vid_embed_BTNM + sqrt_noise * noise",python,selection_mouse +2212,4784813,"jasmine/models/dynamics.py",5691,52,"qrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_",python,selection_mouse +2213,4784814,"jasmine/models/dynamics.py",5691,53,"qrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_B",python,selection_mouse +2214,4784814,"jasmine/models/dynamics.py",5691,54,"qrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BT",python,selection_mouse +2215,4784828,"jasmine/models/dynamics.py",5691,55,"qrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTN",python,selection_mouse +2216,4784886,"jasmine/models/dynamics.py",5691,56,"qrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM",python,selection_mouse +2217,4785304,"jasmine/models/dynamics.py",5691,56,"",python,content +2218,4785376,"jasmine/models/dynamics.py",5690,1,"",python,content +2219,4785773,"jasmine/models/dynamics.py",5690,0,"_",python,content +2220,4785774,"jasmine/models/dynamics.py",5691,0,"",python,selection_keyboard +2221,4786019,"jasmine/models/dynamics.py",5691,0,"a",python,content +2222,4786021,"jasmine/models/dynamics.py",5692,0,"",python,selection_keyboard +2223,4786097,"jasmine/models/dynamics.py",5692,0,"p",python,content +2224,4786099,"jasmine/models/dynamics.py",5693,0,"",python,selection_keyboard +2225,4786214,"jasmine/models/dynamics.py",5693,0,"p",python,content +2226,4786216,"jasmine/models/dynamics.py",5694,0,"",python,selection_keyboard +2227,4786382,"jasmine/models/dynamics.py",5694,0,"l",python,content +2228,4786384,"jasmine/models/dynamics.py",5695,0,"",python,selection_keyboard +2229,4786465,"jasmine/models/dynamics.py",5695,0,"y",python,content +2230,4786466,"jasmine/models/dynamics.py",5696,0,"",python,selection_keyboard +2231,4788289,"jasmine/models/dynamics.py",5696,0,"_",python,content +2232,4788290,"jasmine/models/dynamics.py",5697,0,"",python,selection_keyboard +2233,4790194,"jasmine/models/dynamics.py",5689,0,"",python,selection_command +2234,4790414,"jasmine/models/dynamics.py",5690,0,"",python,selection_command +2235,4790593,"jasmine/models/dynamics.py",5690,0,"s",python,content +2236,4790594,"jasmine/models/dynamics.py",5691,0,"",python,selection_keyboard +2237,4790789,"jasmine/models/dynamics.py",5691,0,"e",python,content +2238,4790789,"jasmine/models/dynamics.py",5692,0,"",python,selection_keyboard +2239,4791201,"jasmine/models/dynamics.py",5692,0,"l",python,content +2240,4791202,"jasmine/models/dynamics.py",5693,0,"",python,selection_keyboard +2241,4791303,"jasmine/models/dynamics.py",5693,0,"f",python,content +2242,4791304,"jasmine/models/dynamics.py",5694,0,"",python,selection_keyboard +2243,4791404,"jasmine/models/dynamics.py",5694,0,".",python,content +2244,4791405,"jasmine/models/dynamics.py",5695,0,"",python,selection_keyboard +2245,4793125,"jasmine/models/dynamics.py",5695,7,"_apply_noise_augmentation",python,content +2246,4793791,"jasmine/models/dynamics.py",5720,0,"()",python,content +2247,4793792,"jasmine/models/dynamics.py",5721,0,"",python,selection_keyboard +2248,4793968,"jasmine/models/dynamics.py",5721,0,"v",python,content +2249,4793968,"jasmine/models/dynamics.py",5722,0,"",python,selection_keyboard +2250,4794065,"jasmine/models/dynamics.py",5722,0,"i",python,content +2251,4794067,"jasmine/models/dynamics.py",5723,0,"",python,selection_keyboard +2252,4794195,"jasmine/models/dynamics.py",5723,0,"d",python,content +2253,4794196,"jasmine/models/dynamics.py",5724,0,"",python,selection_keyboard +2254,4794509,"jasmine/models/dynamics.py",5724,0,"_",python,content +2255,4794510,"jasmine/models/dynamics.py",5725,0,"",python,selection_keyboard +2256,4795931,"jasmine/models/dynamics.py",5721,4,"vid_embed_BTNM",python,content +2257,4796236,"jasmine/models/dynamics.py",5735,0,",",python,content +2258,4796237,"jasmine/models/dynamics.py",5736,0,"",python,selection_keyboard +2259,4796284,"jasmine/models/dynamics.py",5736,0," ",python,content +2260,4796285,"jasmine/models/dynamics.py",5737,0,"",python,selection_keyboard +2261,4797529,"jasmine/models/dynamics.py",5737,0,"r",python,content +2262,4797530,"jasmine/models/dynamics.py",5738,0,"",python,selection_keyboard +2263,4797623,"jasmine/models/dynamics.py",5738,0,"n",python,content +2264,4797624,"jasmine/models/dynamics.py",5739,0,"",python,selection_keyboard +2265,4797743,"jasmine/models/dynamics.py",5739,0,"g",python,content +2266,4797744,"jasmine/models/dynamics.py",5740,0,"",python,selection_keyboard +2267,4798219,"jasmine/models/dynamics.py",5739,0,"",python,selection_command +2268,4798400,"jasmine/models/dynamics.py",5664,0,"",python,selection_command +2269,4799062,"jasmine/models/dynamics.py",5662,0,"",python,selection_command +2270,4799608,"jasmine/models/dynamics.py",5595,68," sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2271,4799992,"jasmine/models/dynamics.py",5544,119," sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2272,4800538,"jasmine/models/dynamics.py",5474,189," one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2273,4800542,"jasmine/models/dynamics.py",5431,232," # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2274,4800546,"jasmine/models/dynamics.py",5430,233,"\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2275,4800628,"jasmine/models/dynamics.py",5369,294," noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2276,4800629,"jasmine/models/dynamics.py",5289,374," noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2277,4800720,"jasmine/models/dynamics.py",5211,452," noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2278,4800895,"jasmine/models/dynamics.py",5144,519," noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2279,4801053,"jasmine/models/dynamics.py",5116,547," ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2280,4801203,"jasmine/models/dynamics.py",5044,619," (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2281,4801359,"jasmine/models/dynamics.py",5004,659," noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2282,4801512,"jasmine/models/dynamics.py",4917,746," # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2283,4801776,"jasmine/models/dynamics.py",4833,830," # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2284,4801924,"jasmine/models/dynamics.py",4762,901," noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2285,4802105,"jasmine/models/dynamics.py",4752,911," )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2286,4802232,"jasmine/models/dynamics.py",4672,991," _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2287,4802460,"jasmine/models/dynamics.py",4628,1035," noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2288,4805922,"jasmine/models/dynamics.py",4561,1102," rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))",python,selection_command +2289,4811155,"jasmine/models/dynamics.py",4561,1103,"",python,content +2290,4811666,"jasmine/models/dynamics.py",4530,0,"",python,selection_command +2291,4811978,"jasmine/models/dynamics.py",4561,0,"",python,selection_command +2292,4812943,"jasmine/models/dynamics.py",4562,0,"",python,selection_command +2293,4813379,"jasmine/models/dynamics.py",4563,0,"",python,selection_command +2294,4813886,"jasmine/models/dynamics.py",4564,0,"",python,selection_command +2295,4813899,"jasmine/models/dynamics.py",4565,0,"",python,selection_command +2296,4813953,"jasmine/models/dynamics.py",4566,0,"",python,selection_command +2297,4813959,"jasmine/models/dynamics.py",4567,0,"",python,selection_command +2298,4814019,"jasmine/models/dynamics.py",4568,0,"",python,selection_command +2299,4814027,"jasmine/models/dynamics.py",4569,0,"",python,selection_command +2300,4814084,"jasmine/models/dynamics.py",4570,0,"",python,selection_command +2301,4814091,"jasmine/models/dynamics.py",4571,0,"",python,selection_command +2302,4814146,"jasmine/models/dynamics.py",4572,0,"",python,selection_command +2303,4814148,"jasmine/models/dynamics.py",4573,0,"",python,selection_command +2304,4814169,"jasmine/models/dynamics.py",4574,0,"",python,selection_command +2305,4814209,"jasmine/models/dynamics.py",4575,0,"",python,selection_command +2306,4814238,"jasmine/models/dynamics.py",4576,0,"",python,selection_command +2307,4814263,"jasmine/models/dynamics.py",4577,0,"",python,selection_command +2308,4814320,"jasmine/models/dynamics.py",4578,0,"",python,selection_command +2309,4814323,"jasmine/models/dynamics.py",4579,0,"",python,selection_command +2310,4814360,"jasmine/models/dynamics.py",4580,0,"",python,selection_command +2311,4814423,"jasmine/models/dynamics.py",4581,0,"",python,selection_command +2312,4814435,"jasmine/models/dynamics.py",4582,0,"",python,selection_command +2313,4814605,"jasmine/models/dynamics.py",4583,0,"",python,selection_command +2314,4814745,"jasmine/models/dynamics.py",4584,0,"",python,selection_command +2315,4815228,"jasmine/models/dynamics.py",4584,0,",",python,content +2316,4815229,"jasmine/models/dynamics.py",4585,0,"",python,selection_keyboard +2317,4815299,"jasmine/models/dynamics.py",4585,0," ",python,content +2318,4815300,"jasmine/models/dynamics.py",4586,0,"",python,selection_keyboard +2319,4815512,"jasmine/models/dynamics.py",4586,0,"n",python,content +2320,4815513,"jasmine/models/dynamics.py",4587,0,"",python,selection_keyboard +2321,4815695,"jasmine/models/dynamics.py",4587,0,"o",python,content +2322,4815697,"jasmine/models/dynamics.py",4588,0,"",python,selection_keyboard +2323,4815772,"jasmine/models/dynamics.py",4588,0,"i",python,content +2324,4815774,"jasmine/models/dynamics.py",4589,0,"",python,selection_keyboard +2325,4816926,"jasmine/models/dynamics.py",4589,0,"s",python,content +2326,4816927,"jasmine/models/dynamics.py",4590,0,"",python,selection_keyboard +2327,4818134,"jasmine/models/dynamics.py",4590,0,"e",python,content +2328,4818135,"jasmine/models/dynamics.py",4591,0,"",python,selection_keyboard +2329,4818766,"jasmine/models/dynamics.py",4591,0,"_",python,content +2330,4818768,"jasmine/models/dynamics.py",4592,0,"",python,selection_keyboard +2331,4819065,"jasmine/models/dynamics.py",4592,0,"l",python,content +2332,4819066,"jasmine/models/dynamics.py",4593,0,"",python,selection_keyboard +2333,4819200,"jasmine/models/dynamics.py",4593,0,"e",python,content +2334,4819200,"jasmine/models/dynamics.py",4594,0,"",python,selection_keyboard +2335,4819348,"jasmine/models/dynamics.py",4594,0,"v",python,content +2336,4819349,"jasmine/models/dynamics.py",4595,0,"",python,selection_keyboard +2337,4819510,"jasmine/models/dynamics.py",4595,0,"e",python,content +2338,4819512,"jasmine/models/dynamics.py",4596,0,"",python,selection_keyboard +2339,4819577,"jasmine/models/dynamics.py",4596,0,"l",python,content +2340,4819579,"jasmine/models/dynamics.py",4597,0,"",python,selection_keyboard +2341,4819884,"jasmine/models/dynamics.py",4597,0,"_",python,content +2342,4819885,"jasmine/models/dynamics.py",4598,0,"",python,selection_keyboard +2343,4820332,"jasmine/models/dynamics.py",4598,0,"e",python,content +2344,4820333,"jasmine/models/dynamics.py",4599,0,"",python,selection_keyboard +2345,4820380,"jasmine/models/dynamics.py",4599,0,"m",python,content +2346,4820381,"jasmine/models/dynamics.py",4600,0,"",python,selection_keyboard +2347,4820569,"jasmine/models/dynamics.py",4600,0,"b",python,content +2348,4820570,"jasmine/models/dynamics.py",4601,0,"",python,selection_keyboard +2349,4820633,"jasmine/models/dynamics.py",4601,0,"e",python,content +2350,4820635,"jasmine/models/dynamics.py",4602,0,"",python,selection_keyboard +2351,4820757,"jasmine/models/dynamics.py",4602,0,"d",python,content +2352,4820759,"jasmine/models/dynamics.py",4603,0,"",python,selection_keyboard +2353,4821991,"jasmine/models/dynamics.py",4603,0,"_",python,content +2354,4821993,"jasmine/models/dynamics.py",4604,0,"",python,selection_keyboard +2355,4822444,"jasmine/models/dynamics.py",4604,0,"B",python,content +2356,4822445,"jasmine/models/dynamics.py",4605,0,"",python,selection_keyboard +2357,4822638,"jasmine/models/dynamics.py",4605,0,"T",python,content +2358,4822639,"jasmine/models/dynamics.py",4606,0,"",python,selection_keyboard +2359,4823065,"jasmine/models/dynamics.py",4606,0,"N",python,content +2360,4823066,"jasmine/models/dynamics.py",4607,0,"",python,selection_keyboard +2361,4823510,"jasmine/models/dynamics.py",4606,1,"",python,content +2362,4823946,"jasmine/models/dynamics.py",4606,0,"1",python,content +2363,4823947,"jasmine/models/dynamics.py",4607,0,"",python,selection_keyboard +2364,4824476,"jasmine/models/dynamics.py",4607,0,"M",python,content +2365,4824477,"jasmine/models/dynamics.py",4608,0,"",python,selection_keyboard +2366,4825130,"jasmine/models/dynamics.py",4607,0,"",python,selection_command +2367,4825208,"jasmine/models/dynamics.py",4561,0,"",python,selection_command +2368,4826133,"jasmine/models/dynamics.py",4561,1,"",python,content +2369,4826168,"jasmine/models/dynamics.py",4569,0,"",python,selection_command +2370,4826227,"jasmine/models/dynamics.py",4538,0,"",python,selection_command +2371,4826629,"jasmine/models/dynamics.py",4539,0,"",python,selection_command +2372,4827126,"jasmine/models/dynamics.py",4540,0,"",python,selection_command +2373,4827149,"jasmine/models/dynamics.py",4541,0,"",python,selection_command +2374,4827214,"jasmine/models/dynamics.py",4542,0,"",python,selection_command +2375,4827215,"jasmine/models/dynamics.py",4543,0,"",python,selection_command +2376,4827276,"jasmine/models/dynamics.py",4544,0,"",python,selection_command +2377,4827277,"jasmine/models/dynamics.py",4545,0,"",python,selection_command +2378,4827774,"jasmine/models/dynamics.py",4544,0,"",python,selection_command +2379,4827989,"jasmine/models/dynamics.py",4544,7,"",python,content +2380,4828262,"jasmine/models/dynamics.py",4544,6,"",python,content +2381,4829552,"jasmine/models/dynamics.py",4544,0,"A",python,content +2382,4829553,"jasmine/models/dynamics.py",4545,0,"",python,selection_keyboard +2383,4829682,"jasmine/models/dynamics.py",4545,0,"p",python,content +2384,4829683,"jasmine/models/dynamics.py",4546,0,"",python,selection_keyboard +2385,4829833,"jasmine/models/dynamics.py",4546,0,"p",python,content +2386,4829834,"jasmine/models/dynamics.py",4547,0,"",python,selection_keyboard +2387,4829950,"jasmine/models/dynamics.py",4547,0,"l",python,content +2388,4829966,"jasmine/models/dynamics.py",4548,0,"",python,selection_keyboard +2389,4830046,"jasmine/models/dynamics.py",4548,0,"y",python,content +2390,4830047,"jasmine/models/dynamics.py",4549,0,"",python,selection_keyboard +2391,4830180,"jasmine/models/dynamics.py",4549,0," ",python,content +2392,4830181,"jasmine/models/dynamics.py",4550,0,"",python,selection_keyboard +2393,4830349,"jasmine/models/dynamics.py",4550,0,"n",python,content +2394,4830350,"jasmine/models/dynamics.py",4551,0,"",python,selection_keyboard +2395,4830514,"jasmine/models/dynamics.py",4551,0,"o",python,content +2396,4830515,"jasmine/models/dynamics.py",4552,0,"",python,selection_keyboard +2397,4830581,"jasmine/models/dynamics.py",4552,0,"i",python,content +2398,4830583,"jasmine/models/dynamics.py",4553,0,"",python,selection_keyboard +2399,4830629,"jasmine/models/dynamics.py",4553,0,"s",python,content +2400,4830630,"jasmine/models/dynamics.py",4554,0,"",python,selection_keyboard +2401,4830829,"jasmine/models/dynamics.py",4554,0,"e",python,content +2402,4830830,"jasmine/models/dynamics.py",4555,0,"",python,selection_keyboard +2403,4830898,"jasmine/models/dynamics.py",4555,0," ",python,content +2404,4830899,"jasmine/models/dynamics.py",4556,0,"",python,selection_keyboard +2405,4831092,"jasmine/models/dynamics.py",4556,0,"a",python,content +2406,4831093,"jasmine/models/dynamics.py",4557,0,"",python,selection_keyboard +2407,4831111,"jasmine/models/dynamics.py",4557,0,"u",python,content +2408,4831112,"jasmine/models/dynamics.py",4558,0,"",python,selection_keyboard +2409,4831324,"jasmine/models/dynamics.py",4558,0,"g",python,content +2410,4831325,"jasmine/models/dynamics.py",4559,0,"",python,selection_keyboard +2411,4831583,"jasmine/models/dynamics.py",4559,0,"m",python,content +2412,4831584,"jasmine/models/dynamics.py",4560,0,"",python,selection_keyboard +2413,4831650,"jasmine/models/dynamics.py",4560,0,"e",python,content +2414,4831650,"jasmine/models/dynamics.py",4561,0,"",python,selection_keyboard +2415,4831776,"jasmine/models/dynamics.py",4561,0,"n",python,content +2416,4831777,"jasmine/models/dynamics.py",4562,0,"",python,selection_keyboard +2417,4831907,"jasmine/models/dynamics.py",4562,0,"t",python,content +2418,4831907,"jasmine/models/dynamics.py",4563,0,"",python,selection_keyboard +2419,4832214,"jasmine/models/dynamics.py",4563,0,"a",python,content +2420,4832215,"jasmine/models/dynamics.py",4564,0,"",python,selection_keyboard +2421,4832344,"jasmine/models/dynamics.py",4564,0,"t",python,content +2422,4832345,"jasmine/models/dynamics.py",4565,0,"",python,selection_keyboard +2423,4832402,"jasmine/models/dynamics.py",4565,0,"i",python,content +2424,4832403,"jasmine/models/dynamics.py",4566,0,"",python,selection_keyboard +2425,4832478,"jasmine/models/dynamics.py",4566,0,"o",python,content +2426,4832479,"jasmine/models/dynamics.py",4567,0,"",python,selection_keyboard +2427,4832628,"jasmine/models/dynamics.py",4567,0,"n",python,content +2428,4832628,"jasmine/models/dynamics.py",4568,0,"",python,selection_keyboard +2429,4833046,"jasmine/models/dynamics.py",4568,0," ",python,content +2430,4833047,"jasmine/models/dynamics.py",4569,0,"",python,selection_keyboard +2431,4833289,"jasmine/models/dynamics.py",4568,0,"",python,selection_command +2432,4833452,"jasmine/models/dynamics.py",4611,0,"",python,selection_command +2433,4833839,"jasmine/models/dynamics.py",4568,0,"",python,selection_command +2434,4834259,"jasmine/models/dynamics.py",4611,0,"",python,selection_command +2435,4834498,"jasmine/models/dynamics.py",4673,0,"\n ",python,content +2436,4834710,"jasmine/models/dynamics.py",4674,8,"",python,content +2437,4837634,"jasmine/models/dynamics.py",4890,0,"",python,selection_mouse +2438,4837636,"jasmine/models/dynamics.py",4889,0,"",python,selection_command +2439,4839740,"jasmine/models/dynamics.py",3934,0,"",python,selection_mouse +2440,4841632,"jasmine/models/dynamics.py",3934,1,"_",python,content +2441,4841888,"jasmine/models/dynamics.py",3935,0,"",python,selection_command +2442,4842067,"jasmine/models/dynamics.py",3936,0,"",python,selection_command +2443,4842257,"jasmine/models/dynamics.py",3937,0,"",python,selection_command +2444,4842621,"jasmine/models/dynamics.py",3937,1,"_",python,content +2445,4842897,"jasmine/models/dynamics.py",3938,0,"",python,selection_command +2446,4843068,"jasmine/models/dynamics.py",3939,0,"",python,selection_command +2447,4843368,"jasmine/models/dynamics.py",3940,0,"",python,selection_command +2448,4843743,"jasmine/models/dynamics.py",3940,1,"_",python,content +2449,4846637,"jasmine/models/dynamics.py",3932,0,"",python,selection_mouse +2450,4846822,"jasmine/models/dynamics.py",3932,3,", _",python,selection_mouse +2451,4846845,"jasmine/models/dynamics.py",3932,4,", _,",python,selection_mouse +2452,4846857,"jasmine/models/dynamics.py",3932,5,", _, ",python,selection_mouse +2453,4846891,"jasmine/models/dynamics.py",3932,6,", _, _",python,selection_mouse +2454,4846892,"jasmine/models/dynamics.py",3932,7,", _, _,",python,selection_mouse +2455,4846906,"jasmine/models/dynamics.py",3932,8,", _, _, ",python,selection_mouse +2456,4846972,"jasmine/models/dynamics.py",3932,9,", _, _, _",python,selection_mouse +2457,4848432,"jasmine/models/dynamics.py",3932,9,"",python,content +2458,4849530,"jasmine/models/dynamics.py",3955,0,"",python,selection_mouse +2459,4849531,"jasmine/models/dynamics.py",3954,0,"",python,selection_command +2460,4850058,"jasmine/models/dynamics.py",3955,0,"",python,selection_command +2461,4850838,"jasmine/models/dynamics.py",3955,0,"[]",python,content +2462,4850839,"jasmine/models/dynamics.py",3956,0,"",python,selection_keyboard +2463,4851033,"jasmine/models/dynamics.py",3956,0,"0",python,content +2464,4851035,"jasmine/models/dynamics.py",3957,0,"",python,selection_keyboard +2465,4851515,"jasmine/models/dynamics.py",3956,0,"",python,selection_command +2466,4866796,"jasmine/models/dynamics.py",3579,0,"",python,selection_mouse +2467,4868998,"jasmine/models/dynamics.py",3549,69," return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2468,4869304,"jasmine/models/dynamics.py",3450,168," noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2469,4869795,"jasmine/models/dynamics.py",3449,169,"\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2470,4869863,"jasmine/models/dynamics.py",3380,238," sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2471,4869871,"jasmine/models/dynamics.py",3329,289," sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2472,4869885,"jasmine/models/dynamics.py",3259,359," one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2473,4869956,"jasmine/models/dynamics.py",3216,402," # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2474,4869958,"jasmine/models/dynamics.py",3215,403,"\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2475,4869981,"jasmine/models/dynamics.py",3154,464," noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2476,4870036,"jasmine/models/dynamics.py",3074,544," noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2477,4870047,"jasmine/models/dynamics.py",2996,622," noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2478,4870066,"jasmine/models/dynamics.py",2929,689," noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2479,4870149,"jasmine/models/dynamics.py",2901,717," ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2480,4870149,"jasmine/models/dynamics.py",2829,789," (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2481,4870150,"jasmine/models/dynamics.py",2789,829," noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2482,4870181,"jasmine/models/dynamics.py",2702,916," # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2483,4870229,"jasmine/models/dynamics.py",2618,1000," # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2484,4870252,"jasmine/models/dynamics.py",2547,1071," noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2485,4870273,"jasmine/models/dynamics.py",2537,1081," )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2486,4870379,"jasmine/models/dynamics.py",2457,1161," _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2487,4870530,"jasmine/models/dynamics.py",2413,1205," noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2488,4870706,"jasmine/models/dynamics.py",2346,1272," rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2489,4870813,"jasmine/models/dynamics.py",2304,1314," B, T, N, M = vid_embed_BTNM.shape\n rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2490,4870952,"jasmine/models/dynamics.py",2242,1376," def _apply_noise_augmentation(self, vid_embed_BTNM, rng):\n B, T, N, M = vid_embed_BTNM.shape\n rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2491,4871127,"jasmine/models/dynamics.py",2241,1377,"\n def _apply_noise_augmentation(self, vid_embed_BTNM, rng):\n B, T, N, M = vid_embed_BTNM.shape\n rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_command +2492,4871296,"jasmine/models/dynamics.py",2241,0,"",python,selection_command +2493,4873786,"jasmine/models/dynamics.py",6694,0,"",python,selection_mouse +2494,4873788,"jasmine/models/dynamics.py",6693,0,"",python,selection_command +2495,4874417,"jasmine/models/dynamics.py",6694,0,"\n\n def _apply_noise_augmentation(self, vid_embed_BTNM, rng):\n B, T, N, M = vid_embed_BTNM.shape\n rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,content +2496,4874466,"jasmine/models/dynamics.py",6695,0,"",python,selection_command +2497,4875287,"jasmine/models/dynamics.py",6701,0,"",python,selection_command +2498,4875790,"jasmine/models/dynamics.py",6763,0,"",python,selection_command +2499,4875830,"jasmine/models/dynamics.py",6805,0,"",python,selection_command +2500,4875852,"jasmine/models/dynamics.py",6872,0,"",python,selection_command +2501,4875872,"jasmine/models/dynamics.py",6916,0,"",python,selection_command +2502,4875936,"jasmine/models/dynamics.py",6996,0,"",python,selection_command +2503,4875937,"jasmine/models/dynamics.py",7006,0,"",python,selection_command +2504,4875959,"jasmine/models/dynamics.py",7077,0,"",python,selection_command +2505,4876023,"jasmine/models/dynamics.py",7161,0,"",python,selection_command +2506,4876023,"jasmine/models/dynamics.py",7248,0,"",python,selection_command +2507,4876080,"jasmine/models/dynamics.py",7288,0,"",python,selection_command +2508,4876081,"jasmine/models/dynamics.py",7360,0,"",python,selection_command +2509,4876116,"jasmine/models/dynamics.py",7388,0,"",python,selection_command +2510,4876150,"jasmine/models/dynamics.py",7455,0,"",python,selection_command +2511,4876172,"jasmine/models/dynamics.py",7533,0,"",python,selection_command +2512,4876229,"jasmine/models/dynamics.py",7613,0,"",python,selection_command +2513,4876243,"jasmine/models/dynamics.py",7669,0,"",python,selection_command +2514,4876264,"jasmine/models/dynamics.py",7675,0,"",python,selection_command +2515,4899922,"jasmine/models/dynamics.py",7580,0,"",python,selection_mouse +2516,4900636,"jasmine/models/dynamics.py",7489,0,"",python,selection_mouse +2517,4902326,"jasmine/models/dynamics.py",7044,0,"",python,selection_mouse +2518,4940066,"jasmine/models/dynamics.py",8611,0,"",python,selection_mouse +2519,4940920,"jasmine/models/dynamics.py",8638,0,"",python,selection_mouse +2520,4940923,"jasmine/models/dynamics.py",8637,0,"",python,selection_command +2521,4941801,"jasmine/models/dynamics.py",8638,0,"",python,selection_mouse +2522,4941807,"jasmine/models/dynamics.py",8637,0,"",python,selection_command +2523,4942411,"jasmine/models/dynamics.py",8606,0,"",python,selection_mouse +2524,4943314,"jasmine/models/dynamics.py",8605,0,"",python,selection_command +2525,4946776,"jasmine/models/dynamics.py",8604,0,"",python,selection_command +2526,4947367,"jasmine/models/dynamics.py",8524,0,"",python,selection_mouse +2527,4947372,"jasmine/models/dynamics.py",8523,0,"",python,selection_command +2528,4948758,"jasmine/models/dynamics.py",8524,0,"\n ",python,content +2529,4951225,"jasmine/models/dynamics.py",8533,0,"v",python,content +2530,4951227,"jasmine/models/dynamics.py",8534,0,"",python,selection_keyboard +2531,4951320,"jasmine/models/dynamics.py",8534,0,"i",python,content +2532,4951322,"jasmine/models/dynamics.py",8535,0,"",python,selection_keyboard +2533,4952145,"jasmine/models/dynamics.py",8535,0,"d",python,content +2534,4952146,"jasmine/models/dynamics.py",8536,0,"",python,selection_keyboard +2535,4952424,"jasmine/models/dynamics.py",8536,0,"_",python,content +2536,4952426,"jasmine/models/dynamics.py",8537,0,"",python,selection_keyboard +2537,4952709,"jasmine/models/dynamics.py",8537,0,"e",python,content +2538,4952710,"jasmine/models/dynamics.py",8538,0,"",python,selection_keyboard +2539,4953006,"jasmine/models/dynamics.py",8538,0,"m",python,content +2540,4953007,"jasmine/models/dynamics.py",8539,0,"",python,selection_keyboard +2541,4953262,"jasmine/models/dynamics.py",8539,0,"b",python,content +2542,4953263,"jasmine/models/dynamics.py",8540,0,"",python,selection_keyboard +2543,4953379,"jasmine/models/dynamics.py",8540,0,"e",python,content +2544,4953381,"jasmine/models/dynamics.py",8541,0,"",python,selection_keyboard +2545,4953484,"jasmine/models/dynamics.py",8541,0,"d",python,content +2546,4953485,"jasmine/models/dynamics.py",8542,0,"",python,selection_keyboard +2547,4954400,"jasmine/models/dynamics.py",8533,9,"vid_embed_BTNM",python,content +2548,4954866,"jasmine/models/dynamics.py",8547,0," ",python,content +2549,4954867,"jasmine/models/dynamics.py",8548,0,"",python,selection_keyboard +2550,4955080,"jasmine/models/dynamics.py",8548,0,"=",python,content +2551,4955081,"jasmine/models/dynamics.py",8549,0,"",python,selection_keyboard +2552,4955171,"jasmine/models/dynamics.py",8549,0," ",python,content +2553,4955172,"jasmine/models/dynamics.py",8550,0,"",python,selection_keyboard +2554,4956272,"jasmine/models/dynamics.py",8550,0,"_",python,content +2555,4956274,"jasmine/models/dynamics.py",8551,0,"",python,selection_keyboard +2556,4956449,"jasmine/models/dynamics.py",8551,0,"a",python,content +2557,4956451,"jasmine/models/dynamics.py",8552,0,"",python,selection_keyboard +2558,4956512,"jasmine/models/dynamics.py",8552,0,"p",python,content +2559,4956513,"jasmine/models/dynamics.py",8553,0,"",python,selection_keyboard +2560,4956660,"jasmine/models/dynamics.py",8553,0,"p",python,content +2561,4956661,"jasmine/models/dynamics.py",8554,0,"",python,selection_keyboard +2562,4957581,"jasmine/models/dynamics.py",8550,0,"s",python,content +2563,4957582,"jasmine/models/dynamics.py",8551,0,"",python,selection_keyboard +2564,4957753,"jasmine/models/dynamics.py",8551,0,"l",python,content +2565,4957754,"jasmine/models/dynamics.py",8552,0,"",python,selection_keyboard +2566,4957780,"jasmine/models/dynamics.py",8552,0,"e",python,content +2567,4957781,"jasmine/models/dynamics.py",8553,0,"",python,selection_keyboard +2568,4958245,"jasmine/models/dynamics.py",8552,1,"",python,content +2569,4958367,"jasmine/models/dynamics.py",8551,1,"",python,content +2570,4958536,"jasmine/models/dynamics.py",8551,0,"e",python,content +2571,4958536,"jasmine/models/dynamics.py",8552,0,"",python,selection_keyboard +2572,4958617,"jasmine/models/dynamics.py",8552,0,"l",python,content +2573,4958618,"jasmine/models/dynamics.py",8553,0,"",python,selection_keyboard +2574,4958670,"jasmine/models/dynamics.py",8553,0,"f",python,content +2575,4958671,"jasmine/models/dynamics.py",8554,0,"",python,selection_keyboard +2576,4958782,"jasmine/models/dynamics.py",8554,0,".",python,content +2577,4958783,"jasmine/models/dynamics.py",8555,0,"",python,selection_keyboard +2578,4960205,"jasmine/models/dynamics.py",8559,0,"l",python,content +2579,4960207,"jasmine/models/dynamics.py",8560,0,"",python,selection_keyboard +2580,4960414,"jasmine/models/dynamics.py",8560,0,"y",python,content +2581,4960415,"jasmine/models/dynamics.py",8561,0,"",python,selection_keyboard +2582,4961000,"jasmine/models/dynamics.py",8555,6,"_apply_noise_augmentation",python,content +2583,4961556,"jasmine/models/dynamics.py",8580,0,"()",python,content +2584,4961557,"jasmine/models/dynamics.py",8581,0,"",python,selection_keyboard +2585,4962174,"jasmine/models/dynamics.py",8581,0,"v",python,content +2586,4962174,"jasmine/models/dynamics.py",8582,0,"",python,selection_keyboard +2587,4962218,"jasmine/models/dynamics.py",8582,0,"o",python,content +2588,4962219,"jasmine/models/dynamics.py",8583,0,"",python,selection_keyboard +2589,4962759,"jasmine/models/dynamics.py",8581,2,"video_tokens_BTN",python,content +2590,4963234,"jasmine/models/dynamics.py",8597,0,",",python,content +2591,4963234,"jasmine/models/dynamics.py",8598,0,"",python,selection_keyboard +2592,4963316,"jasmine/models/dynamics.py",8598,0," ",python,content +2593,4963317,"jasmine/models/dynamics.py",8599,0,"",python,selection_keyboard +2594,4963851,"jasmine/models/dynamics.py",8599,0,"r",python,content +2595,4963852,"jasmine/models/dynamics.py",8600,0,"",python,selection_keyboard +2596,4963927,"jasmine/models/dynamics.py",8600,0,"n",python,content +2597,4963929,"jasmine/models/dynamics.py",8601,0,"",python,selection_keyboard +2598,4981210,"jasmine/models/dynamics.py",8600,1,"",python,content +2599,4981360,"jasmine/models/dynamics.py",8599,1,"",python,content +2600,4981484,"jasmine/models/dynamics.py",8598,1,"",python,content +2601,4981618,"jasmine/models/dynamics.py",8597,1,"",python,content +2602,4983646,"jasmine/models/dynamics.py",8642,0,"",python,selection_mouse +2603,4984216,"jasmine/models/dynamics.py",8590,0,"",python,selection_mouse +2604,4984330,"jasmine/models/dynamics.py",8581,16,"video_tokens_BTN",python,selection_mouse +2605,4985013,"jasmine/models/dynamics.py",8268,0,"",python,selection_mouse +2606,4985583,"jasmine/models/dynamics.py",8161,0,"",python,selection_mouse +2607,4990262,"jasmine/models/dynamics.py",8265,0,"",python,selection_mouse +2608,4991666,"jasmine/train_dynamics.py",0,0,"",python,tab +2609,4993083,"jasmine/train_dynamics.py",18811,0,"",python,selection_mouse +2610,4994007,"jasmine/train_dynamics.py",18632,0,"",python,selection_mouse +2611,4994545,"jasmine/train_dynamics.py",18631,0,"",python,selection_command +2612,4995577,"jasmine/train_dynamics.py",18899,0,"",python,selection_command +2613,4997192,"jasmine/train_dynamics.py",19181,0,"",python,selection_command +2614,4998682,"jasmine/train_dynamics.py",19187,0,"",python,selection_command +2615,4999186,"jasmine/train_dynamics.py",19215,0,"",python,selection_command +2616,4999623,"jasmine/train_dynamics.py",19242,0,"",python,selection_command +2617,5000374,"jasmine/train_dynamics.py",19251,0,"",python,selection_command +2618,5003246,"jasmine/models/dynamics.py",0,0,"",python,tab +2619,5004636,"jasmine/models/dynamics.py",8285,0,"",python,selection_mouse +2620,5008247,"jasmine/models/dynamics.py",8598,0,"",python,selection_mouse +2621,5009013,"jasmine/models/dynamics.py",8597,0,"",python,selection_command +2622,5009263,"jasmine/models/dynamics.py",8597,0,",",python,content +2623,5009264,"jasmine/models/dynamics.py",8598,0,"",python,selection_keyboard +2624,5009383,"jasmine/models/dynamics.py",8598,0," ",python,content +2625,5009384,"jasmine/models/dynamics.py",8599,0,"",python,selection_keyboard +2626,5009573,"jasmine/models/dynamics.py",8599,0,"b",python,content +2627,5009574,"jasmine/models/dynamics.py",8600,0,"",python,selection_keyboard +2628,5009697,"jasmine/models/dynamics.py",8600,0,"a",python,content +2629,5009699,"jasmine/models/dynamics.py",8601,0,"",python,selection_keyboard +2630,5009878,"jasmine/models/dynamics.py",8601,0,"t",python,content +2631,5009879,"jasmine/models/dynamics.py",8602,0,"",python,selection_keyboard +2632,5010048,"jasmine/models/dynamics.py",8602,0,"c",python,content +2633,5010049,"jasmine/models/dynamics.py",8603,0,"",python,selection_keyboard +2634,5010189,"jasmine/models/dynamics.py",8603,0,"h",python,content +2635,5010190,"jasmine/models/dynamics.py",8604,0,"",python,selection_keyboard +2636,5010682,"jasmine/models/dynamics.py",8604,0,"[]",python,content +2637,5010684,"jasmine/models/dynamics.py",8605,0,"",python,selection_keyboard +2638,5011010,"jasmine/models/dynamics.py",8605,0,"""""",python,content +2639,5011011,"jasmine/models/dynamics.py",8606,0,"",python,selection_keyboard +2640,5011939,"jasmine/models/dynamics.py",8606,0,"r",python,content +2641,5011940,"jasmine/models/dynamics.py",8607,0,"",python,selection_keyboard +2642,5012142,"jasmine/models/dynamics.py",8607,0,"n",python,content +2643,5012143,"jasmine/models/dynamics.py",8608,0,"",python,selection_keyboard +2644,5012235,"jasmine/models/dynamics.py",8608,0,"g",python,content +2645,5012236,"jasmine/models/dynamics.py",8609,0,"",python,selection_keyboard +2646,5012767,"jasmine/models/dynamics.py",8608,0,"",python,selection_command +2647,5012986,"jasmine/models/dynamics.py",8655,0,"",python,selection_command +2648,5017542,"jasmine/models/dynamics.py",8547,0,"",python,selection_mouse +2649,5018726,"jasmine/models/dynamics.py",8547,0,",",python,content +2650,5018727,"jasmine/models/dynamics.py",8548,0,"",python,selection_keyboard +2651,5018804,"jasmine/models/dynamics.py",8548,0," ",python,content +2652,5018806,"jasmine/models/dynamics.py",8549,0,"",python,selection_keyboard +2653,5021092,"jasmine/models/dynamics.py",8549,0,"n",python,content +2654,5021094,"jasmine/models/dynamics.py",8550,0,"",python,selection_keyboard +2655,5021270,"jasmine/models/dynamics.py",8550,0,"o",python,content +2656,5021271,"jasmine/models/dynamics.py",8551,0,"",python,selection_keyboard +2657,5021335,"jasmine/models/dynamics.py",8551,0,"i",python,content +2658,5021336,"jasmine/models/dynamics.py",8552,0,"",python,selection_keyboard +2659,5021393,"jasmine/models/dynamics.py",8552,0,"s",python,content +2660,5021394,"jasmine/models/dynamics.py",8553,0,"",python,selection_keyboard +2661,5021538,"jasmine/models/dynamics.py",8553,0,"e",python,content +2662,5021539,"jasmine/models/dynamics.py",8554,0,"",python,selection_keyboard +2663,5021736,"jasmine/models/dynamics.py",8554,0,"_",python,content +2664,5021737,"jasmine/models/dynamics.py",8555,0,"",python,selection_keyboard +2665,5022005,"jasmine/models/dynamics.py",8555,0,"l",python,content +2666,5022007,"jasmine/models/dynamics.py",8556,0,"",python,selection_keyboard +2667,5022100,"jasmine/models/dynamics.py",8556,0,"e",python,content +2668,5022101,"jasmine/models/dynamics.py",8557,0,"",python,selection_keyboard +2669,5022255,"jasmine/models/dynamics.py",8557,0,"v",python,content +2670,5022256,"jasmine/models/dynamics.py",8558,0,"",python,selection_keyboard +2671,5022386,"jasmine/models/dynamics.py",8558,0,"e",python,content +2672,5022387,"jasmine/models/dynamics.py",8559,0,"",python,selection_keyboard +2673,5022419,"jasmine/models/dynamics.py",8559,0,"l",python,content +2674,5022420,"jasmine/models/dynamics.py",8560,0,"",python,selection_keyboard +2675,5022707,"jasmine/models/dynamics.py",8560,0,"_",python,content +2676,5022707,"jasmine/models/dynamics.py",8561,0,"",python,selection_keyboard +2677,5023156,"jasmine/models/dynamics.py",8561,0,"e",python,content +2678,5023157,"jasmine/models/dynamics.py",8562,0,"",python,selection_keyboard +2679,5023223,"jasmine/models/dynamics.py",8562,0,"m",python,content +2680,5023224,"jasmine/models/dynamics.py",8563,0,"",python,selection_keyboard +2681,5023453,"jasmine/models/dynamics.py",8563,0,"b",python,content +2682,5023453,"jasmine/models/dynamics.py",8564,0,"",python,selection_keyboard +2683,5023556,"jasmine/models/dynamics.py",8564,0,"e",python,content +2684,5023557,"jasmine/models/dynamics.py",8565,0,"",python,selection_keyboard +2685,5023651,"jasmine/models/dynamics.py",8565,0,"d",python,content +2686,5023652,"jasmine/models/dynamics.py",8566,0,"",python,selection_keyboard +2687,5023872,"jasmine/models/dynamics.py",8566,0,"_",python,content +2688,5023873,"jasmine/models/dynamics.py",8567,0,"",python,selection_keyboard +2689,5024404,"jasmine/models/dynamics.py",8567,0,"B",python,content +2690,5024404,"jasmine/models/dynamics.py",8568,0,"",python,selection_keyboard +2691,5024604,"jasmine/models/dynamics.py",8568,0,"T",python,content +2692,5024605,"jasmine/models/dynamics.py",8569,0,"",python,selection_keyboard +2693,5025059,"jasmine/models/dynamics.py",8569,0,"1",python,content +2694,5025060,"jasmine/models/dynamics.py",8570,0,"",python,selection_keyboard +2695,5025738,"jasmine/models/dynamics.py",8570,0,"M",python,content +2696,5025739,"jasmine/models/dynamics.py",8571,0,"",python,selection_keyboard +2697,5026318,"jasmine/models/dynamics.py",8570,0,"",python,selection_command +2698,5026464,"jasmine/models/dynamics.py",8679,0,"",python,selection_command +2699,5026940,"jasmine/models/dynamics.py",8726,0,"",python,selection_command +2700,5027664,"jasmine/models/dynamics.py",8725,0,"",python,selection_command +2701,5028166,"jasmine/models/dynamics.py",8724,0,"",python,selection_command +2702,5028198,"jasmine/models/dynamics.py",8723,0,"",python,selection_command +2703,5028257,"jasmine/models/dynamics.py",8722,0,"",python,selection_command +2704,5028257,"jasmine/models/dynamics.py",8721,0,"",python,selection_command +2705,5028316,"jasmine/models/dynamics.py",8720,0,"",python,selection_command +2706,5028317,"jasmine/models/dynamics.py",8719,0,"",python,selection_command +2707,5028369,"jasmine/models/dynamics.py",8718,0,"",python,selection_command +2708,5028370,"jasmine/models/dynamics.py",8717,0,"",python,selection_command +2709,5029259,"jasmine/models/dynamics.py",8717,0,"n",python,content +2710,5029260,"jasmine/models/dynamics.py",8718,0,"",python,selection_keyboard +2711,5029455,"jasmine/models/dynamics.py",8718,0,"o",python,content +2712,5029455,"jasmine/models/dynamics.py",8719,0,"",python,selection_keyboard +2713,5029483,"jasmine/models/dynamics.py",8719,0,"i",python,content +2714,5029484,"jasmine/models/dynamics.py",8720,0,"",python,selection_keyboard +2715,5029536,"jasmine/models/dynamics.py",8720,0,"s",python,content +2716,5029537,"jasmine/models/dynamics.py",8721,0,"",python,selection_keyboard +2717,5029701,"jasmine/models/dynamics.py",8721,0,"e",python,content +2718,5029702,"jasmine/models/dynamics.py",8722,0,"",python,selection_keyboard +2719,5029900,"jasmine/models/dynamics.py",8722,0,"_",python,content +2720,5029900,"jasmine/models/dynamics.py",8723,0,"",python,selection_keyboard +2721,5030419,"jasmine/models/dynamics.py",8723,0,"#",python,content +2722,5030420,"jasmine/models/dynamics.py",8724,0,"",python,selection_keyboard +2723,5030683,"jasmine/models/dynamics.py",8724,0,",",python,content +2724,5030684,"jasmine/models/dynamics.py",8725,0,"",python,selection_keyboard +2725,5030762,"jasmine/models/dynamics.py",8725,0," ",python,content +2726,5030763,"jasmine/models/dynamics.py",8726,0,"",python,selection_keyboard +2727,5031150,"jasmine/models/dynamics.py",8725,1,"",python,content +2728,5031252,"jasmine/models/dynamics.py",8724,1,"",python,content +2729,5031392,"jasmine/models/dynamics.py",8723,1,"",python,content +2730,5031882,"jasmine/models/dynamics.py",8723,0,"l",python,content +2731,5031883,"jasmine/models/dynamics.py",8724,0,"",python,selection_keyboard +2732,5031998,"jasmine/models/dynamics.py",8724,0,"e",python,content +2733,5031999,"jasmine/models/dynamics.py",8725,0,"",python,selection_keyboard +2734,5033262,"jasmine/models/dynamics.py",8725,0,",",python,content +2735,5033263,"jasmine/models/dynamics.py",8726,0,"",python,selection_keyboard +2736,5033286,"jasmine/models/dynamics.py",8726,0," ",python,content +2737,5033287,"jasmine/models/dynamics.py",8727,0,"",python,selection_keyboard +2738,5033757,"jasmine/models/dynamics.py",8726,0,"",python,selection_command +2739,5033907,"jasmine/models/dynamics.py",8725,0,"",python,selection_command +2740,5034176,"jasmine/models/dynamics.py",8725,0,"v",python,content +2741,5034177,"jasmine/models/dynamics.py",8726,0,"",python,selection_keyboard +2742,5034949,"jasmine/models/dynamics.py",8726,0,"e",python,content +2743,5034950,"jasmine/models/dynamics.py",8727,0,"",python,selection_keyboard +2744,5035057,"jasmine/models/dynamics.py",8727,0,"l",python,content +2745,5035058,"jasmine/models/dynamics.py",8728,0,"",python,selection_keyboard +2746,5036465,"jasmine/models/dynamics.py",8717,11,"noise_level_embed_BT1M",python,content +2747,5038326,"jasmine/models/dynamics.py",8660,0,"",python,selection_mouse +2748,5038941,"jasmine/models/dynamics.py",8659,1,"",python,content +2749,5039026,"jasmine/models/dynamics.py",8659,0,"2",python,content +2750,5039027,"jasmine/models/dynamics.py",8660,0,"",python,selection_keyboard +2751,5040429,"jasmine/models/dynamics.py",8831,0,"",python,selection_mouse +2752,5040980,"jasmine/models/dynamics.py",8830,1,"",python,content +2753,5041029,"jasmine/models/dynamics.py",8830,0,"2",python,content +2754,5041030,"jasmine/models/dynamics.py",8831,0,"",python,selection_keyboard +2755,5042402,"jasmine/models/dynamics.py",8879,0,"",python,selection_mouse +2756,5045383,"jasmine/models/dynamics.py",8878,1,"",python,content +2757,5045467,"jasmine/models/dynamics.py",8878,0,"2",python,content +2758,5045468,"jasmine/models/dynamics.py",8879,0,"",python,selection_keyboard +2759,5064019,"jasmine/models/dynamics.py",8878,1,"",python,content +2760,5064923,"jasmine/models/dynamics.py",8878,0,"1",python,content +2761,5064924,"jasmine/models/dynamics.py",8879,0,"",python,selection_keyboard +2762,5065022,"jasmine/models/dynamics.py",8878,0,"",python,selection_command +2763,5065184,"jasmine/models/dynamics.py",8877,0,"",python,selection_command +2764,5065386,"jasmine/models/dynamics.py",8876,0,"",python,selection_command +2765,5065529,"jasmine/models/dynamics.py",8876,0,"1",python,content +2766,5065530,"jasmine/models/dynamics.py",8877,0,"",python,selection_keyboard +2767,5075403,"jasmine/models/dynamics.py",8795,0,"",python,selection_mouse +2768,5075899,"jasmine/models/dynamics.py",8794,1,"",python,content +2769,5076008,"jasmine/models/dynamics.py",8794,0,"2",python,content +2770,5076009,"jasmine/models/dynamics.py",8795,0,"",python,selection_keyboard +2771,5076210,"jasmine/models/dynamics.py",8854,0,"",python,selection_command +2772,5077929,"jasmine/models/dynamics.py",8857,0,"",python,selection_command +2773,5078472,"jasmine/models/dynamics.py",8868,0,"",python,selection_command +2774,5078876,"jasmine/models/dynamics.py",8867,1,"",python,content +2775,5078959,"jasmine/models/dynamics.py",8867,0,"2",python,content +2776,5078961,"jasmine/models/dynamics.py",8868,0,"",python,selection_keyboard +2777,5083414,"jasmine/models/dynamics.py",8831,0,"",python,selection_mouse +2778,5084034,"jasmine/models/dynamics.py",8881,0,"",python,selection_mouse +2779,5172526,"jasmine/models/dynamics.py",8942,0,"",python,selection_mouse +2780,5173194,"jasmine/models/dynamics.py",8941,0,"",python,selection_mouse +2781,5176329,"jasmine/models/dynamics.py",8774,0,"",python,selection_mouse +2782,5176954,"jasmine/models/dynamics.py",8524,0,"",python,selection_mouse +2783,5180751,"jasmine/models/dynamics.py",5941,0,"",python,selection_mouse +2784,5187767,"jasmine/models/dynamics.py",7496,0,"",python,selection_mouse +2785,5187877,"jasmine/models/dynamics.py",7488,17,"noise_level_embed",python,selection_mouse +2786,5191742,"jasmine/models/dynamics.py",2241,0,"",python,selection_mouse +2787,5192371,"jasmine/models/dynamics.py",2240,0,"",python,selection_mouse +2788,5193402,"jasmine/models/dynamics.py",2239,0,"",python,selection_command +2789,5193994,"jasmine/models/dynamics.py",2231,9," )",python,selection_command +2790,5194235,"jasmine/models/dynamics.py",2173,67," self.noise_buckets, self.model_dim, rngs=rngs\n )",python,selection_command +2791,5194381,"jasmine/models/dynamics.py",2129,111," self.noise_level_embed = nnx.Embed(\n self.noise_buckets, self.model_dim, rngs=rngs\n )",python,selection_command +2792,5194649,"jasmine/models/dynamics.py",2129,0,"",python,selection_command +2793,5197958,"jasmine/models/dynamics.py",6684,0,"",python,selection_mouse +2794,5197965,"jasmine/models/dynamics.py",6683,0,"",python,selection_command +2795,5198661,"jasmine/models/dynamics.py",6694,0,"",python,selection_mouse +2796,5198671,"jasmine/models/dynamics.py",6693,0,"",python,selection_command +2797,5199366,"jasmine/models/dynamics.py",6694,0,"\n self.noise_level_embed = nnx.Embed(\n self.noise_buckets, self.model_dim, rngs=rngs\n )",python,content +2798,5199402,"jasmine/models/dynamics.py",6703,0,"",python,selection_command +2799,5226804,"jasmine/models/dynamics.py",1161,0,"",python,selection_mouse +2800,5228176,"jasmine/models/dynamics.py",1114,0,"",python,selection_command +2801,5231240,"jasmine/models/dynamics.py",5834,0,"",python,selection_mouse +2802,5231826,"jasmine/models/dynamics.py",5843,0,"\n self.max_noise_level = max_noise_level\n self.noise_buckets = noise_buckets",python,content +2803,5231846,"jasmine/models/dynamics.py",5852,0,"",python,selection_command +2804,5235494,"jasmine/models/dynamics.py",654,0,"",python,selection_mouse +2805,5236608,"jasmine/models/dynamics.py",622,0,"",python,selection_command +2806,5239652,"jasmine/models/dynamics.py",5428,0,"",python,selection_mouse +2807,5240436,"jasmine/models/dynamics.py",5438,0,"\n max_noise_level: float,\n noise_buckets: int,",python,content +2808,5240445,"jasmine/models/dynamics.py",5447,0,"",python,selection_command +2809,5247955,"jasmine/genie.py",0,0,"",python,tab +2810,5253932,"jasmine/genie.py",4521,0,"",python,selection_mouse +2811,5254698,"jasmine/genie.py",4617,0,"",python,selection_mouse +2812,5255335,"jasmine/genie.py",4564,0,"",python,selection_mouse +2813,5259930,"jasmine/genie.py",5259,0,"",python,selection_mouse +2814,5261490,"jasmine/genie.py",5258,0,"",python,selection_mouse +2815,5262369,"jasmine/genie.py",5263,0,"",python,selection_mouse +2816,5262963,"jasmine/genie.py",5365,0,"",python,selection_mouse +2817,5263527,"jasmine/genie.py",5318,0,"",python,selection_mouse +2818,5264169,"jasmine/genie.py",5266,0,"",python,selection_mouse +2819,5264981,"jasmine/genie.py",5320,0,"",python,selection_mouse +2820,5266980,"jasmine/train_dynamics.py",0,0,"",python,tab +2821,5271877,"jasmine/models/dynamics.py",0,0,"",python,tab +2822,5279056,"jasmine/genie.py",0,0,"",python,tab +2823,5287646,"jasmine/genie.py",13033,0,"",python,selection_mouse +2824,5288602,"jasmine/genie.py",13061,0,"",python,selection_mouse +2825,5289180,"jasmine/genie.py",13037,0,"",python,selection_mouse +2826,5290744,"jasmine/genie.py",13013,0,"",python,selection_mouse +2827,5290748,"jasmine/genie.py",13012,0,"",python,selection_command +2828,5310335,"jasmine/genie.py",13026,0,"",python,selection_command +2829,5310488,"jasmine/genie.py",13039,0,"",python,selection_command +2830,5310620,"jasmine/genie.py",13052,0,"",python,selection_command +2831,5310758,"jasmine/genie.py",13104,0,"",python,selection_command +2832,5310907,"jasmine/genie.py",13163,0,"",python,selection_command +2833,5311031,"jasmine/genie.py",13230,0,"",python,selection_command +2834,5311196,"jasmine/genie.py",13276,0,"",python,selection_command +2835,5311332,"jasmine/genie.py",13277,0,"",python,selection_command +2836,5311608,"jasmine/genie.py",13278,0,"",python,selection_command +2837,5312106,"jasmine/genie.py",13279,0,"",python,selection_command +2838,5312167,"jasmine/genie.py",13280,0,"",python,selection_command +2839,5312190,"jasmine/genie.py",13281,0,"",python,selection_command +2840,5312291,"jasmine/genie.py",13282,0,"",python,selection_command +2841,5313593,"jasmine/genie.py",13360,0,"",python,selection_command +2842,5314300,"jasmine/genie.py",13282,0,"",python,selection_command +2843,5314630,"jasmine/genie.py",13236,0,"",python,selection_command +2844,5314800,"jasmine/genie.py",13169,0,"",python,selection_command +2845,5315288,"jasmine/genie.py",13236,0,"",python,selection_command +2846,5315850,"jasmine/genie.py",13218,45," noise_bucket_idx_B11 = jnp.floor(",python,selection_command +2847,5316056,"jasmine/genie.py",13218,123," noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 / self.max_noise_level) * self.noise_buckets",python,selection_command +2848,5316176,"jasmine/genie.py",13218,155," noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 / self.max_noise_level) * self.noise_buckets\n ).astype(jnp.int32)",python,selection_command +2849,5316313,"jasmine/genie.py",13218,0,"",python,selection_command +2850,5316440,"jasmine/genie.py",13151,0,"",python,selection_command +2851,5316611,"jasmine/genie.py",13092,0,"",python,selection_command +2852,5316900,"jasmine/genie.py",13151,0,"",python,selection_command +2853,5317079,"jasmine/genie.py",13217,0,"\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 / self.max_noise_level) * self.noise_buckets\n ).astype(jnp.int32)",python,content +2854,5317115,"jasmine/genie.py",13230,0,"",python,selection_command +2855,5318001,"jasmine/genie.py",13163,0,"",python,selection_command +2856,5318158,"jasmine/genie.py",13230,0,"",python,selection_command +2857,5318456,"jasmine/genie.py",13231,0,"",python,selection_command +2858,5318594,"jasmine/genie.py",13277,0,"",python,selection_command +2859,5318746,"jasmine/genie.py",13278,0,"",python,selection_command +2860,5318932,"jasmine/genie.py",13279,0,"",python,selection_command +2861,5319097,"jasmine/genie.py",13280,0,"",python,selection_command +2862,5319243,"jasmine/genie.py",13281,0,"",python,selection_command +2863,5320653,"jasmine/genie.py",13282,0,"",python,selection_command +2864,5321155,"jasmine/genie.py",13283,0,"",python,selection_command +2865,5321211,"jasmine/genie.py",13284,0,"",python,selection_command +2866,5321218,"jasmine/genie.py",13285,0,"",python,selection_command +2867,5321279,"jasmine/genie.py",13286,0,"",python,selection_command +2868,5321282,"jasmine/genie.py",13287,0,"",python,selection_command +2869,5321305,"jasmine/genie.py",13288,0,"",python,selection_command +2870,5321359,"jasmine/genie.py",13289,0,"",python,selection_command +2871,5321370,"jasmine/genie.py",13290,0,"",python,selection_command +2872,5321437,"jasmine/genie.py",13291,0,"",python,selection_command +2873,5321438,"jasmine/genie.py",13292,0,"",python,selection_command +2874,5321448,"jasmine/genie.py",13293,0,"",python,selection_command +2875,5321510,"jasmine/genie.py",13294,0,"",python,selection_command +2876,5321517,"jasmine/genie.py",13295,0,"",python,selection_command +2877,5321581,"jasmine/genie.py",13296,0,"",python,selection_command +2878,5321583,"jasmine/genie.py",13297,0,"",python,selection_command +2879,5321717,"jasmine/genie.py",13298,0,"",python,selection_command +2880,5322701,"jasmine/genie.py",13297,1,"",python,content +2881,5323696,"jasmine/genie.py",13297,0,"*",python,content +2882,5323697,"jasmine/genie.py",13298,0,"",python,selection_keyboard +2883,5324253,"jasmine/genie.py",13299,0,"",python,selection_command +2884,5325099,"jasmine/genie.py",13304,0,"",python,selection_command +2885,5325638,"jasmine/genie.py",13304,15,"",python,content +2886,5326354,"jasmine/genie.py",13304,0,"n",python,content +2887,5326355,"jasmine/genie.py",13305,0,"",python,selection_keyboard +2888,5326522,"jasmine/genie.py",13305,0,"o",python,content +2889,5326524,"jasmine/genie.py",13306,0,"",python,selection_keyboard +2890,5326610,"jasmine/genie.py",13306,0,"i",python,content +2891,5326611,"jasmine/genie.py",13307,0,"",python,selection_keyboard +2892,5326614,"jasmine/genie.py",13307,0,"s",python,content +2893,5326615,"jasmine/genie.py",13308,0,"",python,selection_keyboard +2894,5326792,"jasmine/genie.py",13308,0,"e",python,content +2895,5326793,"jasmine/genie.py",13309,0,"",python,selection_keyboard +2896,5326945,"jasmine/genie.py",13309,0,"_",python,content +2897,5326946,"jasmine/genie.py",13310,0,"",python,selection_keyboard +2898,5327173,"jasmine/genie.py",13310,0,"v",python,content +2899,5327174,"jasmine/genie.py",13311,0,"",python,selection_keyboard +2900,5327350,"jasmine/genie.py",13311,0,"u",python,content +2901,5327351,"jasmine/genie.py",13312,0,"",python,selection_keyboard +2902,5327752,"jasmine/genie.py",13311,1,"",python,content +2903,5328018,"jasmine/genie.py",13310,1,"",python,content +2904,5328285,"jasmine/genie.py",13310,0,"b",python,content +2905,5328286,"jasmine/genie.py",13311,0,"",python,selection_keyboard +2906,5328471,"jasmine/genie.py",13311,0,"u",python,content +2907,5328472,"jasmine/genie.py",13312,0,"",python,selection_keyboard +2908,5328893,"jasmine/genie.py",13312,0,"c",python,content +2909,5328894,"jasmine/genie.py",13313,0,"",python,selection_keyboard +2910,5328975,"jasmine/genie.py",13313,0,"k",python,content +2911,5328976,"jasmine/genie.py",13314,0,"",python,selection_keyboard +2912,5329103,"jasmine/genie.py",13314,0,"e",python,content +2913,5329104,"jasmine/genie.py",13315,0,"",python,selection_keyboard +2914,5329322,"jasmine/genie.py",13315,0,"t",python,content +2915,5329324,"jasmine/genie.py",13316,0,"",python,selection_keyboard +2916,5329518,"jasmine/genie.py",13316,0,"s",python,content +2917,5329520,"jasmine/genie.py",13317,0,"",python,selection_keyboard +2918,5330255,"jasmine/genie.py",13319,0,"",python,selection_command +2919,5330445,"jasmine/genie.py",13320,0,"",python,selection_command +2920,5331074,"jasmine/genie.py",13320,5,"",python,content +2921,5331582,"jasmine/genie.py",13320,1,"",python,content +2922,5331747,"jasmine/genie.py",13320,13,"",python,content +2923,5332300,"jasmine/genie.py",13320,0," ",python,content +2924,5332301,"jasmine/genie.py",13321,0,"",python,selection_keyboard +2925,5332519,"jasmine/genie.py",13321,0,"m",python,content +2926,5332520,"jasmine/genie.py",13322,0,"",python,selection_keyboard +2927,5332659,"jasmine/genie.py",13322,0,"a",python,content +2928,5332660,"jasmine/genie.py",13323,0,"",python,selection_keyboard +2929,5332901,"jasmine/genie.py",13323,0,"x",python,content +2930,5332902,"jasmine/genie.py",13324,0,"",python,selection_keyboard +2931,5333296,"jasmine/genie.py",13323,1,"",python,content +2932,5333427,"jasmine/genie.py",13322,1,"",python,content +2933,5333560,"jasmine/genie.py",13321,1,"",python,content +2934,5333676,"jasmine/genie.py",13320,1,"",python,content +2935,5334068,"jasmine/genie.py",13320,0," ",python,content +2936,5334068,"jasmine/genie.py",13321,0,"",python,selection_keyboard +2937,5334218,"jasmine/genie.py",13321,0,"s",python,content +2938,5334218,"jasmine/genie.py",13322,0,"",python,selection_keyboard +2939,5334383,"jasmine/genie.py",13322,0,"e",python,content +2940,5334384,"jasmine/genie.py",13323,0,"",python,selection_keyboard +2941,5334468,"jasmine/genie.py",13323,0,"l",python,content +2942,5334469,"jasmine/genie.py",13324,0,"",python,selection_keyboard +2943,5334570,"jasmine/genie.py",13324,0,"f",python,content +2944,5334572,"jasmine/genie.py",13325,0,"",python,selection_keyboard +2945,5334648,"jasmine/genie.py",13325,0,".",python,content +2946,5334650,"jasmine/genie.py",13326,0,"",python,selection_keyboard +2947,5334968,"jasmine/genie.py",13326,0,"m",python,content +2948,5334969,"jasmine/genie.py",13327,0,"",python,selection_keyboard +2949,5335031,"jasmine/genie.py",13327,0,"a",python,content +2950,5335032,"jasmine/genie.py",13328,0,"",python,selection_keyboard +2951,5335284,"jasmine/genie.py",13328,0,"x",python,content +2952,5335286,"jasmine/genie.py",13329,0,"",python,selection_keyboard +2953,5335501,"jasmine/genie.py",13329,0,"_",python,content +2954,5335501,"jasmine/genie.py",13330,0,"",python,selection_keyboard +2955,5335807,"jasmine/genie.py",13330,0,"n",python,content +2956,5335807,"jasmine/genie.py",13331,0,"",python,selection_keyboard +2957,5335952,"jasmine/genie.py",13331,0,"o",python,content +2958,5335954,"jasmine/genie.py",13332,0,"",python,selection_keyboard +2959,5336037,"jasmine/genie.py",13332,0,"i",python,content +2960,5336038,"jasmine/genie.py",13333,0,"",python,selection_keyboard +2961,5336062,"jasmine/genie.py",13333,0,"s",python,content +2962,5336063,"jasmine/genie.py",13334,0,"",python,selection_keyboard +2963,5336196,"jasmine/genie.py",13334,0,"e",python,content +2964,5336197,"jasmine/genie.py",13335,0,"",python,selection_keyboard +2965,5336314,"jasmine/genie.py",13326,9,"max_noise_level",python,content +2966,5336719,"jasmine/genie.py",13340,0,"",python,selection_command +2967,5337228,"jasmine/genie.py",13372,0,"",python,selection_command +2968,5337369,"jasmine/genie.py",13418,0,"",python,selection_command +2969,5338292,"jasmine/genie.py",13374,124,"",python,content +2970,5338343,"jasmine/genie.py",13386,0,"",python,selection_command +2971,5338863,"jasmine/genie.py",13374,32,"",python,content +2972,5338919,"jasmine/genie.py",13386,0,"",python,selection_command +2973,5340217,"jasmine/genie.py",13480,0,"",python,selection_mouse +2974,5340694,"jasmine/genie.py",13497,0,"",python,selection_mouse +2975,5340695,"jasmine/genie.py",13496,0,"",python,selection_command +2976,5341409,"jasmine/genie.py",13525,0,"",python,selection_mouse +2977,5341562,"jasmine/genie.py",13510,22,"noise_level_embed_BS1M",python,selection_mouse +2978,5348083,"jasmine/genie.py",13373,0,"",python,selection_mouse +2979,5348103,"jasmine/genie.py",13372,0,"",python,selection_command +2980,5349127,"jasmine/genie.py",13483,0,"",python,selection_mouse +2981,5349129,"jasmine/genie.py",13482,0,"",python,selection_command +2982,5349763,"jasmine/genie.py",13497,0,"",python,selection_mouse +2983,5349765,"jasmine/genie.py",13496,0,"",python,selection_command +2984,5350771,"jasmine/genie.py",13907,0,"",python,selection_mouse +2985,5351471,"jasmine/genie.py",13905,0,"",python,selection_mouse +2986,5351637,"jasmine/genie.py",13904,1,"1",python,selection_mouse +2987,5351850,"jasmine/genie.py",13904,4,"1.0 ",python,selection_mouse +2988,5351851,"jasmine/genie.py",13904,5,"1.0 -",python,selection_mouse +2989,5351876,"jasmine/genie.py",13904,6,"1.0 - ",python,selection_mouse +2990,5351876,"jasmine/genie.py",13904,20,"1.0 - unmasked_ratio",python,selection_mouse +2991,5352312,"jasmine/genie.py",13916,0,"",python,selection_mouse +2992,5352312,"jasmine/genie.py",13910,14,"unmasked_ratio",python,selection_mouse +2993,5352513,"jasmine/genie.py",13909,15," unmasked_ratio",python,selection_mouse +2994,5352528,"jasmine/genie.py",13908,16,"- unmasked_ratio",python,selection_mouse +2995,5352581,"jasmine/genie.py",13907,17," - unmasked_ratio",python,selection_mouse +2996,5352581,"jasmine/genie.py",13906,18,"0 - unmasked_ratio",python,selection_mouse +2997,5352637,"jasmine/genie.py",13905,19,".0 - unmasked_ratio",python,selection_mouse +2998,5353210,"jasmine/genie.py",13905,0,"",python,selection_mouse +2999,5353744,"jasmine/genie.py",13649,0,"",python,selection_mouse +3000,5355212,"jasmine/genie.py",13497,0,"",python,selection_mouse +3001,5355224,"jasmine/genie.py",13496,0,"",python,selection_command +3002,5410291,"jasmine/genie.py",13497,0,"",python,selection_mouse +3003,5410295,"jasmine/genie.py",13496,0,"",python,selection_command +3004,5454481,"jasmine/genie.py",13605,0,"",python,selection_mouse +3005,5454616,"jasmine/genie.py",13594,14,"vid_embed_BSNM",python,selection_mouse +3006,5456167,"jasmine/genie.py",13741,0,"",python,selection_mouse +3007,5456262,"jasmine/genie.py",13731,22,"noise_level_embed_BS1M",python,selection_mouse +3008,5457790,"jasmine/genie.py",13609,0,"",python,selection_mouse +3009,5458411,"jasmine/genie.py",13604,0,"",python,selection_mouse +3010,5461121,"jasmine/genie.py",13613,0,"",python,selection_mouse +3011,5461648,"jasmine/genie.py",13620,0,"",python,selection_mouse +3012,5461779,"jasmine/genie.py",13616,11,"expand_dims",python,selection_mouse +3013,5462416,"jasmine/genie.py",13550,0,"",python,selection_mouse +3014,5462546,"jasmine/genie.py",13544,22,"noise_level_embed_B11M",python,selection_mouse +3015,5463059,"jasmine/genie.py",13634,0,"",python,selection_mouse +3016,5463172,"jasmine/genie.py",13628,15,"noise_level_B11",python,selection_mouse +3017,5467044,"jasmine/genie.py",13644,0,"",python,selection_mouse +3018,5468025,"jasmine/genie.py",13623,0,"",python,selection_mouse +3019,5471222,"jasmine/genie.py",13632,0,"",python,selection_mouse +3020,5473242,"jasmine/genie.py",13582,67,"",python,content +3021,5474986,"jasmine/genie.py",13582,0," vid_embed_BSNM += jnp.expand_dims(noise_level_B11, -1)\n",python,content +3022,5475028,"jasmine/genie.py",13632,0,"",python,selection_command +3023,5477769,"jasmine/genie.py",13606,0,"",python,selection_mouse +3024,5502209,"jasmine/genie.py",13635,0,"",python,selection_mouse +3025,5502377,"jasmine/genie.py",13628,15,"noise_level_B11",python,selection_mouse +3026,5507568,"jasmine/genie.py",13238,0,"",python,selection_mouse +3027,5507704,"jasmine/genie.py",13230,20,"noise_bucket_idx_B11",python,selection_mouse +3028,5509078,"jasmine/genie.py",13290,0,"",python,selection_mouse +3029,5509234,"jasmine/genie.py",13281,15,"noise_level_B11",python,selection_mouse +3030,5546230,"jasmine/genie.py",13298,0,"",python,selection_mouse +3031,5546865,"jasmine/genie.py",13373,0,"",python,selection_mouse +3032,5546874,"jasmine/genie.py",13372,0,"",python,selection_command +3033,5547704,"jasmine/genie.py",13483,0,"",python,selection_mouse +3034,5547716,"jasmine/genie.py",13482,0,"",python,selection_command +3035,5548416,"jasmine/genie.py",13497,0,"",python,selection_mouse +3036,5548418,"jasmine/genie.py",13496,0,"",python,selection_command +3037,5641165,"jasmine/genie.py",18025,0,"",python,selection_mouse +3038,5641304,"jasmine/genie.py",18016,11,"noise_level",python,selection_mouse +3039,5649223,"jasmine/genie.py",13037,0,"",python,selection_mouse +3040,5650034,"jasmine/genie.py",13013,0,"",python,selection_mouse +3041,5650042,"jasmine/genie.py",13012,0,"",python,selection_command +3042,5650715,"jasmine/genie.py",13013,0,"\n ",python,content +3043,5651899,"jasmine/genie.py",13014,12,"",python,content +3044,5651995,"jasmine/genie.py",13015,0,"",python,selection_command +3045,5652214,"jasmine/genie.py",13040,0,"",python,selection_command +3046,5652802,"jasmine/genie.py",13040,1,"",python,content +3047,5652845,"jasmine/genie.py",13052,0,"",python,selection_command +3048,5652944,"jasmine/genie.py",13027,0,"",python,selection_command +3049,5660259,"jasmine/genie.py",13131,0,"",python,selection_mouse +3050,5660387,"jasmine/genie.py",13122,11,"noise_level",python,selection_mouse +3051,5661102,"jasmine/genie.py",13139,0,"",python,selection_mouse +3052,5661247,"jasmine/genie.py",13134,7,"reshape",python,selection_mouse +3053,5662909,"jasmine/genie.py",13115,0,"",python,selection_mouse +3054,5663049,"jasmine/genie.py",13104,15,"noise_level_111",python,selection_mouse +3055,5663747,"jasmine/genie.py",13240,0,"",python,selection_mouse +3056,5663889,"jasmine/genie.py",13230,20,"noise_bucket_idx_B11",python,selection_mouse +3057,5664657,"jasmine/genie.py",13240,0,"",python,selection_mouse +3058,5665265,"jasmine/genie.py",13173,0,"",python,selection_mouse +3059,5665402,"jasmine/genie.py",13163,15,"noise_level_B11",python,selection_mouse +3060,5668507,"jasmine/genie.py",13246,0,"",python,selection_mouse +3061,5668628,"jasmine/genie.py",13230,20,"noise_bucket_idx_B11",python,selection_mouse +3062,5672354,"jasmine/genie.py",13402,0,"",python,selection_mouse +3063,5672491,"jasmine/genie.py",13386,22,"noise_level_embed_B11M",python,selection_mouse +3064,5673453,"jasmine/genie.py",13475,0,"",python,selection_mouse +3065,5673599,"jasmine/genie.py",13463,20,"noise_bucket_idx_B11",python,selection_mouse +3066,5675262,"jasmine/genie.py",13527,0,"",python,selection_mouse +3067,5675409,"jasmine/genie.py",13510,22,"noise_level_embed_BS1M",python,selection_mouse +3068,5677711,"jasmine/genie.py",13605,0,"",python,selection_mouse +3069,5677873,"jasmine/genie.py",13594,14,"vid_embed_BSNM",python,selection_mouse +3070,5678773,"jasmine/genie.py",13620,0,"",python,selection_mouse +3071,5680195,"jasmine/genie.py",13649,0,"",python,selection_mouse +3072,5681762,"jasmine/genie.py",13617,0,"",python,selection_mouse +3073,5682646,"jasmine/genie.py",13582,67,"",python,content +3074,5692971,"jasmine/genie.py",13583,0,"",python,selection_command +3075,5693104,"jasmine/genie.py",13631,0,"",python,selection_command +3076,5693631,"jasmine/genie.py",13712,0,"",python,selection_command +3077,5693981,"jasmine/genie.py",13712,13," )",python,selection_command +3078,5694202,"jasmine/genie.py",13631,94," [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3079,5694352,"jasmine/genie.py",13583,142," vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3080,5694485,"jasmine/genie.py",13582,143,"\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3081,5694619,"jasmine/genie.py",13498,227," noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3082,5694766,"jasmine/genie.py",13484,241," )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3083,5694904,"jasmine/genie.py",13447,278," noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3084,5695437,"jasmine/genie.py",13374,351," noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3085,5695466,"jasmine/genie.py",13342,383," ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3086,5695480,"jasmine/genie.py",13264,461," (noise_level_B11 * self.noise_buckets) * self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3087,5695540,"jasmine/genie.py",13218,507," noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) * self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3088,5695587,"jasmine/genie.py",13151,574," noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) * self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3089,5695593,"jasmine/genie.py",13092,633," noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) * self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3090,5695656,"jasmine/genie.py",13040,685," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) * self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3091,5696036,"jasmine/genie.py",13092,633," noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) * self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3092,5696367,"jasmine/genie.py",13040,685," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) * self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3093,5696579,"jasmine/genie.py",13040,0,"",python,selection_command +3094,5733587,"jasmine/genie.py",13039,0,"",python,selection_mouse +3095,5733589,"jasmine/genie.py",13038,0,"",python,selection_command +3096,5749586,"jasmine/genie.py",13321,0,"",python,selection_mouse +3097,5750815,"jasmine/genie.py",13320,0,"",python,selection_command +3098,5751345,"jasmine/genie.py",13319,1,"",python,content +3099,5751921,"jasmine/genie.py",13319,0,"/",python,content +3100,5751922,"jasmine/genie.py",13320,0,"",python,selection_keyboard +3101,5752384,"jasmine/genie.py",13319,0,"",python,selection_command +3102,5756523,"jasmine/models/dynamics.py",0,0,"",python,tab +3103,6074623,"jasmine/train_dynamics.py",0,0,"",python,tab +3104,6075393,"jasmine/genie.py",0,0,"",python,tab +3105,6076576,"jasmine/genie.py",13242,0,"",python,selection_mouse +3106,6077075,"jasmine/genie.py",13241,0,"",python,selection_mouse +3107,6077231,"jasmine/genie.py",13230,20,"noise_bucket_idx_B11",python,selection_mouse +3108,6077842,"jasmine/genie.py",13287,0,"",python,selection_mouse +3109,6077977,"jasmine/genie.py",13281,15,"noise_level_B11",python,selection_mouse +3110,6082582,"jasmine/train_dynamics.py",0,0,"",python,tab +3111,6295115,"jasmine/train_dynamics.py",22462,0,"",python,selection_mouse +3112,6303321,"jasmine/genie.py",0,0,"",python,tab +3113,6313581,"jasmine/genie.py",13281,93,"noise_level_B11 * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n",python,selection_command +3114,6314485,"jasmine/genie.py",13372,0,"",python,selection_command +3115,6314731,"jasmine/genie.py",13296,0,"",python,selection_command +3116,6314908,"jasmine/genie.py",13250,0,"",python,selection_command +3117,6315037,"jasmine/genie.py",13183,0,"",python,selection_command +3118,6315182,"jasmine/genie.py",13124,0,"",python,selection_command +3119,6315325,"jasmine/genie.py",13072,0,"",python,selection_command +3120,6315889,"jasmine/genie.py",13040,51," rng, _rng_noise = jax.random.split(rng)",python,selection_command +3121,6316151,"jasmine/genie.py",13040,110," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)",python,selection_command +3122,6316636,"jasmine/genie.py",13040,177," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))",python,selection_command +3123,6316711,"jasmine/genie.py",13040,223," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(",python,selection_command +3124,6316723,"jasmine/genie.py",13040,301," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) / self.max_noise_level",python,selection_command +3125,6316723,"jasmine/genie.py",13040,333," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_command +3126,6316809,"jasmine/genie.py",13040,406," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(",python,selection_command +3127,6316809,"jasmine/genie.py",13040,443," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11",python,selection_command +3128,6316849,"jasmine/genie.py",13040,457," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )",python,selection_command +3129,6316930,"jasmine/genie.py",13040,541," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))",python,selection_command +3130,6317124,"jasmine/genie.py",13040,542," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n",python,selection_command +3131,6317276,"jasmine/genie.py",13040,590," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(",python,selection_command +3132,6317436,"jasmine/genie.py",13040,671," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2",python,selection_command +3133,6317612,"jasmine/genie.py",13040,685," rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,selection_command +3134,6318085,"jasmine/genie.py",13040,0,"",python,selection_command +3135,6328631,"jasmine/genie.py",22242,0,"",python,selection_mouse +3136,6328634,"jasmine/genie.py",22241,0,"",python,selection_command +3137,6329501,"jasmine/genie.py",21993,0,"",python,selection_mouse +3138,6329502,"jasmine/genie.py",21992,0,"",python,selection_command +3139,6331408,"jasmine/genie.py",21993,0,"\n ",python,content +3140,6331576,"jasmine/genie.py",22006,0,"\n ",python,content +3141,6331577,"jasmine/genie.py",21994,12,"",python,content +3142,6331727,"jasmine/genie.py",22007,0,"\n ",python,content +3143,6331728,"jasmine/genie.py",21995,12,"",python,content +3144,6332098,"jasmine/genie.py",21995,0,"",python,selection_command +3145,6332417,"jasmine/genie.py",21996,12,"",python,content +3146,6332418,"jasmine/genie.py",21995,0," ",python,content +3147,6332682,"jasmine/genie.py",21999,0," ",python,content +3148,6332847,"jasmine/genie.py",22003,0," ",python,content +3149,6333520,"jasmine/genie.py",22007,0,"#",python,content +3150,6333521,"jasmine/genie.py",22008,0,"",python,selection_keyboard +3151,6333683,"jasmine/genie.py",22008,0," ",python,content +3152,6333684,"jasmine/genie.py",22009,0,"",python,selection_keyboard +3153,6333968,"jasmine/genie.py",22009,0,"T",python,content +3154,6333969,"jasmine/genie.py",22010,0,"",python,selection_keyboard +3155,6334152,"jasmine/genie.py",22010,0,"O",python,content +3156,6334153,"jasmine/genie.py",22011,0,"",python,selection_keyboard +3157,6334191,"jasmine/genie.py",22011,0,"D",python,content +3158,6334193,"jasmine/genie.py",22012,0,"",python,selection_keyboard +3159,6334304,"jasmine/genie.py",22012,0,"O",python,content +3160,6334305,"jasmine/genie.py",22013,0,"",python,selection_keyboard +3161,6334399,"jasmine/genie.py",22013,0," ",python,content +3162,6334399,"jasmine/genie.py",22014,0,"",python,selection_keyboard +3163,6334533,"jasmine/genie.py",22014,0,"m",python,content +3164,6334534,"jasmine/genie.py",22015,0,"",python,selection_keyboard +3165,6334722,"jasmine/genie.py",22015,0,"i",python,content +3166,6334723,"jasmine/genie.py",22016,0,"",python,selection_keyboard +3167,6334782,"jasmine/genie.py",22016,0,"h",python,content +3168,6334783,"jasmine/genie.py",22017,0,"",python,selection_keyboard +3169,6334866,"jasmine/genie.py",22017,0,"i",python,content +3170,6334867,"jasmine/genie.py",22018,0,"",python,selection_keyboard +3171,6334965,"jasmine/genie.py",22018,0,"r",python,content +3172,6334966,"jasmine/genie.py",22019,0,"",python,selection_keyboard +3173,6335135,"jasmine/genie.py",22019,0,"\n ",python,content +3174,6336180,"jasmine/genie.py",22020,12,"",python,content +3175,6336350,"jasmine/genie.py",22020,0,"\n rng, _rng_noise = jax.random.split(rng)\n noise_level_111 = noise_level.reshape(1, 1, 1)\n noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\n noise_bucket_idx_B11 = jnp.floor(\n (noise_level_B11 * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\n noise_bucket_idx_B11\n )\n noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\n\n vid_embed_BSNp2M = jnp.concatenate(\n [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\n )",python,content +3176,6336362,"jasmine/genie.py",22033,0,"",python,selection_command +3177,6337426,"jasmine/genie.py",22020,0,"",python,selection_command +3178,6341426,"jasmine/genie.py",22401,0,"",python,selection_mouse +3179,6342545,"jasmine/genie.py",22401,7,"",python,content +3180,6342955,"jasmine/genie.py",22401,0,"c",python,content +3181,6342956,"jasmine/genie.py",22402,0,"",python,selection_keyboard +3182,6343132,"jasmine/genie.py",22402,0,"a",python,content +3183,6343133,"jasmine/genie.py",22403,0,"",python,selection_keyboard +3184,6343262,"jasmine/genie.py",22403,0,"u",python,content +3185,6343263,"jasmine/genie.py",22404,0,"",python,selection_keyboard +3186,6343535,"jasmine/genie.py",22404,0,"s",python,content +3187,6343537,"jasmine/genie.py",22405,0,"",python,selection_keyboard +3188,6343633,"jasmine/genie.py",22405,0,"a",python,content +3189,6343635,"jasmine/genie.py",22406,0,"",python,selection_keyboard +3190,6343721,"jasmine/genie.py",22406,0,"l",python,content +3191,6343723,"jasmine/genie.py",22407,0,"",python,selection_keyboard +3192,6344199,"jasmine/genie.py",22406,0,"",python,selection_command +3193,6345657,"jasmine/genie.py",22477,0,"",python,selection_mouse +3194,6345667,"jasmine/genie.py",22476,0,"",python,selection_command +3195,6346733,"jasmine/genie.py",22477,0,"",python,selection_mouse +3196,6346735,"jasmine/genie.py",22476,0,"",python,selection_command +3197,6347534,"jasmine/genie.py",22400,0,"",python,selection_mouse +3198,6349261,"jasmine/genie.py",22395,0,"",python,selection_mouse +3199,6349414,"jasmine/genie.py",22392,15,"dynamics_causal",python,selection_mouse +3200,6350734,"jasmine/genie.py",22588,0,"",python,selection_mouse +3201,6350888,"jasmine/genie.py",22575,16,"vid_embed_BSNp2M",python,selection_mouse +3202,6352191,"jasmine/genie.py",22588,0,"",python,selection_mouse +3203,6352912,"jasmine/genie.py",22575,16,"vid_embed_BSNp2M",python,selection_mouse +3204,6353861,"jasmine/genie.py",22705,0,"",python,selection_mouse +3205,6353879,"jasmine/genie.py",22704,0,"",python,selection_command +3206,6355219,"jasmine/genie.py",22746,0,"",python,selection_mouse +3207,6356221,"jasmine/genie.py",22734,0,"",python,selection_mouse +3208,6358225,"jasmine/genie.py",22755,0,"",python,selection_mouse +3209,6361743,"jasmine/genie.py",22707,89,"",python,content +3210,6361801,"jasmine/genie.py",22719,0,"",python,selection_command +3211,6363549,"jasmine/genie.py",22706,0,"",python,selection_command +3212,6364027,"jasmine/genie.py",22706,1,"",python,content +3213,6364068,"jasmine/genie.py",22718,0,"",python,selection_command +3214,6364175,"jasmine/genie.py",22754,0,"",python,selection_command +3215,6364590,"jasmine/genie.py",22755,0,"",python,selection_command +3216,6365134,"jasmine/genie.py",22756,0,"",python,selection_command +3217,6365134,"jasmine/genie.py",22757,0,"",python,selection_command +3218,6365141,"jasmine/genie.py",22758,0,"",python,selection_command +3219,6365219,"jasmine/genie.py",22759,0,"",python,selection_command +3220,6365220,"jasmine/genie.py",22760,0,"",python,selection_command +3221,6365233,"jasmine/genie.py",22761,0,"",python,selection_command +3222,6365299,"jasmine/genie.py",22762,0,"",python,selection_command +3223,6365311,"jasmine/genie.py",22763,0,"",python,selection_command +3224,6365336,"jasmine/genie.py",22764,0,"",python,selection_command +3225,6365372,"jasmine/genie.py",22765,0,"",python,selection_command +3226,6365426,"jasmine/genie.py",22766,0,"",python,selection_command +3227,6365429,"jasmine/genie.py",22767,0,"",python,selection_command +3228,6365470,"jasmine/genie.py",22768,0,"",python,selection_command +3229,6365532,"jasmine/genie.py",22769,0,"",python,selection_command +3230,6365538,"jasmine/genie.py",22770,0,"",python,selection_command +3231,6365568,"jasmine/genie.py",22771,0,"",python,selection_command +3232,6365588,"jasmine/genie.py",22772,0,"",python,selection_command +3233,6365611,"jasmine/genie.py",22773,0,"",python,selection_command +3234,6365651,"jasmine/genie.py",22774,0,"",python,selection_command +3235,6365676,"jasmine/genie.py",22775,0,"",python,selection_command +3236,6365776,"jasmine/genie.py",22776,0,"",python,selection_command +3237,6365777,"jasmine/genie.py",22777,0,"",python,selection_command +3238,6365778,"jasmine/genie.py",22778,0,"",python,selection_command +3239,6365779,"jasmine/genie.py",22779,0,"",python,selection_command +3240,6365809,"jasmine/genie.py",22780,0,"",python,selection_command +3241,6365868,"jasmine/genie.py",22781,0,"",python,selection_command +3242,6365869,"jasmine/genie.py",22782,0,"",python,selection_command +3243,6365932,"jasmine/genie.py",22783,0,"",python,selection_command +3244,6365932,"jasmine/genie.py",22784,0,"",python,selection_command +3245,6365975,"jasmine/genie.py",22785,0,"",python,selection_command +3246,6365987,"jasmine/genie.py",22786,0,"",python,selection_command +3247,6366020,"jasmine/genie.py",22787,0,"",python,selection_command +3248,6366053,"jasmine/genie.py",22788,0,"",python,selection_command +3249,6366109,"jasmine/genie.py",22789,0,"",python,selection_command +3250,6366109,"jasmine/genie.py",22790,0,"",python,selection_command +3251,6366147,"jasmine/genie.py",22791,0,"",python,selection_command +3252,6366203,"jasmine/genie.py",22792,0,"",python,selection_command +3253,6366204,"jasmine/genie.py",22793,0,"",python,selection_command +3254,6366231,"jasmine/genie.py",22794,0,"",python,selection_command +3255,6366286,"jasmine/genie.py",22795,0,"",python,selection_command +3256,6366288,"jasmine/genie.py",22796,0,"",python,selection_command +3257,6366327,"jasmine/genie.py",22797,0,"",python,selection_command +3258,6366360,"jasmine/genie.py",22798,0,"",python,selection_command +3259,6366388,"jasmine/genie.py",22799,0,"",python,selection_command +3260,6366482,"jasmine/genie.py",22800,0,"",python,selection_command +3261,6366661,"jasmine/genie.py",22801,0,"",python,selection_command +3262,6367195,"jasmine/genie.py",22800,1,"",python,content +3263,6367354,"jasmine/genie.py",22800,0,"2",python,content +3264,6367356,"jasmine/genie.py",22801,0,"",python,selection_keyboard +3265,6367980,"jasmine/genie.py",22800,0,"",python,selection_command +3266,6381423,"jasmine/genie.py",22736,0,"",python,selection_mouse +3267,6383914,"jasmine/genie.py",22735,0,"",python,selection_mouse +3268,6385082,"jasmine/genie.py",22735,1,"2",python,content +3269,6388670,"jasmine/genie.py",22913,0,"",python,selection_mouse +3270,6390276,"jasmine/genie.py",22913,1,"2",python,content +3271,6395031,"jasmine/train_dynamics.py",0,0,"",python,tab +3272,6397658,"jasmine/models/dynamics.py",0,0,"",python,tab +3273,6400701,"jasmine/models/dynamics.py",9138,0,"",python,selection_mouse +3274,6404193,"jasmine/genie.py",0,0,"",python,tab +3275,6443667,"jasmine/genie.py",22927,0,"",python,selection_mouse +3276,6453151,"jasmine/genie.py",22731,0,"",python,selection_mouse +3277,6453235,"jasmine/genie.py",22718,19,"final_logits_BTNp2V",python,selection_mouse +3278,6466018,"jasmine/genie.py",22938,0,"",python,selection_mouse +3279,6466545,"jasmine/genie.py",22905,0,"",python,selection_mouse +3280,6466685,"jasmine/genie.py",22896,19,"final_logits_BTNp2V",python,selection_mouse +3281,6468134,"jasmine/genie.py",22775,0,"",python,selection_mouse +3282,6468235,"jasmine/genie.py",22774,11,"transformer",python,selection_mouse +3283,6469952,"jasmine/models/dynamics.py",0,0,"",python,tab +3284,6473837,"jasmine/models/dynamics.py",9071,0,"",python,selection_mouse +3285,6473955,"jasmine/models/dynamics.py",9066,11,"transformer",python,selection_mouse +3286,6474888,"jasmine/models/dynamics.py",9061,0,"",python,selection_mouse +3287,6475564,"jasmine/models/dynamics.py",9066,0,"",python,selection_mouse +3288,6475727,"jasmine/models/dynamics.py",9066,11,"transformer",python,selection_mouse +3289,6477008,"jasmine/train_dynamics.py",0,0,"",python,tab +3290,6478644,"jasmine/genie.py",0,0,"",python,tab +3291,6504314,"jasmine/genie.py",22794,0,"",python,selection_mouse +3292,6504442,"jasmine/genie.py",22786,16,"vid_embed_BSNp2M",python,selection_mouse +3293,6505113,"jasmine/genie.py",22780,0,"",python,selection_mouse +3294,6505249,"jasmine/genie.py",22774,11,"transformer",python,selection_mouse +3295,6509650,"jasmine/models/dynamics.py",0,0,"",python,tab +3296,6511591,"jasmine/models/dynamics.py",9074,0,"",python,selection_mouse +3297,6512885,"jasmine/models/dynamics.py",9073,0,"",python,selection_mouse +3298,6513058,"jasmine/models/dynamics.py",6158,0,"",python,selection_command +3299,6515504,"jasmine/models/dynamics.py",6179,0,"",python,selection_mouse +3300,6515968,"jasmine/utils/nn.py",0,0,"import math\nfrom typing import Tuple, Callable, List\n\nfrom flax import nnx\nimport jax\nimport jax.numpy as jnp\nimport einops\n\n\ndef _get_spatiotemporal_positional_encoding(d_model: int, max_len: int = 5000):\n """"""\n Creates a function that applies separate sinusoidal positional encodings to the temporal and spatial dimensions.\n """"""\n pe = jnp.zeros((max_len, d_model))\n position = jnp.arange(0, max_len, dtype=jnp.float32)[:, None]\n div_term = jnp.exp(jnp.arange(0, d_model, 2) * (-math.log(10000.0) / d_model))\n pe = pe.at[:, 0::2].set(jnp.sin(position * div_term))\n pe = pe.at[:, 1::2].set(jnp.cos(position * div_term))\n\n def _encode(x: jax.Array) -> jax.Array:\n """"""\n Args:\n x: The input tensor of shape (Batch, Time, Space, Dimension).\n\n Returns:\n The input tensor with positional encodings added.\n """"""\n assert x.ndim == 4, f""Input must be 4-dimensional, but got shape {x.shape}""\n\n num_timesteps = x.shape[1]\n num_spatial_patches = x.shape[2]\n\n # Temporal positional encoding: (1, T, 1, D)\n temporal_pe = pe[None, :num_timesteps, None, :]\n x = x + temporal_pe\n\n # Spatial positional encoding: (1, 1, S, D)\n spatial_pe = pe[None, None, :num_spatial_patches, :]\n x = x + spatial_pe\n\n return x\n\n return _encode\n\n\nclass STBlock(nnx.Module):\n def __init__(\n self,\n dim: int,\n ffn_dim: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n sow_weights: bool,\n sow_activations: bool,\n ):\n self.dim = dim\n self.ffn_dim = ffn_dim\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.sow_weights = sow_weights\n self.sow_activations = sow_activations\n\n self.spatial_norm = nnx.LayerNorm(\n num_features=self.dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.spatial_attention = nnx.MultiHeadAttention(\n num_heads=self.num_heads,\n in_features=self.dim,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=False\n ),\n rngs=rngs,\n decode=False,\n )\n\n self.temporal_norm = nnx.LayerNorm(\n num_features=self.dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.temporal_attention = nnx.MultiHeadAttention(\n num_heads=self.num_heads,\n in_features=self.dim,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=True\n ),\n rngs=rngs,\n decode=False,\n )\n\n self.ffn_norm = nnx.LayerNorm(\n num_features=self.dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.ffn_dense1 = nnx.Linear(\n in_features=self.dim,\n out_features=self.ffn_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.ffn_dense2 = nnx.Linear(\n in_features=self.ffn_dim,\n out_features=self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n @nnx.remat\n def __call__(self, x_BTNM: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z_BTNM = self.spatial_norm(x_BTNM)\n z_BTNM = self.spatial_attention(z_BTNM, sow_weights=self.sow_weights)\n x_BTNM = x_BTNM + z_BTNM\n\n # --- Temporal attention ---\n x_BNTM = x_BTNM.swapaxes(1, 2)\n z_BNTM = self.temporal_norm(x_BNTM)\n z_BNTM = self.temporal_attention(z_BNTM, sow_weights=self.sow_weights)\n x_BNTM = x_BNTM + z_BNTM\n x_BTNM = x_BNTM.swapaxes(1, 2)\n\n # --- Feedforward ---\n z_BTNM = self.ffn_norm(x_BTNM)\n z_BTND = self.ffn_dense1(z_BTNM)\n z_BTND = jax.nn.gelu(z_BTND)\n z_BTNM = self.ffn_dense2(z_BTND)\n x_BTNM = x_BTNM + z_BTNM\n if self.sow_activations:\n self.sow(nnx.Intermediate, ""activations"", x_BTNM)\n return x_BTNM\n\n\nclass STTransformer(nnx.Module):\n """"""\n Dimension keys:\n B: batch size\n T: number of frames\n N: number of patches per frame\n I: number of input features\n M: model dimension\n D: FFN dimension\n V: vocabulary size\n """"""\n\n def __init__(\n self,\n input_dim: int,\n model_dim: int,\n ffn_dim: int,\n out_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n sow_weights: bool = False,\n sow_activations: bool = False,\n sow_logits: bool = False,\n max_len: int = 5000,\n ):\n self.input_dim = input_dim\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.out_dim = out_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.sow_logits = sow_logits\n self.sow_weights = sow_weights\n self.sow_activations = sow_activations\n\n self.input_norm1 = nnx.LayerNorm(\n num_features=self.input_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.input_dense = nnx.Linear(\n in_features=self.input_dim,\n out_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.input_norm2 = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n\n self.pos_enc = _get_spatiotemporal_positional_encoding(\n self.model_dim, max_len=max_len\n )\n\n self.blocks = []\n for _ in range(self.num_blocks):\n self.blocks.append(\n STBlock(\n dim=self.model_dim,\n ffn_dim=self.ffn_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n sow_weights=self.sow_weights,\n sow_activations=self.sow_activations,\n )\n )\n\n self.output_dense = nnx.Linear(\n in_features=self.model_dim,\n out_features=self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(self, x_BTNI: jax.Array) -> jax.Array:\n x_BTNI = self.input_norm1(x_BTNI)\n x_BTNM = self.input_dense(x_BTNI)\n x_BTNM = self.input_norm2(x_BTNM)\n x_BTNM = self.pos_enc(x_BTNM)\n for block in self.blocks:\n x_BTNM = block(x_BTNM)\n\n x_BTNV = self.output_dense(x_BTNM)\n if self.sow_logits:\n self.sow(nnx.Intermediate, ""logits"", x_BTNV)\n return x_BTNV\n\n\nclass TransformerBlock(nnx.Module):\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n sow_weights: bool,\n sow_activations: bool,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.decode = decode\n self.sow_weights = sow_weights\n self.sow_activations = sow_activations\n\n self.temporal_norm = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.spatial_norm = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.ffn_norm = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.temporal_attention = nnx.MultiHeadAttention(\n num_heads=self.num_heads,\n in_features=self.model_dim,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=True\n ),\n rngs=rngs,\n decode=self.decode,\n )\n self.spatial_attention = nnx.MultiHeadAttention(\n num_heads=self.num_heads,\n in_features=self.model_dim,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=True\n ),\n rngs=rngs,\n decode=self.decode,\n )\n self.ffn_dense1 = nnx.Linear(\n in_features=self.model_dim,\n out_features=self.ffn_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.ffn_dense2 = nnx.Linear(\n in_features=self.ffn_dim,\n out_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n @nnx.remat\n def __call__(\n self, x_BTNM: jax.Array, pos_index: Tuple[jax.Array, jax.Array] | None = None\n ) -> jax.Array:\n # --- Spatial attention ---\n B, T, N, M = x_BTNM.shape\n z_FNM = einops.rearrange(x_BTNM, ""b t n m -> (b t) n m"")\n z_FNM = self.spatial_norm(z_FNM)\n z_FNM = self.spatial_attention(z_FNM, sow_weights=self.sow_weights)\n z_BTNM = einops.rearrange(z_FNM, ""(b t) n m -> b t n m"", t=T)\n x_BTNM = x_BTNM + z_BTNM\n # --- Temporal attention ---\n z_PTM = einops.rearrange(x_BTNM, ""b t n m -> (b n) t m"")\n z_PTM = self.temporal_norm(z_PTM)\n z_PTM = self.temporal_attention(z_PTM, sow_weights=self.sow_weights)\n z_BTNM = einops.rearrange(z_PTM, ""(b n) t m -> b t n m"", n=N)\n x_BTNM = x_BTNM + z_BTNM\n # --- Feedforward ---\n z_BTNM = self.ffn_norm(x_BTNM)\n z_BTND = self.ffn_dense1(z_BTNM)\n z_BTND = jax.nn.gelu(z_BTND)\n z_BTNM = self.ffn_dense2(z_BTND)\n x_BTNM = x_BTNM + z_BTNM\n if self.sow_activations:\n self.sow(nnx.Intermediate, ""activations"", x_BTNM)\n\n return x_BTNM\n\n\nclass Transformer(nnx.Module):\n """"""\n Dimension keys:\n B: batch size\n T: number of frames\n N: number of patches per frame\n I: number of input features\n M: model dimension\n D: FFN dimension\n V: vocabulary size\n F: number of frames in batch\n P: number of patch positions in batch\n """"""\n\n def __init__(\n self,\n input_dim: int,\n model_dim: int,\n ffn_dim: int,\n out_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n sow_logits: bool = False,\n sow_weights: bool = False,\n sow_activations: bool = False,\n max_len: int = 5000,\n ):\n self.input_dim = input_dim\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.out_dim = out_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.sow_logits = sow_logits\n self.sow_weights = sow_weights\n self.sow_activations = sow_activations\n\n self.input_norm1 = nnx.LayerNorm(\n num_features=self.input_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.input_dense = nnx.Linear(\n in_features=self.input_dim,\n out_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.input_norm2 = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n\n self.pos_enc = _get_spatiotemporal_positional_encoding(\n self.model_dim, max_len=max_len\n )\n\n self.blocks: List[TransformerBlock] = []\n for _ in range(self.num_blocks):\n self.blocks.append(\n TransformerBlock(\n model_dim=self.model_dim,\n ffn_dim=self.ffn_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n sow_weights=self.sow_weights,\n sow_activations=self.sow_activations,\n rngs=rngs,\n )\n )\n self.output_dense = nnx.Linear(\n in_features=self.model_dim,\n out_features=self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(\n self, x_BTNI: jax.Array, pos_index: Tuple[jax.Array, jax.Array] | None = None\n ) -> jax.Array:\n x_BTNI = self.input_norm1(x_BTNI)\n x_BTNM = self.input_dense(x_BTNI)\n x_BTNM = self.input_norm2(x_BTNM)\n x_BTNM = self.pos_enc(x_BTNM)\n for block in self.blocks:\n x_BTNM = block(x_BTNM, pos_index)\n\n x_BTNV = self.output_dense(x_BTNM)\n if self.sow_logits:\n self.sow(nnx.Intermediate, ""logits"", x_BTNV)\n return x_BTNV\n\n\ndef normalize(x: jax.Array) -> jax.Array:\n return x / (jnp.linalg.norm(x, ord=2, axis=-1, keepdims=True) + 1e-8)\n\n\nclass VectorQuantizer(nnx.Module):\n """"""\n Dimension keys:\n D: B * T * N\n K: number of latents\n L: latent dimension\n """"""\n\n def __init__(\n self,\n latent_dim: int,\n num_latents: int,\n dropout: float,\n dtype: jnp.dtype,\n rngs: nnx.Rngs,\n ):\n self.latent_dim = latent_dim\n self.num_latents = num_latents\n self.dropout = dropout\n self.dtype = dtype\n\n self.codebook = nnx.Param(\n normalize(\n nnx.initializers.normal(stddev=1)(\n rngs.params(), (self.num_latents, self.latent_dim)\n )\n )\n )\n self.drop = nnx.Dropout(self.dropout, rngs=rngs)\n\n def __call__(\n self, x_DL: jax.Array, training: bool\n ) -> Tuple[jax.Array, jax.Array, jax.Array, jax.Array]:\n # --- Compute distances ---\n x_DL = x_DL.astype(self.dtype)\n codebook = self.codebook.value.astype(self.dtype)\n\n x_DL = normalize(x_DL)\n normalized_codebook_KL = normalize(codebook)\n distance_DK = -jnp.matmul(x_DL, normalized_codebook_KL.T)\n if training:\n distance_DK = self.drop(distance_DK)\n\n # --- Get indices and embeddings ---\n indices_D = jnp.argmin(distance_DK, axis=-1)\n z_DL = codebook[indices_D]\n\n # --- Straight through estimator ---\n z_q_DL = x_DL + jax.lax.stop_gradient(z_DL - x_DL)\n return z_q_DL, z_DL, x_DL, indices_D\n\n def get_codes(self, indices_E: jax.Array) -> jax.Array:\n return self.codebook[indices_E]\n\n\ndef _create_flash_attention_fn(use_flash_attention: bool, is_causal: bool) -> Callable:\n """"""\n Create an attention function that uses flash attention if enabled.\n\n flax.nnx.MultiHeadAttention provides tensors with shape (batch..., length, num_heads, head_dim),\n but jax.nn.dot_product_attention expects (batch, length, num_heads, head_dim). We reshape to\n ensure compatibility. cuDNN's flash attention additionally requires a sequence length that\n is a multiple of 4. We pad the sequence length to the nearest multiple of 4 and mask\n accordingly. Note that cuDNN requires the mask to be broadcast before calling the attention\n function due to strict shape checking.\n """"""\n\n def attention_fn(\n query_BTHD, key_BSHD, value_BSHD, bias=None, mask_B111=None, **kwargs\n ):\n implementation = ""cudnn"" if use_flash_attention else None\n\n def _merge_batch_dims(x):\n return einops.rearrange(x, ""... l h k -> (...) l h k"")\n\n def _pad(x, pad_size):\n return jnp.pad(x, ((0, 0), (0, pad_size), (0, 0), (0, 0)))\n\n original_shape = query_BTHD.shape\n T = query_BTHD.shape[-3]\n S = key_BSHD.shape[-3]\n\n # Pad to nearest multiple of 4\n Q = ((T + 3) // 4) * 4\n pad_size_Q = Q - T\n K = ((S + 3) // 4) * 4\n pad_size_K = K - S\n\n query_BQHD = _pad(_merge_batch_dims(query_BTHD), pad_size_Q)\n key_BKHD = _pad(_merge_batch_dims(key_BSHD), pad_size_K)\n value_BKHD = _pad(_merge_batch_dims(value_BSHD), pad_size_K)\n\n attention_mask = jnp.ones((Q, K), dtype=jnp.bool_)\n attention_mask = attention_mask.at[T:, :].set(False)\n attention_mask = attention_mask.at[:, S:].set(False)\n\n mask_11TS = attention_mask[jnp.newaxis, jnp.newaxis, :, :]\n\n bias_4d = (\n jnp.pad(\n _merge_batch_dims(bias),\n ((0, 0), (0, 0), (0, pad_size_Q), (0, pad_size_K)),\n )\n if bias is not None\n else None\n )\n\n # NOTE: jax.nn.dot_product_attention does not support dropout\n output_4d = jax.nn.dot_product_attention(\n query=query_BQHD,\n key=key_BKHD,\n value=value_BKHD,\n bias=bias_4d,\n mask=mask_11TS,\n implementation=implementation,\n is_causal=is_causal,\n )\n return output_4d[..., :T, :, :].reshape(original_shape)\n\n return attention_fn\n",python,tab +3301,6528970,"jasmine/utils/nn.py",15258,0,"",python,selection_mouse +3302,6529081,"jasmine/utils/nn.py",15253,9,"pos_index",python,selection_mouse +3303,6531981,"jasmine/utils/nn.py",15555,0,"",python,selection_mouse +3304,6533565,"jasmine/utils/nn.py",15548,0,"",python,selection_mouse +3305,6533749,"jasmine/utils/nn.py",15502,0,"",python,selection_command +3306,6535774,"jasmine/utils/nn.py",15521,0,"",python,selection_mouse +3307,6535933,"jasmine/utils/nn.py",14298,0,"",python,selection_command +3308,6537981,"jasmine/utils/nn.py",14435,0,"",python,selection_mouse +3309,6538114,"jasmine/utils/nn.py",8207,0,"",python,selection_command +3310,6543903,"jasmine/utils/nn.py",11152,0,"",python,selection_mouse +3311,6543921,"jasmine/utils/nn.py",11151,0,"",python,selection_command +3312,6544532,"jasmine/utils/nn.py",11083,0,"",python,selection_mouse +3313,6544650,"jasmine/utils/nn.py",11080,9,"pos_index",python,selection_mouse +3314,6562983,"jasmine/models/dynamics.py",0,0,"",python,tab +3315,6562984,"jasmine/models/dynamics.py",6523,0,"",python,selection_mouse +3316,6562999,"jasmine/models/dynamics.py",6522,0,"",python,selection_command +3317,6563777,"jasmine/models/dynamics.py",6546,0,"",python,selection_mouse +3318,6563789,"jasmine/models/dynamics.py",6545,0,"",python,selection_command +3319,6564692,"jasmine/models/dynamics.py",6556,0,"",python,selection_mouse +3320,6564694,"jasmine/models/dynamics.py",6555,0,"",python,selection_command +3321,6566017,"jasmine/utils/nn.py",0,0,"",python,tab +3322,6566018,"jasmine/utils/nn.py",8047,0,"",python,selection_mouse +3323,6566052,"jasmine/utils/nn.py",8046,0,"",python,selection_command +3324,6567261,"jasmine/utils/nn.py",8074,1,"p",python,selection_command +3325,6567435,"jasmine/utils/nn.py",8350,2,"po",python,selection_command +3326,6567560,"jasmine/utils/nn.py",11080,3,"pos",python,selection_command +3327,6567790,"jasmine/utils/nn.py",11080,4,"pos_",python,selection_command +3328,6568125,"jasmine/utils/nn.py",11080,5,"pos_i",python,selection_command +3329,6568197,"jasmine/utils/nn.py",11080,6,"pos_in",python,selection_command +3330,6568337,"jasmine/utils/nn.py",11080,7,"pos_ind",python,selection_command +3331,6568463,"jasmine/utils/nn.py",11080,8,"pos_inde",python,selection_command +3332,6568702,"jasmine/utils/nn.py",11080,9,"pos_index",python,selection_command +3333,6569513,"jasmine/utils/nn.py",15253,9,"pos_index",python,selection_command +3334,6571062,"jasmine/utils/nn.py",15559,9,"pos_index",python,selection_command +3335,6571861,"jasmine/utils/nn.py",11080,9,"pos_index",python,selection_command +3336,6572702,"jasmine/utils/nn.py",15253,9,"pos_index",python,selection_command +3337,6573031,"jasmine/utils/nn.py",15559,9,"pos_index",python,selection_command +3338,6573493,"jasmine/utils/nn.py",11080,9,"pos_index",python,selection_command +3339,6573654,"jasmine/utils/nn.py",15253,9,"pos_index",python,selection_command +3340,6573791,"jasmine/utils/nn.py",15559,9,"pos_index",python,selection_command +3341,6574697,"jasmine/utils/nn.py",11080,9,"pos_index",python,selection_command +3342,6575133,"jasmine/utils/nn.py",15253,9,"pos_index",python,selection_command +3343,6575256,"jasmine/utils/nn.py",15559,9,"pos_index",python,selection_command +3344,6580211,"jasmine/models/dynamics.py",0,0,"",python,tab +3345,6580931,"jasmine/genie.py",0,0,"",python,tab +3346,6582070,"jasmine/genie.py",22851,0,"",python,selection_mouse +3347,6582088,"jasmine/genie.py",22850,0,"",python,selection_command +3348,6741731,"jasmine/genie.py",22851,0,"",python,selection_mouse +3349,6741733,"jasmine/genie.py",22850,0,"",python,selection_command +3350,6784508,"jasmine/genie.py",24292,0,"",python,selection_mouse +3351,6784697,"jasmine/genie.py",24288,6,"arange",python,selection_mouse +3352,6786554,"jasmine/genie.py",24257,0,"",python,selection_mouse +3353,6786684,"jasmine/genie.py",24250,14,"causal_step_fn",python,selection_mouse +3354,6890247,"jasmine/genie.py",22914,0,"",python,selection_mouse +3355,6894731,"jasmine/genie.py",22928,0,"",python,selection_mouse +3356,6895955,"jasmine/genie.py",22927,0,"",python,selection_mouse +3357,6897026,"jasmine/genie.py",22932,0,"",python,selection_mouse +3358,6898200,"jasmine/genie.py",22933,0,"",python,selection_command +3359,6899185,"jasmine/genie.py",22933,0," ",python,content +3360,6899188,"jasmine/genie.py",22934,0,"",python,selection_keyboard +3361,6899885,"jasmine/genie.py",22934,0,"+",python,content +3362,6899886,"jasmine/genie.py",22935,0,"",python,selection_keyboard +3363,6900013,"jasmine/genie.py",22935,0," ",python,content +3364,6900014,"jasmine/genie.py",22936,0,"",python,selection_keyboard +3365,6900165,"jasmine/genie.py",22936,0,"1",python,content +3366,6900165,"jasmine/genie.py",22937,0,"",python,selection_keyboard +3367,6905274,"jasmine/genie.py",22865,0,"",python,selection_mouse +3368,6905783,"jasmine/genie.py",22930,0,"",python,selection_mouse +3369,6905874,"jasmine/genie.py",22927,6,"step_n",python,selection_mouse +3370,6906130,"jasmine/genie.py",22927,7,"step_n ",python,selection_mouse +3371,6906135,"jasmine/genie.py",22927,8,"step_n +",python,selection_mouse +3372,6906193,"jasmine/genie.py",22927,9,"step_n + ",python,selection_mouse +3373,6907404,"jasmine/genie.py",23357,0,"",python,selection_mouse +3374,6910406,"jasmine/genie.py",23472,0,"",python,selection_mouse +3375,6911870,"jasmine/genie.py",23450,0,"",python,selection_mouse +3376,6912016,"jasmine/genie.py",23444,11,"logits_BSNV",python,selection_mouse +3377,6914880,"jasmine/genie.py",23438,0,"",python,selection_mouse +3378,6915584,"jasmine/genie.py",23499,0,"",python,selection_mouse +3379,6916316,"jasmine/genie.py",23613,0,"",python,selection_mouse +3380,6916992,"jasmine/genie.py",23417,0,"",python,selection_mouse +3381,6921814,"jasmine/genie.py",22942,0,"",python,selection_mouse +3382,7147059,"TERMINAL",0,0,"s",,terminal_output +3383,7147129,"TERMINAL",0,0,"o",,terminal_output +3384,7147210,"TERMINAL",0,0,"u",,terminal_output +3385,7147276,"TERMINAL",0,0,"r",,terminal_output +3386,7147878,"TERMINAL",0,0,"",,terminal_output +3387,7148040,"TERMINAL",0,0,"",,terminal_output +3388,7148146,"TERMINAL",0,0,"",,terminal_output +3389,7148282,"TERMINAL",0,0,"",,terminal_output +3390,7148419,"TERMINAL",0,0,"",,terminal_output +3391,7149366,"TERMINAL",0,0,"s",,terminal_output +3392,7149473,"TERMINAL",0,0,"m",,terminal_output +3393,7149576,"TERMINAL",0,0,"i",,terminal_output +3394,7149720,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +3395,7150137,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: nvidia-smihkn0702.localdomain: Sun Sep 28 15:26:34 2025Sun Sep 28 15:26:34 2025\r+-----------------------------------------------------------------------------------------+\r| NVIDIA-SMI 570.133.20Driver Version: 570.133.20 CUDA Version: 12.8 |\r|-----------------------------------------+------------------------+----------------------+\r| GPU NamePersistence-M | Bus-IdDisp.A | Volatile Uncorr. ECC |\r| Fan Temp PerfPwr:Usage/Cap |Memory-Usage | GPU-Util Compute M. |\r|||MIG M. |\r|=========================================+========================+======================|\r| 0 NVIDIA A100-SXM4-40GBOn | 00000000:31:00.0 Off |0 |\r| N/A 45C P057W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r+-----------------------------------------------------------------------------------------+\r| Processes:|\r| GPU GI CIPID Type Process nameGPU Memory |\r|ID IDUsage\t |\r|=========================================================================================|\r| 0 N/A N/A2907G /usr/libexec/Xorg17MiB |\r+-----------------------------------------------------------------------------------------+",,terminal_output +3396,7150943,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +3397,7152991,"jasmine/train_dynamics.py",0,0,"",python,tab +3398,7161649,"jasmine/train_dynamics.py",23085,0,"",python,selection_mouse +3399,7162315,"jasmine/train_dynamics.py",22703,0,"",python,selection_mouse +3400,7162327,"jasmine/train_dynamics.py",22702,0,"",python,selection_command +3401,7163020,"jasmine/train_dynamics.py",22771,0,"",python,selection_mouse +3402,7163564,"jasmine/train_dynamics.py",22956,0,"",python,selection_mouse +3403,7164136,"jasmine/train_dynamics.py",23031,0,"",python,selection_mouse +3404,7164734,"jasmine/train_dynamics.py",23081,0,"",python,selection_mouse +3405,7165423,"jasmine/train_dynamics.py",23093,0,"",python,selection_mouse +3406,7165995,"jasmine/train_dynamics.py",23105,0,"",python,selection_mouse +3407,7166525,"jasmine/train_dynamics.py",23112,0,"",python,selection_mouse +3408,11377846,"jasmine/models/dynamics.py",0,0,"",python,tab +3409,11396251,"jasmine/models/dynamics.py",2721,0,"",python,selection_mouse +3410,11397224,"jasmine/models/dynamics.py",2736,0,"",python,selection_mouse +3411,11397355,"jasmine/models/dynamics.py",2730,15,"max_noise_level",python,selection_mouse +3412,11398493,"jasmine/models/dynamics.py",2755,0,"",python,selection_mouse +3413,11398618,"jasmine/models/dynamics.py",2749,13,"noise_buckets",python,selection_mouse +3414,11400170,"jasmine/models/dynamics.py",2722,0,"",python,selection_mouse +3415,11401196,"jasmine/models/dynamics.py",2737,0,"",python,selection_mouse +3416,11401324,"jasmine/models/dynamics.py",2730,15,"max_noise_level",python,selection_mouse +3417,11402108,"jasmine/models/dynamics.py",2751,0,"",python,selection_mouse +3418,11402243,"jasmine/models/dynamics.py",2749,13,"noise_buckets",python,selection_mouse +3419,11404300,"jasmine/models/dynamics.py",2651,0,"",python,selection_mouse +3420,11404525,"jasmine/models/dynamics.py",2643,11,"noise_level",python,selection_mouse +3421,11405346,"jasmine/models/dynamics.py",2664,0,"",python,selection_mouse +3422,11405526,"jasmine/models/dynamics.py",2657,13,"noise_buckets",python,selection_mouse +3423,11407315,"jasmine/models/dynamics.py",2676,0,"",python,selection_mouse +3424,11407434,"jasmine/models/dynamics.py",2674,15,"max_noise_level",python,selection_mouse +3425,11409614,"jasmine/models/dynamics.py",2928,0,"",python,selection_mouse +3426,11409634,"jasmine/models/dynamics.py",2927,0,"",python,selection_command +3427,11409808,"jasmine/models/dynamics.py",2927,1,")",python,selection_mouse +3428,11409821,"jasmine/models/dynamics.py",2928,0,"",python,selection_command +3429,11409916,"jasmine/models/dynamics.py",2927,1,")",python,selection_mouse +3430,11409917,"jasmine/models/dynamics.py",2925,3,"32)",python,selection_mouse +3431,11409917,"jasmine/models/dynamics.py",2852,76,"l_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3432,11409918,"jasmine/models/dynamics.py",2850,78,"vel_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3433,11409918,"jasmine/models/dynamics.py",2849,79,"evel_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3434,11409918,"jasmine/models/dynamics.py",2848,80,"level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3435,11410048,"jasmine/models/dynamics.py",2847,81,"_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3436,11410048,"jasmine/models/dynamics.py",2845,83,"se_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3437,11410049,"jasmine/models/dynamics.py",2804,124,"ucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3438,11410049,"jasmine/models/dynamics.py",2803,125,"bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3439,11410049,"jasmine/models/dynamics.py",2802,126,"_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3440,11410049,"jasmine/models/dynamics.py",2801,127,"e_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3441,11410050,"jasmine/models/dynamics.py",2800,128,"se_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3442,11410090,"jasmine/models/dynamics.py",2799,129,"ise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3443,11410091,"jasmine/models/dynamics.py",2798,130,"oise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3444,11410147,"jasmine/models/dynamics.py",2797,131,"noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3445,11410147,"jasmine/models/dynamics.py",2796,132," noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3446,11410203,"jasmine/models/dynamics.py",2795,133," noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3447,11410215,"jasmine/models/dynamics.py",2794,134," noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3448,11410274,"jasmine/models/dynamics.py",2793,135," noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3449,11410346,"jasmine/models/dynamics.py",2792,136," noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3450,11410407,"jasmine/models/dynamics.py",2791,137," noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3451,11410492,"jasmine/models/dynamics.py",2790,138," noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)",python,selection_mouse +3452,11411293,"jasmine/models/dynamics.py",2811,0,"",python,selection_mouse +3453,11411868,"jasmine/models/dynamics.py",2927,0,"",python,selection_mouse +3454,11414633,"jasmine/models/dynamics.py",3215,0,"",python,selection_mouse +3455,11416627,"jasmine/models/dynamics.py",3227,0,"",python,selection_mouse +3456,11416898,"jasmine/models/dynamics.py",3227,1,"a",python,selection_mouse +3457,11416951,"jasmine/models/dynamics.py",3227,2,"af",python,selection_mouse +3458,11416952,"jasmine/models/dynamics.py",3227,4,"afe ",python,selection_mouse +3459,11416989,"jasmine/models/dynamics.py",3227,5,"afe s",python,selection_mouse +3460,11417019,"jasmine/models/dynamics.py",3227,6,"afe sq",python,selection_mouse +3461,11417064,"jasmine/models/dynamics.py",3227,7,"afe sqr",python,selection_mouse +3462,11417064,"jasmine/models/dynamics.py",3227,8,"afe sqrt",python,selection_mouse +3463,11417121,"jasmine/models/dynamics.py",3227,9,"afe sqrt:",python,selection_mouse +3464,11417122,"jasmine/models/dynamics.py",3227,10,"afe sqrt: ",python,selection_mouse +3465,11417122,"jasmine/models/dynamics.py",3227,11,"afe sqrt: c",python,selection_mouse +3466,11417180,"jasmine/models/dynamics.py",3227,12,"afe sqrt: cl",python,selection_mouse +3467,11417198,"jasmine/models/dynamics.py",3227,13,"afe sqrt: cli",python,selection_mouse +3468,11417614,"jasmine/models/dynamics.py",3240,0,"",python,selection_mouse +3469,11418463,"jasmine/models/dynamics.py",3290,0,"",python,selection_mouse +3470,11418610,"jasmine/models/dynamics.py",3289,4,"clip",python,selection_mouse +3471,11419635,"jasmine/models/dynamics.py",3320,0,"",python,selection_mouse +3472,11420011,"jasmine/models/dynamics.py",3318,5,"a_min",python,selection_mouse +3473,11420325,"jasmine/models/dynamics.py",3318,6,"a_min=",python,selection_mouse +3474,11420387,"jasmine/models/dynamics.py",3318,7,"a_min=0",python,selection_mouse +3475,11420781,"jasmine/models/dynamics.py",3325,0,"",python,selection_mouse +3476,11421636,"jasmine/models/dynamics.py",3290,0,"",python,selection_mouse +3477,11421793,"jasmine/models/dynamics.py",3289,4,"clip",python,selection_mouse +3478,11431403,"jasmine/train_dynamics.py",0,0,"",python,tab +3479,11442108,"jasmine/train_dynamics.py",2100,0,"",python,selection_mouse +3480,11442946,"jasmine/train_dynamics.py",2066,0,"",python,selection_mouse +3481,11443089,"jasmine/train_dynamics.py",2058,15,"max_noise_level",python,selection_mouse +3482,11753269,"jasmine/train_dynamics.py",1996,0,"",python,selection_mouse +3483,11753282,"jasmine/train_dynamics.py",1995,0,"",python,selection_command +3484,11755791,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +3485,11757252,"TERMINAL",0,0,"q",,terminal_output +3486,11757376,"TERMINAL",0,0,"u",,terminal_output +3487,11757471,"TERMINAL",0,0,"e",,terminal_output +3488,11757537,"TERMINAL",0,0,"u",,terminal_output +3489,11757602,"TERMINAL",0,0,"e",,terminal_output +3490,11757663,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +3491,11757722,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn0702.localdomain: Sun Sep 28 16:43:22 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3527669 accelerat interact tum_cte0 R 2:55:40\t 1 hkn0702",,terminal_output +3492,11758511,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +3493,11760878,"jasmine/models/dynamics.py",0,0,"",python,tab +3494,11773804,"jasmine/models/dynamics.py",3493,0,"",python,selection_mouse +3495,11799785,"jasmine/models/dynamics.py",4728,0,"",python,selection_mouse +3496,11800399,"jasmine/models/dynamics.py",4668,0,"",python,selection_mouse +3497,11802958,"jasmine/models/dynamics.py",4595,0,"",python,selection_mouse +3498,11803092,"jasmine/models/dynamics.py",4591,22,"noise_level_embed_BT1M",python,selection_mouse +3499,11803828,"jasmine/models/dynamics.py",4587,0,"",python,selection_mouse +3500,11803976,"jasmine/models/dynamics.py",4575,14,"vid_embed_BTNM",python,selection_mouse +3501,11910357,"jasmine/models/dynamics.py",2327,0,"",python,selection_mouse +3502,11910842,"jasmine/models/dynamics.py",2267,0,"",python,selection_mouse +3503,11911441,"jasmine/models/dynamics.py",2303,0,"\n ",python,content +3504,11911625,"jasmine/models/dynamics.py",2312,0,"j",python,content +3505,11911626,"jasmine/models/dynamics.py",2313,0,"",python,selection_keyboard +3506,11911735,"jasmine/models/dynamics.py",2313,0,"a",python,content +3507,11911736,"jasmine/models/dynamics.py",2314,0,"",python,selection_keyboard +3508,11911956,"jasmine/models/dynamics.py",2314,0,"x",python,content +3509,11911958,"jasmine/models/dynamics.py",2315,0,"",python,selection_keyboard +3510,11912055,"jasmine/models/dynamics.py",2315,0,".",python,content +3511,11912057,"jasmine/models/dynamics.py",2316,0,"",python,selection_keyboard +3512,11912526,"jasmine/models/dynamics.py",2316,0,"d",python,content +3513,11912527,"jasmine/models/dynamics.py",2317,0,"",python,selection_keyboard +3514,11912733,"jasmine/models/dynamics.py",2317,0,"e",python,content +3515,11912734,"jasmine/models/dynamics.py",2318,0,"",python,selection_keyboard +3516,11912830,"jasmine/models/dynamics.py",2318,0,"b",python,content +3517,11912831,"jasmine/models/dynamics.py",2319,0,"",python,selection_keyboard +3518,11912889,"jasmine/models/dynamics.py",2319,0,"u",python,content +3519,11912891,"jasmine/models/dynamics.py",2320,0,"",python,selection_keyboard +3520,11913049,"jasmine/models/dynamics.py",2320,0,"g",python,content +3521,11913051,"jasmine/models/dynamics.py",2321,0,"",python,selection_keyboard +3522,11913152,"jasmine/models/dynamics.py",2321,0,".",python,content +3523,11913154,"jasmine/models/dynamics.py",2322,0,"",python,selection_keyboard +3524,11913428,"jasmine/models/dynamics.py",2322,0,"b",python,content +3525,11913430,"jasmine/models/dynamics.py",2323,0,"",python,selection_keyboard +3526,11913486,"jasmine/models/dynamics.py",2323,0,"r",python,content +3527,11913487,"jasmine/models/dynamics.py",2324,0,"",python,selection_keyboard +3528,11913662,"jasmine/models/dynamics.py",2324,0,"e",python,content +3529,11913664,"jasmine/models/dynamics.py",2325,0,"",python,selection_keyboard +3530,11913816,"jasmine/models/dynamics.py",2325,0,"a",python,content +3531,11913818,"jasmine/models/dynamics.py",2326,0,"",python,selection_keyboard +3532,11913897,"jasmine/models/dynamics.py",2326,0,"k",python,content +3533,11913898,"jasmine/models/dynamics.py",2327,0,"",python,selection_keyboard +3534,11914127,"jasmine/models/dynamics.py",2327,0,"p",python,content +3535,11914129,"jasmine/models/dynamics.py",2328,0,"",python,selection_keyboard +3536,11914568,"jasmine/models/dynamics.py",2328,0,"o",python,content +3537,11914569,"jasmine/models/dynamics.py",2329,0,"",python,selection_keyboard +3538,11914779,"jasmine/models/dynamics.py",2329,0,"i",python,content +3539,11914779,"jasmine/models/dynamics.py",2330,0,"",python,selection_keyboard +3540,11915330,"jasmine/models/dynamics.py",2330,0,"n",python,content +3541,11915331,"jasmine/models/dynamics.py",2331,0,"",python,selection_keyboard +3542,11915616,"jasmine/models/dynamics.py",2322,9,"breakpoint",python,content +3543,11916233,"jasmine/models/dynamics.py",2332,0,"()",python,content +3544,11916234,"jasmine/models/dynamics.py",2333,0,"",python,selection_keyboard +3545,11916289,"jasmine/models/dynamics.py",2333,1,")",python,content +3546,11916289,"jasmine/models/dynamics.py",2334,0,"",python,selection_keyboard +3547,11916659,"jasmine/models/dynamics.py",2333,0,"",python,selection_command +3548,11917660,"jasmine/models/dynamics.py",2364,0,"",python,selection_command +3549,11918158,"jasmine/models/dynamics.py",2406,0,"",python,selection_command +3550,11918213,"jasmine/models/dynamics.py",2473,0,"",python,selection_command +3551,11918218,"jasmine/models/dynamics.py",2517,0,"",python,selection_command +3552,11918276,"jasmine/models/dynamics.py",2576,0,"",python,selection_command +3553,11918342,"jasmine/models/dynamics.py",2607,0,"",python,selection_command +3554,11919104,"jasmine/models/dynamics.py",2576,0,"",python,selection_command +3555,11919304,"jasmine/models/dynamics.py",2517,0,"",python,selection_command +3556,11919600,"jasmine/models/dynamics.py",2576,0,"",python,selection_command +3557,11920096,"jasmine/models/dynamics.py",2577,0,"\n jax.debug.breakpoint()",python,content +3558,11920107,"jasmine/models/dynamics.py",2586,0,"",python,selection_command +3559,11920307,"jasmine/models/dynamics.py",2617,0,"",python,selection_command +3560,11921851,"jasmine/models/dynamics.py",2679,0,"\n jax.debug.breakpoint()",python,content +3561,11921861,"jasmine/models/dynamics.py",2688,0,"",python,selection_command +3562,11922158,"jasmine/models/dynamics.py",2719,0,"",python,selection_command +3563,11922313,"jasmine/models/dynamics.py",2803,0,"",python,selection_command +3564,11922514,"jasmine/models/dynamics.py",2890,0,"",python,selection_command +3565,11922699,"jasmine/models/dynamics.py",2930,0,"",python,selection_command +3566,11922831,"jasmine/models/dynamics.py",3002,0,"",python,selection_command +3567,11923502,"jasmine/models/dynamics.py",3021,0,"\n jax.debug.breakpoint()",python,content +3568,11923503,"jasmine/models/dynamics.py",3030,0,"",python,selection_command +3569,11923762,"jasmine/models/dynamics.py",3061,0,"",python,selection_command +3570,11923938,"jasmine/models/dynamics.py",3128,0,"",python,selection_command +3571,11924603,"jasmine/models/dynamics.py",3206,0,"",python,selection_command +3572,11924767,"jasmine/models/dynamics.py",3286,0,"",python,selection_command +3573,11926498,"jasmine/models/dynamics.py",3338,0,"\n jax.debug.breakpoint()",python,content +3574,11926506,"jasmine/models/dynamics.py",3347,0,"",python,selection_command +3575,11928468,"jasmine/models/dynamics.py",3555,0,"",python,selection_mouse +3576,11929057,"jasmine/models/dynamics.py",3603,0,"\n jax.debug.breakpoint()",python,content +3577,11929066,"jasmine/models/dynamics.py",3612,0,"",python,selection_command +3578,11929771,"jasmine/models/dynamics.py",3635,0,"",python,selection_command +3579,11929958,"jasmine/models/dynamics.py",3644,0,"",python,selection_command +3580,11930100,"jasmine/models/dynamics.py",3743,0,"",python,selection_command +3581,11930458,"jasmine/models/dynamics.py",3644,0,"",python,selection_command +3582,11930686,"jasmine/models/dynamics.py",3734,0,"\n jax.debug.breakpoint()",python,content +3583,11930693,"jasmine/models/dynamics.py",3743,0,"",python,selection_command +3584,11936222,"jasmine/models/dynamics.py",4804,0,"",python,selection_mouse +3585,11936839,"jasmine/models/dynamics.py",4884,0,"\n ",python,content +3586,11937119,"jasmine/models/dynamics.py",4893,0,"if jnp.any(jnp.isnan(vid_embed_BTNM)): jax.debug.print(""NaNs in vid_embed after noise"")",python,content +3587,11938689,"jasmine/models/dynamics.py",4932,0,"",python,selection_mouse +3588,11939089,"jasmine/models/dynamics.py",4932,0,"\n ",python,content +3589,11940368,"jasmine/models/dynamics.py",4899,0,"",python,selection_mouse +3590,11940939,"jasmine/models/dynamics.py",4897,0,"",python,selection_mouse +3591,11941586,"jasmine/models/dynamics.py",4907,0,"",python,selection_mouse +3592,11942308,"jasmine/models/dynamics.py",4931,0,"",python,selection_mouse +3593,11942888,"jasmine/models/dynamics.py",4978,0,"",python,selection_mouse +3594,11944693,"jasmine/models/dynamics.py",4993,0,"",python,selection_mouse +3595,11946007,"jasmine/models/dynamics.py",4932,0,"",python,selection_mouse +3596,11946532,"jasmine/models/dynamics.py",4931,0,"",python,selection_command +3597,11947396,"jasmine/models/dynamics.py",4885,47," if jnp.any(jnp.isnan(vid_embed_BTNM)): ",python,selection_command +3598,11947611,"jasmine/models/dynamics.py",4885,108," if jnp.any(jnp.isnan(vid_embed_BTNM)): \n jax.debug.print(""NaNs in vid_embed after noise"")",python,selection_command +3599,11947954,"jasmine/models/dynamics.py",4885,0,"",python,selection_command +3600,11948260,"jasmine/models/dynamics.py",4933,0,"",python,selection_command +3601,11948632,"jasmine/models/dynamics.py",4993,0,"\n if jnp.any(jnp.isnan(vid_embed_BTNM)): \n jax.debug.print(""NaNs in vid_embed after noise"")",python,content +3602,11948652,"jasmine/models/dynamics.py",5002,0,"",python,selection_command +3603,11949311,"jasmine/models/dynamics.py",4941,0,"",python,selection_command +3604,11949694,"jasmine/models/dynamics.py",5002,0,"",python,selection_command +3605,11949827,"jasmine/models/dynamics.py",5003,0,"",python,selection_command +3606,11950323,"jasmine/models/dynamics.py",5004,0,"",python,selection_command +3607,11950385,"jasmine/models/dynamics.py",5005,0,"",python,selection_command +3608,11950386,"jasmine/models/dynamics.py",5006,0,"",python,selection_command +3609,11950441,"jasmine/models/dynamics.py",5007,0,"",python,selection_command +3610,11950446,"jasmine/models/dynamics.py",5008,0,"",python,selection_command +3611,11950470,"jasmine/models/dynamics.py",5009,0,"",python,selection_command +3612,11950536,"jasmine/models/dynamics.py",5010,0,"",python,selection_command +3613,11950538,"jasmine/models/dynamics.py",5011,0,"",python,selection_command +3614,11950596,"jasmine/models/dynamics.py",5012,0,"",python,selection_command +3615,11950605,"jasmine/models/dynamics.py",5013,0,"",python,selection_command +3616,11950648,"jasmine/models/dynamics.py",5014,0,"",python,selection_command +3617,11950658,"jasmine/models/dynamics.py",5015,0,"",python,selection_command +3618,11950712,"jasmine/models/dynamics.py",5016,0,"",python,selection_command +3619,11951364,"jasmine/models/dynamics.py",5017,0,"",python,selection_command +3620,11951857,"jasmine/models/dynamics.py",5018,0,"",python,selection_command +3621,11951912,"jasmine/models/dynamics.py",5019,0,"",python,selection_command +3622,11951912,"jasmine/models/dynamics.py",5020,0,"",python,selection_command +3623,11951940,"jasmine/models/dynamics.py",5021,0,"",python,selection_command +3624,11952067,"jasmine/models/dynamics.py",5022,0,"",python,selection_command +3625,11952218,"jasmine/models/dynamics.py",5023,0,"",python,selection_command +3626,11952574,"jasmine/models/dynamics.py",5023,14,"",python,content +3627,11953056,"jasmine/models/dynamics.py",5023,0,"n",python,content +3628,11953057,"jasmine/models/dynamics.py",5024,0,"",python,selection_keyboard +3629,11953256,"jasmine/models/dynamics.py",5024,0,"o",python,content +3630,11953258,"jasmine/models/dynamics.py",5025,0,"",python,selection_keyboard +3631,11953353,"jasmine/models/dynamics.py",5025,0,"i",python,content +3632,11953355,"jasmine/models/dynamics.py",5026,0,"",python,selection_keyboard +3633,11953365,"jasmine/models/dynamics.py",5026,0,"s",python,content +3634,11953366,"jasmine/models/dynamics.py",5027,0,"",python,selection_keyboard +3635,11953534,"jasmine/models/dynamics.py",5027,0,"e",python,content +3636,11953535,"jasmine/models/dynamics.py",5028,0,"",python,selection_keyboard +3637,11953740,"jasmine/models/dynamics.py",5028,0,":",python,content +3638,11953741,"jasmine/models/dynamics.py",5029,0,"",python,selection_keyboard +3639,11954359,"jasmine/models/dynamics.py",5028,1,"",python,content +3640,11954688,"jasmine/models/dynamics.py",5028,0,"_",python,content +3641,11954690,"jasmine/models/dynamics.py",5029,0,"",python,selection_keyboard +3642,11955007,"jasmine/models/dynamics.py",5029,0,"l",python,content +3643,11955008,"jasmine/models/dynamics.py",5030,0,"",python,selection_keyboard +3644,11955128,"jasmine/models/dynamics.py",5030,0,"e",python,content +3645,11955129,"jasmine/models/dynamics.py",5031,0,"",python,selection_keyboard +3646,11955290,"jasmine/models/dynamics.py",5031,0,"v",python,content +3647,11955291,"jasmine/models/dynamics.py",5032,0,"",python,selection_keyboard +3648,11955507,"jasmine/models/dynamics.py",5032,0,"l",python,content +3649,11955508,"jasmine/models/dynamics.py",5033,0,"",python,selection_keyboard +3650,11955867,"jasmine/models/dynamics.py",5032,1,"",python,content +3651,11956234,"jasmine/models/dynamics.py",5023,9,"noise_level_embed_BT1M",python,content +3652,11957290,"jasmine/models/dynamics.py",5040,0,"",python,selection_mouse +3653,11957426,"jasmine/models/dynamics.py",5023,22,"noise_level_embed_BT1M",python,selection_mouse +3654,11958286,"jasmine/models/dynamics.py",5093,0,"",python,selection_mouse +3655,11958429,"jasmine/models/dynamics.py",5087,9,"vid_embed",python,selection_mouse +3656,11959217,"jasmine/models/dynamics.py",5087,9,"n",python,content +3657,11959218,"jasmine/models/dynamics.py",5088,0,"",python,selection_keyboard +3658,11959401,"jasmine/models/dynamics.py",5088,0,"o",python,content +3659,11959403,"jasmine/models/dynamics.py",5089,0,"",python,selection_keyboard +3660,11959469,"jasmine/models/dynamics.py",5089,0,"i",python,content +3661,11959471,"jasmine/models/dynamics.py",5090,0,"",python,selection_keyboard +3662,11959587,"jasmine/models/dynamics.py",5090,0,"s",python,content +3663,11959589,"jasmine/models/dynamics.py",5091,0,"",python,selection_keyboard +3664,11959786,"jasmine/models/dynamics.py",5091,0,"e",python,content +3665,11959787,"jasmine/models/dynamics.py",5092,0,"",python,selection_keyboard +3666,11960251,"jasmine/models/dynamics.py",5092,0,"_",python,content +3667,11960253,"jasmine/models/dynamics.py",5093,0,"",python,selection_keyboard +3668,11960521,"jasmine/models/dynamics.py",5093,0,"l",python,content +3669,11960522,"jasmine/models/dynamics.py",5094,0,"",python,selection_keyboard +3670,11960732,"jasmine/models/dynamics.py",5094,0,"e",python,content +3671,11960732,"jasmine/models/dynamics.py",5095,0,"",python,selection_keyboard +3672,11960920,"jasmine/models/dynamics.py",5095,0,"v",python,content +3673,11960921,"jasmine/models/dynamics.py",5096,0,"",python,selection_keyboard +3674,11961040,"jasmine/models/dynamics.py",5096,0,"e",python,content +3675,11961041,"jasmine/models/dynamics.py",5097,0,"",python,selection_keyboard +3676,11961140,"jasmine/models/dynamics.py",5097,0,"l",python,content +3677,11961140,"jasmine/models/dynamics.py",5098,0,"",python,selection_keyboard +3678,11961367,"jasmine/models/dynamics.py",5098,0,"_",python,content +3679,11961368,"jasmine/models/dynamics.py",5099,0,"",python,selection_keyboard +3680,11961598,"jasmine/models/dynamics.py",5099,0,"e",python,content +3681,11961599,"jasmine/models/dynamics.py",5100,0,"",python,selection_keyboard +3682,11961686,"jasmine/models/dynamics.py",5100,0,"m",python,content +3683,11961687,"jasmine/models/dynamics.py",5101,0,"",python,selection_keyboard +3684,11961922,"jasmine/models/dynamics.py",5101,0,"b",python,content +3685,11961923,"jasmine/models/dynamics.py",5102,0,"",python,selection_keyboard +3686,11961999,"jasmine/models/dynamics.py",5102,0,"e",python,content +3687,11962000,"jasmine/models/dynamics.py",5103,0,"",python,selection_keyboard +3688,11962082,"jasmine/models/dynamics.py",5103,0,"d",python,content +3689,11962083,"jasmine/models/dynamics.py",5104,0,"",python,selection_keyboard +3690,11976868,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +3691,11979988,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +3692,11979989,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0702 jasmine]$ \r(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +3693,11984338,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +3694,11985362,"TERMINAL",0,0,"s",,terminal_output +3695,11985415,"TERMINAL",0,0,"h",,terminal_output +3696,11985477,"TERMINAL",0,0," ",,terminal_output +3697,11985979,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",,terminal_output +3698,11986312,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=06:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\n# source .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\n# lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\r\n# tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\r\n\r\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3518963\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=250 \\r\n --log_checkpoint_interval=250 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-noise-lvl-default-$slurm_job_id \\r\n --tags dyn breakout noise-lvl default \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 100 \\r\n --wsd_decay_steps 1000 \\r\n --num_steps 5000 \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --lam_checkpoint $lam_checkpoint \\r\n --val_interval 1000 \\r\n --eval_full_frame \\r\n",,terminal_output +3699,11986464,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=3602395\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0702\r\nSLURM_JOB_START_TIME=1759060062\r\nSLURM_STEP_NODELIST=hkn0702\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759070862\r\nSLURM_PMI2_SRUN_PORT=38045\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3527669\r\nSLURM_PTY_PORT=34957\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=36\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e8.hkn0702\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=156\r\nSLURM_NODELIST=hkn0702\r\nSLURM_SRUN_COMM_PORT=35753\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3527669\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0702\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=35753\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0702\r\n",,terminal_output +3700,11986598,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +3701,11991300,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +3702,11997423,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +3703,11997768,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +3704,11998576,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20250928_164722-j3t59oxb\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run breakout-dyn-noise-lvl-default-3527669\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/j3t59oxb\r\n",,terminal_output +3705,11998840,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26561024, 'lam': 16900976, 'tokenizer': 33750256, 'total': 77212256}\r\n",,terminal_output +3706,12000243,"TERMINAL",0,0,"WARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/interactive/3527669/000250 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/interactive/3527669/000250) to end with "".orbax-checkpoint-tmp"".\r\n",,terminal_output +3707,12001367,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +3708,12017187,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 790, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 620, in main\r\n compiled = train_step.lower(optimizer, first_batch).compile()\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/compilation.py"", line 476, in lower\r\n lowered = self.jitted_fn.lower(*pure_args, **pure_kwargs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/compilation.py"", line 126, in __call__\r\n out = self.f(*args, **kwargs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 476, in train_step\r\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/graph.py"", line 2045, in update_context_manager_wrapper\r\n return f(*args, **kwargs)\r\n ^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/autodiff.py"", line 163, in grad_wrapper\r\n fn_out = gradded_fn(*pure_args)\r\n ^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/autodiff.py"", line 88, in __call__\r\n out = self.f(*args)\r\n ^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 474, in loss_fn\r\n return dynamics_loss_fn(model, inputs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 462, in dynamics_loss_fn\r\n outputs = model(inputs)\r\n ^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py"", line 196, in __call__\r\n dyna_logits_BTNV, dyna_mask = self.dynamics(outputs)\r\n ^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/dynamics.py"", line 137, in __call__\r\n if jnp.any(jnp.isnan(vid_embed_BTNM)):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\njax.errors.TracerBoolConversionError: Attempted boolean conversion of traced array with shape bool[].\r\nThe error occurred while tracing the function train_step at /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py:468 for jit. This concrete value was not available in Python because it depends on the values of the arguments optimizer.states[0][2], optimizer.states[0][4], optimizer.states[0][401], optimizer.states[0][402], optimizer.states[0][403], optimizer.states[0][404], optimizer.states[0][405], optimizer.states[0][406], optimizer.states[0][407], optimizer.states[0][408], optimizer.states[0][409], optimizer.states[0][410], optimizer.states[0][411], optimizer.states[0][412], optimizer.states[0][413], optimizer.states[0][414], optimizer.states[0][415], optimizer.states[0][416], optimizer.states[0][417], optimizer.states[0][418], optimizer.states[0][419], optimizer.states[0][420], optimizer.states[0][421], optimizer.states[0][422], optimizer.states[0][423], optimizer.states[0][424], optimizer.states[0][425], optimizer.states[0][426], optimizer.states[0][427], optimizer.states[0][428], optimizer.states[0][429], optimizer.states[0][430], optimizer.states[0][431], optimizer.states[0][432], optimizer.states[0][433], optimizer.states[0][434], optimizer.states[0][435], optimizer.states[0][436], optimizer.states[0][437], optimizer.states[0][438], optimizer.states[0][439], optimizer.states[0][440], optimizer.states[0][441], optimizer.states[0][442], optimizer.states[0][443], optimizer.states[0][444], optimizer.states[0][445], optimizer.states[0][446], optimizer.states[0][447], optimizer.states[0][448], optimizer.states[0][449], optimizer.states[0][450], optimizer.states[0][451], optimizer.states[0][452], optimizer.states[0][453], optimizer.states[0][454], optimizer.states[0][455], optimizer.states[0][456], optimizer.states[0][457], optimizer.states[0][458], optimizer.states[0][459], optimizer.states[0][460], optimizer.states[0][461], optimizer.states[0][462], optimizer.states[0][463], optimizer.states[0][464], optimizer.states[0][465], optimizer.states[0][466], optimizer.states[0][467], optimizer.states[0][468], optimizer.states[0][469], optimizer.states[0][470], optimizer.states[0][471], optimizer.states[0][472], optimizer.states[0][473], optimizer.states[0][474], optimizer.states[0][475], optimizer.states[0][476], optimizer.states[0][477], optimizer.states[0][478], optimizer.states[0][479], optimizer.states[0][480], optimizer.states[0][481], optimizer.states[0][482], optimizer.states[0][483], optimizer.states[0][484], optimizer.states[0][485], optimizer.states[0][486], optimizer.states[0][487], optimizer.states[0][488], optimizer.states[0][489], optimizer.states[0][490], optimizer.states[0][491], optimizer.states[0][492], optimizer.states[0][493], optimizer.states[0][494], optimizer.states[0][495], optimizer.states[0][496], optimizer.states[0][497], optimizer.states[0][498], optimizer.states[0][499], optimizer.states[0][500], optimizer.states[0][501], optimizer.states[0][502], optimizer.states[0][503], optimizer.states[0][504], optimizer.states[0][505], optimizer.states[0][506], optimizer.states[0][507], optimizer.states[0][508], optimizer.states[0][509], optimizer.states[0][510], optimizer.states[0][511], optimizer.states[0][512], optimizer.states[0][513], inputs['rng'], and inputs['videos'].\r\nSee https://docs.jax.dev/en/latest/errors.html#jax.errors.TracerBoolConversionError\r\n--------------------\r\nFor simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n",,terminal_output +3709,12018824,"TERMINAL",0,0,"wandb: \r\nwandb: 🚀 View run breakout-dyn-noise-lvl-default-3527669 at: https://wandb.ai/instant-uv/jafar/runs/j3t59oxb\r\nwandb: Find logs at: ../../../../../hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20250928_164722-j3t59oxb/logs\r\n",,terminal_output +3710,12018948,"TERMINAL",0,0,"W0928 16:47:43.645570 3694452 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugonly job_name: ""jax_worker"": CANCELLED: CANCELLED\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_status:1, grpc_message:""CANCELLED""} [type.googleapis.com/tensorflow.DerivedStatus='']\r\n",,terminal_output +3711,12019556,"TERMINAL",0,0,"/usr/lib64/python3.12/multiprocessing/resource_tracker.py:254: UserWarning: resource_tracker: There appear to be 20 leaked shared_memory objects to clean up at shutdown\r\n warnings.warn('resource_tracker: There appear to be %d '\r\n",,terminal_output +3712,12019928,"TERMINAL",0,0,"srun: error: hkn0702: task 0: Exited with exit code 1\r\n]0;tum_cte0515@hkn0702:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0702 jasmine]$ ",,terminal_output +3713,12044394,"TERMINAL",0,0,"salloc: Job 3527669 has exceeded its time limit and its allocation has been revoked.\nslurmstepd: error: *** STEP 3527669.interactive ON hkn0702 CANCELLED AT 2025-09-28T16:48:09 DUE TO TIME LIMIT ***\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\n",,terminal_output +3714,12073960,"TERMINAL",0,0,"srun: error: hkn0702: task 0: Killed\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +3715,12076100,"TERMINAL",0,0,"salloc --time=03:00:00 --partition=accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +3716,12076150,"TERMINAL",0,0,"]633;Csalloc: Pending job allocation 3527864\r\nsalloc: job 3527864 queued and waiting for resources\r\n",,terminal_output +3717,12077807,"TERMINAL",0,0,"^Csalloc: Job allocation 3527864 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +3718,12078942,"TERMINAL",0,0,"queu",,terminal_command +3719,12079000,"TERMINAL",0,0,"]633;Cbash: queu: command not found...\r\n",,terminal_output +3720,12079710,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +3721,12080590,"TERMINAL",0,0,"queue",,terminal_command +3722,12080662,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Sun Sep 28 16:48:45 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3527669 accelerat interact tum_cte0 CG 3:00:27\t 1 hkn0702",,terminal_output +3723,12081704,"TERMINAL",0,0,"6",,terminal_output +3724,12082030,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +3725,12083567,"TERMINAL",0,0,"idling",,terminal_command +3726,12083630,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Sun Sep 28 16:48:48 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 25 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated:\t 6 nodes idle\rPartition dev_accelerated-h100 :\t 1 nodes idle\rPartition accelerated-h100:\t 5 nodes idle\rPartition large:\t 6 nodes idle\rPartition accelerated-h200:\t 2 nodes idle",,terminal_output +3727,12084669,"TERMINAL",0,0,"95",,terminal_output +3728,12085122,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +3729,12092599,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +3730,12092654,"TERMINAL",0,0,"]633;Csalloc: Granted job allocation 3527866\r\n",,terminal_output +3731,12092761,"TERMINAL",0,0,"salloc: Waiting for resource configuration\r\n",,terminal_output +3732,12111891,"TERMINAL",0,0,"s",,terminal_output +3733,12112013,"TERMINAL",0,0,"o",,terminal_output +3734,12112091,"TERMINAL",0,0,"ur",,terminal_output +3735,12112335,"TERMINAL",0,0,"c",,terminal_output +3736,12112487,"TERMINAL",0,0,"e",,terminal_output +3737,12112610,"TERMINAL",0,0," .",,terminal_output +3738,12112820,"TERMINAL",0,0,"v",,terminal_output +3739,12112986,"TERMINAL",0,0,"e",,terminal_output +3740,12113210,"TERMINAL",0,0,"n",,terminal_output +3741,12113463,"TERMINAL",0,0,"\t",,terminal_output +3742,12113830,"TERMINAL",0,0,"b",,terminal_output +3743,12114009,"TERMINAL",0,0,"in",,terminal_output +3744,12114223,"TERMINAL",0,0,"\t",,terminal_output +3745,12114595,"TERMINAL",0,0,"a",,terminal_output +3746,12114754,"TERMINAL",0,0,"c",,terminal_output +3747,12114981,"TERMINAL",0,0,"t",,terminal_output +3748,12115350,"TERMINAL",0,0,"\t",,terminal_output +3749,12115564,"TERMINAL",0,0,"\t",,terminal_output +3750,12116214,"TERMINAL",0,0,"",,terminal_output +3751,12116591,"TERMINAL",0,0,"\r\n",,terminal_output +3752,12119819,"TERMINAL",0,0,"salloc: Nodes hkn0402 are ready for job\r\n",,terminal_output +3753,12119965,"TERMINAL",0,0,"source .ven\tbin\tact\t\r\n",,terminal_output +3754,12120701,"TERMINAL",0,0,"]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h[tum_cte0515@hkn0402 jasmine]$ source .venv/bin/activate\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +3755,12147335,"jasmine/train_dynamics.py",0,0,"",python,tab +3756,12149106,"jasmine/models/dynamics.py",0,0,"",python,tab +3757,12150584,"jasmine/models/dynamics.py",5035,0,"",python,selection_mouse +3758,12151367,"jasmine/models/dynamics.py",5091,0,"",python,selection_mouse +3759,12151450,"jasmine/models/dynamics.py",5090,0,"",python,selection_command +3760,12152216,"jasmine/models/dynamics.py",5050,68," jax.debug.print(""NaNs in noise_level_embed after noise"")",python,selection_command +3761,12152459,"jasmine/models/dynamics.py",4994,124," if jnp.any(jnp.isnan(noise_level_embed_BT1M)): \n jax.debug.print(""NaNs in noise_level_embed after noise"")",python,selection_command +3762,12152610,"jasmine/models/dynamics.py",4933,185," jax.debug.print(""NaNs in vid_embed after noise"")\n if jnp.any(jnp.isnan(noise_level_embed_BT1M)): \n jax.debug.print(""NaNs in noise_level_embed after noise"")",python,selection_command +3763,12152761,"jasmine/models/dynamics.py",4885,233," if jnp.any(jnp.isnan(vid_embed_BTNM)): \n jax.debug.print(""NaNs in vid_embed after noise"")\n if jnp.any(jnp.isnan(noise_level_embed_BT1M)): \n jax.debug.print(""NaNs in noise_level_embed after noise"")",python,selection_command +3764,12153122,"jasmine/models/dynamics.py",4893,0,"",python,selection_command +3765,12153829,"jasmine/models/dynamics.py",5062,0,"#",python,content +3766,12153830,"jasmine/models/dynamics.py",5002,0,"#",python,content +3767,12153830,"jasmine/models/dynamics.py",4945,0,"#",python,content +3768,12153830,"jasmine/models/dynamics.py",4893,0,"#",python,content +3769,12153831,"jasmine/models/dynamics.py",4894,0,"",python,selection_keyboard +3770,12153906,"jasmine/models/dynamics.py",5066,0," ",python,content +3771,12153906,"jasmine/models/dynamics.py",5005,0," ",python,content +3772,12153906,"jasmine/models/dynamics.py",4947,0," ",python,content +3773,12153906,"jasmine/models/dynamics.py",4894,0," ",python,content +3774,12153907,"jasmine/models/dynamics.py",4895,0,"",python,selection_keyboard +3775,12154185,"jasmine/models/dynamics.py",4894,0,"",python,selection_command +3776,12164698,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_output +3777,12165592,"TERMINAL",0,0,"alloc --time=01:00:00 --partition=accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_output +3778,12165906,"TERMINAL",0,0,"\ridling",,terminal_output +3779,12166403,"TERMINAL",0,0,"queue",,terminal_output +3780,12167189,"TERMINAL",0,0,"sbatch_dir slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss",,terminal_output +3781,12169008,"TERMINAL",0,0,"\r",,terminal_output +3782,12169657,"TERMINAL",0,0,"",,terminal_output +3783,12171344,"TERMINAL",0,0,"sh",,terminal_output +3784,12171482,"TERMINAL",0,0," ",,terminal_output +3785,12171726,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",,terminal_output +3786,12172550,"TERMINAL",0,0,"\rslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=06:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\n# source .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\n# lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\r\n# tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\r\n\r\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3518963\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=250 \\r\n --log_checkpoint_interval=250 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-noise-lvl-default-$slurm_job_id \\r\n --tags dyn breakout noise-lvl default \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 100 \\r\n --wsd_decay_steps 1000 \\r\n --num_steps 5000 \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --lam_checkpoint $lam_checkpoint \\r\n --val_interval 1000 \\r\n --eval_full_frame \\r\n",,terminal_output +3787,12172756,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=217802\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0402\r\nSLURM_JOB_START_TIME=1759070937\r\nSLURM_STEP_NODELIST=hkn0402\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759074537\r\nSLURM_PMI2_SRUN_PORT=36803\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3527866\r\nSLURM_PTY_PORT=42519\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=36\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0402\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=139\r\nSLURM_NODELIST=hkn0402\r\nSLURM_SRUN_COMM_PORT=45011\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3527866\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0402\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=45011\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0402\r\n",,terminal_output +3788,12172866,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +3789,12185704,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +3790,12191827,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +3791,12192212,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +3792,12193373,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20250928_165036-64wqq56h\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run breakout-dyn-noise-lvl-default-3527866\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/64wqq56h\r\n",,terminal_output +3793,12193678,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26561024, 'lam': 16900976, 'tokenizer': 33750256, 'total': 77212256}\r\n",,terminal_output +3794,12198285,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +3795,12203765,"jasmine/train_dynamics.py",0,0,"",python,tab +3796,12206265,"jasmine/genie.py",0,0,"",python,tab +3797,12209488,"jasmine/models/dynamics.py",0,0,"",python,tab +3798,12213842,"jasmine/models/dynamics.py",4332,0,"",python,selection_mouse +3799,12213992,"jasmine/models/dynamics.py",4331,10,"mask_limit",python,selection_mouse +3800,12225317,"jasmine/models/dynamics.py",2577,0,"",python,selection_mouse +3801,12225320,"jasmine/models/dynamics.py",2576,0,"",python,selection_command +3802,12225962,"jasmine/models/dynamics.py",2537,0,"",python,selection_mouse +3803,12238080,"jasmine/models/dynamics.py",0,0,"",python,tab +3804,12280204,"TERMINAL",0,0,"Total memory size: 2.8 GB, Output size: 0.9 GB, Temp size: 1.9 GB, Argument size: 0.9 GB, Host temp size: 0.0 GB.\r\n",,terminal_output +3805,12280316,"TERMINAL",0,0,"FLOPs: 5.902e+10, Bytes: 5.203e+10 (48.5 GB), Intensity: 1.1 FLOPs/byte\r\nStarting training from step 0...\r\n",,terminal_output +3806,12284505,"TERMINAL",0,0,"Entering jdb:\r\n",,terminal_output +3807,12406857,"TERMINAL",0,0,"l",,terminal_output +3808,12407278,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/dynamics.py(110)\r\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\r\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\r\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\r\n jax.debug.breakpoint()\r\n \r\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\r\n-> jax.debug.breakpoint()\r\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M\r\n \r\n",,terminal_output +3809,12412403,"jasmine/models/dynamics.py",0,0,"",python,tab +3810,12432800,"TERMINAL",0,0,"c",,terminal_output +3811,12433067,"TERMINAL",0,0,"\r\n",,terminal_output +3812,12434548,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n",,terminal_output +3813,12437000,"TERMINAL",0,0,"l",,terminal_output +3814,12437231,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/dynamics.py(107)\r\n jax.debug.breakpoint()\r\n \r\n # safe sqrt: clip argument to >= 0\r\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\r\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\r\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\r\n-> jax.debug.breakpoint()\r\n \r\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\r\n jax.debug.breakpoint()\r\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M\r\n \r\n",,terminal_output +3815,12456999,"jasmine/models/dynamics.py",0,0,"",python,tab +3816,12459967,"jasmine/models/dynamics.py",2457,0,"",python,selection_mouse +3817,12460105,"jasmine/models/dynamics.py",2452,13,"noise_level_B",python,selection_mouse +3818,12465278,"TERMINAL",0,0,"s",,terminal_output +3819,12465475,"TERMINAL",0,0,"e",,terminal_output +3820,12465536,"TERMINAL",0,0,"l",,terminal_output +3821,12465784,"TERMINAL",0,0,"f",,terminal_output +3822,12466214,"TERMINAL",0,0," ",,terminal_output +3823,12466904,"TERMINAL",0,0,"f",,terminal_output +3824,12466967,"TERMINAL",0,0,".",,terminal_output +3825,12467260,"TERMINAL",0,0,"n",,terminal_output +3826,12467443,"TERMINAL",0,0,"o",,terminal_output +3827,12467520,"TERMINAL",0,0,"is",,terminal_output +3828,12467765,"TERMINAL",0,0,"e",,terminal_output +3829,12467910,"TERMINAL",0,0,"_",,terminal_output +3830,12468174,"TERMINAL",0,0,"b",,terminal_output +3831,12468326,"TERMINAL",0,0,"u",,terminal_output +3832,12468389,"TERMINAL",0,0,"c",,terminal_output +3833,12468534,"TERMINAL",0,0,"k",,terminal_output +3834,12468595,"TERMINAL",0,0,"e",,terminal_output +3835,12468743,"TERMINAL",0,0,"t",,terminal_output +3836,12468805,"TERMINAL",0,0,"s",,terminal_output +3837,12468958,"TERMINAL",0,0,"\r\n(jdb) 10\r\n",,terminal_output +3838,12470799,"jasmine/models/dynamics.py",0,0,"",python,tab +3839,12475530,"TERMINAL",0,0,"s",,terminal_output +3840,12475753,"TERMINAL",0,0,"el",,terminal_output +3841,12476003,"TERMINAL",0,0,".",,terminal_output +3842,12476324,"TERMINAL",0,0," ",,terminal_output +3843,12476387,"TERMINAL",0,0,"d",,terminal_output +3844,12476554,"TERMINAL",0,0,".",,terminal_output +3845,12477088,"TERMINAL",0,0," ",,terminal_output +3846,12477204,"TERMINAL",0,0," ",,terminal_output +3847,12477320,"TERMINAL",0,0,"f",,terminal_output +3848,12477427,"TERMINAL",0,0,".",,terminal_output +3849,12478071,"TERMINAL",0,0,"m",,terminal_output +3850,12478188,"TERMINAL",0,0,"a",,terminal_output +3851,12478420,"TERMINAL",0,0,"x",,terminal_output +3852,12478622,"TERMINAL",0,0,"_",,terminal_output +3853,12478927,"TERMINAL",0,0,"n",,terminal_output +3854,12479087,"TERMINAL",0,0,"o",,terminal_output +3855,12479148,"TERMINAL",0,0,"i",,terminal_output +3856,12479279,"TERMINAL",0,0,"s",,terminal_output +3857,12479770,"TERMINAL",0,0,"e",,terminal_output +3858,12480080,"TERMINAL",0,0,"_",,terminal_output +3859,12480409,"TERMINAL",0,0,"l",,terminal_output +3860,12480515,"TERMINAL",0,0,"e",,terminal_output +3861,12480666,"TERMINAL",0,0,"v",,terminal_output +3862,12480786,"TERMINAL",0,0,"e",,terminal_output +3863,12480890,"TERMINAL",0,0,"l",,terminal_output +3864,12481093,"TERMINAL",0,0,"\r\n(jdb) 0.7\r\n",,terminal_output +3865,12482920,"jasmine/models/dynamics.py",0,0,"",python,tab +3866,12530650,"jasmine/models/dynamics.py",2608,0,"",python,selection_mouse +3867,12530650,"jasmine/models/dynamics.py",2607,0,"",python,selection_command +3868,12536975,"jasmine/models/dynamics.py",0,0,"",python,tab +3869,12539065,"jasmine/models/dynamics.py",2462,0,"",python,selection_mouse +3870,12539196,"jasmine/models/dynamics.py",2452,13,"noise_level_B",python,selection_mouse +3871,12543144,"TERMINAL",0,0,"noise_level_B",,terminal_output +3872,12546054,"TERMINAL",0,0,"[",,terminal_output +3873,12546268,"TERMINAL",0,0,"0",,terminal_output +3874,12546513,"TERMINAL",0,0,"]",,terminal_output +3875,12546844,"TERMINAL",0,0,"\r\n",,terminal_output +3876,12546905,"TERMINAL",0,0,"(jdb) Array(0.45255098, dtype=float32)\r\n",,terminal_output +3877,12548931,"TERMINAL",0,0,"^[[A",,terminal_output +3878,12550790,"TERMINAL",0,0," ",,terminal_output +3879,12550996,"TERMINAL",0,0," ",,terminal_output +3880,12551141,"TERMINAL",0,0,"  ",,terminal_output +3881,12552180,"TERMINAL",0,0,"noise_level_B",,terminal_output +3882,12552866,"TERMINAL",0,0,"[",,terminal_output +3883,12553022,"TERMINAL",0,0,"1",,terminal_output +3884,12553231,"TERMINAL",0,0,"}",,terminal_output +3885,12553594,"TERMINAL",0,0,"\r\n(jdb) *** SyntaxError: closing parenthesis '}' does not match opening parenthesis '['\r\n",,terminal_output +3886,12555231,"TERMINAL",0,0,"noise_level_B",,terminal_output +3887,12555664,"TERMINAL",0,0,"\r\n(jdb) Array([0.45255098, 0.6441636 , 0.66346925, 0.04603618, 0.06802309,\r\n 0.69920456, 0.23514807, 0.65358114, 0.49310675, 0.6987329 ,\r\n 0.29514727, 0.00263699, 0.40009692, 0.6722459 , 0.54544824,\r\n 0.69171315, 0.09180859, 0.634662 , 0.59807974, 0.660335 ,\r\n 0.43653592, 0.20926596, 0.04501296, 0.4408583 , 0.46784565,\r\n 0.2517042 , 0.23913348, 0.12439712, 0.56089723, 0.66213465,\r\n 0.30742902, 0.33678833, 0.4094986 , 0.08725767, 0.40686068,\r\n 0.38637325, 0.54127145, 0.37343186, 0.6557393 , 0.20142 ,\r\n 0.55386937, 0.5326104 , 0.22468646, 0.34933808, 0.45658964,\r\n 0.53491926, 0.42469105, 0.31144363, 0.685411 , 0.4874776 ,\r\n 0.6796607 , 0.15456446, 0.63261634, 0.17223041, 0.09819267,\r\n 0.09345566, 0.20636612, 0.56945485, 0.33953947, 0.1999843 ,\r\n 0.32493833, 0.6805619 , 0.34506497, 0.68235236, 0.1110582 ,\r\n 0.5945545 , 0.14432307, 0.60123146, 0.00452096, 0.47235835,\r\n 0.20066756, 0.642303 , 0.2481875 , 0.56008446, 0.34156114,\r\n 0.648798 , 0.32805106, 0.37387052, 0.20767422, 0.07023842,\r\n 0.61710835, 0.52920985, 0.0995904 , 0.09819601, 0.31743416,\r\n 0.2991825 , 0.3547224 , 0.14432673, 0.31014135, 0.08719108,\r\n 0.1480295 , 0.4119163 , 0.4997636 , 0.20711906, 0.658858 ,\r\n 0.3466018 , 0.15562956, 0.5944861 , 0.4517534 , 0.64904135,\r\n 0.60707915, 0.49129245, 0.23407486, 0.5858251 , 0.24955001,\r\n 0.30913532, 0.10704726, 0.1676466 , 0.36603883, 0.538971 ,\r\n 0.4430276 , 0.1924243 , 0.5491045 , 0.25620812, 0.33868474,\r\n 0.20711271, 0.5150012 , 0.0739684 , 0.21971789, 0.3666449 ], dtype=float32)\r\n",,terminal_output +3888,12559864,"TERMINAL",0,0,"l",,terminal_output +3889,12560022,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/dynamics.py(107)\r\n jax.debug.breakpoint()\r\n \r\n # safe sqrt: clip argument to >= 0\r\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\r\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\r\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\r\n-> jax.debug.breakpoint()\r\n \r\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\r\n jax.debug.breakpoint()\r\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M\r\n \r\n",,terminal_output +3890,12566924,"TERMINAL",0,0,"sqrt_noise",,terminal_output +3891,12567541,"TERMINAL",0,0,"\r\n(jdb) Array([[[[0.67271906]]],\r\n\r\n\r\n [[[0.80259806]]],\r\n\r\n\r\n [[[0.8145362 ]]],\r\n\r\n\r\n [[[0.21456043]]],\r\n\r\n\r\n [[[0.26081234]]],\r\n\r\n\r\n [[[0.8361845 ]]],\r\n\r\n\r\n [[[0.48492068]]],\r\n\r\n\r\n [[[0.80844367]]],\r\n\r\n\r\n [[[0.7022156 ]]],\r\n\r\n\r\n [[[0.83590245]]],\r\n\r\n\r\n [[[0.5432746 ]]],\r\n\r\n\r\n [[[0.05135166]]],\r\n\r\n\r\n [[[0.6325321 ]]],\r\n\r\n\r\n [[[0.81990606]]],\r\n\r\n\r\n [[[0.7385447 ]]],\r\n\r\n\r\n [[[0.83169293]]],\r\n\r\n\r\n [[[0.30299932]]],\r\n\r\n\r\n [[[0.7966567 ]]],\r\n\r\n\r\n [[[0.77335614]]],\r\n\r\n\r\n [[[0.81261 ]]],\r\n\r\n\r\n [[[0.66070867]]],\r\n\r\n\r\n [[[0.45745596]]],\r\n\r\n\r\n [[[0.21216258]]],\r\n\r\n\r\n [[[0.6639716 ]]],\r\n\r\n\r\n [[[0.68399245]]],\r\n\r\n\r\n [[[0.5017013 ]]],\r\n\r\n\r\n [[[0.48901275]]],\r\n\r\n\r\n [[[0.35269976]]],\r\n\r\n\r\n [[[0.74893075]]],\r\n\r\n\r\n [[[0.81371653]]],\r\n\r\n\r\n [[[0.5544628 ]]],\r\n\r\n\r\n [[[0.58033466]]],\r\n\r\n\r\n [[[0.6399208 ]]],\r\n\r\n\r\n [[[0.2953941 ]]],\r\n\r\n\r\n [[[0.6378563 ]]],\r\n\r\n\r\n [[[0.6215893 ]]],\r\n\r\n\r\n [[[0.7357115 ]]],\r\n\r\n\r\n [[[0.6110907 ]]],\r\n\r\n\r\n [[[0.8097773 ]]],\r\n\r\n\r\n [[[0.4487984 ]]],\r\n\r\n\r\n [[[0.744224 ]]],\r\n\r\n\r\n [[[0.7298016 ]]],\r\n\r\n\r\n [[[0.47401103]]],\r\n\r\n\r\n [[[0.5910483 ]]],\r\n\r\n\r\n [[[0.67571414]]],\r\n\r\n\r\n [[[0.7313818 ]]],\r\n\r\n\r\n [[[0.6516832 ]]],\r\n\r\n\r\n [[[0.5580714 ]]],\r\n\r\n\r\n [[[0.8278955 ]]],\r\n\r\n\r\n [[[0.69819593]]],\r\n\r\n\r\n [[[0.8244154 ]]],\r\n\r\n\r\n [[[0.39314684]]],\r\n\r\n\r\n [[[0.79537183]]],\r\n\r\n\r\n [[[0.41500652]]],\r\n\r\n\r\n [[[0.31335708]]],\r\n\r\n\r\n [[[0.3057052 ]]],\r\n\r\n\r\n [[[0.45427537]]],\r\n\r\n\r\n [[[0.75462234]]],\r\n\r\n\r\n [[[0.58270013]]],\r\n\r\n\r\n [[[0.44719604]]],\r\n\r\n\r\n [[[0.5700336 ]]],\r\n\r\n\r\n [[[0.8249617 ]]],\r\n\r\n\r\n [[[0.5874223 ]]],\r\n\r\n\r\n [[[0.8260462 ]]],\r\n\r\n\r\n [[[0.33325395]]],\r\n\r\n\r\n [[[0.7710736 ]]],\r\n\r\n\r\n [[[0.37989876]]],\r\n\r\n\r\n [[[0.77539116]]],\r\n\r\n\r\n [[[0.06723811]]],\r\n\r\n\r\n [[[0.6872833 ]]],\r\n\r\n\r\n [[[0.44795933]]],\r\n\r\n\r\n [[[0.80143803]]],\r\n\r\n\r\n [[[0.4981842 ]]],\r\n\r\n\r\n [[[0.74838793]]],\r\n\r\n\r\n [[[0.5844323 ]]],\r\n\r\n\r\n [[[0.80547994]]],\r\n\r\n\r\n [[[0.5727574 ]]],\r\n\r\n\r\n [[[0.61144954]]],\r\n\r\n\r\n [[[0.45571285]]],\r\n\r\n\r\n [[[0.26502532]]],\r\n\r\n\r\n [[[0.78556246]]],\r\n\r\n\r\n [[[0.72746813]]],\r\n\r\n\r\n [[[0.31557947]]],\r\n\r\n\r\n [[[0.31336242]]],\r\n\r\n\r\n [[[0.56341296]]],\r\n\r\n\r\n [[[0.5469758 ]]],\r\n\r\n\r\n [[[0.59558576]]],\r\n\r\n\r\n [[[0.37990358]]],\r\n\r\n\r\n [[[0.55690336]]],\r\n\r\n\r\n [[[0.29528135]]],\r\n\r\n\r\n [[[0.38474604]]],\r\n\r\n\r\n [[[0.641807 ]]],\r\n\r\n\r\n [[[0.70693964]]],\r\n\r\n\r\n [[[0.45510334]]],\r\n\r\n\r\n [[[0.81170064]]],\r\n\r\n\r\n [[[0.58872896]]],\r\n\r\n\r\n [[[0.39449912]]],\r\n\r\n\r\n [[[0.77102923]]],\r\n\r\n\r\n [[[0.672126 ]]],\r\n\r\n\r\n [[[0.80563104]]],\r\n\r\n\r\n [[[0.7791528 ]]],\r\n\r\n\r\n [[[0.70092255]]],\r\n\r\n\r\n [[[0.48381284]]],\r\n\r\n\r\n [[[0.7653921 ]]],\r\n\r\n\r\n [[[0.4995498 ]]],\r\n\r\n\r\n [[[0.55599934]]],\r\n\r\n\r\n [[[0.32718077]]],\r\n\r\n\r\n [[[0.4094467 ]]],\r\n\r\n\r\n [[[0.6050114 ]]],\r\n\r\n\r\n [[[0.7341464 ]]],\r\n\r\n\r\n [[[0.66560316]]],\r\n\r\n\r\n [[[0.43866193]]],\r\n\r\n\r\n [[[0.74101585]]],\r\n\r\n\r\n [[[0.50617003]]],\r\n\r\n\r\n [[[0.5819663 ]]],\r\n\r\n\r\n [[[0.45509636]]],\r\n\r\n\r\n [[[0.7176358 ]]],\r\n\r\n\r\n [[[0.27197132]]],\r\n\r\n\r\n [[[0.46874073]]],\r\n\r\n\r\n [[[0.6055121 ]]]], dtype=float32)\r\n",,terminal_output +3892,12597573,"TERMINAL",0,0,"sqrt_one_minus",,terminal_output +3893,12598040,"TERMINAL",0,0,"\r\n(jdb) Array([[[[0.73989797]]],\r\n\r\n\r\n [[[0.59652025]]],\r\n\r\n\r\n [[[0.5801127 ]]],\r\n\r\n\r\n [[[0.9767107 ]]],\r\n\r\n\r\n [[[0.9653895 ]]],\r\n\r\n\r\n [[[0.5484482 ]]],\r\n\r\n\r\n [[[0.87455815]]],\r\n\r\n\r\n [[[0.5885736 ]]],\r\n\r\n\r\n [[[0.71196437]]],\r\n\r\n\r\n [[[0.548878 ]]],\r\n\r\n\r\n [[[0.839555 ]]],\r\n\r\n\r\n [[[0.99868065]]],\r\n\r\n\r\n [[[0.7745341 ]]],\r\n\r\n\r\n [[[0.5724981 ]]],\r\n\r\n\r\n [[[0.6742045 ]]],\r\n\r\n\r\n [[[0.55523586]]],\r\n\r\n\r\n [[[0.9529907 ]]],\r\n\r\n\r\n [[[0.604432 ]]],\r\n\r\n\r\n [[[0.6339718 ]]],\r\n\r\n\r\n [[[0.58280784]]],\r\n\r\n\r\n [[[0.7506424 ]]],\r\n\r\n\r\n [[[0.8892323 ]]],\r\n\r\n\r\n [[[0.97723436]]],\r\n\r\n\r\n [[[0.7477578 ]]],\r\n\r\n\r\n [[[0.7294891 ]]],\r\n\r\n\r\n [[[0.8650409 ]]],\r\n\r\n\r\n [[[0.8722766 ]]],\r\n\r\n\r\n [[[0.93573654]]],\r\n\r\n\r\n [[[0.6626483 ]]],\r\n\r\n\r\n [[[0.5812619 ]]],\r\n\r\n\r\n [[[0.8322085 ]]],\r\n\r\n\r\n [[[0.8143781 ]]],\r\n\r\n\r\n [[[0.7684409 ]]],\r\n\r\n\r\n [[[0.9553755 ]]],\r\n\r\n\r\n [[[0.7701554 ]]],\r\n\r\n\r\n [[[0.78334326]]],\r\n\r\n\r\n [[[0.677295 ]]],\r\n\r\n\r\n [[[0.79156053]]],\r\n\r\n\r\n [[[0.58673733]]],\r\n\r\n\r\n [[[0.893633 ]]],\r\n\r\n\r\n [[[0.6679301 ]]],\r\n\r\n\r\n [[[0.68365896]]],\r\n\r\n\r\n [[[0.8805189 ]]],\r\n\r\n\r\n [[[0.80663615]]],\r\n\r\n\r\n [[[0.7371637 ]]],\r\n\r\n\r\n [[[0.6819683 ]]],\r\n\r\n\r\n [[[0.7584912 ]]],\r\n\r\n\r\n [[[0.829793 ]]],\r\n\r\n\r\n [[[0.56088233]]],\r\n\r\n\r\n [[[0.7159067 ]]],\r\n\r\n\r\n [[[0.56598526]]],\r\n\r\n\r\n [[[0.9194757 ]]],\r\n\r\n\r\n [[[0.6061218 ]]],\r\n\r\n\r\n [[[0.9098185 ]]],\r\n\r\n\r\n [[[0.9496353 ]]],\r\n\r\n\r\n [[[0.9521262 ]]],\r\n\r\n\r\n [[[0.89086133]]],\r\n\r\n\r\n [[[0.6561594 ]]],\r\n\r\n\r\n [[[0.8126872 ]]],\r\n\r\n\r\n [[[0.89443594]]],\r\n\r\n\r\n [[[0.82162136]]],\r\n\r\n\r\n [[[0.5651885 ]]],\r\n\r\n\r\n [[[0.8092806 ]]],\r\n\r\n\r\n [[[0.5636024 ]]],\r\n\r\n\r\n [[[0.9428371 ]]],\r\n\r\n\r\n [[[0.63674605]]],\r\n\r\n\r\n [[[0.9250281 ]]],\r\n\r\n\r\n [[[0.63148123]]],\r\n\r\n\r\n [[[0.99773693]]],\r\n\r\n\r\n [[[0.72638947]]],\r\n\r\n\r\n [[[0.89405394]]],\r\n\r\n\r\n [[[0.5980778 ]]],\r\n\r\n\r\n [[[0.8670712 ]]],\r\n\r\n\r\n [[[0.6632613 ]]],\r\n\r\n\r\n [[[0.81144243]]],\r\n\r\n\r\n [[[0.592623 ]]],\r\n\r\n\r\n [[[0.8197249 ]]],\r\n\r\n\r\n [[[0.7912834 ]]],\r\n\r\n\r\n [[[0.8901268 ]]],\r\n\r\n\r\n [[[0.96424145]]],\r\n\r\n\r\n [[[0.6187824 ]]],\r\n\r\n\r\n [[[0.6861415 ]]],\r\n\r\n\r\n [[[0.94889915]]],\r\n\r\n\r\n [[[0.9496336 ]]],\r\n\r\n\r\n [[[0.8261754 ]]],\r\n\r\n\r\n [[[0.8371484 ]]],\r\n\r\n\r\n [[[0.80329174]]],\r\n\r\n\r\n [[[0.92502606]]],\r\n\r\n\r\n [[[0.8305773 ]]],\r\n\r\n\r\n [[[0.9554103 ]]],\r\n\r\n\r\n [[[0.92302245]]],\r\n\r\n\r\n [[[0.76686615]]],\r\n\r\n\r\n [[[0.7072739 ]]],\r\n\r\n\r\n [[[0.8904386 ]]],\r\n\r\n\r\n [[[0.5840736 ]]],\r\n\r\n\r\n [[[0.8083305 ]]],\r\n\r\n\r\n [[[0.9188963 ]]],\r\n\r\n\r\n [[[0.63679975]]],\r\n\r\n\r\n [[[0.74043673]]],\r\n\r\n\r\n [[[0.59241766]]],\r\n\r\n\r\n [[[0.626834 ]]],\r\n\r\n\r\n [[[0.71323735]]],\r\n\r\n\r\n [[[0.8751715 ]]],\r\n\r\n\r\n [[[0.6435642 ]]],\r\n\r\n\r\n [[[0.86628515]]],\r\n\r\n\r\n [[[0.83118266]]],\r\n\r\n\r\n [[[0.9449617 ]]],\r\n\r\n\r\n [[[0.912334 ]]],\r\n\r\n\r\n [[[0.7962168 ]]],\r\n\r\n\r\n [[[0.67899114]]],\r\n\r\n\r\n [[[0.7463058 ]]],\r\n\r\n\r\n [[[0.89865214]]],\r\n\r\n\r\n [[[0.6714875 ]]],\r\n\r\n\r\n [[[0.8624337 ]]],\r\n\r\n\r\n [[[0.81321293]]],\r\n\r\n\r\n [[[0.8904422 ]]],\r\n\r\n\r\n [[[0.6964186 ]]],\r\n\r\n\r\n [[[0.96230537]]],\r\n\r\n\r\n [[[0.88333577]]],\r\n\r\n\r\n [[[0.7958361 ]]]], dtype=float32)\r\n",,terminal_output +3894,12618984,"TERMINAL",0,0,"l",,terminal_output +3895,12619191,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/dynamics.py(107)\r\n jax.debug.breakpoint()\r\n \r\n # safe sqrt: clip argument to >= 0\r\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\r\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\r\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\r\n-> jax.debug.breakpoint()\r\n \r\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\r\n jax.debug.breakpoint()\r\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M\r\n \r\n",,terminal_output +3896,12624951,"jasmine/train_dynamics.py",0,0,"",python,tab +3897,12625973,"jasmine/models/dynamics.py",0,0,"",python,tab +3898,12630715,"jasmine/models/dynamics.py",2625,0,"",python,selection_mouse +3899,12630859,"jasmine/models/dynamics.py",2617,10,"noise_BTNM",python,selection_mouse +3900,12642002,"jasmine/models/dynamics.py",3156,0,"",python,selection_mouse +3901,12642159,"jasmine/models/dynamics.py",3153,4,"self",python,selection_mouse +3902,12643017,"jasmine/models/dynamics.py",3160,0,"",python,selection_mouse +3903,12643166,"jasmine/models/dynamics.py",3158,17,"noise_level_embed",python,selection_mouse +3904,12653059,"jasmine/models/dynamics.py",3182,0,"",python,selection_mouse +3905,12653223,"jasmine/models/dynamics.py",3176,20,"noise_bucket_idx_B11",python,selection_mouse +3906,12657373,"jasmine/models/dynamics.py",3154,0,"",python,selection_mouse +3907,12657506,"jasmine/models/dynamics.py",3153,4,"self",python,selection_mouse +3908,12657721,"jasmine/models/dynamics.py",3153,5,"self.",python,selection_mouse +3909,12657738,"jasmine/models/dynamics.py",3153,22,"self.noise_level_embed",python,selection_mouse +3910,12661260,"TERMINAL",0,0,"self.noise_level_embed",,terminal_output +3911,12662046,"TERMINAL",0,0,"(",,terminal_output +3912,12662422,"TERMINAL",0,0,"0",,terminal_output +3913,12663036,"TERMINAL",0,0,")",,terminal_output +3914,12663526,"TERMINAL",0,0,"\r\n(jdb) *** AttributeError: 'int' object has no attribute 'dtype'\r\n",,terminal_output +3915,12666853,"TERMINAL",0,0,"o",,terminal_output +3916,12668550,"TERMINAL",0,0," ",,terminal_output +3917,12669041,"TERMINAL",0,0,"z",,terminal_output +3918,12669467,"TERMINAL",0,0,"e",,terminal_output +3919,12669796,"TERMINAL",0,0,"r",,terminal_output +3920,12670404,"TERMINAL",0,0,"o",,terminal_output +3921,12670525,"TERMINAL",0,0," ",,terminal_output +3922,12670718,"TERMINAL",0,0,"=",,terminal_output +3923,12670836,"TERMINAL",0,0," ",,terminal_output +3924,12673139,"TERMINAL",0,0,"j",,terminal_output +3925,12673244,"TERMINAL",0,0,"a",,terminal_output +3926,12673417,"TERMINAL",0,0,"x",,terminal_output +3927,12673600,"TERMINAL",0,0,"-",,terminal_output +3928,12674209,"TERMINAL",0,0," ",,terminal_output +3929,12674446,"TERMINAL",0,0,".",,terminal_output +3930,12677213,"TERMINAL",0,0,"bash",,terminal_focus +3931,12677719,"TERMINAL",0,0,"srun",,terminal_focus +3932,12678149,"jasmine/train_dynamics.py",0,0,"",python,tab +3933,12679765,"jasmine/genie.py",0,0,"",python,tab +3934,12681021,"jasmine/models/dynamics.py",0,0,"",python,tab +3935,12682777,"jasmine/models/dynamics.py",3259,0,"",python,selection_mouse +3936,12682954,"jasmine/models/dynamics.py",3240,22,"noise_level_embed_B11M",python,selection_mouse +3937,12685231,"jasmine/models/dynamics.py",2635,0,"",python,selection_mouse +3938,12686343,"jasmine/models/dynamics.py",2600,0,"",python,selection_mouse +3939,12690057,"TERMINAL",0,0,"A",,terminal_output +3940,12690306,"TERMINAL",0,0,"r",,terminal_output +3941,12690454,"TERMINAL",0,0,"r",,terminal_output +3942,12690660,"TERMINAL",0,0,"a",,terminal_output +3943,12690876,"TERMINAL",0,0,"y",,terminal_output +3944,12691300,"TERMINAL",0,0,"(",,terminal_output +3945,12692371,"TERMINAL",0,0,"0",,terminal_output +3946,12693514,"TERMINAL",0,0,")",,terminal_output +3947,12694089,"TERMINAL",0,0,"\r\n(jdb) *** SyntaxError: invalid syntax\r\n",,terminal_output +3948,12699578,"jasmine/train_dynamics.py",0,0,"",python,tab +3949,12703499,"jasmine/models/dynamics.py",0,0,"",python,tab +3950,12710533,"TERMINAL",0,0,"n",,terminal_output +3951,12710714,"TERMINAL",0,0,"o",,terminal_output +3952,12710781,"TERMINAL",0,0,"i",,terminal_output +3953,12710884,"TERMINAL",0,0,"s",,terminal_output +3954,12711027,"TERMINAL",0,0,"e",,terminal_output +3955,12711240,"TERMINAL",0,0,"_",,terminal_output +3956,12711493,"TERMINAL",0,0,"l",,terminal_output +3957,12711556,"TERMINAL",0,0,"e",,terminal_output +3958,12711710,"TERMINAL",0,0,"v",,terminal_output +3959,12711869,"TERMINAL",0,0,"e",,terminal_output +3960,12711932,"TERMINAL",0,0,"l",,terminal_output +3961,12712169,"TERMINAL",0,0,"_",,terminal_output +3962,12712409,"TERMINAL",0,0,"e",,terminal_output +3963,12712471,"TERMINAL",0,0,"m",,terminal_output +3964,12712708,"TERMINAL",0,0,"b",,terminal_output +3965,12712771,"TERMINAL",0,0,"e",,terminal_output +3966,12712893,"TERMINAL",0,0,"d",,terminal_output +3967,12713094,"TERMINAL",0,0,"\t",,terminal_output +3968,12713882,"TERMINAL",0,0,"",,terminal_output +3969,12715070,"TERMINAL",0,0,"_",,terminal_output +3970,12715439,"TERMINAL",0,0,"B",,terminal_output +3971,12716616,"TERMINAL",0,0,"1",,terminal_output +3972,12716766,"TERMINAL",0,0,"1",,terminal_output +3973,12717600,"TERMINAL",0,0,"M",,terminal_output +3974,12717945,"TERMINAL",0,0,"\r\n(jdb) Array([[[[ 4.4392552e-02, -2.9844549e-02, -3.1102074e-02, ...,\r\n -1.2073218e-02, 4.3277056e-03, -6.1354764e-02]]],\r\n\r\n\r\n [[[ 1.1847667e-01, -1.0830977e-02, -1.6250851e-02, ...,\r\n -2.4750691e-02, -2.7230620e-02, -8.5215392e-03]]],\r\n\r\n\r\n [[[ 1.1847667e-01, -1.0830977e-02, -1.6250851e-02, ...,\r\n -2.4750691e-02, -2.7230620e-02, -8.5215392e-03]]],\r\n\r\n\r\n ...,\r\n\r\n\r\n [[[-5.4778680e-02, 4.7858335e-02, 1.9920882e-02, ...,\r\n 5.6935117e-02, -1.5326574e-02, 2.6951581e-02]]],\r\n\r\n\r\n [[[ 1.6207712e-02, -4.6082556e-02, 3.1736858e-05, ...,\r\n 9.7126283e-02, 2.2778701e-02, 4.0227242e-02]]],\r\n\r\n\r\n [[[ 3.6209282e-02, -5.9933860e-02, 2.8886722e-02, ...,\r\n -1.1514775e-02, -2.0673620e-02, 2.2546995e-02]]]], dtype=float32)\r\n",,terminal_output +3975,12725948,"jasmine/models/dynamics.py",0,0,"",python,tab +3976,12764262,"jasmine/models/dynamics.py",3052,0,"",python,selection_mouse +3977,12764263,"jasmine/models/dynamics.py",3051,0,"",python,selection_command +3978,12814446,"jasmine/models/dynamics.py",0,0,"",python,tab +3979,12816343,"jasmine/models/dynamics.py",3169,0,"",python,selection_mouse +3980,12816513,"jasmine/models/dynamics.py",3158,17,"noise_level_embed",python,selection_mouse +3981,12823453,"jasmine/models/dynamics.py",4421,0,"",python,selection_command +3982,12827338,"jasmine/models/dynamics.py",3222,0,"",python,selection_mouse +3983,12827537,"jasmine/models/dynamics.py",3206,22,"noise_level_embed_BT1M",python,selection_mouse +3984,12831764,"jasmine/models/dynamics.py",4809,0,"",python,selection_mouse +3985,12831943,"jasmine/models/dynamics.py",4808,22,"noise_level_embed_BT1M",python,selection_mouse +3986,12834175,"jasmine/models/dynamics.py",5127,0,"",python,selection_mouse +3987,12835731,"jasmine/models/dynamics.py",5127,0,"k",python,content +3988,12835733,"jasmine/models/dynamics.py",5128,0,"",python,selection_keyboard +3989,12836295,"jasmine/models/dynamics.py",5127,1,"",python,content +3990,12836580,"jasmine/models/dynamics.py",5056,0,"",python,selection_command +3991,12836946,"jasmine/models/dynamics.py",4998,129,"",python,content +3992,12837102,"jasmine/models/dynamics.py",4935,0,"",python,selection_command +3993,12837470,"jasmine/models/dynamics.py",4885,113,"",python,content +3994,12840069,"jasmine/models/dynamics.py",5193,0,"",python,selection_mouse +3995,12840226,"jasmine/models/dynamics.py",5182,22,"noise_level_embed_BT1M",python,selection_mouse +3996,12889786,"jasmine/models/dynamics.py",2905,0,"",python,selection_mouse +3997,12889925,"jasmine/models/dynamics.py",2890,18,"noise_bucket_idx_B",python,selection_mouse +3998,12893742,"TERMINAL",0,0,"noise_bucket_idx_B",,terminal_output +3999,12894038,"TERMINAL",0,0,"\r\n(jdb) Array([6, 9, 9, 0, 0, 9, 3, 9, 7, 9, 4, 0, 5, 9, 7, 9, 1, 9, 8, 9, 6, 2,\r\n 0, 6, 6, 3, 3, 1, 8, 9, 4, 4, 5, 1, 5, 5, 7, 5, 9, 2, 7, 7, 3, 4,\r\n 6, 7, 6, 4, 9, 6, 9, 2, 9, 2, 1, 1, 2, 8, 4, 2, 4, 9, 4, 9, 1, 8,\r\n 2, 8, 0, 6, 2, 9, 3, 8, 4, 9, 4, 5, 2, 1, 8, 7, 1, 1, 4, 4, 5, 2,\r\n 4, 1, 2, 5, 7, 2, 9, 4, 2, 8, 6, 9, 8, 7, 3, 8, 3, 4, 1, 2, 5, 7,\r\n 6, 2, 7, 3, 4, 2, 7, 1, 3, 5], dtype=int32)\r\n",,terminal_output +4000,12906415,"jasmine/train_dynamics.py",0,0,"",python,tab +4001,12908002,"TERMINAL",0,0,"bash",,terminal_focus +4002,12908661,"TERMINAL",0,0,"srun",,terminal_focus +4003,12910061,"jasmine/train_dynamics.py",0,0,"",python,tab +4004,12911724,"jasmine/genie.py",0,0,"",python,tab +4005,12917770,"jasmine/models/dynamics.py",0,0,"",python,tab +4006,12944000,"jasmine/train_dynamics.py",0,0,"",python,tab +4007,12963349,"jasmine/train_dynamics.py",22407,0,"",python,selection_mouse +4008,12963515,"jasmine/train_dynamics.py",22405,3,"rng",python,selection_mouse +4009,13056201,"jasmine/train_dynamics.py",15973,0,"",python,selection_mouse +4010,13056877,"jasmine/train_dynamics.py",15899,0,"",python,selection_mouse +4011,13056890,"jasmine/train_dynamics.py",15898,0,"",python,selection_command +4012,13092691,"jasmine/models/dynamics.py",0,0,"",python,tab +4013,13110859,"jasmine/models/dynamics.py",3052,0,"",python,selection_mouse +4014,13110861,"jasmine/models/dynamics.py",3051,0,"",python,selection_command +4015,13111621,"jasmine/models/dynamics.py",3021,0,"",python,selection_mouse +4016,13111624,"jasmine/models/dynamics.py",3020,0,"",python,selection_command +4017,13112807,"jasmine/models/dynamics.py",3052,0,"",python,selection_mouse +4018,13112808,"jasmine/models/dynamics.py",3051,0,"",python,selection_command +4019,13127737,"jasmine/models/dynamics.py",3634,0,"",python,selection_mouse +4020,13127738,"jasmine/models/dynamics.py",3633,0,"",python,selection_command +4021,13128356,"jasmine/models/dynamics.py",3684,0,"",python,selection_mouse +4022,13129053,"jasmine/models/dynamics.py",3668,0,"",python,selection_mouse +4023,13129167,"jasmine/models/dynamics.py",3644,30,"noise_augmented_vid_embed_BTNM",python,selection_mouse +4024,13132062,"jasmine/models/dynamics.py",3516,0,"",python,selection_command +4025,13140011,"TERMINAL",0,0,"l",,terminal_output +4026,13140207,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/dynamics.py(107)\r\n jax.debug.breakpoint()\r\n \r\n # safe sqrt: clip argument to >= 0\r\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\r\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\r\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\r\n-> jax.debug.breakpoint()\r\n \r\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\r\n jax.debug.breakpoint()\r\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M\r\n \r\n",,terminal_output +4027,13143637,"TERMINAL",0,0,"s",,terminal_output +4028,13143864,"TERMINAL",0,0,"q",,terminal_output +4029,13144123,"TERMINAL",0,0,"r",,terminal_output +4030,13144256,"TERMINAL",0,0,"t",,terminal_output +4031,13144517,"TERMINAL",0,0,"_",,terminal_output +4032,13145252,"TERMINAL",0,0,"m",,terminal_output +4033,13145443,"TERMINAL",0,0,"i",,terminal_output +4034,13145506,"TERMINAL",0,0,"n",,terminal_output +4035,13146039,"TERMINAL",0,0,"u",,terminal_output +4036,13146774,"TERMINAL",0,0," ",,terminal_output +4037,13146933,"TERMINAL",0,0," ",,terminal_output +4038,13147066,"TERMINAL",0,0," ",,terminal_output +4039,13147187,"TERMINAL",0,0," ",,terminal_output +4040,13147492,"TERMINAL",0,0,"o",,terminal_output +4041,13147644,"TERMINAL",0,0,"n",,terminal_output +4042,13147780,"TERMINAL",0,0,"e",,terminal_output +4043,13147973,"TERMINAL",0,0,"_",,terminal_output +4044,13148237,"TERMINAL",0,0,"m",,terminal_output +4045,13148411,"TERMINAL",0,0,"i",,terminal_output +4046,13148472,"TERMINAL",0,0,"n",,terminal_output +4047,13148706,"TERMINAL",0,0,"u",,terminal_output +4048,13148846,"TERMINAL",0,0,"s",,terminal_output +4049,13149011,"TERMINAL",0,0,"\r\n(jdb) Array([[[[0.73989797]]],\r\n\r\n\r\n [[[0.59652025]]],\r\n\r\n\r\n [[[0.5801127 ]]],\r\n\r\n\r\n [[[0.9767107 ]]],\r\n\r\n\r\n [[[0.9653895 ]]],\r\n\r\n\r\n [[[0.5484482 ]]],\r\n\r\n\r\n [[[0.87455815]]],\r\n\r\n\r\n [[[0.5885736 ]]],\r\n\r\n\r\n [[[0.71196437]]],\r\n\r\n\r\n [[[0.548878 ]]],\r\n\r\n\r\n [[[0.839555 ]]],\r\n\r\n\r\n [[[0.99868065]]],\r\n\r\n\r\n [[[0.7745341 ]]],\r\n\r\n\r\n [[[0.5724981 ]]],\r\n\r\n\r\n [[[0.6742045 ]]],\r\n\r\n\r\n [[[0.55523586]]],\r\n\r\n\r\n [[[0.9529907 ]]],\r\n\r\n\r\n [[[0.604432 ]]],\r\n\r\n\r\n [[[0.6339718 ]]],\r\n\r\n\r\n [[[0.58280784]]],\r\n\r\n\r\n [[[0.7506424 ]]],\r\n\r\n\r\n [[[0.8892323 ]]],\r\n\r\n\r\n [[[0.97723436]]],\r\n\r\n\r\n [[[0.7477578 ]]],\r\n\r\n\r\n [[[0.7294891 ]]],\r\n\r\n\r\n [[[0.8650409 ]]],\r\n\r\n\r\n [[[0.8722766 ]]],\r\n\r\n\r\n [[[0.93573654]]],\r\n\r\n\r\n [[[0.6626483 ]]],\r\n\r\n\r\n [[[0.5812619 ]]],\r\n\r\n\r\n [[[0.8322085 ]]],\r\n\r\n\r\n [[[0.8143781 ]]],\r\n\r\n\r\n [[[0.7684409 ]]],\r\n\r\n\r\n [[[0.9553755 ]]],\r\n\r\n\r\n [[[0.7701554 ]]],\r\n\r\n\r\n [[[0.78334326]]],\r\n\r\n\r\n [[[0.677295 ]]],\r\n\r\n\r\n [[[0.79156053]]],\r\n\r\n\r\n [[[0.58673733]]],\r\n\r\n\r\n [[[0.893633 ]]],\r\n\r\n\r\n [[[0.6679301 ]]],\r\n\r\n\r\n [[[0.68365896]]],\r\n\r\n\r\n [[[0.8805189 ]]],\r\n\r\n\r\n [[[0.80663615]]],\r\n\r\n\r\n [[[0.7371637 ]]],\r\n\r\n\r\n [[[0.6819683 ]]],\r\n\r\n\r\n [[[0.7584912 ]]],\r\n\r\n\r\n [[[0.829793 ]]],\r\n\r\n\r\n [[[0.56088233]]],\r\n\r\n\r\n [[[0.7159067 ]]],\r\n\r\n\r\n [[[0.56598526]]],\r\n\r\n\r\n [[[0.9194757 ]]],\r\n\r\n\r\n [[[0.6061218 ]]],\r\n\r\n\r\n [[[0.9098185 ]]],\r\n\r\n\r\n [[[0.9496353 ]]],\r\n\r\n\r\n [[[0.9521262 ]]],\r\n\r\n\r\n [[[0.89086133]]],\r\n\r\n\r\n [[[0.6561594 ]]],\r\n\r\n\r\n [[[0.8126872 ]]],\r\n\r\n\r\n [[[0.89443594]]],\r\n\r\n\r\n [[[0.82162136]]],\r\n\r\n\r\n [[[0.5651885 ]]],\r\n\r\n\r\n [[[0.8092806 ]]],\r\n\r\n\r\n [[[0.5636024 ]]],\r\n\r\n\r\n [[[0.9428371 ]]],\r\n\r\n\r\n [[[0.63674605]]],\r\n\r\n\r\n [[[0.9250281 ]]],\r\n\r\n\r\n [[[0.63148123]]],\r\n\r\n\r\n [[[0.99773693]]],\r\n\r\n\r\n [[[0.72638947]]],\r\n\r\n\r\n [[[0.89405394]]],\r\n\r\n\r\n [[[0.5980778 ]]],\r\n\r\n\r\n [[[0.8670712 ]]],\r\n\r\n\r\n [[[0.6632613 ]]],\r\n\r\n\r\n [[[0.81144243]]],\r\n\r\n\r\n [[[0.592623 ]]],\r\n\r\n\r\n [[[0.8197249 ]]],\r\n\r\n\r\n [[[0.7912834 ]]],\r\n\r\n\r\n [[[0.8901268 ]]],\r\n\r\n\r\n [[[0.96424145]]],\r\n\r\n\r\n [[[0.6187824 ]]],\r\n\r\n\r\n [[[0.6861415 ]]],\r\n\r\n\r\n [[[0.94889915]]],\r\n\r\n\r\n [[[0.9496336 ]]],\r\n\r\n\r\n [[[0.8261754 ]]],\r\n\r\n\r\n [[[0.8371484 ]]],\r\n\r\n\r\n [[[0.80329174]]],\r\n\r\n\r\n [[[0.92502606]]],\r\n\r\n\r\n [[[0.8305773 ]]],\r\n\r\n\r\n [[[0.9554103 ]]],\r\n\r\n\r\n [[[0.92302245]]],\r\n\r\n\r\n [[[0.76686615]]],\r\n\r\n\r\n [[[0.7072739 ]]],\r\n\r\n\r\n [[[0.8904386 ]]],\r\n\r\n\r\n [[[0.5840736 ]]],\r\n\r\n\r\n [[[0.8083305 ]]],\r\n\r\n\r\n [[[0.9188963 ]]],\r\n\r\n\r\n [[[0.63679975]]],\r\n\r\n\r\n [[[0.74043673]]],\r\n\r\n\r\n [[[0.59241766]]],\r\n\r\n\r\n [[[0.626834 ]]],\r\n\r\n\r\n [[[0.71323735]]],\r\n\r\n\r\n [[[0.8751715 ]]],\r\n\r\n\r\n [[[0.6435642 ]]],\r\n\r\n\r\n [[[0.86628515]]],\r\n\r\n\r\n [[[0.83118266]]],\r\n\r\n\r\n [[[0.9449617 ]]],\r\n\r\n\r\n [[[0.912334 ]]],\r\n\r\n\r\n [[[0.7962168 ]]],\r\n\r\n\r\n [[[0.67899114]]],\r\n\r\n\r\n [[[0.7463058 ]]],\r\n\r\n\r\n [[[0.89865214]]],\r\n\r\n\r\n [[[0.6714875 ]]],\r\n\r\n\r\n [[[0.8624337 ]]],\r\n\r\n\r\n [[[0.81321293]]],\r\n\r\n\r\n [[[0.8904422 ]]],\r\n\r\n\r\n [[[0.6964186 ]]],\r\n\r\n\r\n [[[0.96230537]]],\r\n\r\n\r\n [[[0.88333577]]],\r\n\r\n\r\n [[[0.7958361 ]]]], dtype=float32)\r\n",,terminal_output +4050,13155452,"TERMINAL",0,0,"l",,terminal_output +4051,13155560,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/dynamics.py(107)\r\n jax.debug.breakpoint()\r\n \r\n # safe sqrt: clip argument to >= 0\r\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\r\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\r\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\r\n-> jax.debug.breakpoint()\r\n \r\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\r\n jax.debug.breakpoint()\r\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M\r\n \r\n",,terminal_output +4052,13161360,"TERMINAL",0,0,"noise_level_B111",,terminal_output +4053,13162500,"TERMINAL",0,0,"\r\n(jdb) Array([[[[0.45255098]]],\r\n\r\n\r\n [[[0.6441636 ]]],\r\n\r\n\r\n [[[0.66346925]]],\r\n\r\n\r\n [[[0.04603618]]],\r\n\r\n\r\n [[[0.06802309]]],\r\n\r\n\r\n [[[0.69920456]]],\r\n\r\n\r\n [[[0.23514807]]],\r\n\r\n\r\n [[[0.65358114]]],\r\n\r\n\r\n [[[0.49310675]]],\r\n\r\n\r\n [[[0.6987329 ]]],\r\n\r\n\r\n [[[0.29514727]]],\r\n\r\n\r\n [[[0.00263699]]],\r\n\r\n\r\n [[[0.40009692]]],\r\n\r\n\r\n [[[0.6722459 ]]],\r\n\r\n\r\n [[[0.54544824]]],\r\n\r\n\r\n [[[0.69171315]]],\r\n\r\n\r\n [[[0.09180859]]],\r\n\r\n\r\n [[[0.634662 ]]],\r\n\r\n\r\n [[[0.59807974]]],\r\n\r\n\r\n [[[0.660335 ]]],\r\n\r\n\r\n [[[0.43653592]]],\r\n\r\n\r\n [[[0.20926596]]],\r\n\r\n\r\n [[[0.04501296]]],\r\n\r\n\r\n [[[0.4408583 ]]],\r\n\r\n\r\n [[[0.46784565]]],\r\n\r\n\r\n [[[0.2517042 ]]],\r\n\r\n\r\n [[[0.23913348]]],\r\n\r\n\r\n [[[0.12439712]]],\r\n\r\n\r\n [[[0.56089723]]],\r\n\r\n\r\n [[[0.66213465]]],\r\n\r\n\r\n [[[0.30742902]]],\r\n\r\n\r\n [[[0.33678833]]],\r\n\r\n\r\n [[[0.4094986 ]]],\r\n\r\n\r\n [[[0.08725767]]],\r\n\r\n\r\n [[[0.40686068]]],\r\n\r\n\r\n [[[0.38637325]]],\r\n\r\n\r\n [[[0.54127145]]],\r\n\r\n\r\n [[[0.37343186]]],\r\n\r\n\r\n [[[0.6557393 ]]],\r\n\r\n\r\n [[[0.20142 ]]],\r\n\r\n\r\n [[[0.55386937]]],\r\n\r\n\r\n [[[0.5326104 ]]],\r\n\r\n\r\n [[[0.22468646]]],\r\n\r\n\r\n [[[0.34933808]]],\r\n\r\n\r\n [[[0.45658964]]],\r\n\r\n\r\n [[[0.53491926]]],\r\n\r\n\r\n [[[0.42469105]]],\r\n\r\n\r\n [[[0.31144363]]],\r\n\r\n\r\n [[[0.685411 ]]],\r\n\r\n\r\n [[[0.4874776 ]]],\r\n\r\n\r\n [[[0.6796607 ]]],\r\n\r\n\r\n [[[0.15456446]]],\r\n\r\n\r\n [[[0.63261634]]],\r\n\r\n\r\n [[[0.17223041]]],\r\n\r\n\r\n [[[0.09819267]]],\r\n\r\n\r\n [[[0.09345566]]],\r\n\r\n\r\n [[[0.20636612]]],\r\n\r\n\r\n [[[0.56945485]]],\r\n\r\n\r\n [[[0.33953947]]],\r\n\r\n\r\n [[[0.1999843 ]]],\r\n\r\n\r\n [[[0.32493833]]],\r\n\r\n\r\n [[[0.6805619 ]]],\r\n\r\n\r\n [[[0.34506497]]],\r\n\r\n\r\n [[[0.68235236]]],\r\n\r\n\r\n [[[0.1110582 ]]],\r\n\r\n\r\n [[[0.5945545 ]]],\r\n\r\n\r\n [[[0.14432307]]],\r\n\r\n\r\n [[[0.60123146]]],\r\n\r\n\r\n [[[0.00452096]]],\r\n\r\n\r\n [[[0.47235835]]],\r\n\r\n\r\n [[[0.20066756]]],\r\n\r\n\r\n [[[0.642303 ]]],\r\n\r\n\r\n [[[0.2481875 ]]],\r\n\r\n\r\n [[[0.56008446]]],\r\n\r\n\r\n [[[0.34156114]]],\r\n\r\n\r\n [[[0.648798 ]]],\r\n\r\n\r\n [[[0.32805106]]],\r\n\r\n\r\n [[[0.37387052]]],\r\n\r\n\r\n [[[0.20767422]]],\r\n\r\n\r\n [[[0.07023842]]],\r\n\r\n\r\n [[[0.61710835]]],\r\n\r\n\r\n [[[0.52920985]]],\r\n\r\n\r\n [[[0.0995904 ]]],\r\n\r\n\r\n [[[0.09819601]]],\r\n\r\n\r\n [[[0.31743416]]],\r\n\r\n\r\n [[[0.2991825 ]]],\r\n\r\n\r\n [[[0.3547224 ]]],\r\n\r\n\r\n [[[0.14432673]]],\r\n\r\n\r\n [[[0.31014135]]],\r\n\r\n\r\n [[[0.08719108]]],\r\n\r\n\r\n [[[0.1480295 ]]],\r\n\r\n\r\n [[[0.4119163 ]]],\r\n\r\n\r\n [[[0.4997636 ]]],\r\n\r\n\r\n [[[0.20711906]]],\r\n\r\n\r\n [[[0.658858 ]]],\r\n\r\n\r\n [[[0.3466018 ]]],\r\n\r\n\r\n [[[0.15562956]]],\r\n\r\n\r\n [[[0.5944861 ]]],\r\n\r\n\r\n [[[0.4517534 ]]],\r\n\r\n\r\n [[[0.64904135]]],\r\n\r\n\r\n [[[0.60707915]]],\r\n\r\n\r\n [[[0.49129245]]],\r\n\r\n\r\n [[[0.23407486]]],\r\n\r\n\r\n [[[0.5858251 ]]],\r\n\r\n\r\n [[[0.24955001]]],\r\n\r\n\r\n [[[0.30913532]]],\r\n\r\n\r\n [[[0.10704726]]],\r\n\r\n\r\n [[[0.1676466 ]]],\r\n\r\n\r\n [[[0.36603883]]],\r\n\r\n\r\n [[[0.538971 ]]],\r\n\r\n\r\n [[[0.4430276 ]]],\r\n\r\n\r\n [[[0.1924243 ]]],\r\n\r\n\r\n [[[0.5491045 ]]],\r\n\r\n\r\n [[[0.25620812]]],\r\n\r\n\r\n [[[0.33868474]]],\r\n\r\n\r\n [[[0.20711271]]],\r\n\r\n\r\n [[[0.5150012 ]]],\r\n\r\n\r\n [[[0.0739684 ]]],\r\n\r\n\r\n [[[0.21971789]]],\r\n\r\n\r\n [[[0.3666449 ]]]], dtype=float32)\r\n",,terminal_output +4054,13257722,"jasmine/train_dynamics.py",0,0,"",python,tab +4055,13259153,"jasmine/genie.py",0,0,"",python,tab +4056,13259811,"jasmine/models/dynamics.py",0,0,"",python,tab +4057,13263153,"jasmine/models/dynamics.py",3835,0,"",python,selection_mouse +4058,13263168,"jasmine/models/dynamics.py",3834,0,"",python,selection_command +4059,13263289,"jasmine/models/dynamics.py",3834,2,"M\n",python,selection_mouse +4060,13263289,"jasmine/models/dynamics.py",3835,1,"\n",python,selection_command +4061,13263352,"jasmine/models/dynamics.py",3765,70,"\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4062,13263352,"jasmine/models/dynamics.py",3370,465,"\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4063,13263364,"jasmine/models/dynamics.py",3120,715," noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4064,13263438,"jasmine/models/dynamics.py",2994,841," ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4065,13263438,"jasmine/models/dynamics.py",2922,913," (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4066,13263439,"jasmine/models/dynamics.py",2882,953," noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4067,13263479,"jasmine/models/dynamics.py",2795,1040," # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4068,13263545,"jasmine/models/dynamics.py",2680,1155," jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4069,13263562,"jasmine/models/dynamics.py",2609,1226," noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4070,13263623,"jasmine/models/dynamics.py",2578,1257," jax.debug.breakpoint()\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4071,13263623,"jasmine/models/dynamics.py",2568,1267," )\n jax.debug.breakpoint()\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4072,13263646,"jasmine/models/dynamics.py",2488,1347," _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n jax.debug.breakpoint()\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4073,13263710,"jasmine/models/dynamics.py",2377,1458," rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n jax.debug.breakpoint()\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4074,13263775,"jasmine/models/dynamics.py",2444,1391," noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n jax.debug.breakpoint()\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4075,13263787,"jasmine/models/dynamics.py",2578,1257," jax.debug.breakpoint()\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4076,13263802,"jasmine/models/dynamics.py",3024,811," jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4077,13263822,"jasmine/models/dynamics.py",3282,553," noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4078,13263842,"jasmine/models/dynamics.py",3370,465,"\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4079,13263853,"jasmine/models/dynamics.py",3422,413,"one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4080,13263869,"jasmine/models/dynamics.py",3492,343,"sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4081,13263891,"jasmine/models/dynamics.py",3544,291,"qrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4082,13263902,"jasmine/models/dynamics.py",3614,221,"x.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4083,13263978,"jasmine/models/dynamics.py",3615,220,".debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4084,13264032,"jasmine/models/dynamics.py",3546,289,"t_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4085,13264033,"jasmine/models/dynamics.py",3495,340,"t_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4086,13264033,"jasmine/models/dynamics.py",3382,453,"afe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4087,13264048,"jasmine/models/dynamics.py",3370,465,"\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4088,13264069,"jasmine/models/dynamics.py",2885,950," noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4089,13264090,"jasmine/models/dynamics.py",2609,1226," noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4090,13264102,"jasmine/models/dynamics.py",2488,1347," _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n jax.debug.breakpoint()\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4091,13264121,"jasmine/models/dynamics.py",2377,1458," rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n jax.debug.breakpoint()\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4092,13264133,"jasmine/models/dynamics.py",2335,1500," B, T, N, M = vid_embed_BTNM.shape\n rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n jax.debug.breakpoint()\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4093,13264360,"jasmine/models/dynamics.py",2304,1531," jax.debug.breakpoint()\n B, T, N, M = vid_embed_BTNM.shape\n rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n jax.debug.breakpoint()\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4094,13264444,"jasmine/models/dynamics.py",2242,1593," def _apply_noise_augmentation(self, vid_embed_BTNM, rng):\n jax.debug.breakpoint()\n B, T, N, M = vid_embed_BTNM.shape\n rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\n noise_level_B = jax.random.uniform(\n _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\n )\n jax.debug.breakpoint()\n noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\n jax.debug.breakpoint()\n # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\n # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\n noise_bucket_idx_B = jnp.floor(\n (noise_level_B * self.noise_buckets) / self.max_noise_level\n ).astype(jnp.int32)\n jax.debug.breakpoint()\n noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\n noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\n noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\n noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\n jax.debug.breakpoint()\n\n # safe sqrt: clip argument to >= 0\n one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\n sqrt_one_minus = jnp.sqrt(one_minus_noise)\n sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\n jax.debug.breakpoint()\n\n noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\n jax.debug.breakpoint()\n return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M",python,selection_mouse +4095,13349333,"jasmine/models/dynamics.py",5215,0,"",python,selection_mouse +4096,13392609,"jasmine/models/dynamics.py",4138,0,"",python,selection_mouse +4097,13392612,"jasmine/models/dynamics.py",4137,0,"",python,selection_command +4098,13393348,"jasmine/models/dynamics.py",3973,0,"",python,selection_mouse +4099,13393349,"jasmine/models/dynamics.py",3972,0,"",python,selection_command +4100,13436426,"jasmine/models/dynamics.py",4119,0,"",python,selection_mouse +4101,13438830,"jasmine/models/dynamics.py",3765,0,"",python,selection_mouse +4102,13438841,"jasmine/models/dynamics.py",3764,0,"",python,selection_command +4103,13444107,"jasmine/models/dynamics.py",3668,0,"",python,selection_mouse +4104,13444461,"jasmine/models/dynamics.py",3765,0,"",python,selection_mouse +4105,13444467,"jasmine/models/dynamics.py",3764,0,"",python,selection_command +4106,13444958,"jasmine/models/dynamics.py",3765,0,"\n ",python,content +4107,13445445,"jasmine/models/dynamics.py",3774,0," assert not jnp.isnan(self.max_noise_level)\n assert self.max_noise_level > 0\n assert not jnp.isnan(noise_level_B).any()\n assert not jnp.isnan(vid_embed_BTNM).any()",python,content +4108,13446538,"jasmine/models/dynamics.py",3778,0,"",python,selection_mouse +4109,13446884,"jasmine/models/dynamics.py",3774,4,"",python,content +4110,13447297,"jasmine/models/dynamics.py",3774,0,"^",python,content +4111,13447298,"jasmine/models/dynamics.py",3775,0,"",python,selection_keyboard +4112,13447838,"jasmine/models/dynamics.py",3774,1,"",python,content +4113,13447839,"jasmine/models/dynamics.py",3774,0,"",python,selection_keyboard +4114,13448245,"jasmine/models/dynamics.py",3773,0,"",python,selection_command +4115,13448422,"jasmine/models/dynamics.py",3824,0,"",python,selection_command +4116,13448942,"jasmine/models/dynamics.py",3817,35," assert self.max_noise_level > 0",python,selection_command +4117,13449138,"jasmine/models/dynamics.py",3817,81," assert self.max_noise_level > 0\n assert not jnp.isnan(noise_level_B).any()",python,selection_command +4118,13449269,"jasmine/models/dynamics.py",3817,128," assert self.max_noise_level > 0\n assert not jnp.isnan(noise_level_B).any()\n assert not jnp.isnan(vid_embed_BTNM).any()",python,selection_command +4119,13449655,"jasmine/models/dynamics.py",3821,0,"",python,selection_command +4120,13449800,"jasmine/models/dynamics.py",3903,0," ",python,content +4121,13449801,"jasmine/models/dynamics.py",3857,0," ",python,content +4122,13449801,"jasmine/models/dynamics.py",3821,0," ",python,content +4123,13450280,"jasmine/models/dynamics.py",3824,0,"",python,selection_command +4124,13451054,"jasmine/models/dynamics.py",3855,0,"",python,selection_mouse +4125,13451692,"jasmine/models/dynamics.py",3815,0,"",python,selection_mouse +4126,13452205,"jasmine/models/dynamics.py",3816,0,"",python,selection_command +4127,13452431,"jasmine/models/dynamics.py",3816,0,",",python,content +4128,13452432,"jasmine/models/dynamics.py",3817,0,"",python,selection_keyboard +4129,13452526,"jasmine/models/dynamics.py",3817,0," ",python,content +4130,13452527,"jasmine/models/dynamics.py",3818,0,"",python,selection_keyboard +4131,13454601,"jasmine/models/dynamics.py",3818,0,"""""",python,content +4132,13454602,"jasmine/models/dynamics.py",3819,0,"",python,selection_keyboard +4133,13455408,"jasmine/models/dynamics.py",3819,0,"M",python,content +4134,13455409,"jasmine/models/dynamics.py",3820,0,"",python,selection_keyboard +4135,13455703,"jasmine/models/dynamics.py",3820,0,"a",python,content +4136,13455704,"jasmine/models/dynamics.py",3821,0,"",python,selection_keyboard +4137,13455913,"jasmine/models/dynamics.py",3821,0,"x",python,content +4138,13455914,"jasmine/models/dynamics.py",3822,0,"",python,selection_keyboard +4139,13456000,"jasmine/models/dynamics.py",3822,0," ",python,content +4140,13456001,"jasmine/models/dynamics.py",3823,0,"",python,selection_keyboard +4141,13456142,"jasmine/models/dynamics.py",3823,0,"n",python,content +4142,13456144,"jasmine/models/dynamics.py",3824,0,"",python,selection_keyboard +4143,13456295,"jasmine/models/dynamics.py",3824,0,"o",python,content +4144,13456297,"jasmine/models/dynamics.py",3825,0,"",python,selection_keyboard +4145,13456361,"jasmine/models/dynamics.py",3825,0,"i",python,content +4146,13456363,"jasmine/models/dynamics.py",3826,0,"",python,selection_keyboard +4147,13456392,"jasmine/models/dynamics.py",3826,0,"s",python,content +4148,13456393,"jasmine/models/dynamics.py",3827,0,"",python,selection_keyboard +4149,13456561,"jasmine/models/dynamics.py",3827,0,"e",python,content +4150,13456562,"jasmine/models/dynamics.py",3828,0,"",python,selection_keyboard +4151,13456622,"jasmine/models/dynamics.py",3828,0," ",python,content +4152,13456624,"jasmine/models/dynamics.py",3829,0,"",python,selection_keyboard +4153,13456833,"jasmine/models/dynamics.py",3829,0,"l",python,content +4154,13456835,"jasmine/models/dynamics.py",3830,0,"",python,selection_keyboard +4155,13456917,"jasmine/models/dynamics.py",3830,0,"e",python,content +4156,13456918,"jasmine/models/dynamics.py",3831,0,"",python,selection_keyboard +4157,13457466,"jasmine/models/dynamics.py",3831,0,"v",python,content +4158,13457467,"jasmine/models/dynamics.py",3832,0,"",python,selection_keyboard +4159,13457569,"jasmine/models/dynamics.py",3832,0,"e",python,content +4160,13457570,"jasmine/models/dynamics.py",3833,0,"",python,selection_keyboard +4161,13457719,"jasmine/models/dynamics.py",3833,0,"l",python,content +4162,13457720,"jasmine/models/dynamics.py",3834,0,"",python,selection_keyboard +4163,13457767,"jasmine/models/dynamics.py",3834,0," ",python,content +4164,13457768,"jasmine/models/dynamics.py",3835,0,"",python,selection_keyboard +4165,13457891,"jasmine/models/dynamics.py",3835,0,"i",python,content +4166,13457892,"jasmine/models/dynamics.py",3836,0,"",python,selection_keyboard +4167,13457994,"jasmine/models/dynamics.py",3836,0,"s",python,content +4168,13457996,"jasmine/models/dynamics.py",3837,0,"",python,selection_keyboard +4169,13458033,"jasmine/models/dynamics.py",3837,0," ",python,content +4170,13458034,"jasmine/models/dynamics.py",3838,0,"",python,selection_keyboard +4171,13458799,"jasmine/models/dynamics.py",3838,0,"N",python,content +4172,13458800,"jasmine/models/dynamics.py",3839,0,"",python,selection_keyboard +4173,13459137,"jasmine/models/dynamics.py",3839,0,"a",python,content +4174,13459138,"jasmine/models/dynamics.py",3840,0,"",python,selection_keyboard +4175,13459582,"jasmine/models/dynamics.py",3840,0,"N",python,content +4176,13459584,"jasmine/models/dynamics.py",3841,0,"",python,selection_keyboard +4177,13460461,"jasmine/models/dynamics.py",3930,0,"",python,selection_mouse +4178,13460858,"jasmine/models/dynamics.py",3882,0,"",python,selection_mouse +4179,13461528,"jasmine/models/dynamics.py",3882,0,",",python,content +4180,13461530,"jasmine/models/dynamics.py",3883,0,"",python,selection_keyboard +4181,13461595,"jasmine/models/dynamics.py",3883,0," ",python,content +4182,13461596,"jasmine/models/dynamics.py",3884,0,"",python,selection_keyboard +4183,13461868,"jasmine/models/dynamics.py",3884,0,"""""",python,content +4184,13461869,"jasmine/models/dynamics.py",3885,0,"",python,selection_keyboard +4185,13463666,"jasmine/models/dynamics.py",3885,0,"M",python,content +4186,13463668,"jasmine/models/dynamics.py",3886,0,"",python,selection_keyboard +4187,13463827,"jasmine/models/dynamics.py",3886,0,"a",python,content +4188,13463828,"jasmine/models/dynamics.py",3887,0,"",python,selection_keyboard +4189,13464178,"jasmine/models/dynamics.py",3887,0,"x",python,content +4190,13464179,"jasmine/models/dynamics.py",3888,0,"",python,selection_keyboard +4191,13464371,"jasmine/models/dynamics.py",3888,0," ",python,content +4192,13464372,"jasmine/models/dynamics.py",3889,0,"",python,selection_keyboard +4193,13464839,"jasmine/models/dynamics.py",3889,0,"n",python,content +4194,13464839,"jasmine/models/dynamics.py",3890,0,"",python,selection_keyboard +4195,13465027,"jasmine/models/dynamics.py",3890,0,"o",python,content +4196,13465028,"jasmine/models/dynamics.py",3891,0,"",python,selection_keyboard +4197,13465072,"jasmine/models/dynamics.py",3891,0,"i",python,content +4198,13465072,"jasmine/models/dynamics.py",3892,0,"",python,selection_keyboard +4199,13465189,"jasmine/models/dynamics.py",3892,0,"s",python,content +4200,13465190,"jasmine/models/dynamics.py",3893,0,"",python,selection_keyboard +4201,13465362,"jasmine/models/dynamics.py",3893,0,"e",python,content +4202,13465363,"jasmine/models/dynamics.py",3894,0,"",python,selection_keyboard +4203,13465422,"jasmine/models/dynamics.py",3894,0," ",python,content +4204,13465423,"jasmine/models/dynamics.py",3895,0,"",python,selection_keyboard +4205,13465532,"jasmine/models/dynamics.py",3895,0,"l",python,content +4206,13465532,"jasmine/models/dynamics.py",3896,0,"",python,selection_keyboard +4207,13465627,"jasmine/models/dynamics.py",3896,0,"e",python,content +4208,13465628,"jasmine/models/dynamics.py",3897,0,"",python,selection_keyboard +4209,13465787,"jasmine/models/dynamics.py",3897,0,"v",python,content +4210,13465788,"jasmine/models/dynamics.py",3898,0,"",python,selection_keyboard +4211,13465931,"jasmine/models/dynamics.py",3898,0,"e",python,content +4212,13465932,"jasmine/models/dynamics.py",3899,0,"",python,selection_keyboard +4213,13466008,"jasmine/models/dynamics.py",3899,0,"l",python,content +4214,13466008,"jasmine/models/dynamics.py",3900,0,"",python,selection_keyboard +4215,13466121,"jasmine/models/dynamics.py",3900,0," ",python,content +4216,13466122,"jasmine/models/dynamics.py",3901,0,"",python,selection_keyboard +4217,13466244,"jasmine/models/dynamics.py",3901,0,"i",python,content +4218,13466245,"jasmine/models/dynamics.py",3902,0,"",python,selection_keyboard +4219,13466347,"jasmine/models/dynamics.py",3902,0,"s",python,content +4220,13466348,"jasmine/models/dynamics.py",3903,0,"",python,selection_keyboard +4221,13466464,"jasmine/models/dynamics.py",3903,0," ",python,content +4222,13466465,"jasmine/models/dynamics.py",3904,0,"",python,selection_keyboard +4223,13467331,"jasmine/models/dynamics.py",3904,0,"s",python,content +4224,13467332,"jasmine/models/dynamics.py",3905,0,"",python,selection_keyboard +4225,13467402,"jasmine/models/dynamics.py",3905,0,"m",python,content +4226,13467403,"jasmine/models/dynamics.py",3906,0,"",python,selection_keyboard +4227,13467565,"jasmine/models/dynamics.py",3906,0,"a",python,content +4228,13467566,"jasmine/models/dynamics.py",3907,0,"",python,selection_keyboard +4229,13467646,"jasmine/models/dynamics.py",3907,0,"l",python,content +4230,13467647,"jasmine/models/dynamics.py",3908,0,"",python,selection_keyboard +4231,13467796,"jasmine/models/dynamics.py",3908,0,"l",python,content +4232,13467797,"jasmine/models/dynamics.py",3909,0,"",python,selection_keyboard +4233,13467894,"jasmine/models/dynamics.py",3909,0,"e",python,content +4234,13467895,"jasmine/models/dynamics.py",3910,0,"",python,selection_keyboard +4235,13467982,"jasmine/models/dynamics.py",3910,0,"r",python,content +4236,13467985,"jasmine/models/dynamics.py",3911,0,"",python,selection_keyboard +4237,13468057,"jasmine/models/dynamics.py",3911,0," ",python,content +4238,13468060,"jasmine/models/dynamics.py",3912,0,"",python,selection_keyboard +4239,13468429,"jasmine/models/dynamics.py",3912,0,"t",python,content +4240,13468430,"jasmine/models/dynamics.py",3913,0,"",python,selection_keyboard +4241,13468609,"jasmine/models/dynamics.py",3913,0,"h",python,content +4242,13468610,"jasmine/models/dynamics.py",3914,0,"",python,selection_keyboard +4243,13468697,"jasmine/models/dynamics.py",3914,0,"a",python,content +4244,13468699,"jasmine/models/dynamics.py",3915,0,"",python,selection_keyboard +4245,13468764,"jasmine/models/dynamics.py",3915,0,"n",python,content +4246,13468765,"jasmine/models/dynamics.py",3916,0,"",python,selection_keyboard +4247,13468912,"jasmine/models/dynamics.py",3916,0," ",python,content +4248,13468914,"jasmine/models/dynamics.py",3917,0,"",python,selection_keyboard +4249,13470025,"jasmine/models/dynamics.py",3917,0,"0",python,content +4250,13470026,"jasmine/models/dynamics.py",3918,0,"",python,selection_keyboard +4251,13471079,"jasmine/models/dynamics.py",3819,0,"",python,selection_mouse +4252,13471731,"jasmine/models/dynamics.py",3819,0,"[",python,content +4253,13471732,"jasmine/models/dynamics.py",3820,0,"",python,selection_keyboard +4254,13472014,"jasmine/models/dynamics.py",3820,0,"D",python,content +4255,13472015,"jasmine/models/dynamics.py",3821,0,"",python,selection_keyboard +4256,13472103,"jasmine/models/dynamics.py",3821,0,"E",python,content +4257,13472105,"jasmine/models/dynamics.py",3822,0,"",python,selection_keyboard +4258,13472284,"jasmine/models/dynamics.py",3822,0,"B",python,content +4259,13472285,"jasmine/models/dynamics.py",3823,0,"",python,selection_keyboard +4260,13472564,"jasmine/models/dynamics.py",3823,0,"G",python,content +4261,13472565,"jasmine/models/dynamics.py",3824,0,"",python,selection_keyboard +4262,13472943,"jasmine/models/dynamics.py",3823,1,"",python,content +4263,13473907,"jasmine/models/dynamics.py",3823,0,"U",python,content +4264,13473908,"jasmine/models/dynamics.py",3824,0,"",python,selection_keyboard +4265,13473988,"jasmine/models/dynamics.py",3824,0,"G",python,content +4266,13473989,"jasmine/models/dynamics.py",3825,0,"",python,selection_keyboard +4267,13474702,"jasmine/models/dynamics.py",3825,0,"]",python,content +4268,13474702,"jasmine/models/dynamics.py",3826,0,"",python,selection_keyboard +4269,13475141,"jasmine/models/dynamics.py",3826,0," ",python,content +4270,13475142,"jasmine/models/dynamics.py",3827,0,"",python,selection_keyboard +4271,13476475,"jasmine/models/dynamics.py",3826,0,"",python,selection_mouse +4272,13476602,"jasmine/models/dynamics.py",3825,1,"]",python,selection_mouse +4273,13476619,"jasmine/models/dynamics.py",3824,2,"G]",python,selection_mouse +4274,13476674,"jasmine/models/dynamics.py",3823,3,"UG]",python,selection_mouse +4275,13476683,"jasmine/models/dynamics.py",3822,4,"BUG]",python,selection_mouse +4276,13476737,"jasmine/models/dynamics.py",3821,5,"EBUG]",python,selection_mouse +4277,13476791,"jasmine/models/dynamics.py",3820,6,"DEBUG]",python,selection_mouse +4278,13477353,"jasmine/models/dynamics.py",3819,7,"[DEBUG]",python,selection_mouse +4279,13480181,"jasmine/models/dynamics.py",3893,0,"",python,selection_mouse +4280,13480452,"jasmine/models/dynamics.py",3893,0,"[DEBUG]",python,content +4281,13481064,"jasmine/models/dynamics.py",3900,0," ",python,content +4282,13481065,"jasmine/models/dynamics.py",3901,0,"",python,selection_keyboard +4283,13481634,"jasmine/models/dynamics.py",3985,0,"",python,selection_mouse +4284,13482523,"jasmine/models/dynamics.py",3985,0,",",python,content +4285,13482524,"jasmine/models/dynamics.py",3986,0,"",python,selection_keyboard +4286,13482581,"jasmine/models/dynamics.py",3986,0," ",python,content +4287,13482582,"jasmine/models/dynamics.py",3987,0,"",python,selection_keyboard +4288,13482969,"jasmine/models/dynamics.py",3987,0,"""""",python,content +4289,13482970,"jasmine/models/dynamics.py",3988,0,"",python,selection_keyboard +4290,13483635,"jasmine/models/dynamics.py",3988,0,"[DEBUG]",python,content +4291,13484746,"jasmine/models/dynamics.py",3995,0," ",python,content +4292,13484747,"jasmine/models/dynamics.py",3996,0,"",python,selection_keyboard +4293,13485631,"jasmine/models/dynamics.py",3996,0,"n",python,content +4294,13485632,"jasmine/models/dynamics.py",3997,0,"",python,selection_keyboard +4295,13485805,"jasmine/models/dynamics.py",3997,0,"o",python,content +4296,13485806,"jasmine/models/dynamics.py",3998,0,"",python,selection_keyboard +4297,13485886,"jasmine/models/dynamics.py",3998,0,"i",python,content +4298,13485888,"jasmine/models/dynamics.py",3999,0,"",python,selection_keyboard +4299,13485975,"jasmine/models/dynamics.py",3999,0,"s",python,content +4300,13485976,"jasmine/models/dynamics.py",4000,0,"",python,selection_keyboard +4301,13486137,"jasmine/models/dynamics.py",4000,0,"e",python,content +4302,13486137,"jasmine/models/dynamics.py",4001,0,"",python,selection_keyboard +4303,13486319,"jasmine/models/dynamics.py",4001,0,"_",python,content +4304,13486320,"jasmine/models/dynamics.py",4002,0,"",python,selection_keyboard +4305,13486574,"jasmine/models/dynamics.py",4002,0,"l",python,content +4306,13486575,"jasmine/models/dynamics.py",4003,0,"",python,selection_keyboard +4307,13486631,"jasmine/models/dynamics.py",4003,0,"e",python,content +4308,13486632,"jasmine/models/dynamics.py",4004,0,"",python,selection_keyboard +4309,13486782,"jasmine/models/dynamics.py",4004,0,"v",python,content +4310,13486783,"jasmine/models/dynamics.py",4005,0,"",python,selection_keyboard +4311,13486914,"jasmine/models/dynamics.py",4005,0,"e",python,content +4312,13486916,"jasmine/models/dynamics.py",4006,0,"",python,selection_keyboard +4313,13487001,"jasmine/models/dynamics.py",4006,0,"l",python,content +4314,13487001,"jasmine/models/dynamics.py",4007,0,"",python,selection_keyboard +4315,13487446,"jasmine/models/dynamics.py",4007,0,"_",python,content +4316,13487447,"jasmine/models/dynamics.py",4008,0,"",python,selection_keyboard +4317,13487836,"jasmine/models/dynamics.py",4008,0,"B",python,content +4318,13487837,"jasmine/models/dynamics.py",4009,0,"",python,selection_keyboard +4319,13488039,"jasmine/models/dynamics.py",4009,0," ",python,content +4320,13488040,"jasmine/models/dynamics.py",4010,0,"",python,selection_keyboard +4321,13488150,"jasmine/models/dynamics.py",4010,0,"i",python,content +4322,13488151,"jasmine/models/dynamics.py",4011,0,"",python,selection_keyboard +4323,13488250,"jasmine/models/dynamics.py",4011,0,"s",python,content +4324,13488251,"jasmine/models/dynamics.py",4012,0,"",python,selection_keyboard +4325,13488354,"jasmine/models/dynamics.py",4012,0," ",python,content +4326,13488355,"jasmine/models/dynamics.py",4013,0,"",python,selection_keyboard +4327,13488970,"jasmine/models/dynamics.py",4013,0,"N",python,content +4328,13488971,"jasmine/models/dynamics.py",4014,0,"",python,selection_keyboard +4329,13489154,"jasmine/models/dynamics.py",4014,0,"a",python,content +4330,13489155,"jasmine/models/dynamics.py",4015,0,"",python,selection_keyboard +4331,13489490,"jasmine/models/dynamics.py",4015,0,"N",python,content +4332,13489491,"jasmine/models/dynamics.py",4016,0,"",python,selection_keyboard +4333,13490292,"jasmine/models/dynamics.py",4124,0,"",python,selection_mouse +4334,13491214,"jasmine/models/dynamics.py",4068,0,"",python,selection_mouse +4335,13492661,"jasmine/models/dynamics.py",4068,0,",",python,content +4336,13492662,"jasmine/models/dynamics.py",4069,0,"",python,selection_keyboard +4337,13493058,"jasmine/models/dynamics.py",4069,0," ",python,content +4338,13493059,"jasmine/models/dynamics.py",4070,0,"",python,selection_keyboard +4339,13493890,"jasmine/models/dynamics.py",4070,0,"""""",python,content +4340,13493891,"jasmine/models/dynamics.py",4071,0,"",python,selection_keyboard +4341,13494807,"jasmine/models/dynamics.py",4071,0,"D",python,content +4342,13494808,"jasmine/models/dynamics.py",4072,0,"",python,selection_keyboard +4343,13494952,"jasmine/models/dynamics.py",4072,0,"E",python,content +4344,13494953,"jasmine/models/dynamics.py",4073,0,"",python,selection_keyboard +4345,13495023,"jasmine/models/dynamics.py",4073,0,"B",python,content +4346,13495024,"jasmine/models/dynamics.py",4074,0,"",python,selection_keyboard +4347,13495154,"jasmine/models/dynamics.py",4074,0,"U",python,content +4348,13495155,"jasmine/models/dynamics.py",4075,0,"",python,selection_keyboard +4349,13495327,"jasmine/models/dynamics.py",4075,0,"G",python,content +4350,13495327,"jasmine/models/dynamics.py",4076,0,"",python,selection_keyboard +4351,13496708,"jasmine/models/dynamics.py",4076,0,"]",python,content +4352,13496709,"jasmine/models/dynamics.py",4077,0,"",python,selection_keyboard +4353,13498127,"jasmine/models/dynamics.py",4071,0,"[",python,content +4354,13498130,"jasmine/models/dynamics.py",4072,0,"",python,selection_keyboard +4355,13499276,"jasmine/models/dynamics.py",4079,0,"",python,selection_mouse +4356,13500093,"jasmine/models/dynamics.py",4141,0,"",python,selection_command +4357,13500805,"jasmine/models/dynamics.py",4079,0,"",python,selection_command +4358,13501175,"jasmine/models/dynamics.py",4078,0,"",python,selection_command +4359,13501876,"jasmine/models/dynamics.py",4078,0,",",python,content +4360,13501877,"jasmine/models/dynamics.py",4079,0,"",python,selection_keyboard +4361,13502274,"jasmine/models/dynamics.py",4078,1,"",python,content +4362,13502423,"jasmine/models/dynamics.py",4078,0," ",python,content +4363,13502424,"jasmine/models/dynamics.py",4079,0,"",python,selection_keyboard +4364,13502738,"jasmine/models/dynamics.py",4079,0,"v",python,content +4365,13502739,"jasmine/models/dynamics.py",4080,0,"",python,selection_keyboard +4366,13503105,"jasmine/models/dynamics.py",4080,0,"i",python,content +4367,13503106,"jasmine/models/dynamics.py",4081,0,"",python,selection_keyboard +4368,13503240,"jasmine/models/dynamics.py",4081,0,"d",python,content +4369,13503241,"jasmine/models/dynamics.py",4082,0,"",python,selection_keyboard +4370,13503468,"jasmine/models/dynamics.py",4082,0,"_",python,content +4371,13503469,"jasmine/models/dynamics.py",4083,0,"",python,selection_keyboard +4372,13503673,"jasmine/models/dynamics.py",4083,0,"e",python,content +4373,13503674,"jasmine/models/dynamics.py",4084,0,"",python,selection_keyboard +4374,13503784,"jasmine/models/dynamics.py",4084,0,"m",python,content +4375,13503785,"jasmine/models/dynamics.py",4085,0,"",python,selection_keyboard +4376,13503996,"jasmine/models/dynamics.py",4085,0,"b",python,content +4377,13503997,"jasmine/models/dynamics.py",4086,0,"",python,selection_keyboard +4378,13504053,"jasmine/models/dynamics.py",4086,0,"e",python,content +4379,13504054,"jasmine/models/dynamics.py",4087,0,"",python,selection_keyboard +4380,13504168,"jasmine/models/dynamics.py",4087,0,"d",python,content +4381,13504168,"jasmine/models/dynamics.py",4088,0,"",python,selection_keyboard +4382,13504389,"jasmine/models/dynamics.py",4088,0,"_",python,content +4383,13504390,"jasmine/models/dynamics.py",4089,0,"",python,selection_keyboard +4384,13504860,"jasmine/models/dynamics.py",4089,0,"B",python,content +4385,13504861,"jasmine/models/dynamics.py",4090,0,"",python,selection_keyboard +4386,13505132,"jasmine/models/dynamics.py",4090,0,"T",python,content +4387,13505133,"jasmine/models/dynamics.py",4091,0,"",python,selection_keyboard +4388,13505357,"jasmine/models/dynamics.py",4091,0,"N",python,content +4389,13505358,"jasmine/models/dynamics.py",4092,0,"",python,selection_keyboard +4390,13505869,"jasmine/models/dynamics.py",4092,0,"M",python,content +4391,13505870,"jasmine/models/dynamics.py",4093,0,"",python,selection_keyboard +4392,13506087,"jasmine/models/dynamics.py",4093,0," ",python,content +4393,13506088,"jasmine/models/dynamics.py",4094,0,"",python,selection_keyboard +4394,13506363,"jasmine/models/dynamics.py",4094,0,"a",python,content +4395,13506365,"jasmine/models/dynamics.py",4095,0,"",python,selection_keyboard +4396,13506639,"jasmine/models/dynamics.py",4094,1,"",python,content +4397,13506838,"jasmine/models/dynamics.py",4094,0,"h",python,content +4398,13506839,"jasmine/models/dynamics.py",4095,0,"",python,selection_keyboard +4399,13506889,"jasmine/models/dynamics.py",4095,0,"a",python,content +4400,13506890,"jasmine/models/dynamics.py",4096,0,"",python,selection_keyboard +4401,13506951,"jasmine/models/dynamics.py",4096,0,"s",python,content +4402,13506952,"jasmine/models/dynamics.py",4097,0,"",python,selection_keyboard +4403,13507062,"jasmine/models/dynamics.py",4097,0," ",python,content +4404,13507064,"jasmine/models/dynamics.py",4098,0,"",python,selection_keyboard +4405,13507522,"jasmine/models/dynamics.py",4098,0,"N",python,content +4406,13507524,"jasmine/models/dynamics.py",4099,0,"",python,selection_keyboard +4407,13507996,"jasmine/models/dynamics.py",4099,0,"a",python,content +4408,13507997,"jasmine/models/dynamics.py",4100,0,"",python,selection_keyboard +4409,13508325,"jasmine/models/dynamics.py",4100,0,"N",python,content +4410,13508326,"jasmine/models/dynamics.py",4101,0,"",python,selection_keyboard +4411,13509270,"jasmine/models/dynamics.py",4011,0,"",python,selection_mouse +4412,13509431,"jasmine/models/dynamics.py",4010,2,"is",python,selection_mouse +4413,13509856,"jasmine/models/dynamics.py",4010,2,"h",python,content +4414,13509857,"jasmine/models/dynamics.py",4011,0,"",python,selection_keyboard +4415,13509905,"jasmine/models/dynamics.py",4011,0,"a",python,content +4416,13509906,"jasmine/models/dynamics.py",4012,0,"",python,selection_keyboard +4417,13509986,"jasmine/models/dynamics.py",4012,0,"s",python,content +4418,13509987,"jasmine/models/dynamics.py",4013,0,"",python,selection_keyboard +4419,13514879,"jasmine/models/dynamics.py",3462,0,"",python,selection_mouse +4420,13515015,"jasmine/models/dynamics.py",3455,16,"noise_level_B111",python,selection_mouse +4421,13530906,"jasmine/models/dynamics.py",3804,0,"",python,selection_mouse +4422,13531599,"jasmine/models/dynamics.py",3765,0,"",python,selection_mouse +4423,13532872,"jasmine/models/dynamics.py",3764,0,"",python,selection_command +4424,13533630,"jasmine/models/dynamics.py",3735,31,"",python,content +4425,13533662,"jasmine/models/dynamics.py",3743,0,"",python,selection_command +4426,13533764,"jasmine/models/dynamics.py",3644,0,"",python,selection_command +4427,13534113,"jasmine/models/dynamics.py",3743,0,"",python,selection_command +4428,13534376,"jasmine/models/dynamics.py",3819,0,"\n ",python,content +4429,13535034,"jasmine/models/dynamics.py",3824,4,"",python,content +4430,13535381,"jasmine/models/dynamics.py",3820,4,"",python,content +4431,13535548,"jasmine/models/dynamics.py",3819,1,"",python,content +4432,13535818,"jasmine/models/dynamics.py",3818,0,"",python,selection_command +4433,13536050,"jasmine/models/dynamics.py",3719,0,"",python,selection_command +4434,13536726,"jasmine/models/dynamics.py",3734,0,"\n ",python,content +4435,13536992,"jasmine/models/dynamics.py",3735,8,"",python,content +4436,13537262,"jasmine/models/dynamics.py",3636,0,"",python,selection_command +4437,13537486,"jasmine/models/dynamics.py",3735,0,"",python,selection_command +4438,13537991,"jasmine/models/dynamics.py",3736,0,"",python,selection_command +4439,13538014,"jasmine/models/dynamics.py",3821,0,"",python,selection_command +4440,13538042,"jasmine/models/dynamics.py",3906,0,"",python,selection_command +4441,13538147,"jasmine/models/dynamics.py",3989,0,"",python,selection_command +4442,13538303,"jasmine/models/dynamics.py",4074,0,"",python,selection_command +4443,13538557,"jasmine/models/dynamics.py",3989,0,"",python,selection_command +4444,13538740,"jasmine/models/dynamics.py",4073,0,"\n ",python,content +4445,13538946,"jasmine/models/dynamics.py",4074,8,"",python,content +4446,13539075,"jasmine/models/dynamics.py",3989,0,"",python,selection_command +4447,13539580,"jasmine/models/dynamics.py",3906,0,"",python,selection_command +4448,13539594,"jasmine/models/dynamics.py",3821,0,"",python,selection_command +4449,13539663,"jasmine/models/dynamics.py",3736,0,"",python,selection_command +4450,13539663,"jasmine/models/dynamics.py",3735,0,"",python,selection_command +4451,13539684,"jasmine/models/dynamics.py",3636,0,"",python,selection_command +4452,13539750,"jasmine/models/dynamics.py",3635,0,"",python,selection_command +4453,13539872,"jasmine/models/dynamics.py",3604,0,"",python,selection_command +4454,13540344,"jasmine/models/dynamics.py",3604,31,"",python,content +4455,13540434,"jasmine/models/dynamics.py",3535,0,"",python,selection_command +4456,13540609,"jasmine/models/dynamics.py",3484,0,"",python,selection_command +4457,13540742,"jasmine/models/dynamics.py",3414,0,"",python,selection_command +4458,13540930,"jasmine/models/dynamics.py",3371,0,"",python,selection_command +4459,13541023,"jasmine/models/dynamics.py",3370,0,"",python,selection_command +4460,13541190,"jasmine/models/dynamics.py",3339,0,"",python,selection_command +4461,13541501,"jasmine/models/dynamics.py",3339,31,"",python,content +4462,13541612,"jasmine/models/dynamics.py",3278,0,"",python,selection_command +4463,13541810,"jasmine/models/dynamics.py",3198,0,"",python,selection_command +4464,13541928,"jasmine/models/dynamics.py",3120,0,"",python,selection_command +4465,13542078,"jasmine/models/dynamics.py",3053,0,"",python,selection_command +4466,13542224,"jasmine/models/dynamics.py",3022,0,"",python,selection_command +4467,13542593,"jasmine/models/dynamics.py",3022,31,"",python,content +4468,13542606,"jasmine/models/dynamics.py",3030,0,"",python,selection_command +4469,13542693,"jasmine/models/dynamics.py",3002,0,"",python,selection_command +4470,13542861,"jasmine/models/dynamics.py",2930,0,"",python,selection_command +4471,13543006,"jasmine/models/dynamics.py",2890,0,"",python,selection_command +4472,13543163,"jasmine/models/dynamics.py",2803,0,"",python,selection_command +4473,13543302,"jasmine/models/dynamics.py",2719,0,"",python,selection_command +4474,13543451,"jasmine/models/dynamics.py",2688,0,"",python,selection_command +4475,13543866,"jasmine/models/dynamics.py",2680,31,"",python,content +4476,13543884,"jasmine/models/dynamics.py",2688,0,"",python,selection_command +4477,13543957,"jasmine/models/dynamics.py",2617,0,"",python,selection_command +4478,13544133,"jasmine/models/dynamics.py",2586,0,"",python,selection_command +4479,13544589,"jasmine/models/dynamics.py",2578,31,"",python,content +4480,13544602,"jasmine/models/dynamics.py",2586,0,"",python,selection_command +4481,13544660,"jasmine/models/dynamics.py",2576,0,"",python,selection_command +4482,13544809,"jasmine/models/dynamics.py",2496,0,"",python,selection_command +4483,13544972,"jasmine/models/dynamics.py",2452,0,"",python,selection_command +4484,13545129,"jasmine/models/dynamics.py",2385,0,"",python,selection_command +4485,13545290,"jasmine/models/dynamics.py",2343,0,"",python,selection_command +4486,13545607,"jasmine/models/dynamics.py",2312,0,"",python,selection_command +4487,13545907,"jasmine/models/dynamics.py",2304,31,"",python,content +4488,13545931,"jasmine/models/dynamics.py",2312,0,"",python,selection_command +4489,13554206,"jasmine/models/dynamics.py",3847,0,"",python,selection_mouse +4490,13555042,"jasmine/models/dynamics.py",3991,0,"",python,selection_mouse +4491,13555044,"jasmine/models/dynamics.py",3990,0,"",python,selection_command +4492,13557795,"TERMINAL",0,0,"q",,terminal_output +4493,13558453,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3527866.0 task 0: running\r\n",,terminal_output +4494,13560796,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3527866.0 task 0: running\r\n",,terminal_output +4495,13561167,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3527866.0\r\n",,terminal_output +4496,13561229,"TERMINAL",0,0,"srun: forcing job termination\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-4:\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\nTraceback (most recent call last):\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\nslurmstepd: error: *** STEP 3527866.0 ON hkn0402 CANCELLED AT 2025-09-28T17:13:25 ***\r\n",,terminal_output +4497,13561751,"TERMINAL",0,0,"(jdb) ",,terminal_output +4498,13561862,"TERMINAL",0,0,"]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +4499,13561926,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +4500,13564298,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",,terminal_output +4501,13567342,"jasmine/train_dynamics.py",0,0,"",python,tab +4502,13567814,"jasmine/train_dynamics.py",16147,0,"",python,selection_mouse +4503,13568968,"jasmine/train_dynamics.py",15917,0,"",python,selection_mouse +4504,13569454,"jasmine/train_dynamics.py",15888,0,"",python,selection_mouse +4505,13570061,"jasmine/train_dynamics.py",15873,0,"",python,selection_command +4506,13570450,"jasmine/train_dynamics.py",15873,0,"#",python,content +4507,13570452,"jasmine/train_dynamics.py",15874,0,"",python,selection_keyboard +4508,13570557,"jasmine/train_dynamics.py",15874,0," ",python,content +4509,13570558,"jasmine/train_dynamics.py",15875,0,"",python,selection_keyboard +4510,13571091,"jasmine/train_dynamics.py",15874,0,"",python,selection_command +4511,13572838,"jasmine/train_dynamics.py",16537,0,"",python,selection_mouse +4512,13573386,"jasmine/train_dynamics.py",16531,0,"\n",python,content +4513,13573400,"jasmine/train_dynamics.py",16532,0," ",python,content +4514,13574594,"jasmine/train_dynamics.py",16532,4,"",python,content +4515,13574713,"jasmine/train_dynamics.py",16531,1,"",python,content +4516,13575302,"jasmine/train_dynamics.py",16532,0,"",python,selection_command +4517,13577196,"jasmine/train_dynamics.py",16536,0,"",python,selection_command +4518,13577753,"jasmine/train_dynamics.py",16536,0," ",python,content +4519,13577754,"jasmine/train_dynamics.py",16537,0,"",python,selection_keyboard +4520,13577996,"jasmine/train_dynamics.py",16537,0,"#",python,content +4521,13577997,"jasmine/train_dynamics.py",16538,0,"",python,selection_keyboard +4522,13578299,"jasmine/train_dynamics.py",16537,1,"",python,content +4523,13578425,"jasmine/train_dynamics.py",16536,1,"",python,content +4524,13578643,"jasmine/train_dynamics.py",16536,0,"#",python,content +4525,13578643,"jasmine/train_dynamics.py",16537,0,"",python,selection_keyboard +4526,13578752,"jasmine/train_dynamics.py",16537,0," ",python,content +4527,13578753,"jasmine/train_dynamics.py",16538,0,"",python,selection_keyboard +4528,13579488,"jasmine/train_dynamics.py",16537,0,"",python,selection_command +4529,13582148,"TERMINAL",0,0,"ource .venv/bin/activate\r\n\r",,terminal_output +4530,13582544,"TERMINAL",0,0,"h slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",,terminal_output +4531,13582708,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=06:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\n# source .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\n# lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\r\n# tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\r\n\r\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3518963\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=250 \\r\n --log_checkpoint_interval=250 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-noise-lvl-default-$slurm_job_id \\r\n --tags dyn breakout noise-lvl default \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 100 \\r\n --wsd_decay_steps 1000 \\r\n --num_steps 5000 \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --lam_checkpoint $lam_checkpoint \\r\n --val_interval 1000 \\r\n --eval_full_frame \\r\n",,terminal_output +4532,13582861,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=217802\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0402\r\nSLURM_JOB_START_TIME=1759070937\r\nSLURM_STEP_NODELIST=hkn0402\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759074537\r\nSLURM_PMI2_SRUN_PORT=36803\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3527866\r\nSLURM_PTY_PORT=42519\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=36\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0402\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=139\r\nSLURM_NODELIST=hkn0402\r\nSLURM_SRUN_COMM_PORT=45011\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3527866\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0402\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=45011\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0402\r\n",,terminal_output +4533,13583039,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +4534,13585745,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +4535,13591857,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +4536,13592174,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +4537,13592980,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20250928_171356-6r8e2qaj\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run breakout-dyn-noise-lvl-default-3527866\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/6r8e2qaj\r\n",,terminal_output +4538,13593117,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26561024, 'lam': 16900976, 'tokenizer': 33750256, 'total': 77212256}\r\n",,terminal_output +4539,13595965,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +4540,13607932,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 790, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 620, in main\r\n compiled = train_step.lower(optimizer, first_batch).compile()\r\n ^^^^^^^^^^^^^^^^\r\nAttributeError: 'function' object has no attribute 'lower'\r\n",,terminal_output +4541,13609266,"TERMINAL",0,0,"wandb: \r\nwandb: 🚀 View run breakout-dyn-noise-lvl-default-3527866 at: https://wandb.ai/instant-uv/jafar/runs/6r8e2qaj\r\nwandb: Find logs at: ../../../../../hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20250928_171356-6r8e2qaj/logs\r\n",,terminal_output +4542,13609370,"TERMINAL",0,0,"W0928 17:14:14.034988 226316 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugonly job_name: ""jax_worker"": CANCELLED: CANCELLED\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_message:""CANCELLED"", grpc_status:1} [type.googleapis.com/tensorflow.DerivedStatus='']\r\n",,terminal_output +4543,13609942,"TERMINAL",0,0,"/usr/lib64/python3.12/multiprocessing/resource_tracker.py:254: UserWarning: resource_tracker: There appear to be 4 leaked shared_memory objects to clean up at shutdown\r\n warnings.warn('resource_tracker: There appear to be %d '\r\n",,terminal_output +4544,13610292,"TERMINAL",0,0,"srun: error: hkn0402: task 0: Exited with exit code 1\r\n]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +4545,13619659,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"import os\n\n\nos.environ.setdefault(""XLA_PYTHON_CLIENT_MEM_FRACTION"", ""0.98"")\n\nfrom dataclasses import dataclass, field\nimport itertools\nfrom typing import cast, Optional\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\nfrom utils.train_utils import (\n get_lr_schedule,\n count_parameters_by_component,\n print_mem_stats,\n print_compiled_memory_stats,\n print_compiled_cost_analysis,\n)\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n # Tokenizer\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_type: str = ""maskgit"" # supported options: maskgit, causal\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n max_noise_level: float = 0.7\n noise_buckets: int = 10\n dropout: float = 0.0\n mask_limit: float = 0.5\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n use_gt_actions: bool = False\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n eval_full_frame: bool = False\n val_maskgit_steps: int = 25\n val_temperature: float = 1\n val_sample_argmax: bool = False\n wandb_id: str = """"\n\n\ndef build_model(args: Args, rng: jax.Array) -> tuple[Genie, jax.Array]:\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_actions=args.num_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n use_gt_actions=args.use_gt_actions,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n max_noise_level=args.max_noise_level,\n noise_buckets=args.noise_buckets,\n mask_limit=args.mask_limit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n decode=False,\n rngs=rngs,\n )\n if args.use_gt_actions:\n assert (\n not args.lam_checkpoint\n ), ""Cannot use LAM when using ground-truth actions.""\n else:\n assert genie.lam is not None\n del genie.lam.decoder\n return genie, rng\n\n\ndef build_optimizer(genie: Genie, args: Args) -> nnx.ModelAndOptimizer:\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.param_dtype, # moments in full precision\n )\n optimizer = nnx.ModelAndOptimizer(genie, tx)\n return optimizer\n\n\ndef build_mesh_and_sharding(\n num_devices: int,\n) -> tuple[Mesh, NamedSharding, NamedSharding, NamedSharding]:\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n actions_sharding = NamedSharding(mesh, PartitionSpec(""data"", None))\n return mesh, replicated_sharding, videos_sharding, actions_sharding\n\n\ndef shard_optimizer_states(\n optimizer: nnx.ModelAndOptimizer, replicated_sharding: NamedSharding\n) -> None:\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n\ndef build_dataloader(args: Args, data_dir: str) -> grain.DataLoaderIterator:\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(data_dir, x)\n for x in os.listdir(data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n return grain_iterator\n\n\ndef build_checkpoint_manager(args: Args) -> Optional[ocp.CheckpointManager]:\n if args.restore_ckpt or args.save_ckpt:\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(\n ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler\n ),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(\n ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler\n ),\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n return checkpoint_manager\n else:\n return None\n\n\ndef restore_or_initialize_components(\n args: Args,\n checkpoint_manager: Optional[ocp.CheckpointManager],\n optimizer: nnx.ModelAndOptimizer,\n train_iterator: grain.DataLoaderIterator,\n rng: jax.Array,\n replicated_sharding: NamedSharding,\n val_iterator: Optional[grain.DataLoaderIterator],\n restore_step: Optional[int] = None,\n) -> tuple[\n int,\n nnx.ModelAndOptimizer,\n grain.DataLoaderIterator,\n grain.DataLoaderIterator,\n jax.Array,\n]:\n step = 0\n if checkpoint_manager and restore_step is None:\n restore_step = checkpoint_manager.latest_step()\n if args.restore_ckpt:\n assert checkpoint_manager is not None\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n if val_iterator:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n val_dataloader_state=grain.checkpoint.CheckpointRestore(val_iterator), # type: ignore\n )\n else:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n )\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(), args=restore_args\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n train_iterator = restored[""train_dataloader_state""]\n if val_iterator:\n val_iterator = restored[""val_dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n rng, _rng = jax.random.split(rng)\n optimizer = restore_genie_components(optimizer, replicated_sharding, _rng, args)\n return step, optimizer, train_iterator, val_iterator, rng\n\n\ndef _calculate_step_metrics(\n outputs: dict[str, jax.Array],\n gt: jax.Array,\n num_actions: int,\n num_patch_latents: int,\n) -> tuple[jax.Array, dict]:\n mask = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n gt_val = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt_val, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt_val, recon)).mean()\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]),\n size=num_patch_latents,\n fill_value=0,\n )\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n psnr=psnr,\n ssim=ssim,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n if ""lam_indices"" in outputs.keys():\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]),\n size=num_actions,\n fill_value=0,\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n metrics[""codebook_usage_lam""] = codebook_usage_lam\n return ce_loss, metrics\n\n\ndef main(args: Args) -> None:\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n genie, rng = build_model(args, rng)\n _, params, _ = nnx.split(genie, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n optimizer = build_optimizer(genie, args)\n del genie\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n _, replicated_sharding, videos_sharding, actions_sharding = build_mesh_and_sharding(\n num_devices\n )\n\n shard_optimizer_states(optimizer, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n checkpoint_manager = build_checkpoint_manager(args)\n\n # --- Create DataLoaderIterator from dataloader ---\n train_iterator = build_dataloader(args, args.data_dir)\n val_iterator = None\n if args.val_data_dir:\n val_iterator = build_dataloader(args, args.val_data_dir)\n\n # --- Restore checkpoint ---\n step, optimizer, train_iterator, val_iterator, rng = (\n restore_or_initialize_components(\n args,\n checkpoint_manager,\n optimizer,\n train_iterator,\n rng,\n replicated_sharding,\n val_iterator,\n )\n )\n\n # --- Define loss and train step (close over args) ---\n def dynamics_loss_fn(\n model: Genie,\n inputs: dict,\n ) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs)\n ce_loss, metrics = _calculate_step_metrics(\n outputs, gt, args.num_actions, args.num_patch_latents\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n # @nnx.jit(donate_argnums=0)\n def train_step(\n optimizer: nnx.ModelAndOptimizer, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(model: Genie) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n model.train()\n return dynamics_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n optimizer.model\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return loss, recon, metrics\n\n # @nnx.jit\n def val_step(genie: Genie, inputs: dict) -> dict:\n """"""Evaluate model and compute metrics""""""\n genie.eval()\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n (loss, (recon, metrics)) = dynamics_loss_fn(genie, inputs)\n val_output = {""loss"": loss, ""recon"": recon, ""metrics"": metrics}\n\n # --- Evaluate full frame prediction (sampling) ---\n if args.eval_full_frame:\n inputs[""videos""] = gt.astype(args.dtype)\n tokenizer_outputs = genie.tokenizer.vq_encode(\n inputs[""videos""], training=False\n )\n tokens_full_frame = tokenizer_outputs[""indices""]\n lam_indices_E = None\n if not args.use_gt_actions:\n lam_indices_E = genie.vq_encode(inputs, training=False)\n inputs[""latent_actions""] = lam_indices_E\n inputs[""videos""] = inputs[""videos""][\n :, :-1\n ] # remove last frame for generation\n recon_full_frame, logits_full_frame = genie.sample(\n batch=inputs,\n seq_len=args.seq_len,\n noise_level=0.,\n temperature=args.val_temperature,\n sample_argmax=args.val_sample_argmax,\n maskgit_steps=args.val_maskgit_steps,\n )\n # Calculate metrics for the last frame only\n step_outputs = {\n ""recon"": recon_full_frame[:, -1],\n ""token_logits"": logits_full_frame[:, -1],\n ""video_tokens"": tokens_full_frame[:, -1],\n ""mask"": jnp.ones_like(tokens_full_frame[:, -1]),\n }\n if lam_indices_E is not None:\n lam_indices_B = lam_indices_E.reshape((-1, args.seq_len - 1))[:, -1]\n step_outputs[""lam_indices""] = lam_indices_B\n\n loss_full_frame, metrics_full_frame = _calculate_step_metrics(\n step_outputs, gt[:, -1], args.num_actions, args.num_patch_latents\n )\n val_output.update(\n {\n ""loss_full_frame"": loss_full_frame,\n ""recon_full_frame"": recon_full_frame,\n ""metrics_full_frame"": metrics_full_frame,\n }\n )\n return val_output\n\n def calculate_validation_metrics(val_dataloader, genie, rng):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n loss_full_frame_per_step = []\n metrics_full_frame_per_step = []\n batch = None\n recon = None\n recon_full_frame = None\n for batch in val_dataloader:\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n val_outputs = val_step(genie, batch)\n loss_per_step.append(val_outputs[""loss""])\n metrics_per_step.append(val_outputs[""metrics""])\n recon = val_outputs[""recon""]\n if args.eval_full_frame:\n loss_full_frame_per_step.append(val_outputs[""loss_full_frame""])\n metrics_full_frame_per_step.append(val_outputs[""metrics_full_frame""])\n recon_full_frame = val_outputs[""recon_full_frame""]\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(\n f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}""\n )\n\n val_metrics = {\n f""val_{key}"": np.mean([float(m[key]) for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n val_metrics[""val_loss""] = np.mean(loss_per_step)\n if args.eval_full_frame:\n val_metrics_full_frame = {\n f""val_full_frame_{key}"": np.mean(\n [float(m[key]) for m in metrics_full_frame_per_step]\n )\n for key in metrics_full_frame_per_step[0].keys()\n }\n val_metrics.update(val_metrics_full_frame)\n val_metrics[""val_full_frame_loss""] = np.mean(loss_full_frame_per_step)\n return val_metrics, batch, recon, recon_full_frame\n\n # --- TRAIN LOOP ---\n dataloader_train = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in train_iterator\n )\n dataloader_val = None\n if val_iterator:\n dataloader_val = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in val_iterator\n )\n if jax.process_index() == 0:\n first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader_train = itertools.chain([first_batch], dataloader_train)\n print(f""Starting training from step {step}..."")\n first_step = step\n while step < args.num_steps:\n for batch in dataloader_train:\n # --- Train step ---\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n loss, recon, metrics = train_step(optimizer, batch)\n if step == first_step:\n print_mem_stats(""After params initialized"")\n step += 1\n\n # --- Validation loss ---\n val_results = {}\n if dataloader_val and step % args.val_interval == 0:\n rng, _rng_mask_val = jax.random.split(rng, 2)\n print(""Calculating validation metrics..."")\n val_metrics, val_gt_batch, val_recon, val_recon_full_frame = (\n calculate_validation_metrics(\n dataloader_val, optimizer.model, _rng_mask_val\n )\n )\n print(f""Step {step}, validation loss: {val_metrics['val_loss']}"")\n val_results = {\n ""metrics"": val_metrics,\n ""gt_batch"": val_gt_batch,\n ""recon"": val_recon,\n ""full_frame"": val_recon_full_frame,\n }\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n log_dict = {""loss"": loss, ""step"": step, **metrics}\n if val_results:\n log_dict.update(val_results[""metrics""])\n wandb.log(log_dict)\n if step % args.log_image_interval == 0:\n gt_seq = batch[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if val_results:\n val_results[""gt_seq_val""] = (\n val_results[""gt_batch""][""videos""][0].astype(jnp.float32)\n / 255.0\n )\n val_results[""recon_seq_val""] = val_results[""recon""][0].clip(\n 0, 1\n )\n val_comparison_seq = jnp.concatenate(\n (val_results[""gt_seq_val""], val_results[""recon_seq_val""]),\n axis=1,\n )\n val_results[""val_comparison_seq""] = einops.rearrange(\n val_comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if args.eval_full_frame:\n val_results[""full_frame_seq_val""] = val_results[\n ""full_frame""\n ][0].clip(0, 1)\n val_results[""val_full_frame_comparison_seq""] = (\n jnp.concatenate(\n (\n val_results[""gt_seq_val""],\n val_results[""full_frame_seq_val""],\n ),\n axis=1,\n )\n )\n val_results[""val_full_frame_comparison_seq""] = (\n einops.rearrange(\n val_results[""val_full_frame_comparison_seq""] * 255,\n ""t h w c -> h (t w) c"",\n )\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n if val_results:\n log_images.update(\n dict(\n val_image=wandb.Image(\n np.asarray(\n val_results[""gt_seq_val""][args.seq_len - 1]\n )\n ),\n val_recon=wandb.Image(\n np.asarray(\n val_results[""recon_seq_val""][\n args.seq_len - 1\n ]\n )\n ),\n val_true_vs_recon=wandb.Image(\n np.asarray(\n val_results[""val_comparison_seq""].astype(\n np.uint8\n )\n )\n ),\n )\n )\n if args.eval_full_frame:\n log_images.update(\n dict(\n val_full_frame=wandb.Image(\n np.asarray(\n val_results[""full_frame_seq_val""][\n args.seq_len - 1\n ]\n )\n ),\n val_true_vs_full_frame=wandb.Image(\n np.asarray(\n val_results[\n ""val_full_frame_comparison_seq""\n ].astype(np.uint8)\n )\n ),\n )\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n assert checkpoint_manager is not None\n optimizer_state = nnx.state(optimizer)\n if val_iterator:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n val_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n val_iterator # type: ignore\n ),\n )\n else:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n )\n checkpoint_manager.save(step, args=ckpt_manager_args)\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n if checkpoint_manager:\n checkpoint_manager.close()\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n main(args)\n",python,tab +4546,13619660,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",21891,0,"",python,selection_command +4547,13619791,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +4548,13621911,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +4549,13623225,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +4550,13623721,"jasmine/train_dynamics.py",0,0,"",python,tab +4551,13624367,"jasmine/train_dynamics.py",16397,0,"",python,selection_mouse +4552,13626729,"jasmine/train_dynamics.py",21921,0,"",python,selection_command +4553,13635029,"jasmine/train_dynamics.py",21891,69," compiled = train_step.lower(optimizer, first_batch).compile()",python,selection_command +4554,13635260,"jasmine/train_dynamics.py",21891,133," compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())",python,selection_command +4555,13635378,"jasmine/train_dynamics.py",21891,196," compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())",python,selection_command +4556,13635637,"jasmine/train_dynamics.py",21899,0,"",python,selection_command +4557,13636191,"jasmine/train_dynamics.py",22033,0,"#",python,content +4558,13636191,"jasmine/train_dynamics.py",21969,0,"#",python,content +4559,13636191,"jasmine/train_dynamics.py",21899,0,"#",python,content +4560,13636192,"jasmine/train_dynamics.py",21900,0,"",python,selection_keyboard +4561,13636282,"jasmine/train_dynamics.py",22036,0," ",python,content +4562,13636282,"jasmine/train_dynamics.py",21971,0," ",python,content +4563,13636282,"jasmine/train_dynamics.py",21900,0," ",python,content +4564,13636283,"jasmine/train_dynamics.py",21901,0,"",python,selection_keyboard +4565,13636631,"jasmine/train_dynamics.py",21900,0,"",python,selection_command +4566,13646052,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",,terminal_output +4567,13646279,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=06:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\n# source .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\n# lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\r\n# tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\r\n\r\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3518963\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=250 \\r\n --log_checkpoint_interval=250 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-noise-lvl-default-$slurm_job_id \\r\n --tags dyn breakout noise-lvl default \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 100 \\r\n --wsd_decay_steps 1000 \\r\n --num_steps 5000 \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --lam_checkpoint $lam_checkpoint \\r\n --val_interval 1000 \\r\n --eval_full_frame \\r\n",,terminal_output +4568,13646410,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=217802\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0402\r\nSLURM_JOB_START_TIME=1759070937\r\nSLURM_STEP_NODELIST=hkn0402\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759074537\r\nSLURM_PMI2_SRUN_PORT=36803\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3527866\r\nSLURM_PTY_PORT=42519\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=36\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0402\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=139\r\nSLURM_NODELIST=hkn0402\r\nSLURM_SRUN_COMM_PORT=45011\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3527866\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0402\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=45011\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0402\r\n",,terminal_output +4569,13646514,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +4570,13649201,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +4571,13655139,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +4572,13655446,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +4573,13656215,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20250928_171500-r4rqbg5n\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run breakout-dyn-noise-lvl-default-3527866\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/r4rqbg5n\r\n",,terminal_output +4574,13656325,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26561024, 'lam': 16900976, 'tokenizer': 33750256, 'total': 77212256}\r\n",,terminal_output +4575,13658725,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +4576,13670626,"TERMINAL",0,0,"Starting training from step 0...\r\n",,terminal_output +4577,13748427,"TERMINAL",0,0,"\r\nMemstats: After params initialized:\r\n\tUsing (GB) 1.11 / 38.7 (2.868217%) on cuda:0\r\n",,terminal_output +4578,13774104,"TERMINAL",0,0,"bash",,terminal_focus +4579,13774960,"TERMINAL",0,0,"srun",,terminal_focus +4580,13788696,"jasmine/train_dynamics.py",0,0,"",python,tab +4581,13796440,"jasmine/models/dynamics.py",0,0,"",python,tab +4582,13797325,"jasmine/models/dynamics.py",3959,0,"",python,selection_mouse +4583,13797852,"jasmine/models/dynamics.py",3837,0,"",python,selection_mouse +4584,13800091,"jasmine/models/dynamics.py",3888,0,"",python,selection_mouse +4585,13801520,"jasmine/models/dynamics.py",3866,0,"",python,selection_mouse +4586,13802624,"jasmine/models/dynamics.py",3803,84," assert not jnp.isnan(vid_embed_BTNM).any(), ""[DEBUG] vid_embed_BTNM has NaN""",python,selection_command +4587,13802835,"jasmine/models/dynamics.py",3720,167," assert not jnp.isnan(noise_level_B).any(), ""[DEBUG] noise_level_B has NaN""\n assert not jnp.isnan(vid_embed_BTNM).any(), ""[DEBUG] vid_embed_BTNM has NaN""",python,selection_command +4588,13802961,"jasmine/models/dynamics.py",3635,252," assert self.max_noise_level > 0, ""[DEBUG] Max noise level is smaller than 0""\n assert not jnp.isnan(noise_level_B).any(), ""[DEBUG] noise_level_B has NaN""\n assert not jnp.isnan(vid_embed_BTNM).any(), ""[DEBUG] vid_embed_BTNM has NaN""",python,selection_command +4589,13803108,"jasmine/models/dynamics.py",3550,337," assert not jnp.isnan(self.max_noise_level), ""[DEBUG] Max noise level is NaN""\n assert self.max_noise_level > 0, ""[DEBUG] Max noise level is smaller than 0""\n assert not jnp.isnan(noise_level_B).any(), ""[DEBUG] noise_level_B has NaN""\n assert not jnp.isnan(vid_embed_BTNM).any(), ""[DEBUG] vid_embed_BTNM has NaN""",python,selection_command +4590,13804124,"jasmine/models/dynamics.py",3558,0,"",python,selection_command +4591,13821752,"jasmine/models/dynamics.py",3550,0,"",python,selection_command +4592,13823982,"jasmine/models/dynamics.py",3550,0," assert not jnp.isnan(self.max_noise_level), f""[DEBUG] Max noise level is NaN, value: {self.max_noise_level}""\n",python,content +4593,13824271,"jasmine/models/dynamics.py",3667,0," assert self.max_noise_level > 0, f""[DEBUG] Max noise level is smaller than 0, value: {self.max_noise_level}""\n",python,content +4594,13824600,"jasmine/models/dynamics.py",3784,0," assert not jnp.isnan(noise_level_B).any(), f""[DEBUG] noise_level_B has NaN, values: {noise_level_B}""\n",python,content +4595,13824925,"jasmine/models/dynamics.py",3893,0," assert not jnp.isnan(vid_embed_BTNM).any(), f""[DEBUG] vid_embed_BTNM has NaN, values: {vid_embed_BTNM}""\n",python,content +4596,13824928,"jasmine/models/dynamics.py",4005,338,"",python,content +4597,13894040,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3527866.2 task 0: running\r\n",,terminal_output +4598,13894449,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3527866.2\r\nsrun: forcing job termination\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-1:\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nTraceback (most recent call last):\r\nKeyboardInterrupt\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nTraceback (most recent call last):\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\nTraceback (most recent call last):\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3527866.2 ON hkn0402 CANCELLED AT 2025-09-28T17:18:59 ***\r\n",,terminal_output +4599,13894593,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3527866.2\r\nsrun: job abort in progress\r\n",,terminal_output +4600,13894747,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3527866.2\r\n",,terminal_output +4601,13894827,"TERMINAL",0,0,"]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +4602,13933381,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",,terminal_output +4603,13935162,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +4604,13935324,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine]$ sh slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",,terminal_output +4605,13936797,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",84,0,"",shellscript,selection_mouse +4606,13937871,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",83,1,"",shellscript,content +4607,13938001,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",82,1,"",shellscript,content +4608,13938848,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",82,0,"4",shellscript,content +4609,13938849,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",83,0,"",shellscript,selection_keyboard +4610,13938906,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",83,0,"9",shellscript,content +4611,13938907,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",84,0,"",shellscript,selection_keyboard +4612,13939210,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",83,1,"",shellscript,content +4613,13939362,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",83,0,"8",shellscript,content +4614,13939363,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",84,0,"",shellscript,selection_keyboard +4615,13941614,"TERMINAL",0,0,"\r\n\r\r\n\r\r\n\r",,terminal_output +4616,13941762,"TERMINAL",0,0,"",,terminal_output +4617,13942505,"TERMINAL",0,0,"s \r",,terminal_output +4618,13942723,"TERMINAL",0,0,"y",,terminal_output +4619,13942773,"TERMINAL",0,0,"n",,terminal_output +4620,13942993,"TERMINAL",0,0,"c",,terminal_output +4621,13943454,"TERMINAL",0,0,"-",,terminal_output +4622,13943812,"TERMINAL",0,0,"r",,terminal_output +4623,13943901,"TERMINAL",0,0,"u",,terminal_output +4624,13944135,"TERMINAL",0,0,"n",,terminal_output +4625,13944321,"TERMINAL",0,0,"n",,terminal_output +4626,13944677,"TERMINAL",0,0,"e",,terminal_output +4627,13944783,"TERMINAL",0,0,"r",,terminal_output +4628,13944935,"TERMINAL",0,0,"-",,terminal_output +4629,13945099,"TERMINAL",0,0,"2",,terminal_output +4630,13945396,"TERMINAL",0,0,"\r\n[?2004l\rsending incremental file list\r\n",,terminal_output +4631,13950785,"TERMINAL",0,0,"jasmine/\r\njasmine/genie.py\r\njasmine/sample.py\r\njasmine/train_dynamics.py\r\njasmine/models/dynamics.py\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/train_dyn_single_gpu.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/train_dyn_single_gpu_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/train_dyn_single_gpu_gt_actions.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/train_dyn_single_gpu_gt_actions_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/train_dyn_single_gpu_smaller_lr_50k.sh\r\n",,terminal_output +4632,13951082,"TERMINAL",0,0,"\r\nsent 135,267 bytes received 495 bytes 24,684.00 bytes/sec\r\ntotal size is 28,031,762 speedup is 206.48\r\n]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +4633,13952578,"TERMINAL",0,0,"r \r",,terminal_output +4634,13952734,"TERMINAL",0,0,"u",,terminal_output +4635,13952798,"TERMINAL",0,0,"n",,terminal_output +4636,13952935,"TERMINAL",0,0,"n",,terminal_output +4637,13953024,"TERMINAL",0,0,"e",,terminal_output +4638,13953083,"TERMINAL",0,0,"r",,terminal_output +4639,13953212,"TERMINAL",0,0,"-",,terminal_output +4640,13953274,"TERMINAL",0,0,"2",,terminal_output +4641,13953602,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_2[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +4642,13954338,"TERMINAL",0,0,"s",,terminal_output +4643,13954417,"TERMINAL",0,0,"b",,terminal_output +4644,13954522,"TERMINAL",0,0,"a",,terminal_output +4645,13954580,"TERMINAL",0,0,"t",,terminal_output +4646,13954710,"TERMINAL",0,0,"c",,terminal_output +4647,13954762,"TERMINAL",0,0,"h",,terminal_output +4648,13954878,"TERMINAL",0,0," ",,terminal_output +4649,13955295,"TERMINAL",0,0,"[DEBUG]",,terminal_output +4650,13956064,"TERMINAL",0,0,"",,terminal_output +4651,13957285,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +4652,13963815,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +4653,13963816,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",1960,0,"",shellscript,selection_mouse +4654,13964363,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",1967,0,"",shellscript,selection_mouse +4655,13965065,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",1970,0,"",shellscript,selection_mouse +4656,13965712,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",1968,0,"",shellscript,selection_mouse +4657,13966461,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",1969,0,"",shellscript,selection_command +4658,13966757,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",1962,7,"",shellscript,content +4659,13967889,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",1962,0,"default",shellscript,content +4660,13973999,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu copy.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --job-name=train_dyn_default_breakout_longer\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n# source .venv/bin/activate\n\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\n# lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\n# tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\n\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3518963\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\n\nenv | grep SLURM\n\nexport PYTHONUNBUFFERED=1\n\nsrun python jasmine/train_dynamics.py \\n --save_ckpt \\n --image_height=10 \\n --image_width=10 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=120 \\n --init_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=250 \\n --log_checkpoint_interval=250 \\n --dyna_type=maskgit \\n --log \\n --name=breakout-dyn-noise-lvl-default-$slurm_job_id \\n --tags dyn breakout noise-lvl default \\n --entity instant-uv \\n --project jafar \\n --patch_size 4 \\n --lam_patch_size 4 \\n --warmup_steps 100 \\n --wsd_decay_steps 1000 \\n --num_steps 5000 \\n --data_dir $array_records_dir_train \\n --val_data_dir $array_records_dir_val \\n --tokenizer_checkpoint $tokenizer_checkpoint \\n --lam_checkpoint $lam_checkpoint \\n --val_interval 1000 \\n --eval_full_frame \\n",shellscript,tab +4661,13982061,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --job-name=train_dyn_default_breakout_longer\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n# source .venv/bin/activate\n\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\n# lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\n# tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\n\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3518963\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3518963\n\nenv | grep SLURM\n\nexport PYTHONUNBUFFERED=1\n\nsrun python jasmine/train_dynamics.py \\n --save_ckpt \\n --image_height=10 \\n --image_width=10 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=120 \\n --init_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=250 \\n --log_checkpoint_interval=250 \\n --dyna_type=maskgit \\n --log \\n --name=breakout-dyn-noise-lvl-default-$slurm_job_id \\n --tags dyn breakout noise-lvl default \\n --entity instant-uv \\n --project jafar \\n --patch_size 4 \\n --lam_patch_size 4 \\n --warmup_steps 100 \\n --wsd_decay_steps 1000 \\n --num_steps 5000 \\n --data_dir $array_records_dir_train \\n --val_data_dir $array_records_dir_val \\n --tokenizer_checkpoint $tokenizer_checkpoint \\n --lam_checkpoint $lam_checkpoint \\n --val_interval 1000 \\n --eval_full_frame \\n",shellscript,tab +4662,13984492,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2427,0,"",shellscript,selection_mouse +4663,13986226,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1939,0,"",shellscript,selection_mouse +4664,13986946,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1939,8,"",shellscript,content +4665,13987334,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1939,1,"",shellscript,content +4666,13987602,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1939,3,"",shellscript,content +4667,13987895,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1939,1,"",shellscript,content +4668,13988205,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1939,5,"",shellscript,content +4669,13988448,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1939,1,"",shellscript,content +4670,13989070,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1939,3,"",shellscript,content +4671,13989519,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1939,1,"",shellscript,content +4672,13989825,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1939,7,"",shellscript,content +4673,13991552,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1939,0,"b",shellscript,content +4674,13991554,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1940,0,"",shellscript,selection_keyboard +4675,13991606,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1940,0,"r",shellscript,content +4676,13991607,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1941,0,"",shellscript,selection_keyboard +4677,13991754,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1941,0,"e",shellscript,content +4678,13991755,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1942,0,"",shellscript,selection_keyboard +4679,13991996,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1942,0,"a",shellscript,content +4680,13991997,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1943,0,"",shellscript,selection_keyboard +4681,13992108,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1943,0,"k",shellscript,content +4682,13992109,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1944,0,"",shellscript,selection_keyboard +4683,13992226,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1944,0,"o",shellscript,content +4684,13992227,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1945,0,"",shellscript,selection_keyboard +4685,13992332,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1945,0,"u",shellscript,content +4686,13992333,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1946,0,"",shellscript,selection_keyboard +4687,13992439,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1946,0,"t",shellscript,content +4688,13992439,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1947,0,"",shellscript,selection_keyboard +4689,13992576,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1947,0,"-",shellscript,content +4690,13992577,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1948,0,"",shellscript,selection_keyboard +4691,13992950,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1948,0,"a",shellscript,content +4692,13992951,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1949,0,"",shellscript,selection_keyboard +4693,13993259,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1948,1,"",shellscript,content +4694,13993495,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1948,0,"n",shellscript,content +4695,13993496,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1949,0,"",shellscript,selection_keyboard +4696,13993589,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1949,0,"a",shellscript,content +4697,13993590,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1950,0,"",shellscript,selection_keyboard +4698,13993656,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1950,0,"n",shellscript,content +4699,13993657,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1951,0,"",shellscript,selection_keyboard +4700,13993873,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1951,0,"-",shellscript,content +4701,13993874,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1952,0,"",shellscript,selection_keyboard +4702,13994945,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1952,0,"a",shellscript,content +4703,13994946,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1953,0,"",shellscript,selection_keyboard +4704,13995058,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1953,0,"s",shellscript,content +4705,13995058,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1954,0,"",shellscript,selection_keyboard +4706,13995249,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1954,0,"s",shellscript,content +4707,13995250,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1955,0,"",shellscript,selection_keyboard +4708,13995458,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1955,0,"e",shellscript,content +4709,13995459,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1956,0,"",shellscript,selection_keyboard +4710,13995536,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1956,0,"r",shellscript,content +4711,13995537,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1957,0,"",shellscript,selection_keyboard +4712,13995641,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1957,0,"t",shellscript,content +4713,13995641,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1958,0,"",shellscript,selection_keyboard +4714,13995857,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1958,0,"i",shellscript,content +4715,13995857,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1959,0,"",shellscript,selection_keyboard +4716,13995936,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1959,0,"o",shellscript,content +4717,13995936,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1960,0,"",shellscript,selection_keyboard +4718,13996108,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1960,0,"n",shellscript,content +4719,13996108,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1961,0,"",shellscript,selection_keyboard +4720,13996335,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1961,0,"-",shellscript,content +4721,13996336,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1962,0,"",shellscript,selection_keyboard +4722,13996657,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1962,0,"r",shellscript,content +4723,13996658,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1963,0,"",shellscript,selection_keyboard +4724,13996791,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1963,0,"u",shellscript,content +4725,13996791,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1964,0,"",shellscript,selection_keyboard +4726,13996871,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1964,0,"n",shellscript,content +4727,13996872,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1965,0,"",shellscript,selection_keyboard +4728,13997238,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1964,0,"",shellscript,selection_command +4729,13998685,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1997,0,"",shellscript,selection_mouse +4730,13998857,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1997,1,"b",shellscript,selection_mouse +4731,13998877,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1997,2,"br",shellscript,selection_mouse +4732,13998932,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1997,3,"bre",shellscript,selection_mouse +4733,13998940,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1997,4,"brea",shellscript,selection_mouse +4734,13998958,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1997,5,"break",shellscript,selection_mouse +4735,13998975,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1997,6,"breako",shellscript,selection_mouse +4736,13999031,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1997,7,"breakou",shellscript,selection_mouse +4737,13999038,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",1997,8,"breakout",shellscript,selection_mouse +4738,14000617,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2023,0,"",shellscript,selection_mouse +4739,14001464,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2016,7,"",shellscript,content +4740,14002258,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2016,0,"d",shellscript,content +4741,14002259,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2017,0,"",shellscript,selection_keyboard +4742,14002459,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2017,0,"e",shellscript,content +4743,14002460,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2018,0,"",shellscript,selection_keyboard +4744,14002557,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2018,0,"b",shellscript,content +4745,14002558,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2019,0,"",shellscript,selection_keyboard +4746,14002639,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2019,0,"u",shellscript,content +4747,14002640,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2020,0,"",shellscript,selection_keyboard +4748,14002754,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2020,0,"g",shellscript,content +4749,14002755,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2021,0,"",shellscript,selection_keyboard +4750,14002945,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2020,0,"",shellscript,selection_command +4751,14007808,"TERMINAL",0,0,"",,terminal_output +4752,14008362,"TERMINAL",0,0,"s",,terminal_output +4753,14008601,"TERMINAL",0,0,"yn",,terminal_output +4754,14008809,"TERMINAL",0,0,"c",,terminal_output +4755,14008871,"TERMINAL",0,0,"-",,terminal_output +4756,14009094,"TERMINAL",0,0,"r",,terminal_output +4757,14009158,"TERMINAL",0,0,"u",,terminal_output +4758,14009277,"TERMINAL",0,0,"n",,terminal_output +4759,14009402,"TERMINAL",0,0,"n",,terminal_output +4760,14009465,"TERMINAL",0,0,"e",,terminal_output +4761,14009527,"TERMINAL",0,0,"r",,terminal_output +4762,14009611,"TERMINAL",0,0,"-",,terminal_output +4763,14009746,"TERMINAL",0,0,"2",,terminal_output +4764,14009891,"TERMINAL",0,0,"\r\n[?2004l\rsending incremental file list\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh\r\n",,terminal_output +4765,14009967,"TERMINAL",0,0,"\r\nsent 38,090 bytes received 256 bytes 76,692.00 bytes/sec\r\ntotal size is 28,034,183 speedup is 731.08\r\n]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_2[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +4766,14012680,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +4767,14015735,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +4768,14019420,"TERMINAL",0,0,"s",,terminal_output +4769,14019487,"TERMINAL",0,0,"b",,terminal_output +4770,14019607,"TERMINAL",0,0,"a",,terminal_output +4771,14019714,"TERMINAL",0,0,"t",,terminal_output +4772,14019783,"TERMINAL",0,0,"c",,terminal_output +4773,14019887,"TERMINAL",0,0,"h",,terminal_output +4774,14019949,"TERMINAL",0,0," ",,terminal_output +4775,14020246,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",,terminal_output +4776,14023348,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh\r\n[?2004l\rSubmitted batch job 3527891\r\n]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_2[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +4777,14024486,"TERMINAL",0,0,"q",,terminal_output +4778,14024559,"TERMINAL",0,0,"u",,terminal_output +4779,14024663,"TERMINAL",0,0,"e",,terminal_output +4780,14024725,"TERMINAL",0,0,"u",,terminal_output +4781,14024791,"TERMINAL",0,0,"e",,terminal_output +4782,14024906,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: squeue --mehkn0402.localdomain: Sun Sep 28 17:21:09 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3527891 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3527866 dev_accel interact tum_cte0 R32:12\t 1 hkn0402",,terminal_output +4783,14025935,"TERMINAL",0,0,"103",,terminal_output +4784,14026759,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_2[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +4785,14027954,"TERMINAL",0,0,"d",,terminal_output +4786,14028076,"TERMINAL",0,0,"e",,terminal_output +4787,14028210,"TERMINAL",0,0,"v",,terminal_output +4788,14028283,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +4789,14031963,"jasmine/train_dynamics.py",0,0,"",python,tab +4790,14036154,"jasmine/train_dynamics.py",16546,0,"",python,selection_mouse +4791,14036156,"jasmine/train_dynamics.py",16545,0,"",python,selection_command +4792,14037000,"jasmine/train_dynamics.py",16536,0,"",python,selection_command +4793,14037907,"jasmine/train_dynamics.py",16536,1,"",python,content +4794,14038055,"jasmine/train_dynamics.py",16536,1,"",python,content +4795,14038368,"jasmine/train_dynamics.py",16535,0,"",python,selection_command +4796,14044309,"jasmine/train_dynamics.py",15880,0,"",python,selection_mouse +4797,14044888,"jasmine/train_dynamics.py",15873,0,"",python,selection_command +4798,14045240,"jasmine/train_dynamics.py",15873,1,"",python,content +4799,14045441,"jasmine/train_dynamics.py",15873,1,"",python,content +4800,14045750,"jasmine/train_dynamics.py",15872,0,"",python,selection_command +4801,14051045,"jasmine/models/dynamics.py",0,0,"",python,tab +4802,14053116,"jasmine/models/dynamics.py",3613,0,"",python,selection_command +4803,14053419,"jasmine/models/dynamics.py",3730,0,"",python,selection_command +4804,14053597,"jasmine/models/dynamics.py",3847,0,"",python,selection_command +4805,14053708,"jasmine/models/dynamics.py",3956,0,"",python,selection_command +4806,14055517,"jasmine/models/dynamics.py",3893,111," assert not jnp.isnan(vid_embed_BTNM).any(), f""[DEBUG] vid_embed_BTNM has NaN, values: {vid_embed_BTNM}""",python,selection_command +4807,14055703,"jasmine/models/dynamics.py",3784,220," assert not jnp.isnan(noise_level_B).any(), f""[DEBUG] noise_level_B has NaN, values: {noise_level_B}""\n assert not jnp.isnan(vid_embed_BTNM).any(), f""[DEBUG] vid_embed_BTNM has NaN, values: {vid_embed_BTNM}""",python,selection_command +4808,14055849,"jasmine/models/dynamics.py",3667,337," assert self.max_noise_level > 0, f""[DEBUG] Max noise level is smaller than 0, value: {self.max_noise_level}""\n assert not jnp.isnan(noise_level_B).any(), f""[DEBUG] noise_level_B has NaN, values: {noise_level_B}""\n assert not jnp.isnan(vid_embed_BTNM).any(), f""[DEBUG] vid_embed_BTNM has NaN, values: {vid_embed_BTNM}""",python,selection_command +4809,14056003,"jasmine/models/dynamics.py",3550,454," assert not jnp.isnan(self.max_noise_level), f""[DEBUG] Max noise level is NaN, value: {self.max_noise_level}""\n assert self.max_noise_level > 0, f""[DEBUG] Max noise level is smaller than 0, value: {self.max_noise_level}""\n assert not jnp.isnan(noise_level_B).any(), f""[DEBUG] noise_level_B has NaN, values: {noise_level_B}""\n assert not jnp.isnan(vid_embed_BTNM).any(), f""[DEBUG] vid_embed_BTNM has NaN, values: {vid_embed_BTNM}""",python,selection_command +4810,14056272,"jasmine/models/dynamics.py",3550,455,"",python,content +4811,14056927,"jasmine/models/dynamics.py",3550,1,"",python,content +4812,14056942,"jasmine/models/dynamics.py",3558,0,"",python,selection_command +4813,14058500,"jasmine/models/dynamics.py",3449,0,"",python,selection_mouse +4814,14059084,"jasmine/models/dynamics.py",3423,0,"",python,selection_mouse +4815,14060698,"jasmine/models/dynamics.py",3126,0,"",python,selection_mouse +4816,14061602,"jasmine/models/dynamics.py",3046,0,"",python,selection_mouse +4817,14061729,"jasmine/models/dynamics.py",3034,17,"noise_level_embed",python,selection_mouse +4818,14062945,"jasmine/models/dynamics.py",3053,0,"",python,selection_mouse +4819,14063083,"jasmine/models/dynamics.py",3052,20,"noise_bucket_idx_B11",python,selection_mouse +4820,14064774,"jasmine/models/dynamics.py",2968,0,"",python,selection_mouse +4821,14064911,"jasmine/models/dynamics.py",2960,18,"noise_bucket_idx_B",python,selection_mouse +4822,14069297,"jasmine/models/dynamics.py",2456,0,"",python,selection_mouse +4823,14069304,"jasmine/models/dynamics.py",2455,0,"",python,selection_command +4824,14070162,"jasmine/models/dynamics.py",2394,0,"",python,selection_mouse +4825,14071134,"jasmine/models/dynamics.py",2412,0,"\n ",python,content +4826,14071466,"jasmine/models/dynamics.py",2421,0,"e",python,content +4827,14071467,"jasmine/models/dynamics.py",2422,0,"",python,selection_keyboard +4828,14071656,"jasmine/models/dynamics.py",2422,0,"p",python,content +4829,14071659,"jasmine/models/dynamics.py",2423,0,"",python,selection_keyboard +4830,14071776,"jasmine/models/dynamics.py",2423,0,"s",python,content +4831,14071777,"jasmine/models/dynamics.py",2424,0,"",python,selection_keyboard +4832,14072006,"jasmine/models/dynamics.py",2424,0," ",python,content +4833,14072007,"jasmine/models/dynamics.py",2425,0,"",python,selection_keyboard +4834,14072212,"jasmine/models/dynamics.py",2425,0,"=",python,content +4835,14072213,"jasmine/models/dynamics.py",2426,0,"",python,selection_keyboard +4836,14072326,"jasmine/models/dynamics.py",2426,0," ",python,content +4837,14072328,"jasmine/models/dynamics.py",2427,0,"",python,selection_keyboard +4838,14073405,"jasmine/models/dynamics.py",2427,0,"1",python,content +4839,14073406,"jasmine/models/dynamics.py",2428,0,"",python,selection_keyboard +4840,14074324,"jasmine/models/dynamics.py",2428,0,"3",python,content +4841,14074325,"jasmine/models/dynamics.py",2429,0,"",python,selection_keyboard +4842,14074428,"jasmine/models/dynamics.py",2429,0,"-",python,content +4843,14074429,"jasmine/models/dynamics.py",2430,0,"",python,selection_keyboard +4844,14074997,"jasmine/models/dynamics.py",2429,1,"",python,content +4845,14075094,"jasmine/models/dynamics.py",2428,1,"",python,content +4846,14075268,"jasmine/models/dynamics.py",2427,1,"",python,content +4847,14075827,"jasmine/models/dynamics.py",2427,0,"1",python,content +4848,14075828,"jasmine/models/dynamics.py",2428,0,"",python,selection_keyboard +4849,14076203,"jasmine/models/dynamics.py",2428,0,"e",python,content +4850,14076204,"jasmine/models/dynamics.py",2429,0,"",python,selection_keyboard +4851,14076463,"jasmine/models/dynamics.py",2429,0,"-",python,content +4852,14076465,"jasmine/models/dynamics.py",2430,0,"",python,selection_keyboard +4853,14076733,"jasmine/models/dynamics.py",2430,0,"3",python,content +4854,14076734,"jasmine/models/dynamics.py",2431,0,"",python,selection_keyboard +4855,14079327,"jasmine/models/dynamics.py",2555,0,"",python,selection_mouse +4856,14080231,"jasmine/models/dynamics.py",2555,0," ",python,content +4857,14080233,"jasmine/models/dynamics.py",2556,0,"",python,selection_keyboard +4858,14080350,"jasmine/models/dynamics.py",2556,0,"-",python,content +4859,14080351,"jasmine/models/dynamics.py",2557,0,"",python,selection_keyboard +4860,14080425,"jasmine/models/dynamics.py",2557,0," ",python,content +4861,14080426,"jasmine/models/dynamics.py",2558,0,"",python,selection_keyboard +4862,14081060,"jasmine/models/dynamics.py",2558,0,"e",python,content +4863,14081061,"jasmine/models/dynamics.py",2559,0,"",python,selection_keyboard +4864,14081142,"jasmine/models/dynamics.py",2559,0,"p",python,content +4865,14081144,"jasmine/models/dynamics.py",2560,0,"",python,selection_keyboard +4866,14081238,"jasmine/models/dynamics.py",2560,0,"s",python,content +4867,14081239,"jasmine/models/dynamics.py",2561,0,"",python,selection_keyboard +4868,14082269,"jasmine/models/dynamics.py",2806,0,"",python,selection_mouse +4869,14086162,"jasmine/models/dynamics.py",2453,0,"",python,selection_mouse +4870,14086808,"jasmine/models/dynamics.py",2431,0,"",python,selection_mouse +4871,14102330,"jasmine/models/dynamics.py",2571,0,"",python,selection_mouse +4872,14103130,"jasmine/models/dynamics.py",2550,0,"",python,selection_mouse +4873,14103290,"jasmine/models/dynamics.py",2540,15,"max_noise_level",python,selection_mouse +4874,14110280,"jasmine/models/dynamics.py",2428,0,"",python,selection_mouse +4875,14116921,"jasmine/models/dynamics.py",2427,1,"",python,content +4876,14116981,"jasmine/models/dynamics.py",2427,0,"5",python,content +4877,14116983,"jasmine/models/dynamics.py",2428,0,"",python,selection_keyboard +4878,14117262,"jasmine/models/dynamics.py",2429,0,"",python,selection_command +4879,14117446,"jasmine/models/dynamics.py",2430,0,"",python,selection_command +4880,14117621,"jasmine/models/dynamics.py",2431,0,"",python,selection_command +4881,14127287,"jasmine/models/dynamics.py",3283,0,"",python,selection_mouse +4882,14127441,"jasmine/models/dynamics.py",3282,1,"0",python,selection_mouse +4883,14128030,"jasmine/models/dynamics.py",3330,0,"",python,selection_mouse +4884,14128193,"jasmine/models/dynamics.py",3325,16,"noise_level_B111",python,selection_mouse +4885,14129292,"jasmine/models/dynamics.py",3393,0,"",python,selection_mouse +4886,14129425,"jasmine/models/dynamics.py",3388,15,"one_minus_noise",python,selection_mouse +4887,14129948,"jasmine/models/dynamics.py",3449,0,"",python,selection_mouse +4888,14130105,"jasmine/models/dynamics.py",3444,16,"noise_level_B111",python,selection_mouse +4889,14151891,"TERMINAL",0,0,"g",,terminal_output +4890,14151963,"TERMINAL",0,0,"i",,terminal_output +4891,14152031,"TERMINAL",0,0,"t",,terminal_output +4892,14152134,"TERMINAL",0,0," ",,terminal_output +4893,14152287,"TERMINAL",0,0,"s",,terminal_output +4894,14152412,"TERMINAL",0,0,"t",,terminal_output +4895,14152746,"TERMINAL",0,0,"",,terminal_output +4896,14152811,"TERMINAL",0,0,"",,terminal_output +4897,14153288,"TERMINAL",0,0,"d",,terminal_output +4898,14153358,"TERMINAL",0,0,"i",,terminal_output +4899,14153466,"TERMINAL",0,0,"f",,terminal_output +4900,14153627,"TERMINAL",0,0,"f\r\n[?2004l\r[?1h=\r",,terminal_output +4901,14153761,"TERMINAL",0,0,"diff --git a/jasmine/genie.py b/jasmine/genie.py\r\nindex 061cebd..54da37e 100644\r\n--- a/jasmine/genie.py\r\n+++ b/jasmine/genie.py\r\n@@ -207,16 +207,20 @@ class Genie(nnx.Module):\r\n self,\r\n batch: Dict[str, jax.Array],\r\n seq_len: int,\r\n+ noise_level: float = 0.,\r\n temperature: float = 1,\r\n sample_argmax: bool = False,\r\n maskgit_steps: int = 25,\r\n ) -> tuple[jax.Array, jax.Array]:\r\n+ assert (\r\n+ noise_level < self.max_noise_level\r\n+ ), ""Noise level must me smaller than max_noise_level.""\r\n if self.dyna_type == ""maskgit"":\r\n return self.sample_maskgit(\r\n- batch, seq_len, 0.0, maskgit_steps, temperature, sample_argmax\r\n+ batch, seq_len, noise_level, maskgit_steps, temperature, sample_argmax\r\n )\r\n elif self.dyna_type == ""causal"":\r\n- return self.sample_causal(batch, seq_len, temperature, sample_argmax)\r\n+ return self.sample_causal(batch, seq_len, noise_level, temperature, sample_ argmax)\r\n else:\r\n raise ValueError(f""Dynamics model type unknown: {self.dyna_type}"")\r\n \r\n@@ -255,9 +259,6 @@ class Genie(nnx.Module):\r\n P: S * N\r\n """"""\r\n assert isinstance(self.dynamics, DynamicsMaskGIT)\r\n- assert (\r\n- noise_level < self.max_noise_level\r\n- ), ""Noise level must me smaller than max_noise_level.""\r\n:",,terminal_output +4902,14156120,"TERMINAL",0,0,"\r # --- Encode videos and actions ---\r\n:",,terminal_output +4903,14160459,"TERMINAL",0,0,"\r videos_BTHWC = batch[""videos""]\r\n:",,terminal_output +4904,14160652,"TERMINAL",0,0,"\r tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\r\n:",,terminal_output +4905,14160892,"TERMINAL",0,0,"\r@@ -328,19 +329,18 @@ class Genie(nnx.Module):\r\n:",,terminal_output +4906,14162056,"TERMINAL",0,0,"\r act_embed_BS1M = jnp.reshape(\r\n:",,terminal_output +4907,14162642,"TERMINAL",0,0,"\r act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\r\n:",,terminal_output +4908,14163159,"TERMINAL",0,0,"\r )\r\n:\r- # TODO mihir\r\n:\r \r\n:",,terminal_output +4909,14163217,"TERMINAL",0,0,"\r+ # TODO mihir\r\n:",,terminal_output +4910,14163388,"TERMINAL",0,0,"\r rng, _rng_noise = jax.random.split(rng)\r\n:",,terminal_output +4911,14163766,"TERMINAL",0,0,"\r noise_level_111 = noise_level.reshape(1, 1, 1)\r\n:",,terminal_output +4912,14164006,"TERMINAL",0,0,"\r noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\r\n:",,terminal_output +4913,14165092,"TERMINAL",0,0,"\r noise_bucket_idx_B11 = jnp.floor(\r\n:\r- (noise_level_B11 / self.max_noise_level) * self.noise_buckets\r\n:\r+ (noise_level_B11 * self.noise_buckets) / self.max_noise_level\r\n:\r ).astype(jnp.int32)\r\n:\r noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\r\n:\r noise_bucket_idx_B11\r\n:\r )\r\n:\r noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\r\n:\r- vid_embed_BSNM += jnp.expand_dims(noise_level_B11, -1)\r\n:\r \r\n:\r vid_embed_BSNp2M = jnp.concatenate(\r\n:\r [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\r\n:\r@@ -441,6 +441,7 @@ class Genie(nnx.Module):\r\n:\r self,\r\n:\r batch: Dict[str, jax.Array],\r\n:\r seq_len: int,\r\n:\r+ noise_level: float,\r\n:\r temperature: float = 1,\r\n:\r sample_argmax: bool = False,\r\n:\r ) -> tuple[jax.Array, jax.Array]:\r\n:\r@@ -528,12 +529,28 @@ class Genie(nnx.Module):\r\n:\r act_embed_BS1M = jnp.reshape(\r\n:",,terminal_output +4914,14165152,"TERMINAL",0,0,"\r act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\r\n:",,terminal_output +4915,14166233,"TERMINAL",0,0,"\r )\r\n:",,terminal_output +4916,14167280,"TERMINAL",0,0,"\r- vid_embed_BSNp1M = jnp.concatenate([act_embed_BS1M, vid_embed_BSNM], axis=2 :\r)\r\n:\r- final_logits_BTNp1V = (\r\n:\r- dynamics_causal.transformer(vid_embed_BSNp1M, (step_t, step_n))\r\n:\r+\r\n:\r+ # TODO mihir\r\n:\r+\r\n:\r+ rng, _rng_noise = jax.random.split(rng)\r\n:\r+ noise_level_111 = noise_level.reshape(1, 1, 1)\r\n:\r+ noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\r\n:\r+ noise_bucket_idx_B11 = jnp.floor(\r\n:\r+ (noise_level_B11 * self.noise_buckets) / self.max_noise_level\r\n:\r+ ).astype(jnp.int32)\r\n:\r+ noise_level_embed_B11M = dynamics_causal.noise_level_embed(\r\n:\r+ noise_bucket_idx_B11\r\n:\r+ )\r\n:\r+ noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\r\n:\r+\r\n:\r+ vid_embed_BSNp2M = jnp.concatenate(\r\n:\r+ [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\r\n:",,terminal_output +4917,14167332,"TERMINAL",0,0,"\r+ )\r\n:",,terminal_output +4918,14167391,"TERMINAL",0,0,"\r+ final_logits_BTNp2V = (\r\n:",,terminal_output +4919,14167787,"TERMINAL",0,0,"\r+ dynamics_causal.transformer(vid_embed_BSNp2M, (step_t, step_n))\r\n:",,terminal_output +4920,14168264,"TERMINAL",0,0,"\r / temperature\r\n:",,terminal_output +4921,14169223,"TERMINAL",0,0,"\r )\r\n:\r- final_logits_BV = final_logits_BTNp1V[:, step_t, step_n, :]\r\n:\r+ final_logits_BV = final_logits_BTNp2V[:, step_t, step_n + 1, :]\r\n:\r \r\n:\r # --- Sample new tokens for final frame ---\r\n:\r if sample_argmax:\r\n:\rdiff --git a/jasmine/models/dynamics.py b/jasmine/models/dynamics.py\r\n:\rindex 5ce4dcd..0a69649 100644\r\n:\r--- a/jasmine/models/dynamics.py\r\n:\r+++ b/jasmine/models/dynamics.py\r\n:\r@@ -78,6 +78,33 @@ class DynamicsMaskGIT(nnx.Module):\r\n:\r self.noise_buckets, self.model_dim, rngs=rngs\r\n:\r )\r\n:\r \r\n:\r+ def _apply_noise_augmentation(self, vid_embed_BTNM, rng):\r\n:\r+ B, T, N, M = vid_embed_BTNM.shape\r\n:\r+ rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\r\n:\r+ eps = 5e-3\r\n:\r+ noise_level_B = jax.random.uniform(\r\n:\r+ _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level - eps\r\n:\r+ )\r\n:\r+ noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\r\n:\r+ # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\r\n:\r+ # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\r\n:\r+ noise_bucket_idx_B = jnp.floor(\r\n:\r+ (noise_level_B * self.noise_buckets) / self.max_noise_level\r\n:\r+ ).astype(jnp.int32)\r\n:\r+ noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\r\n:\r+ noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\r\n:\r+ noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\r\n:\r+ noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\r\n:\r+\r\n:",,terminal_output +4922,14170176,"TERMINAL",0,0,"\r+ # safe sqrt: clip argument to >= 0\r\n:",,terminal_output +4923,14170707,"TERMINAL",0,0,"\r+ one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\r\n:\r+ sqrt_one_minus = jnp.sqrt(one_minus_noise)\r\n:\r+ sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\r\n:",,terminal_output +4924,14170768,"TERMINAL",0,0,"\r+\r\n:",,terminal_output +4925,14170945,"TERMINAL",0,0,"\r+ noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * :",,terminal_output +4926,14171543,"TERMINAL",0,0,"\r noise_BTNM\r\n:\r+\r\n:\r+ return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M\r\n:\r+\r\n:",,terminal_output +4927,14171605,"TERMINAL",0,0,"\r def __call__(\r\n:",,terminal_output +4928,14176941,"TERMINAL",0,0,"\r self,\r\n:",,terminal_output +4929,14178208,"TERMINAL",0,0,"\r batch: Dict[str, jax.Array],\r\n:\r@@ -87,7 +114,7 @@ class DynamicsMaskGIT(nnx.Module):\r\n:\r latent_actions_BTm11L = batch[""latent_actions""]\r\n:\r vid_embed_BTNM = self.patch_embed(video_tokens_BTN)\r\n:\r \r\n:\r- B, T, N, M = vid_embed_BTNM.shape\r\n:\r+ B = vid_embed_BTNM.shape[0]\r\n:\r rng, _rng_prob, *_rngs_mask = jax.random.split(batch[""mask_rng""], B + 2)\r\n:\r mask_prob = jax.random.uniform(_rng_prob, shape=(B,), minval=self.mask_limit)\r\n:\r per_sample_shape = vid_embed_BTNM.shape[1:-1]\r\n:\r@@ -100,28 +127,9 @@ class DynamicsMaskGIT(nnx.Module):\r\n:\r jnp.expand_dims(mask, -1), self.mask_token.value, vid_embed_BTNM\r\n:\r )\r\n:\r \r\n:\r- # --- Sample noise ---\r\n:\r- rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\r\n:\r- noise_level_B = jax.random.uniform(\r\n:\r- _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\r\n:\r- )\r\n:\r- noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\r\n:\r- # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\r\n:\r- # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\r\n:\r- noise_bucket_idx_B = jnp.floor(\r\n:\r- (noise_level_B * self.noise_buckets) / self.max_noise_level\r\n:\r- ).astype(jnp.int32)\r\n:\r- noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\r\n:\r- noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\r\n:",,terminal_output +4930,14178266,"TERMINAL",0,0,"\r- noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\r\n:",,terminal_output +4931,14179052,"TERMINAL",0,0,"\r- noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\r\n:",,terminal_output +4932,14179658,"TERMINAL",0,0,"\r-\r\n:\r- # safe sqrt: clip argument to >= 0\r\n:\r- one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\r\n:\r- sqrt_one_minus = jnp.sqrt(one_minus_noise)\r\n:\r- sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\r\n:",,terminal_output +4933,14180299,"TERMINAL",0,0,"\r+ # --- Apply noise augmentation ---\r\n:",,terminal_output +4934,14180959,"TERMINAL",0,0,"\r+ vid_embed_BTNM, noise_level_embed_BT1M = self._apply_noise_augmentation(vid_emb :\red_BTNM, rng)\r\n:\r \r\n:\r- vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * noise_BTNM\r\n:\r # --- Predict transition ---\r\n:\r act_embed_BTm11M = self.action_up(latent_actions_BTm11L)\r\n:\r padded_act_embed_BT1M = jnp.pad(\r\n:\r@@ -147,6 +155,8 @@ class DynamicsCausal(nnx.Module):\r\n:",,terminal_output +4935,14181021,"TERMINAL",0,0,"\r num_blocks: int,\r\n:",,terminal_output +4936,14181904,"TERMINAL",0,0,"\r num_heads: int,\r\n:",,terminal_output +4937,14182369,"TERMINAL",0,0,"\r dropout: float,\r\n:",,terminal_output +4938,14183057,"TERMINAL",0,0,"\r+ max_noise_level: float,\r\n:\r+ noise_buckets: int,\r\n:\r param_dtype: jnp.dtype,\r\n:\r dtype: jnp.dtype,\r\n:\r use_flash_attention: bool,\r\n:\r@@ -160,6 +170,8 @@ class DynamicsCausal(nnx.Module):\r\n:\r self.num_blocks = num_blocks\r\n:\r self.num_heads = num_heads\r\n:\r self.dropout = dropout\r\n:\r+ self.max_noise_level = max_noise_level\r\n:\r+ self.noise_buckets = noise_buckets\r\n:\r self.param_dtype = param_dtype\r\n:\r self.dtype = dtype\r\n:\r self.use_flash_attention = use_flash_attention\r\n:\r@@ -187,6 +199,34 @@ class DynamicsCausal(nnx.Module):\r\n:\r dtype=self.dtype,\r\n:\r rngs=rngs,\r\n:\r )\r\n:\r+ self.noise_level_embed = nnx.Embed(\r\n:\r+ self.noise_buckets, self.model_dim, rngs=rngs\r\n:",,terminal_output +4939,14185301,"TERMINAL",0,0,"\r+ )\r\n:",,terminal_output +4940,14186535,"TERMINAL",0,0,"\r+\r\n:\r+ def _apply_noise_augmentation(self, vid_embed_BTNM, rng):\r\n:\r+ B, T, N, M = vid_embed_BTNM.shape\r\n:\r+ rng, _rng_noise_lvl, _rng_noise = jax.random.split(rng, 3)\r\n:\r+ noise_level_B = jax.random.uniform(\r\n:\r+ _rng_noise_lvl, shape=(B,), minval=0.0, maxval=self.max_noise_level\r\n:\r+ )\r\n:\r+ noise_BTNM = jax.random.normal(_rng_noise, shape=(B, T, N, M))\r\n:\r+ # We calculate `(noise_level * noise_buckets) / max_noise_level` instead of\r\n:\r+ # `(noise_level_B / max_noise_level) * noise_buckets` for numerical stability.\r\n:\r+ noise_bucket_idx_B = jnp.floor(\r\n:\r+ (noise_level_B * self.noise_buckets) / self.max_noise_level\r\n:\r+ ).astype(jnp.int32)\r\n:\r+ noise_bucket_idx_B11 = noise_bucket_idx_B.reshape(B, 1, 1)\r\n:\r+ noise_level_embed_B11M = self.noise_level_embed(noise_bucket_idx_B11)\r\n:\r+ noise_level_embed_BT1M = jnp.tile(noise_level_embed_B11M, (1, T, 1, 1))\r\n:\r+ noise_level_B111 = noise_level_B.reshape(B, 1, 1, 1)\r\n:\r+\r\n:\r+ # safe sqrt: clip argument to >= 0\r\n:\r+ one_minus_noise = jnp.clip(1.0 - noise_level_B111, a_min=0.0)\r\n:\r+ sqrt_one_minus = jnp.sqrt(one_minus_noise)\r\n:\r+ sqrt_noise = jnp.sqrt(jnp.clip(noise_level_B111, a_min=0.0))\r\n:\r+\r\n:\r+ noise_augmented_vid_embed_BTNM = sqrt_one_minus * vid_embed_BTNM + sqrt_noise * :\r noise_BTNM\r\n:\r+ return noise_augmented_vid_embed_BTNM, noise_level_embed_BT1M\r\n:",,terminal_output +4941,14186591,"TERMINAL",0,0,"\r \r\n:",,terminal_output +4942,14188128,"TERMINAL",0,0,"\r def __call__(\r\n:",,terminal_output +4943,14188683,"TERMINAL",0,0,"\r self,\r\n:\r@@ -199,9 +239,10 @@ class DynamicsCausal(nnx.Module):\r\n:\r padded_act_embed_BT1M = jnp.pad(\r\n:",,terminal_output +4944,14189422,"TERMINAL",0,0,"\r act_embed_BTm11M, ((0, 0), (1, 0), (0, 0), (0, 0))\r\n:\r )\r\n:\r- vid_embed_BTNp1M = jnp.concatenate(\r\n:\r- [padded_act_embed_BT1M, vid_embed_BTNM], axis=2\r\n:\r+ vid_embed_BTNM, noise_level_embed_BT1M = self._apply_noise_augmentation(video_t :\rokens_BTN, batch[""rng""])\r\n:\r+ vid_embed_BTNp2M = jnp.concatenate(\r\n:\r+ [padded_act_embed_BT1M, noise_level_embed_BT1M, vid_embed_BTNM], axis=2\r\n:\r )\r\n:\r- logits_BTNp1V = self.transformer(vid_embed_BTNp1M)\r\n:\r- logits_BTNV = logits_BTNp1V[:, :, :-1]\r\n:\r+ logits_BTNp2V = self.transformer(vid_embed_BTNp2M)\r\n:\r+ logits_BTNV = logits_BTNp2V[:, :, 1:-1]\r\n:\r return logits_BTNV, jnp.ones_like(video_tokens_BTN)\r\n:\rdiff --git a/jasmine/train_dynamics.py b/jasmine/train_dynamics.py\r\n:\rindex 2734952..20568ff 100644\r\n:\r--- a/jasmine/train_dynamics.py\r\n:\r+++ b/jasmine/train_dynamics.py\r\n:\r@@ -506,11 +506,12 @@ def main(args: Args) -> None:\r\n:\r :, :-1\r\n:\r ] # remove last frame for generation\r\n:\r recon_full_frame, logits_full_frame = genie.sample(\r\n:\r- inputs,\r\n:\r- args.seq_len,\r\n:",,terminal_output +4945,14197232,"TERMINAL",0,0,"\r- args.val_temperature,\r\n:",,terminal_output +4946,14198428,"TERMINAL",0,0,"\r- args.val_sample_argmax,\r\n:\r- args.val_maskgit_steps,\r\n:\r+ batch=inputs,\r\n:\r+ seq_len=args.seq_len,\r\n:\r+ noise_level=0.,\r\n:\r+ temperature=args.val_temperature,\r\n:\r+ sample_argmax=args.val_sample_argmax,\r\n:\r+ maskgit_steps=args.val_maskgit_steps,\r\n:\r )\r\n:\r # Calculate metrics for the last frame only\r\n:\r step_outputs = {\r\n:\r@@ -616,9 +617,9 @@ def main(args: Args) -> None:\r\n:\r if jax.process_index() == 0:\r\n:\r first_batch = next(dataloader_train)\r\n:\r first_batch[""rng""] = rng # type: ignore\r\n:\r- compiled = train_step.lower(optimizer, first_batch).compile()\r\n:\r- print_compiled_memory_stats(compiled.memory_analysis())\r\n:\r- print_compiled_cost_analysis(compiled.cost_analysis())\r\n:\r+ # compiled = train_step.lower(optimizer, first_batch).compile()\r\n:\r+ # print_compiled_memory_stats(compiled.memory_analysis())\r\n:\r+ # print_compiled_cost_analysis(compiled.cost_analysis())\r\n:\r # Do not skip the first batch during training\r\n:\r dataloader_train = itertools.chain([first_batch], dataloader_train)\r\n:\r print(f""Starting training from step {step}..."")\r\n:\r\r(END)",,terminal_output +4947,14198487,"TERMINAL",0,0,"\r\r(END)",,terminal_output +4948,14204809,"TERMINAL",0,0,"\r\r(END)",,terminal_output +4949,14205416,"TERMINAL",0,0,"\r\r(END)\r\r(END)\r\r(END)\r\r(END)\r\r(END)",,terminal_output +4950,14205473,"TERMINAL",0,0,"\r\r(END)",,terminal_output +4951,14205666,"TERMINAL",0,0,"\r\r(END)",,terminal_output +4952,14205778,"TERMINAL",0,0,"\r\r(END)",,terminal_output +4953,14205950,"TERMINAL",0,0,"\r\r(END)",,terminal_output +4954,14212620,"jasmine/train_dynamics.py",0,0,"",python,tab +4955,14212621,"jasmine/train_dynamics.py",21760,512," if jax.process_index() == 0:\n first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n # compiled = train_step.lower(optimizer, first_batch).compile()\n # print_compiled_memory_stats(compiled.memory_analysis())\n # print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader_train = itertools.chain([first_batch], dataloader_train)\n print(f""Starting training from step {step}..."")\n",python,selection_command +4956,14213513,"jasmine/train_dynamics.py",22291,0,"",python,selection_mouse +4957,14214125,"jasmine/train_dynamics.py",22111,0,"",python,selection_mouse +4958,14214762,"jasmine/train_dynamics.py",22090,53," # Do not skip the first batch during training",python,selection_command +4959,14214945,"jasmine/train_dynamics.py",22025,118," # print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training",python,selection_command +4960,14215090,"jasmine/train_dynamics.py",21959,184," # print_compiled_memory_stats(compiled.memory_analysis())\n # print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training",python,selection_command +4961,14215215,"jasmine/train_dynamics.py",21887,256," # compiled = train_step.lower(optimizer, first_batch).compile()\n # print_compiled_memory_stats(compiled.memory_analysis())\n # print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training",python,selection_command +4962,14215364,"jasmine/train_dynamics.py",21838,305," first_batch[""rng""] = rng # type: ignore\n # compiled = train_step.lower(optimizer, first_batch).compile()\n # print_compiled_memory_stats(compiled.memory_analysis())\n # print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training",python,selection_command +4963,14215694,"jasmine/train_dynamics.py",21887,256," # compiled = train_step.lower(optimizer, first_batch).compile()\n # print_compiled_memory_stats(compiled.memory_analysis())\n # print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training",python,selection_command +4964,14215909,"jasmine/train_dynamics.py",21895,0,"",python,selection_command +4965,14216357,"jasmine/train_dynamics.py",22098,1,"",python,content +4966,14216357,"jasmine/train_dynamics.py",22033,1,"",python,content +4967,14216357,"jasmine/train_dynamics.py",21967,1,"",python,content +4968,14216357,"jasmine/train_dynamics.py",21895,1,"",python,content +4969,14216508,"jasmine/train_dynamics.py",22095,1,"",python,content +4970,14216509,"jasmine/train_dynamics.py",22031,1,"",python,content +4971,14216509,"jasmine/train_dynamics.py",21966,1,"",python,content +4972,14216509,"jasmine/train_dynamics.py",21895,1,"",python,content +4973,14216693,"jasmine/train_dynamics.py",21894,0,"",python,selection_command +4974,14218693,"jasmine/train_dynamics.py",22091,0," #",python,content +4975,14218694,"jasmine/train_dynamics.py",22029,0,"# ",python,content +4976,14218705,"jasmine/train_dynamics.py",21964,0," #",python,content +4977,14218706,"jasmine/train_dynamics.py",21895,0,"# ",python,content +4978,14218717,"jasmine/train_dynamics.py",21895,0,"",python,selection_command +4979,14219877,"jasmine/train_dynamics.py",21887,71," # compiled = train_step.lower(optimizer, first_batch).compile()",python,selection_command +4980,14220194,"jasmine/train_dynamics.py",21887,137," # compiled = train_step.lower(optimizer, first_batch).compile()\n # print_compiled_memory_stats(compiled.memory_analysis())",python,selection_command +4981,14220341,"jasmine/train_dynamics.py",21887,202," # compiled = train_step.lower(optimizer, first_batch).compile()\n # print_compiled_memory_stats(compiled.memory_analysis())\n # print_compiled_cost_analysis(compiled.cost_analysis())",python,selection_command +4982,14220777,"jasmine/train_dynamics.py",21895,0,"",python,selection_command +4983,14221227,"jasmine/train_dynamics.py",22033,1,"",python,content +4984,14221228,"jasmine/train_dynamics.py",21967,1,"",python,content +4985,14221228,"jasmine/train_dynamics.py",21895,1,"",python,content +4986,14221424,"jasmine/train_dynamics.py",22031,1,"",python,content +4987,14221424,"jasmine/train_dynamics.py",21966,1,"",python,content +4988,14221424,"jasmine/train_dynamics.py",21895,1,"",python,content +4989,14221594,"jasmine/train_dynamics.py",21894,0,"",python,selection_command +4990,14223989,"jasmine/train_dynamics.py",22105,0,"",python,selection_mouse +4991,14228701,"TERMINAL",0,0,"\r[?1l>]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +4992,14229531,"TERMINAL",0,0,"r",,terminal_output +4993,14229594,"TERMINAL",0,0,"u",,terminal_output +4994,14229705,"TERMINAL",0,0,"n",,terminal_output +4995,14229820,"TERMINAL",0,0,"n",,terminal_output +4996,14229884,"TERMINAL",0,0,"e",,terminal_output +4997,14229988,"TERMINAL",0,0,"r",,terminal_output +4998,14230251,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs]$ ",,terminal_output +4999,14236074,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",0,0,"",shellscript,tab +5000,14236181,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine_jobs]$ ",,terminal_output +5001,14245056,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2188,0,"",shellscript,selection_mouse +5002,14245660,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh",2192,0,"",shellscript,selection_mouse +5003,14250037,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine_jobs]$ ",,terminal_output +5004,14250947,"TERMINAL",0,0,"ru",,terminal_output +5005,14251065,"TERMINAL",0,0,"n",,terminal_output +5006,14251186,"TERMINAL",0,0,"n",,terminal_output +5007,14251246,"TERMINAL",0,0,"e",,terminal_output +5008,14251317,"TERMINAL",0,0,"r",,terminal_output +5009,14251478,"TERMINAL",0,0,"-",,terminal_output +5010,14251538,"TERMINAL",0,0,"2",,terminal_output +5011,14251882,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_2[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +5012,14252763,"TERMINAL",0,0,"v",,terminal_output +5013,14252828,"TERMINAL",0,0,"i",,terminal_output +5014,14252943,"TERMINAL",0,0,"m",,terminal_output +5015,14253006,"TERMINAL",0,0," ",,terminal_output +5016,14254760,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +5017,14254853,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ v\r\n\rim ",,terminal_output +5018,14256247,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh",0,0,"",shellscript,tab +5019,14257403,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh",0,0,"",shellscript,tab +5020,14259348,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",0,0,"",shellscript,tab +5021,14263400,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5022,14263652,"TERMINAL",0,0,"\rslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh\r\n[?2004l\r[?1049h[>4;2m[?1h=[?2004h[?1004h[?12h[?12l[?25lc]10;?]11;?#!/usr/bin/env bash#SBATCH --nodes=1#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=06:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scraatch/tum_ind3695-jafa_ws_shared/logs/logs_mmihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scrattch/tum_ind3695-jafa_ws_shared/logs/logs_miihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakkout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min beforee timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\n# source .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspaace/scratch/tum_ind3695-jafa_ws_shared/dataa_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspacee/scratch/tum_ind3695-jafa_ws_shared/data_bbreakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME1,1Top[?25h",,terminal_output +5023,14263707,"TERMINAL",0,0,"P+q436f\P+q6b75\P+q6b64\P+q6b72\P+q6b6c\P+q2332\P+q2334\P+q2569\P+q2a37\P+q6b31\[?12$p[?25l/3333/3333 [?25h[?25l/f6f6/e3e3 [?25h",,terminal_output +5024,14265206,"TERMINAL",0,0,"[?25l#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=06:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihirr/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir//breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\n# source .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breeakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakkout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakkout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR1,1Top[?25h",,terminal_output +5025,14266143,"TERMINAL",0,0,"[?25lj 2,0-1[?25h",,terminal_output +5026,14267840,"TERMINAL",0,0,"[?25lj 3,1 [?25h[?25lj 4[?25h[?25lj 5[?25h[?25lj 6[?25h[?25lj 7[?25h[?25lj 8[?25h[?25lj 9[?25h[?25lj 10,1[?25h[?25lj 1[?25h[?25lj 2[?25h[?25lj 3[?25h[?25lj 4,0-1[?25h[?25lj 5,1 [?25h[?25lj 6[?25h[?25lj 7,0-1[?25h[?25lj 8,1 [?25h[?25lj 9[?25h[?25lj 20[?25h[?25lj 1,0-1[?25h[?25lj 2,1 [?25h[?25lj 3[?25h[?25lj 4,0-1[?25h[?25lj 5,1 [?25h[?25lj 6[?25h[?25lj # lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breeakout/lam/interactive/351257627,0-15%[?25h[?25lj # tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoinnts/breakout/tokenizer/interactive/351250228,110%[?25h[?25lj 9[?25h[?25lj \r\n30,0-113%[?25h[?25lj lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakkout/lam/interactive/351896331,118%[?25h[?25lj tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpointss/breakout/tokenizer/interactive/351896332,123%[?25h[?25lj 3,0-1[?25h[?25lj env | grep SLURM34,127%[?25h[?25lj srun python jasmine/train_dynamics.py \\r\n --save_ckpt \35,132%[?25h[?25lj \r\n--image_height=10 \36,0-135%[?25h[?25lj \r\n--image_width=10 \37,137%[?25h[?25lj \r\n--ckpt_dir $CHECKPOINT_DIR \38,0-140%[?25h[?25lj \r\n--batch_size=120 \39,143%[?25h[?25lj \r\n--init_lr=0 \40,145%[?25h[?25lj \r\n--max_lr=3e-5 \41,148%[?25h[?25lj \r\n--log_image_interval=1000 \42,151%[?25h[?25lj \r\n--log_checkpoint_interval=1000 \43,154%[?25h[?25lj \r\n--dyna_type=maskgit \44,156%[?25h",,terminal_output +5027,14268084,"TERMINAL",0,0,"[?25lj --log \\r\n --name=breakout-dyn-50ksteps-3e-5-noise-lvl-$slurm_job_id \45,161%[?25h",,terminal_output +5028,14268300,"TERMINAL",0,0,"[?25lj 6[?25h",,terminal_output +5029,14268885,"TERMINAL",0,0,"[?25l--tags dyn breakout 50ksteps 3e-5 noise-lvl \\r\n --entity instant-uv \47,165%[?25h[?25lj 8[?25h[?25lj \r\n--project jafar \49,168%[?25h[?25lj \r\n--patch_size 4 \50,171%[?25h[?25lj \r\n--lam_patch_size 4 \51,174%[?25h",,terminal_output +5030,14269038,"TERMINAL",0,0,"[?25lj \r\n--warmup_steps 1000 \52,177%[?25h",,terminal_output +5031,14269203,"TERMINAL",0,0,"[?25lj --wsd_decay_steps 10000 \\r\n --num_steps 50000 \53,182%[?25h",,terminal_output +5032,14269345,"TERMINAL",0,0,"[?25lj 4[?25h",,terminal_output +5033,14269515,"TERMINAL",0,0,"[?25lj \r\n--data_dir $array_records_dir_train \55,185%[?25h",,terminal_output +5034,14269621,"TERMINAL",0,0,"[?25lj \r\n--val_data_dir $array_records_dir_val \56,188%[?25h",,terminal_output +5035,14269814,"TERMINAL",0,0,"[?25lk 5[?25h",,terminal_output +5036,14270005,"TERMINAL",0,0,"[?25lk 4[?25h",,terminal_output +5037,14270150,"TERMINAL",0,0,"[?25lk 3[?25h",,terminal_output +5038,14270290,"TERMINAL",0,0,"[?25lk 2[?25h",,terminal_output +5039,14270437,"TERMINAL",0,0,"[?25lk 1[?25h",,terminal_output +5040,14270768,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +5041,14271307,"TERMINAL",0,0,"[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h",,terminal_output +5042,14271677,"TERMINAL",0,0,"[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 10[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h",,terminal_output +5043,14271734,"TERMINAL",0,0,"[?25ll 8[?25h",,terminal_output +5044,14271997,"TERMINAL",0,0,"[?25ll 9[?25h",,terminal_output +5045,14272309,"TERMINAL",0,0,"[?25ll 20[?25h",,terminal_output +5046,14272451,"TERMINAL",0,0,"[?25ll 1[?25h",,terminal_output +5047,14272735,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +5048,14272883,"TERMINAL",0,0,"[?25ll 3[?25h",,terminal_output +5049,14273022,"TERMINAL",0,0,"[?25ll 4[?25h",,terminal_output +5050,14273174,"TERMINAL",0,0,"[?25ll 5[?25h",,terminal_output +5051,14273843,"TERMINAL",0,0,"[?25ll 6[?25h",,terminal_output +5052,14274448,"TERMINAL",0,0,"[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 30[?25h[?25ll 1[?25h[?25ll 2[?25h",,terminal_output +5053,14274510,"TERMINAL",0,0,"[?25ll 3[?25h",,terminal_output +5054,14274934,"TERMINAL",0,0,"[?25ll 4[?25h",,terminal_output +5055,14275168,"TERMINAL",0,0,"[?25ll 5[?25h",,terminal_output +5056,14275771,"TERMINAL",0,0,"[?25ll 6[?25h",,terminal_output +5057,14275925,"TERMINAL",0,0,"[?25ll 7[?25h",,terminal_output +5058,14276085,"TERMINAL",0,0,"[?25ll 8[?25h",,terminal_output +5059,14276221,"TERMINAL",0,0,"[?25ll 9[?25h",,terminal_output +5060,14276549,"TERMINAL",0,0,"[?25ld[?25h[?25lw -lvl-$slurm_job_id \[?25h",,terminal_output +5061,14276877,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +5062,14276939,"TERMINAL",0,0,"[?25lw lvl-$slurm_job_id \[?25h",,terminal_output +5063,14277283,"TERMINAL",0,0,"[?25ld[?25h[?25lw -$slurm_job_id \[?25h",,terminal_output +5064,14277590,"TERMINAL",0,0,"[?25li -- INSERT --51,3988%[?25h",,terminal_output +5065,14277845,"TERMINAL",0,0,"[?25ln-$slurm_job_id \40[?25h",,terminal_output +5066,14277999,"TERMINAL",0,0,"[?25la-$slurm_job_id \1[?25h",,terminal_output +5067,14278227,"TERMINAL",0,0,"[?25ln-$slurm_job_id \2[?25h",,terminal_output +5068,14278724,"TERMINAL",0,0,"[?25l--$slurm_job_id \3[?25h",,terminal_output +5069,14278843,"TERMINAL",0,0,"[?25ld-$slurm_job_id \4[?25h",,terminal_output +5070,14279061,"TERMINAL",0,0,"[?25le-$slurm_job_id \5[?25h",,terminal_output +5071,14279131,"TERMINAL",0,0,"[?25lb-$slurm_job_id \6[?25h",,terminal_output +5072,14279278,"TERMINAL",0,0,"[?25lu-$slurm_job_id \7[?25h",,terminal_output +5073,14279387,"TERMINAL",0,0,"[?25lg-$slurm_job_id \8[?25h",,terminal_output +5074,14279829,"TERMINAL",0,0,"[?25l^[",,terminal_output +5075,14279945,"TERMINAL",0,0," 51,4788%[?25h[?25lj 2[?25h",,terminal_output +5076,14280392,"TERMINAL",0,0,"[?25ll 8[?25h",,terminal_output +5077,14280676,"TERMINAL",0,0,"[?25li -- INSERT --52,4888%[?25h",,terminal_output +5078,14280989,"TERMINAL",0,0,"[?25l \5[?25h",,terminal_output +5079,14281382,"TERMINAL",0,0,"[?25l \4[?25h",,terminal_output +5080,14281899,"TERMINAL",0,0,"[?25l \39[?25h",,terminal_output +5081,14282437,"TERMINAL",0,0,"[?25ln \40[?25h",,terminal_output +5082,14282495,"TERMINAL",0,0,"[?25la \1[?25h",,terminal_output +5083,14282580,"TERMINAL",0,0,"[?25ln \2[?25h",,terminal_output +5084,14282832,"TERMINAL",0,0,"[?25l- \3[?25h",,terminal_output +5085,14283045,"TERMINAL",0,0,"[?25ld \4[?25h",,terminal_output +5086,14283243,"TERMINAL",0,0,"[?25le \5[?25h",,terminal_output +5087,14283341,"TERMINAL",0,0,"[?25lb \6[?25h",,terminal_output +5088,14283402,"TERMINAL",0,0,"[?25lu \7[?25h",,terminal_output +5089,14283464,"TERMINAL",0,0,"[?25lg \8[?25h",,terminal_output +5090,14283879,"TERMINAL",0,0,"[?25l^[",,terminal_output +5091,14283985,"TERMINAL",0,0," 52,4788%[?25h",,terminal_output +5092,14287148,"TERMINAL",0,0,"[?25l::[?25h",,terminal_output +5093,14287314,"TERMINAL",0,0,"w",,terminal_output +5094,14287762,"TERMINAL",0,0,"q",,terminal_output +5095,14288010,"TERMINAL",0,0,"\r[?25l[?2004l[>4;m[?25h[>4;m[?1049l]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_2[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +5096,14289675,"TERMINAL",0,0,"vim slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5097,14290133,"TERMINAL",0,0,"",,terminal_output +5098,14291040,"TERMINAL",0,0,"\r",,terminal_output +5099,14291443,"TERMINAL",0,0,"",,terminal_output +5100,14291879,"TERMINAL",0,0," slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5101,14292038,"TERMINAL",0,0," slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5102,14292175,"TERMINAL",0,0," slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5103,14292375,"TERMINAL",0,0,"s slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5104,14292524,"TERMINAL",0,0,"b slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5105,14292587,"TERMINAL",0,0,"a slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5106,14292717,"TERMINAL",0,0,"t slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5107,14292858,"TERMINAL",0,0,"c slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5108,14292920,"TERMINAL",0,0,"h slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5109,14295542,"TERMINAL",0,0,"runner-2\r\n\r",,terminal_output +5110,14296363,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5111,14296563,"TERMINAL",0,0,"\r\n\r",,terminal_output +5112,14297180,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5113,14297628,"TERMINAL",0,0,"",,terminal_output +5114,14297795,"TERMINAL",0,0,"",,terminal_output +5115,14298332,"TERMINAL",0,0,"",,terminal_output +5116,14298457,"TERMINAL",0,0,"",,terminal_output +5117,14298697,"TERMINAL",0,0,"\r\r",,terminal_output +5118,14299101,"TERMINAL",0,0,"",,terminal_output +5119,14299670,"TERMINAL",0,0," slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5120,14299829,"TERMINAL",0,0,"",,terminal_output +5121,14300295,"TERMINAL",0,0,"v slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5122,14300347,"TERMINAL",0,0,"i slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5123,14300443,"TERMINAL",0,0,"m slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5124,14300669,"TERMINAL",0,0,"\r\n\r\r\n[?2004l\r[?1049h[>4;2m[?1h=[?2004h[?1004h[?12h[?12l[?25lc]10;?]11;?lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakkout/lam/interactive/3518963tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpointss/breakout/tokenizer/interactive/3518963\r\n\r\nenv | grep SLURM\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=1000 \\r\n --log_checkpoint_interval=1000 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-50ksteps-3e-5-nan-debug-$slurm_job_id \\r\n --tags dyn breakout 50ksteps 3e-5 nan-debug \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 1000 \\r\n --wsd_decay_steps 10000 \\r\n --num_steps 50000 \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --lam_checkpoint $lam_checkpoint \\r\n --val_interval 1000 \\r\n --eval_full_frame \52,47Bot[?25hP+q436f\P+q6b75\P+q6b64\P+q6b72\P+q6b6c\P+q2332\P+q2334\P+q2569\P+q2a37\P+q6b31\[?12$p[?25l/3333/3333 [?25h[?25l/f6f6/e3e3 [?25h",,terminal_output +5125,14301272,"TERMINAL",0,0,"[?25lk 1[?25h",,terminal_output +5126,14302525,"TERMINAL",0,0,"[?25lk 0,11[?25h[?25lk 49,25[?25h[?25lk 8,36[?25h[?25lk 7,31[?25h[?25lk 6,19[?25h[?25lk 5,17[?25h[?25lk 4,22[?25h[?25lk 3,3[?25h[?25lk 2,2[?25h[?25lk 1,23[?25h[?25lk 0,17[?25h[?25lk 39,39[?25h[?25lk 8,0-1[?25h[?25lk 7,16 [?25h[?25lk 6,0-1[?25h[?25lk # tokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoinnts/breakout/tokenizer/interactive/351250235,4793%[?25h[?25lk # lam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breeakout/lam/interactive/351257634,4788%[?25h[?25lk 33,0-185%[?25h[?25lk CHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakkout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR32,4777%[?25h[?25lk 31,4774%[?25h[?25lk slurm_job_id=$SLURM_JOB_ID30,0-171%[?25h[?25lk job_name=$SLURM_JOB_NAME29,2468%[?25h[?25lk array_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakkout/breakout_episodes_perfect/val28,4761%[?25h[?25lk 7,0-1[?25h[?25lk array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breeakout/breakout_episodes_perfect/train26,2656%[?25h[?25lk 5,24[?25h[?25lk 24,0-154%[?25h",,terminal_output +5127,14302994,"TERMINAL",0,0,"[?25lk module unload devel/cuda/12.4\r\n# source .venv/bin/activate23,4748%[?25h[?25lk module unload mpi/openmpi/5.022,4743%[?25h[?25lk cat $021,0-140%[?25h[?25lk # Log the sbatch script20,2737%[?25h[?25lk 19,2935%[?25h[?25lk #SBATCH --signal=b:usr1@300 # 5 min before timeout18,2932%[?25h[?25lk #SBATCH --requeue17,0-129%[?25h[?25lk #SBATCH --job-name=train_dyn_default_breakout_longer16,627%[?25h[?25lk #SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir//breakout/dyn/%x_%j.log15,2323%[?25h[?25lk 4,0-1[?25h[?25lk #SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihirr/breakout/dyn/%x_%j.log@@@ 13,4720%[?25h[?25lk 2,1[?25h[?25lk #SBATCH --gres=gpu:111,4718%[?25h[?25lk #SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=510,4713%[?25h[?25lk #SBATCH --ntasks-per-node=1\r\n#SBATCH --time=06:00:00@@@ 9,478%[?25h[?25lk #SBATCH --nodes=18,205%[?25h",,terminal_output +5128,14303382,"TERMINAL",0,0,"[?25lj 9,47[?25h",,terminal_output +5129,14303794,"TERMINAL",0,0,"[?25lk 8,20[?25h",,terminal_output +5130,14304386,"TERMINAL",0,0,"[?25lk @@@ 7,252%[?25h",,terminal_output +5131,14304566,"TERMINAL",0,0,"[?25lk #!/usr/bin/env bash6,31Top[?25h",,terminal_output +5132,14304949,"TERMINAL",0,0,"[?25lk 5,23[?25h",,terminal_output +5133,14305453,"TERMINAL",0,0,"[?25lh 2[?25h",,terminal_output +5134,14305960,"TERMINAL",0,0,"[?25lh 1[?25h[?25lh 0[?25h[?25lh 19[?25h",,terminal_output +5135,14306022,"TERMINAL",0,0,"[?25lh 8[?25h",,terminal_output +5136,14306085,"TERMINAL",0,0,"[?25lh 7[?25h",,terminal_output +5137,14306274,"TERMINAL",0,0,"[?25lh 6[?25h",,terminal_output +5138,14306495,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +5139,14306559,"TERMINAL",0,0,"[?25lw :00:00[?25h",,terminal_output +5140,14306680,"TERMINAL",0,0,"[?25li -- INSERT --5,16Top[?25h",,terminal_output +5141,14308852,"TERMINAL",0,0,"[?25l4:00:007[?25h",,terminal_output +5142,14308905,"TERMINAL",0,0,"[?25l8:00:008[?25h",,terminal_output +5143,14309350,"TERMINAL",0,0,"[?25l^[",,terminal_output +5144,14309458,"TERMINAL",0,0," 5,17Top[?25h",,terminal_output +5145,14309715,"TERMINAL",0,0,"[?25l::[?25h",,terminal_output +5146,14309780,"TERMINAL",0,0,"W",,terminal_output +5147,14309904,"TERMINAL",0,0,"q",,terminal_output +5148,14310127,"TERMINAL",0,0,"\r[?25lE492: Not an editor command: Wq5,17Top[?25h",,terminal_output +5149,14310801,"TERMINAL",0,0,"[?25l::[?25h",,terminal_output +5150,14311333,"TERMINAL",0,0,"w",,terminal_output +5151,14311483,"TERMINAL",0,0,"q",,terminal_output +5152,14311829,"TERMINAL",0,0,"\r[?25l[?2004l[>4;m[?25h[>4;m[?1049l]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_2[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +5153,14312915,"TERMINAL",0,0,"vim slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5154,14313322,"TERMINAL",0,0,"",,terminal_output +5155,14313889,"TERMINAL",0,0,"",,terminal_output +5156,14314233,"TERMINAL",0,0,"\r",,terminal_output +5157,14314707,"TERMINAL",0,0,"im slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5158,14314844,"TERMINAL",0,0,"m slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5159,14315090,"TERMINAL",0,0," slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5160,14315386,"TERMINAL",0,0,"s slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5161,14315517,"TERMINAL",0,0,"b slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5162,14315580,"TERMINAL",0,0,"a slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5163,14315709,"TERMINAL",0,0,"t slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5164,14315840,"TERMINAL",0,0,"c slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.shh slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5165,14316125,"TERMINAL",0,0,"\r\n\r\r\n[?2004l\rSubmitted batch job 3527904\r\n]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_2[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +5166,14318692,"TERMINAL",0,0,"qu",,terminal_output +5167,14318806,"TERMINAL",0,0,"e",,terminal_output +5168,14318859,"TERMINAL",0,0,"u",,terminal_output +5169,14319456,"TERMINAL",0,0,"e",,terminal_output +5170,14319622,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: squeue --mehkn0402.localdomain: Sun Sep 28 17:26:04 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3527904 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3527891 accelerat train_dy tum_cte0 R\t3:26\t 1 hkn07303527866 dev_accel interact tum_cte0 R37:07\t 1 hkn0402",,terminal_output +5171,14320677,"TERMINAL",0,0,"578",,terminal_output +5172,14321191,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_2[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +5173,14322872,"TERMINAL",0,0,"r",,terminal_output +5174,14322933,"TERMINAL",0,0,"u",,terminal_output +5175,14323197,"TERMINAL",0,0,"n",,terminal_output +5176,14323506,"TERMINAL",0,0,"n",,terminal_output +5177,14323568,"TERMINAL",0,0,"e",,terminal_output +5178,14323681,"TERMINAL",0,0,"r",,terminal_output +5179,14323855,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs]$ ",,terminal_output +5180,14324981,"TERMINAL",0,0,"s",,terminal_output +5181,14325218,"TERMINAL",0,0,"y",,terminal_output +5182,14325280,"TERMINAL",0,0,"n",,terminal_output +5183,14325638,"TERMINAL",0,0,"c",,terminal_output +5184,14325759,"TERMINAL",0,0,"-",,terminal_output +5185,14326434,"TERMINAL",0,0,"r",,terminal_output +5186,14326496,"TERMINAL",0,0,"u",,terminal_output +5187,14326610,"TERMINAL",0,0,"n",,terminal_output +5188,14326909,"TERMINAL",0,0,"n",,terminal_output +5189,14326966,"TERMINAL",0,0,"e",,terminal_output +5190,14327384,"TERMINAL",0,0,"r",,terminal_output +5191,14327915,"TERMINAL",0,0,"\r\n[?2004l\rsending incremental file list\r\n",,terminal_output +5192,14329143,"TERMINAL",0,0,"jasmine/genie.py\r\njasmine/train_dynamics.py\r\njasmine/models/dynamics.py\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh\r\n",,terminal_output +5193,14329300,"TERMINAL",0,0,"\r\nsent 112,508 bytes received 355 bytes 75,242.00 bytes/sec\r\ntotal size is 28,033,742 speedup is 248.39\r\n]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs]$ ",,terminal_output +5194,14338458,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",0,0,"",shellscript,tab +5195,14338534,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine_jobs]$ ",,terminal_output +5196,14343129,"TERMINAL",0,0,"s\r\n\r",,terminal_output +5197,14343235,"TERMINAL",0,0,"b",,terminal_output +5198,14343297,"TERMINAL",0,0,"a",,terminal_output +5199,14343370,"TERMINAL",0,0,"t",,terminal_output +5200,14343476,"TERMINAL",0,0,"c",,terminal_output +5201,14343640,"TERMINAL",0,0," ",,terminal_output +5202,14344124,"TERMINAL",0,0,"",,terminal_output +5203,14344334,"TERMINAL",0,0,"h",,terminal_output +5204,14344412,"TERMINAL",0,0," ",,terminal_output +5205,14344662,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5206,14346908,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh\r\n[?2004l\rSubmitted batch job 3527905\r\n]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs]$ ",,terminal_output +5207,14348625,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine_jobs]$ ",,terminal_output +5208,14349461,"TERMINAL",0,0,"d",,terminal_output +5209,14349524,"TERMINAL",0,0,"e",,terminal_output +5210,14349702,"TERMINAL",0,0,"v",,terminal_output +5211,14349775,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +5212,14350182,"TERMINAL",0,0,"g",,terminal_output +5213,14350256,"TERMINAL",0,0,"i",,terminal_output +5214,14350320,"TERMINAL",0,0,"t",,terminal_output +5215,14350433,"TERMINAL",0,0," ",,terminal_output +5216,14350674,"TERMINAL",0,0,"d",,terminal_output +5217,14350788,"TERMINAL",0,0,"i",,terminal_output +5218,14350913,"TERMINAL",0,0,"f",,terminal_output +5219,14351057,"TERMINAL",0,0,"f",,terminal_output +5220,14351123,"TERMINAL",0,0,"\r\n[?2004l\r[?1h=\rdiff --git a/jasmine/genie.py b/jasmine/genie.py\r\nindex 061cebd..54da37e 100644\r\n--- a/jasmine/genie.py\r\n+++ b/jasmine/genie.py\r\n@@ -207,16 +207,20 @@ class Genie(nnx.Module):\r\n self,\r\n batch: Dict[str, jax.Array],\r\n seq_len: int,\r\n+ noise_level: float = 0.,\r\n temperature: float = 1,\r\n sample_argmax: bool = False,\r\n maskgit_steps: int = 25,\r\n ) -> tuple[jax.Array, jax.Array]:\r\n+ assert (\r\n+ noise_level < self.max_noise_level\r\n+ ), ""Noise level must me smaller than max_noise_level.""\r\n if self.dyna_type == ""maskgit"":\r\n return self.sample_maskgit(\r\n- batch, seq_len, 0.0, maskgit_steps, temperature, sample_argmax\r\n+ batch, seq_len, noise_level, maskgit_steps, temperature, sample_argmax\r\n )\r\n elif self.dyna_type == ""causal"":\r\n- return self.sample_causal(batch, seq_len, temperature, sample_argmax)\r\n+ return self.sample_causal(batch, seq_len, noise_level, temperature, sample_ argmax)\r\n else:\r\n raise ValueError(f""Dynamics model type unknown: {self.dyna_type}"")\r\n \r\n@@ -255,9 +259,6 @@ class Genie(nnx.Module):\r\n P: S * N\r\n """"""\r\n assert isinstance(self.dynamics, DynamicsMaskGIT)\r\n- assert (\r\n- noise_level < self.max_noise_level\r\n- ), ""Noise level must me smaller than max_noise_level.""\r\n:",,terminal_output +5221,14352347,"TERMINAL",0,0,"\r # --- Encode videos and actions ---\r\n:",,terminal_output +5222,14353088,"TERMINAL",0,0,"\r videos_BTHWC = batch[""videos""]\r\n:\r tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\r\n:\r@@ -328,19 +329,18 @@ class Genie(nnx.Module):\r\n:\r act_embed_BS1M = jnp.reshape(\r\n:\r act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\r\n:\r )\r\n:\r- # TODO mihir\r\n:\r \r\n:\r+ # TODO mihir\r\n:",,terminal_output +5223,14353728,"TERMINAL",0,0,"\r rng, _rng_noise = jax.random.split(rng)\r\n:\r noise_level_111 = noise_level.reshape(1, 1, 1)\r\n:\r noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\r\n:\r noise_bucket_idx_B11 = jnp.floor(\r\n:\r- (noise_level_B11 / self.max_noise_level) * self.noise_buckets\r\n:\r+ (noise_level_B11 * self.noise_buckets) / self.max_noise_level\r\n:\r ).astype(jnp.int32)\r\n:\r noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\r\n:\r noise_bucket_idx_B11\r\n:\r )\r\n:\r noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\r\n:\r- vid_embed_BSNM += jnp.expand_dims(noise_level_B11, -1)\r\n:\r \r\n:\r vid_embed_BSNp2M = jnp.concatenate(\r\n:\r [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\r\n:\r@@ -441,6 +441,7 @@ class Genie(nnx.Module):\r\n:\r self,\r\n:\r batch: Dict[str, jax.Array],\r\n:\r seq_len: int,\r\n:\r+ noise_level: float,\r\n:\r temperature: float = 1,\r\n:",,terminal_output +5224,14354491,"TERMINAL",0,0,"\r sample_argmax: bool = False,\r\n:",,terminal_output +5225,14355138,"TERMINAL",0,0,"\r ) -> tuple[jax.Array, jax.Array]:\r\n:\r@@ -528,12 +529,28 @@ class Genie(nnx.Module):\r\n:\r act_embed_BS1M = jnp.reshape(\r\n:\r act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\r\n:\r )\r\n:\r- vid_embed_BSNp1M = jnp.concatenate([act_embed_BS1M, vid_embed_BSNM], axis=2 :\r)\r\n:",,terminal_output +5226,14355195,"TERMINAL",0,0,"\r- final_logits_BTNp1V = (\r\n:",,terminal_output +5227,14355351,"TERMINAL",0,0,"\r- dynamics_causal.transformer(vid_embed_BSNp1M, (step_t, step_n))\r\n:",,terminal_output +5228,14356222,"TERMINAL",0,0,"\r+\r\n:\r+ # TODO mihir\r\n:\r+\r\n:\r+ rng, _rng_noise = jax.random.split(rng)\r\n:\r+ noise_level_111 = noise_level.reshape(1, 1, 1)\r\n:\r+ noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\r\n:\r+ noise_bucket_idx_B11 = jnp.floor(\r\n:\r+ (noise_level_B11 * self.noise_buckets) / self.max_noise_level\r\n:\r+ ).astype(jnp.int32)\r\n:\r+ noise_level_embed_B11M = dynamics_causal.noise_level_embed(\r\n:\r+ noise_bucket_idx_B11\r\n:\r+ )\r\n:\r+ noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\r\n:",,terminal_output +5229,14358062,"TERMINAL",0,0,"\r+\r\n:",,terminal_output +5230,14358890,"TERMINAL",0,0,"\r+ vid_embed_BSNp2M = jnp.concatenate(\r\n:\r+ [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\r\n:\r+ )\r\n:\r+ final_logits_BTNp2V = (\r\n:\r+ dynamics_causal.transformer(vid_embed_BSNp2M, (step_t, step_n))\r\n:\r / temperature\r\n:\r )\r\n:\r- final_logits_BV = final_logits_BTNp1V[:, step_t, step_n, :]\r\n:\r+ final_logits_BV = final_logits_BTNp2V[:, step_t, step_n + 1, :]\r\n:\r \r\n:\r # --- Sample new tokens for final frame ---\r\n:\r if sample_argmax:\r\n:\rdiff --git a/jasmine/models/dynamics.py b/jasmine/models/dynamics.py\r\n:\rindex 5ce4dcd..0a69649 100644\r\n:",,terminal_output +5231,14358946,"TERMINAL",0,0,"\r--- a/jasmine/models/dynamics.py\r\n:",,terminal_output +5232,14359271,"TERMINAL",0,0,"\r+++ b/jasmine/models/dynamics.py\r\n:",,terminal_output +5233,14360880,"TERMINAL",0,0,"\r@@ -78,6 +78,33 @@ class DynamicsMaskGIT(nnx.Module):\r\n:",,terminal_output +5234,14361306,"TERMINAL",0,0,"\rM )\r\n\r:",,terminal_output +5235,14362448,"TERMINAL",0,0,"\rM act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\r\n\r:\rM act_embed_BS1M = jnp.reshape(\r\n\r:\rM@@ -528,12 +529,28 @@ class Genie(nnx.Module):\r\n\r:\rM ) -> tuple[jax.Array, jax.Array]:\r\n\r:\rM sample_argmax: bool = False,\r\n\r:\rM temperature: float = 1,\r\n\r:\rM+ noise_level: float,\r\n\r:\rM seq_len: int,\r\n\r:\rM batch: Dict[str, jax.Array],\r\n\r:\rM self,\r\n\r:\rM@@ -441,6 +441,7 @@ class Genie(nnx.Module):\r\n\r:\rM [act_embed_BS1M, noise_level_embed_BS1M, vid_embed_BSNM], axis=2\r\n\r:\rM vid_embed_BSNp2M = jnp.concatenate(\r\n\r:\rM \r\n\r:\rM- vid_embed_BSNM += jnp.expand_dims(noise_level_B11, -1)\r\n\r:\rM noise_level_embed_BS1M = jnp.tile(noise_level_embed_B11M, (1, S, 1, 1))\r\n\r:\rM )\r\n\r:\rM noise_bucket_idx_B11\r\n\r:\rM noise_level_embed_B11M = dynamics_maskgit.noise_level_embed(\r\n\r:\rM ).astype(jnp.int32)\r\n\r:\rM+ (noise_level_B11 * self.noise_buckets) / self.max_noise_level\r\n\r:\rM- (noise_level_B11 / self.max_noise_level) * self.noise_buckets\r\n\r:\rM noise_bucket_idx_B11 = jnp.floor(\r\n\r:",,terminal_output +5236,14362722,"TERMINAL",0,0,"\rM noise_level_B11 = jnp.tile(noise_level_111, (B, 1, 1))\r\n\r:\rM noise_level_111 = noise_level.reshape(1, 1, 1)\r\n\r:\rM rng, _rng_noise = jax.random.split(rng)\r\n\r:\rM+ # TODO mihir\r\n\r:\rM \r\n\r:\rM- # TODO mihir\r\n\r:\rM )\r\n\r:\rM act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\r\n\r:\rM act_embed_BS1M = jnp.reshape(\r\n\r:\rM@@ -328,19 +329,18 @@ class Genie(nnx.Module):\r\n\r:",,terminal_output +5237,14362785,"TERMINAL",0,0,"\rM tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\r\n\r:",,terminal_output +5238,14363033,"TERMINAL",0,0,"\rM videos_BTHWC = batch[""videos""]\r\n\r:",,terminal_output +5239,14363672,"TERMINAL",0,0,"\rM # --- Encode videos and actions ---\r\n\r:\rM- ), ""Noise level must me smaller than max_noise_level.""\r\n\r:\rM- noise_level < self.max_noise_level\r\n\r:\rM- assert (\r\n\r:\rM assert isinstance(self.dynamics, DynamicsMaskGIT)\r\n\r:\rM """"""\r\n\r:\rM P: S * N\r\n\r:",,terminal_output +5240,14367394,"TERMINAL",0,0,"\r+ noise_level: float,\r\n:",,terminal_output +5241,14367936,"TERMINAL",0,0,"\r[?1l>]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +5242,14410644,"TERMINAL",0,0,"q",,terminal_output +5243,14410757,"TERMINAL",0,0,"u",,terminal_output +5244,14410817,"TERMINAL",0,0,"e",,terminal_output +5245,14410881,"TERMINAL",0,0,"u",,terminal_output +5246,14410985,"TERMINAL",0,0,"e",,terminal_output +5247,14411134,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: squeue --mehkn0402.localdomain: Sun Sep 28 17:27:35 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3527904 accelerat train_dy tum_cte0 R\t0:56\t 1 hkn04293527905 accelerat train_dy tum_cte0 R\t0:56\t 1 hkn05343527891 accelerat train_dy tum_cte0 R\t4:57\t 1 hkn07303527866 dev_accel interact tum_cte0 R38:38\t 1 hkn0402",,terminal_output +5248,14412132,"TERMINAL",0,0,"67789",,terminal_output +5249,14412955,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +5250,14487039,"jasmine/train_dynamics.py",0,0,"",python,tab +5251,14488349,"jasmine/train_dynamics.py",22075,0,"",python,selection_mouse +5252,14495336,"jasmine/models/dynamics.py",0,0,"",python,tab +5253,14504649,"jasmine/models/dynamics.py",3315,0,"",python,selection_mouse +5254,14513663,"jasmine/models/dynamics.py",3513,0,"",python,selection_mouse +5255,14514309,"jasmine/models/dynamics.py",3442,0,"",python,selection_mouse +5256,14514945,"jasmine/models/dynamics.py",3280,0,"",python,selection_mouse +5257,14515518,"jasmine/models/dynamics.py",3215,0,"",python,selection_mouse +5258,14516677,"jasmine/models/dynamics.py",2996,0,"",python,selection_mouse +5259,14516834,"jasmine/models/dynamics.py",2985,18,"noise_bucket_idx_B",python,selection_mouse +5260,14539054,"jasmine/models/dynamics.py",3324,0,"",python,selection_mouse +5261,14539605,"jasmine/models/dynamics.py",3332,0,"",python,selection_mouse +5262,14540163,"jasmine/models/dynamics.py",3325,16,"noise_level_B111",python,selection_mouse +5263,14543106,"jasmine/models/dynamics.py",2835,0,"",python,selection_mouse +5264,14543773,"jasmine/models/dynamics.py",2836,0,"",python,selection_mouse +5265,14543920,"jasmine/models/dynamics.py",2822,18,"noise_bucket_idx_B",python,selection_mouse +5266,14544682,"jasmine/models/dynamics.py",2953,0,"",python,selection_mouse +5267,14545923,"jasmine/models/dynamics.py",2953,0,"\n ",python,content +5268,14546431,"jasmine/models/dynamics.py",2962,0,"\n ",python,content +5269,14546431,"jasmine/models/dynamics.py",2954,8,"",python,content +5270,14547029,"jasmine/models/dynamics.py",2954,0,"",python,selection_command +5271,14547715,"jasmine/models/dynamics.py",2955,8,"",python,content +5272,14547715,"jasmine/models/dynamics.py",2954,0,"\n",python,content +5273,14548169,"jasmine/models/dynamics.py",2955,0," ",python,content +5274,14548604,"jasmine/models/dynamics.py",2959,0," ",python,content +5275,14567164,"jasmine/models/dynamics.py",2955,0,"",python,selection_command +5276,14569532,"jasmine/models/dynamics.py",2955,0," noise_bucket_idx_B = jnp.clip(noise_bucket_idx_B, 0, self.noise_buckets - 1)\n",python,content +5277,14569537,"jasmine/models/dynamics.py",3040,9,"",python,content +5278,14617060,"jasmine/models/dynamics.py",3249,0,"",python,selection_mouse +5279,14618209,"jasmine/models/dynamics.py",2953,0,"",python,selection_mouse +5280,14618886,"jasmine/models/dynamics.py",2954,0,"",python,selection_mouse +5281,14621399,"jasmine/models/dynamics.py",2431,0,"",python,selection_mouse +5282,14621804,"jasmine/models/dynamics.py",2430,0,"",python,selection_command +5283,14622452,"jasmine/models/dynamics.py",2413,19,"",python,content +5284,14622468,"jasmine/models/dynamics.py",2421,0,"",python,selection_command +5285,14622595,"jasmine/models/dynamics.py",2465,0,"",python,selection_command +5286,14622925,"jasmine/models/dynamics.py",2542,0,"",python,selection_command +5287,14623490,"jasmine/models/dynamics.py",2541,1,"",python,content +5288,14623637,"jasmine/models/dynamics.py",2540,1,"",python,content +5289,14623787,"jasmine/models/dynamics.py",2539,1,"",python,content +5290,14623887,"jasmine/models/dynamics.py",2538,1,"",python,content +5291,14624036,"jasmine/models/dynamics.py",2537,1,"",python,content +5292,14624174,"jasmine/models/dynamics.py",2536,1,"",python,content +5293,14625550,"jasmine/models/dynamics.py",2964,0,"",python,selection_mouse +5294,14626489,"jasmine/models/dynamics.py",2955,0,"",python,selection_mouse +5295,14627231,"jasmine/models/dynamics.py",2959,0,"",python,selection_mouse +5296,14627812,"jasmine/models/dynamics.py",2960,0,"",python,selection_mouse +5297,14628481,"jasmine/models/dynamics.py",2969,0,"",python,selection_mouse +5298,14629799,"jasmine/models/dynamics.py",2993,0,"",python,selection_mouse +5299,14630960,"jasmine/models/dynamics.py",2951,0,"",python,selection_mouse +5300,14631116,"jasmine/models/dynamics.py",2938,18,"noise_bucket_idx_B",python,selection_mouse +5301,14635473,"jasmine/models/dynamics.py",3015,0,"",python,selection_mouse +5302,14636020,"jasmine/models/dynamics.py",2952,0,"",python,selection_mouse +5303,14636157,"jasmine/models/dynamics.py",2938,18,"noise_bucket_idx_B",python,selection_mouse +5304,14657621,"jasmine/models/dynamics.py",3138,0,"",python,selection_mouse +5305,14661352,"jasmine/models/dynamics.py",3406,0,"",python,selection_mouse +5306,14662155,"jasmine/models/dynamics.py",3405,1,"",python,content +5307,14662369,"jasmine/models/dynamics.py",3405,1,"",python,content +5308,14663086,"jasmine/models/dynamics.py",3464,0,"",python,selection_command +5309,14663261,"jasmine/models/dynamics.py",3515,0,"",python,selection_command +5310,14665533,"jasmine/models/dynamics.py",3522,1,"",python,content +5311,14665685,"jasmine/models/dynamics.py",3522,1,"",python,content +5312,14667884,"jasmine/models/dynamics.py",3301,0,"",python,selection_mouse +5313,14683424,"jasmine/models/dynamics.py",3298,0,"",python,selection_mouse +5314,14688434,"jasmine/models/dynamics.py",4006,0,"",python,selection_mouse +5315,14698057,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",0,0,"",shellscript,tab +5316,14704566,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh",1627,0,"",shellscript,selection_mouse +5317,14709228,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",0,0,"",shellscript,tab +5318,14711937,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2014,0,"",shellscript,selection_mouse +5319,14712459,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1950,0,"",shellscript,selection_mouse +5320,14713369,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1950,1,"-",shellscript,selection_command +5321,14713781,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1950,1,"-",shellscript,selection_command +5322,14714049,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1950,0,"",shellscript,selection_command +5323,14714871,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2014,0,"-",shellscript,content +5324,14714872,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1950,0,"-",shellscript,content +5325,14714873,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1951,0,"",shellscript,selection_keyboard +5326,14715135,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2016,0,"b",shellscript,content +5327,14715135,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1951,0,"b",shellscript,content +5328,14715136,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1952,0,"",shellscript,selection_keyboard +5329,14715353,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2018,0,"u",shellscript,content +5330,14715354,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1952,0,"u",shellscript,content +5331,14715354,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1953,0,"",shellscript,selection_keyboard +5332,14715371,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2020,0,"c",shellscript,content +5333,14715371,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1953,0,"c",shellscript,content +5334,14715372,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1954,0,"",shellscript,selection_keyboard +5335,14715558,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2022,0,"k",shellscript,content +5336,14715559,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1954,0,"k",shellscript,content +5337,14715559,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1955,0,"",shellscript,selection_keyboard +5338,14715644,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2024,0,"e",shellscript,content +5339,14715644,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1955,0,"e",shellscript,content +5340,14715645,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1956,0,"",shellscript,selection_keyboard +5341,14715904,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2026,0,"t",shellscript,content +5342,14715904,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1956,0,"t",shellscript,content +5343,14715905,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1957,0,"",shellscript,selection_keyboard +5344,14716004,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2028,0,"-",shellscript,content +5345,14716004,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1957,0,"-",shellscript,content +5346,14716004,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1958,0,"",shellscript,selection_keyboard +5347,14716583,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2030,0,"c",shellscript,content +5348,14716586,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1958,0,"c",shellscript,content +5349,14716586,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1959,0,"",shellscript,selection_keyboard +5350,14716853,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2032,0,"l",shellscript,content +5351,14716854,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1959,0,"l",shellscript,content +5352,14716854,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1960,0,"",shellscript,selection_keyboard +5353,14717039,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2034,0,"i",shellscript,content +5354,14717039,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1960,0,"i",shellscript,content +5355,14717039,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1961,0,"",shellscript,selection_keyboard +5356,14717190,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2036,0,"p",shellscript,content +5357,14717191,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1961,0,"p",shellscript,content +5358,14717191,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1962,0,"",shellscript,selection_keyboard +5359,14717455,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1961,0,"",shellscript,selection_command +5360,14717702,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1960,0,"",shellscript,selection_command +5361,14718203,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1959,0,"",shellscript,selection_command +5362,14718217,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1958,0,"",shellscript,selection_command +5363,14718287,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1957,0,"",shellscript,selection_command +5364,14718287,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1956,0,"",shellscript,selection_command +5365,14718322,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1955,0,"",shellscript,selection_command +5366,14718501,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1954,0,"",shellscript,selection_command +5367,14718635,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1953,0,"",shellscript,selection_command +5368,14718781,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1952,0,"",shellscript,selection_command +5369,14718926,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1951,0,"",shellscript,selection_command +5370,14719157,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1950,0,"",shellscript,selection_command +5371,14719351,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2026,0,"",shellscript,selection_command +5372,14719495,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2026,1,"",shellscript,content +5373,14720265,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2026,0," ",shellscript,content +5374,14720266,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2027,0,"",shellscript,selection_keyboard +5375,14720634,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2026,0,"",shellscript,selection_command +5376,14722615,"jasmine/models/dynamics.py",0,0,"",python,tab +5377,14724099,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +5378,14732227,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",0,0,"",shellscript,tab +5379,14732262,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +5380,14742474,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +5381,14743236,"TERMINAL",0,0,"s",,terminal_output +5382,14743345,"TERMINAL",0,0,"b",,terminal_output +5383,14743413,"TERMINAL",0,0,"a",,terminal_output +5384,14743517,"TERMINAL",0,0,"t",,terminal_output +5385,14743656,"TERMINAL",0,0,"c",,terminal_output +5386,14743980,"TERMINAL",0,0,"",,terminal_output +5387,14744132,"TERMINAL",0,0,"",,terminal_output +5388,14744267,"TERMINAL",0,0,"",,terminal_output +5389,14744804,"TERMINAL",0,0,"s",,terminal_output +5390,14745069,"TERMINAL",0,0,"y",,terminal_output +5391,14745213,"TERMINAL",0,0,"n",,terminal_output +5392,14745430,"TERMINAL",0,0,"c",,terminal_output +5393,14745491,"TERMINAL",0,0,"-",,terminal_output +5394,14745671,"TERMINAL",0,0,"r",,terminal_output +5395,14745823,"TERMINAL",0,0,"u",,terminal_output +5396,14745884,"TERMINAL",0,0,"n",,terminal_output +5397,14746017,"TERMINAL",0,0,"n",,terminal_output +5398,14746153,"TERMINAL",0,0,"er",,terminal_output +5399,14746258,"TERMINAL",0,0,"-",,terminal_output +5400,14746590,"TERMINAL",0,0,"3",,terminal_output +5401,14747372,"TERMINAL",0,0,"\r\n[?2004l\rsending incremental file list\r\n",,terminal_output +5402,14747477,"TERMINAL",0,0,"./\r\n.gitignore\r\nREADME.md\r\ndiff.diff\r\ndiff2.diff\r\nkiller.sh\r\nkiller_partition.sh\r\nlog.log\r\npyproject.toml\r\ntest.py\r\nuv.lock\r\n",,terminal_output +5403,14748318,"TERMINAL",0,0,"input_pipeline/\r\ninput_pipeline/generate_breakout_dataset_agent.py\r\njasmine/\r\njasmine/__init__.py\r\njasmine/genie.py\r\njasmine/sample.py\r\njasmine/train_dynamics.py\r\njasmine/train_lam.py\r\njasmine/train_tokenizer.py\r\njasmine/models/\r\njasmine/models/dynamics.py\r\njasmine/models/lam.py\r\njasmine/models/tokenizer.py\r\njasmine/utils/\r\njasmine/utils/dataloader.py\r\njasmine/utils/nn.py\r\njasmine/utils/preprocess.py\r\njasmine/utils/train_utils.py\r\nmodels/\r\nslurm/dev/alfred/\r\nslurm/dev/alfred/berlin/\r\nslurm/dev/alfred/berlin/gt_actions/\r\nslurm/dev/alfred/berlin/gt_actions/sample_causal_32gpus.sbatch\r\nslurm/dev/alfred/berlin/gt_actions/sample_darkness_filter.sbatch\r\nslurm/dev/alfred/berlin/gt_actions/train_dynacmis_on_3nodes_2gpu.sbatch\r\nslurm/dev/alfred/berlin/gt_actions/train_dynacmis_on_8gpu.sbatch\r\nslurm/dev/alfred/berlin/gt_actions/train_dynacmis_on_more_than4.sbatch\r\nslurm/dev/alfred/berlin/gt_actions/train_dynacmis_overfit.sbatch\r\nslurm/dev/alfred/berlin/gt_actions/train_dynacmis_overfit_2gpus.sbatch\r\nslurm/dev/alfred/berlin/gt_actions/train_dynacmis_overfit_4gpu.sbatch\r\nslurm/dev/alfred/berlin/gt_actions/train_dynacmis_overfit_4gpu_to_8gpu.sbatch\r\nslurm/dev/alfred/berlin/gt_actions/train_tok_topology_one_gpu copy.sbatch\r\nslurm/dev/alfred/berlin/gt_actions/train_tok_topology_one_gpu.sbatch\r\nslurm/dev/alfred/berlin/gt_actions/train_tok_topology_restore_to_on_gpu.sbatch\r\nslurm/dev/alfred/berlin/gt_actions/train_tok_topology_two_gpus.sbatch\r\nslurm/dev/alfred/berlin/gt_sampling/\r\nslurm/dev/alfred/berlin/gt_sampling/sampling_mock.sbatch\r\nslurm/dev/alfred/berlin/job_requeueing/\r\nslurm/dev/alfred/berlin/job_requeueing/cpu_requeue_dev.sbatch\r\nslurm/dev/alfred/berlin/job_requeueing/train_lam_chain_dev.sbatch\r\nslurm/dev/alfred/berlin/job_requeueing/train_lam_requeue_dev.sbatch\r\nslurm/dev/alfred/berlin/job_requeueing/train_lam_requeue_dev_gemini.sbatch\r\nslurm/dev/alfred/berlin/job_requeueing/dynamic_lr_tuning/\r\nslurm/dev/alfred/berlin/job_requeueing/dynamic_lr_tuning/train_dynacmis_lr_general.sbatch\r\nslurm/dev/alfred/berlin/job_requeueing/dynamic_validation/\r\nslurm/dev/alfred/berlin/job_requeueing/dynamic_validation/train_dynacmis_lr_general.sbatch\r\nslurm/dev/alfred/berlin/job_requeueing/lam_lr_tuning/\r\nslurm/dev/alfred/berlin/job_requeueing/lam_lr_tuning/train_lam_lr_general.sbatch\r\nslurm/dev/alfred/berlin/job_requeueing/tokenizer_lr_tuning/\r\nslurm/dev/alfred/berlin/job_requeueing/tokenizer_lr_tuning/train_tokenizer_lr_general.sbatch\r\nslurm/dev/alfred/berlin/restore_ckpt/\r\nslurm/dev/alfred/berlin/restore_ckpt/restore_ckpt_single.sbatch\r\nslurm/dev/alfred/berlin/restore_ckpt/restore_multi.sbatch\r\nslurm/dev/alfred/berlin/restore_dynamics_non_cotraining/\r\nslurm/dev/alfred/berlin/restore_dynamics_non_cotraining/coinrun_dynamics_batch_size_144_3e-5_invest_nan_restore_40k.sbatch\r\nslurm/dev/alfred/berlin/test_franz_pr/\r\nslurm/dev/alfred/berlin/test_franz_pr/train_dynacmis_overfit.sbatch\r\nslurm/dev/alfred/berlin/test_franz_pr/train_lam_overfit.sbatch\r\nslurm/dev/alfred/berlin/test_franz_pr/train_tokenizer_overfit.sbatch\r\nslurm/dev/alfred/berlin/topology/\r\nslurm/dev/alfred/berlin/topology/sample_overfit_single_gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/train_dynacmis_overfit_1.sbatch\r\nslurm/dev/alfred/berlin/topology/train_dynacmis_overfit_1_gt_actions.sbatch\r\nslurm/dev/alfred/berlin/topology/train_dynacmis_overfit_1_gt_actions_noise.sbatch\r\nslurm/dev/alfred/berlin/topology/train_dynacmis_overfit_1_noise.sbatch\r\nslurm/dev/alfred/berlin/topology/train_dynacmis_overfit_2_nodes_2_gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/train_dynacmis_overfit_2_nodes_4_gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/train_dynacmis_overfit_4gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/train_lam_overfit.sbatch\r\nslurm/dev/alfred/berlin/topology/train_tokenizer_overfit_1gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/train_tokenizer_overfit_2_gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/train_tokenizer_overfit_4_gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/train_tokenizer_overfit_single_record.sbatch\r\nslurm/dev/alfred/berlin/topology/train_tokenizer_overfit_single_record_requeue.sbatch\r\nslurm/dev/alfred/berlin/topology/train_tokenizer_restore_1gpu_to_1gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/train_tokenizer_restore_1gpu_to_2gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/train_tokenizer_restore_2gpu_to_1gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/train_tokenizer_restore_2gpu_to_2gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/nnx/\r\nslurm/dev/alfred/berlin/topology/nnx/train_tokenizer_overfit_1.sbatch\r\nslurm/dev/alfred/berlin/topology/nnx/train_tokenizer_overfit_2_gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/nnx/train_tokenizer_restore_2gpu_to_1gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/nnx/train_tokenizer_restore_2gpu_to_2gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/prennx/\r\nslurm/dev/alfred/berlin/topology/prennx/train_tokenizer_overfit_2_gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/prennx/train_tokenizer_restore_2gpu_to_1gpu.sbatch\r\nslurm/dev/alfred/berlin/topology/prennx/train_tokenizer_restore_2gpu_to_2gpu.sbatch\r\nslurm/dev/alfred/berlin/topology_restore_fix/\r\nslurm/dev/alfred/berlin/topology_restore_fix/train_tokenizer_overfit_4_gpu.sbatch\r\nslurm/dev/alfred/berlin/topology_restore_fix/train_tokenizer_overfit_4_to_1_gpu.sbatch\r\nslurm/dev/alfred/berlin/train_dyn_dev/\r\nslurm/dev/alfred/berlin/train_dyn_dev/train_dynacmis_overfit.sbatch\r\nslurm/dev/alfred/berlin/train_lam_dev/\r\nslurm/dev/alfred/berlin/train_lam_dev/train_lam.sbatch\r\nslurm/dev/alfred/berlin/train_tok_dev/\r\nslurm/dev/alfred/berlin/train_tok_dev/train_tok.sbatch\r\nslurm/dev/alfred/berlin/train_tok_dev/train_tok_overfit.sbatch\r\nslurm/dev/alfred/berlin/validation/tokenizer_lr_tuning/\r\nslurm/dev/alfred/berlin/validation/tokenizer_lr_tuning/train_tokenizer_lr_general.sbatch\r\nslurm/dev/alfred/berlin/validation/tokenizer_without_optimizer/\r\nslurm/dev/alfred/berlin/validation/tokenizer_without_optimizer/train_tokenizer_lr_general.sbatch\r\nslurm/dev/alfred/helmholtz_cluster/\r\nslurm/dev/alfred/helmholtz_cluster/train_tok_dev/\r\nslurm/dev/alfred/helmholtz_cluster/train_tok_dev/train_tok.sbatch\r\nslurm/dev/alfred/helmholtz_cluster/train_tok_dev/train_tok_overfit.sbatch\r\nslurm/dev/alfred/horeka/\r\nslurm/dev/alfred/horeka/input_pipeline_ws/\r\nslurm/dev/alfred/horeka/input_pipeline_ws/actions/\r\nslurm/dev/alfred/horeka/input_pipeline_ws/actions/download_actions.sbatch\r\nslurm/dev/alfred/horeka/input_pipeline_ws/actions/download_actions_all.sh\r\nslurm/dev/alfred/horeka/jobs_cur/\r\nslurm/dev/alfred/horeka/jobs_cur/atari/\r\nslurm/dev/alfred/horeka/jobs_cur/atari/sample_causal.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/atari/sample_maskgit.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/atari/train_dynamics_causal.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/atari/train_dynamics_maskgit.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/atari/train_tokenizer_lr_1e-4.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/\r\nslurm/dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/dev_dyn_gt_actions.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/dev_dyn_gt_actions_dev_single.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/dyn_gt_actions.sh\r\nslurm/dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/dyn_gt_actions_ablation.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/dyn_gt_actions_ablation_baseline.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/dyn_gt_actions_ablation_one_node.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/dyn_gt_actions_ablation_single_gpu copy.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/dyn_gt_actions_ablation_single_gpu.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/dyn_gt_actions_maskgit_causal.sh\r\nslurm/dev/alfred/horeka/jobs_cur/tokenizer/\r\nslurm/dev/alfred/horeka/jobs_cur/tokenizer/train_tokenizer_37M_single_gpu.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/tokenizer/train_tokenizer_37M_single_node.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/\r\nslurm/dev/alfred/horeka/jobs_old/allocate/\r\nslurm/dev/alfred/horeka/jobs_old/allocate/cpu.sh\r\nslurm/dev/alfred/horeka/jobs_old/allocate/multigpu_gpu.sh\r\nslurm/dev/alfred/horeka/jobs_old/allocate/single_gpu.sh\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/adjusted_lr/\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/adjusted_lr/train_tokenizer_16_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/adjusted_lr/train_tokenizer_1_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/adjusted_lr/train_tokenizer_1_nodes.sh\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/adjusted_lr/train_tokenizer_2_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/adjusted_lr/train_tokenizer_32_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/adjusted_lr/train_tokenizer_4_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/adjusted_lr/train_tokenizer_4_nodes.sh\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/adjusted_lr/train_tokenizer_4_nodes_frequent_chkpt.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/adjusted_lr/train_tokenizer_64_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/adjusted_lr/train_tokenizer_8_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/const_lr/\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/const_lr/train_tokenizer_16_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/const_lr/train_tokenizer_1_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/const_lr/train_tokenizer_2_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/const_lr/train_tokenizer_32_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/const_lr/train_tokenizer_4_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/const_lr/train_tokenizer_8_nodes.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/oai_subset/\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/oai_subset/train_tokenizer_1_nodes.sh\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/oai_subset/train_tokenizer_2_nodes.sh\r\nslurm/dev/alfred/horeka/jobs_old/batchsize_scaling/oai_subset/train_tokenizer_2_nodes_samples_500.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/checkpoint_fix/\r\nslurm/dev/alfred/horeka/jobs_old/checkpoint_fix/train_tokenizer.sh\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/train_tokenizer_coinrun.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/base/\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/base/train_dynamics_coinrun.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/base/train_lam_coinrun.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/base/train_tokenizer_coinrun.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/latent_action_ablation/\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/latent_action_ablation/train_dynamics_coinrun.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/latent_action_ablation/train_dynamics_coinrun.sh\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/latent_action_ablation/train_lam_12.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/latent_action_ablation/train_lam_24.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/latent_action_ablation/train_lam_48.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/latent_action_ablation/train_lam_6.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/latent_action_ablation/train_lam_6.sh\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/latent_action_ablation/train_tokenizer_coinrun.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/coinrun/latent_action_ablation/train_tokenizer_coinrun.sh\r\nslurm/dev/alfred/horeka/jobs_old/dyn_gt_actions_ablation/\r\nslurm/dev/alfred/horeka/jobs_old/dyn_gt_actions_ablation/dyn_gt_actions_ablation.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/dyn_gt_actions_ablation/dyn_gt_actions_ablation_baseline.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/dyn_gt_actions_ablation/dyn_gt_actions_ablation_dev.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/dyn_gt_actions_ablation/masked_lim_yolo.sh\r\nslurm/dev/alfred/horeka/jobs_old/generate_single_samples/\r\nslurm/dev/alfred/horeka/jobs_old/generate_single_samples/generate_samples_50k.sh\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/download_10xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/download_6xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/download_7xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/download_8xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/download_9xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/download_index_json.sh\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/mp4_to_array_record_open_ai_6xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/mp4_to_array_record_open_ai_dev.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/mp4_to_npy_open_ai_10xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/mp4_to_npy_open_ai_6xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/mp4_to_npy_open_ai_7xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/mp4_to_npy_open_ai_8xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_local/mp4_to_npy_open_ai_9xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/actions_download/\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/actions_download/download_actions.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/actions_download/download_actions_all.sh\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/mp4_to_arrayrecords_w_actions/\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/mp4_to_arrayrecords_w_actions/preproc_w_actions.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/videos/\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/videos/download_10xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/videos/download_6xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/videos/download_7xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/videos/download_8xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/videos/download_9xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/videos/download_index_json.sh\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/videos/mp4_to_array_record_open_ai.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/videos/mp4_to_array_record_open_ai_chunked.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/videos/mp4_to_array_record_open_ai_dev.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/videos/mp4_to_array_record_open_ai_to_fast.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/input_pipeline_ws/videos/mp4_to_npy_open_ai.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/job_chaining/\r\nslurm/dev/alfred/horeka/jobs_old/job_chaining/chain_example.sh\r\nslurm/dev/alfred/horeka/jobs_old/job_requeueing/\r\nslurm/dev/alfred/horeka/jobs_old/job_requeueing/train_lam_chain_dev.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/job_requeueing/train_lam_requeue_dev.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/job_requeueing/example_tokenizer_lr_tuning/\r\nslurm/dev/alfred/horeka/jobs_old/job_requeueing/example_tokenizer_lr_tuning/lr_tuning_tokenizer.sh\r\nslurm/dev/alfred/horeka/jobs_old/job_requeueing/example_tokenizer_lr_tuning/train_tokenizer_lr_general.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/job_requeueing/lr_tuning/\r\nslurm/dev/alfred/horeka/jobs_old/job_requeueing/lr_tuning/tokenizer_lr_tuning.py\r\nslurm/dev/alfred/horeka/jobs_old/job_requeueing/lr_tuning/tokenizer/\r\nslurm/dev/alfred/horeka/jobs_old/job_requeueing/lr_tuning/tokenizer/lr_tuning_tokenizer.sh\r\nslurm/dev/alfred/horeka/jobs_old/job_requeueing/lr_tuning/tokenizer/train_tokenizer_lr_general.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/masked_lim/\r\nslurm/dev/alfred/horeka/jobs_old/masked_lim/masked_lim_dev.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/masked_lim/masked_lim_yolo.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/masked_lim/masked_lim_yolo.sh\r\nslurm/dev/alfred/horeka/jobs_old/masked_lim_noise/\r\nslurm/dev/alfred/horeka/jobs_old/masked_lim_noise/masked_lim_dev.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/masked_lim_noise/masked_lim_yolo.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/masked_lim_noise/masked_lim_yolo.sh\r\nslurm/dev/alfred/horeka/jobs_old/overfit_minecraft_single_sample/\r\nslurm/dev/alfred/horeka/jobs_old/overfit_minecraft_single_sample/train_dynamics_overfit_sample.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_minecraft_single_sample/train_dynamics_overfit_sample.sh\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/train_lam_dev.sh\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/train_lam_init.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/train_lam_samples_12.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/train_lam_samples_12288.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/train_lam_samples_1536.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/train_lam_samples_24576.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/train_lam_samples_3072.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/train_lam_samples_384.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/train_lam_samples_49152.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/train_lam_samples_6144.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/train_lam_samples_96.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/from_ckpt/\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/from_ckpt/train_lam_dev.sh\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/lam/from_ckpt/train_lam_samples_12.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/train_tokenizer_dev.sh\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/train_tokenizer_samples_12.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/train_tokenizer_samples_12288.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/train_tokenizer_samples_1536.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/train_tokenizer_samples_24576.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/train_tokenizer_samples_3072.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/train_tokenizer_samples_384.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/train_tokenizer_samples_49152.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/train_tokenizer_samples_6144.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/train_tokenizer_samples_96.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/from_ckpt/\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/from_ckpt/train_tokenizer_samples_12.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_run_ds_oai/tokenizer/from_ckpt/train_tokenizer_samples_12.sh\r\nslurm/dev/alfred/horeka/jobs_old/overfit_sample/\r\nslurm/dev/alfred/horeka/jobs_old/overfit_sample/train_dynamics_overfit_sample.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_sample/train_lam_overfit_sample.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_sample/train_tokenizer_overfit_sample.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_sample/train_tokenizer_overfit_sample_size_0.6_mio.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_sample/train_tokenizer_overfit_sample_size_0_5.sh\r\nslurm/dev/alfred/horeka/jobs_old/overfit_sample/train_tokenizer_overfit_sample_size_21_mio.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_sample/train_tokenizer_overfit_sample_size_2_mio.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_sample/train_tokenizer_overfit_sample_size_9_mio.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/overfit_sample/train_tokenizer_overfit_sample_size_small_mio.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/mp4_to_npy_10xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/mp4_to_npy_6xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/mp4_to_npy_7xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/mp4_to_npy_8xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/mp4_to_npy_9xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/mp4_to_npy_open_ai.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/mp4_to_npy_test.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/npy_to_tfrecord.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/npy_to_tfrecord_10xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/npy_to_tfrecord_6xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/npy_to_tfrecord_7xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/npy_to_tfrecord_8xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/npy_to_tfrecord_9xx.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/preprocess_video_splitter.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/preprocess_video_to_npy.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/preprocess/preprocess_video_to_npy_test.sh\r\nslurm/dev/alfred/horeka/jobs_old/procgen/\r\nslurm/dev/alfred/horeka/jobs_old/procgen/cp_script.sh\r\nslurm/dev/alfred/horeka/jobs_old/procgen/data_gen_gym_acrobot.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/procgen/data_gen_gym_carracing.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/procgen/data_gen_gym_coinrun.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/procgen/data_gen_gym_coinrun.sh\r\nslurm/dev/alfred/horeka/jobs_old/procgen/data_gen_gym_mountaincar copy.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/procgen/data_gen_gym_mountaincar.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/procgen/data_gen_gym_multi.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/procgen/data_gen_gym_pendulum.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/rsync/\r\nslurm/dev/alfred/horeka/jobs_old/rsync/rsync.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/rsync/rsync_tf_records.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/sample_jafar/\r\nslurm/dev/alfred/horeka/jobs_old/sample_jafar/sample_coinrun.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/sampling/\r\nslurm/dev/alfred/horeka/jobs_old/sampling/sample_coinrun.sh\r\nslurm/dev/alfred/horeka/jobs_old/sampling/sample_knoms.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/sampling/sample_knoms.sh\r\nslurm/dev/alfred/horeka/jobs_old/sampling/sample_knoms_mihir.sh\r\nslurm/dev/alfred/horeka/jobs_old/train_dyn/\r\nslurm/dev/alfred/horeka/jobs_old/train_dyn/train_dyn_knoms_full.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_dyn_dev/\r\nslurm/dev/alfred/horeka/jobs_old/train_dyn_dev/train_dyn.sh\r\nslurm/dev/alfred/horeka/jobs_old/train_dyn_dev/train_dyn_checkpt_loading_test_dev.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_dyn_dev/train_dyn_dev.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_dyn_dev/train_dyn_single_batch.sh\r\nslurm/dev/alfred/horeka/jobs_old/train_lam/\r\nslurm/dev/alfred/horeka/jobs_old/train_lam/train_lam_full.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_lam_dev/\r\nslurm/dev/alfred/horeka/jobs_old/train_lam_dev/train_lam.sh\r\nslurm/dev/alfred/horeka/jobs_old/train_lam_dev/train_lam_dev.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_lam_dev/train_lam_full_dev.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_lam_dev/train_lam_single_batch.sh\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer/\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer/train_lam_oai_dev copy.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer/train_lam_oai_dev.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer/train_tokenizer_knoms_overfit_single_batch.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer/train_tokenizer_knoms_overfit_single_sample.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer/train_tokenizer_knoms_overfit_tfrecord_10.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer/train_tokenizer_knoms_overfit_tfrecord_full.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer_dev/\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer_dev/train_tokenizer.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer_dev/train_tokenizer.sh\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer_dev/train_tokenizer_copy.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer_dev/train_tokenizer_h100.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer_dev/train_tokenizer_overfit_tfrecord_10.sh\r\nslurm/dev/alfred/horeka/jobs_old/train_tokenizer_dev/train_tokenizer_single_batch.sh\r\nslurm/dev/alfred/horeka/jobs_old/validation/\r\nslurm/dev/alfred/horeka/jobs_old/validation/tokenizer_lr_tuning/\r\nslurm/dev/alfred/horeka/jobs_old/validation/tokenizer_lr_tuning/lr_tuning_tokenizer.sh\r\nslurm/dev/alfred/horeka/jobs_old/validation/tokenizer_lr_tuning/train_tokenizer_lr_general.sbatch\r\nslurm/dev/alfred/horeka/jobs_old/validation/tokenizer_without_optimizer/\r\nslurm/dev/alfred/horeka/jobs_old/validation/tokenizer_without_optimizer/lr_tuning_tokenizer.sh\r\nslurm/dev/alfred/horeka/jobs_old/validation/tokenizer_without_optimizer/train_tokenizer_lr_general.sbatch\r\nslurm/dev/alfred/parallel_lam_dynamics_training/\r\nslurm/dev/alfred/parallel_lam_dynamics_training/train_lam_init_params.sbatch\r\nslurm/dev/alfred/parallel_lam_dynamics_training/train_parallel_lam_dynamics.sbatch\r\nslurm/dev/alfred/parallel_lam_dynamics_training/train_parallel_lam_dynamics_lr_1e-4.sbatch\r\nslurm/dev/alfred/parallel_lam_dynamics_training/train_parallel_lam_dynamics_mock.sbatch\r\nslurm/dev/mihir/cremers/\r\nslurm/dev/mihir/cremers/coinrun_train_tokenizer.sbatch\r\nslurm/dev/mihir/cremers/train_tokenizer.sbatch\r\nslurm/dev/mihir/cremers/train_tokenizer_overfit_batch_coinrun.sbatch\r\nslurm/dev/mihir/horeka/\r\nslurm/dev/mihir/horeka/generate_dataset_10m.sh\r\nslurm/dev/mihir/horeka/train_tokenizer.sh\r\nslurm/dev/mihir/horeka/train_tokenizer_coinrun.sbatch\r\nslurm/dev/mihir/horeka/causal_fit_modelsizes/\r\nslurm/dev/mihir/horeka/causal_fit_modelsizes/train_dynamics_new_arch_180M.sbatch\r\nslurm/dev/mihir/horeka/causal_fit_modelsizes/train_dynamics_new_arch_255M.sbatch\r\nslurm/dev/mihir/horeka/causal_fit_modelsizes/train_dynamics_new_arch_356M.sbatch\r\nslurm/dev/mihir/horeka/causal_fit_modelsizes/train_dynamics_new_arch_500M.sbatch\r\nslurm/dev/mihir/horeka/causal_fit_modelsizes/train_dynamics_new_arch_80M.sbatch\r\nslurm/dev/mihir/horeka/chunked_ablation/\r\nslurm/dev/mihir/horeka/chunked_ablation/train_dynamics_chunked_dataset.sbatch\r\nslurm/dev/mihir/horeka/chunked_ablation/train_dynamics_unchunked_dataset.sbatch\r\nslurm/dev/mihir/horeka/overfit_batch/\r\nslurm/dev/mihir/horeka/overfit_batch/train_lam_overfit_batch.sbatch\r\nslurm/dev/mihir/horeka/overfit_batch/train_tokenizer_overfit_batch.sbatch\r\nslurm/dev/mihir/horeka/overfit_batch_tiny/\r\nslurm/dev/mihir/horeka/overfit_batch_tiny/train_dynamics_overfit_batch_big_lr.sbatch\r\nslurm/dev/mihir/horeka/overfit_batch_tiny/train_dynamics_overfit_batch_mid_lr.sbatch\r\nslurm/dev/mihir/horeka/overfit_batch_tiny/train_dynamics_overfit_batch_smol_lr.sbatch\r\nslurm/dev/mihir/horeka/overfit_batch_tiny/train_lam_overfit_batch.sbatch\r\nslurm/dev/mihir/horeka/overfit_batch_tiny/train_tokenizer_overfit_batch.sbatch\r\nslurm/dev/mihir/horeka/overfit_sample/causal/\r\nslurm/dev/mihir/horeka/overfit_sample/causal/dynamics_overfit_sample.sbatch\r\nslurm/dev/mihir/horeka/overfit_sample/causal/dynamics_overfit_sample_gaussian_noise.sbatch\r\nslurm/dev/mihir/horeka/overfit_sample/maskgit/\r\nslurm/dev/mihir/horeka/overfit_sample/maskgit/train_lam_overfit_sample.sbatch\r\nslurm/dev/mihir/horeka/overfit_sample/maskgit/train_tokenizer_overfit_sample.sbatch\r\nslurm/dev/mihir/horeka/overfit_sample_tiny/\r\nslurm/dev/mihir/horeka/overfit_sample_tiny/train_dynamics_overfit_sample_big_lr-2.sbatch\r\nslurm/dev/mihir/horeka/overfit_sample_tiny/train_dynamics_overfit_sample_big_lr.sbatch\r\nslurm/dev/mihir/horeka/overfit_sample_tiny/train_dynamics_overfit_sample_mid_lr.sbatch\r\nslurm/dev/mihir/horeka/overfit_sample_tiny/train_dynamics_overfit_sample_smol_lr.sbatch\r\nslurm/dev/mihir/horeka/overfit_sample_tiny/train_lam_overfit_sample.sbatch\r\nslurm/dev/mihir/horeka/overfit_sample_tiny/train_tokenizer_overfit_sample.sbatch\r\nslurm/dev/mihir/horeka/yolo-runs/\r\nslurm/dev/mihir/horeka/yolo-runs/sampling.sh\r\nslurm/dev/mihir/horeka/yolo-runs/sampling_dev.sh\r\nslurm/dev/mihir/horeka/yolo-runs/tester.sh\r\nslurm/dev/mihir/horeka/yolo-runs/train_dynamics_maskgit_speedrun.sbatch\r\nslurm/dev/mihir/horeka/yolo-runs/train_dynamics_new_arch.sbatch\r\nslurm/dev/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch\r\nslurm/dev/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch\r\nslurm/dev/mihir/horeka/yolo-runs/train_tokenizer_maskgit_speedrun.sbatch\r\nslurm/jobs/alfred/\r\nslurm/jobs/alfred/berlin/\r\nslurm/jobs/alfred/berlin/atari/\r\nslurm/jobs/alfred/berlin/atari/atari_dynamics/\r\nslurm/jobs/alfred/berlin/atari/atari_dynamics/atari_dyn_maskgit_causal.sh\r\nslurm/jobs/alfred/berlin/atari/atari_dynamics/atari_dynamics.sbatch\r\nslurm/jobs/alfred/berlin/atari/atari_sampling/\r\nslurm/jobs/alfred/berlin/atari/atari_sampling/sample_causal.sbatch\r\nslurm/jobs/alfred/berlin/atari/atari_sampling/sample_maskgit.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/coinrun_lam_baseline.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/coinrun_lam_ffn_2048_num_block_4.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/coinrun_lam_mix_precision_ffn_2048_num_block_4.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/coinrun_lam_mixed_prec.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/coinrun_lam_wsd.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/coinrun_lam_wsd_3e-5_3e-6.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/coinrun_lam_wsd_3e-6.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/coinrun_lam_wsd_8e-6.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/wsd/\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/wsd/coinrun_lam_wsd_3e-5_3e-6.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/wsd/coinrun_lam_wsd_3e-6.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_ablation/wsd/coinrun_lam_wsd_8e-6.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_bigrun/\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_bigrun/coinrun_dynamics_reproduction.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_bigrun/coinrun_dynamics_reproduction_cotrain.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_bigrun/coinrun_lam_big_run.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_bigrun/coinrun_tokenizer_repoduction_ffn_512_n_blocks_8_full_prec.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_bigrun/generate_data.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_bigrun/sample.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_bigrun/sample_cotrain.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_data/\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_data/generate_data.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_data/generate_data_500mio.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_data/generate_data_npy.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_dynamics/\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_dynamics/coinrun_dyn_maskgit_causal.sh\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_dynamics/coinrun_dynamics.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_batch_size_144.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_batch_size_144_3e-5.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_batch_size_144_3e-5_invest_nan.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_batch_size_144_3e-5_invest_nan_restore_40k.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_batch_size_144_cotraining.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_batch_size_144_cotraining_3e-5.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_maskgit_overfit.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_lam/\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_lam/lam_coinrun.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction/\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_dynamics_reproduction.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_dynamics_reproduction_cotrain.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_full_prec.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_full_prec_cosine.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_full_prec_w_restore.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_mix_prec_cosine.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_mix_prec_cosine_min_init_lt_3e-6.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_tokenizer_repoduction.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction/sample.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction/sample_cotrain.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_dynamics_reproduction.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_dynamics_reproduction_cotrain.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_dc_0.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_ffn_512.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_ffn_512_num_blocks_8.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_ffn_512_num_blocks_8_full_prec.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_lower_lr.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_lower_lr_0.5x.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_lower_lr_3e6_1e5_0.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_lower_lr_3e6_3e5_0.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_tokenizer_repoduction.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_tokenizer_repoduction_ffn_512_n_blocks_8.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_tokenizer_repoduction_ffn_512_n_blocks_8_full_prec.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/generate_data.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/sample.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/sample_cotrain.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_sampling/\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_sampling/sample_causal.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_sampling/sample_maskgit.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_sampling/sample_maskgit_cotrain_3e-5.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_sampling/sample_maskgit_non_cotrain_3e-5.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/coinrun_sampling/sample_maskgit_overfit.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/w_val/\r\nslurm/jobs/alfred/berlin/coinrun/w_val/lam_baseline.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/w_val/lam_baseline_no_print.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/w_val/lam_batch_100.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/w_val/lam_batch_200.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/w_val/lam_jasmine_default.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/w_val/tokenizer_baseline.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/w_val/tokenizer_baseline_debug.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/w_val/tokenizer_baseline_debug_500m.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/w_val/tokenizer_batch_100.sbatch\r\nslurm/jobs/alfred/berlin/coinrun/w_val/tokenizer_batch_200.sbatch\r\nslurm/jobs/alfred/berlin/downlaod/\r\nslurm/jobs/alfred/berlin/downlaod/download_actions.sbatch\r\nslurm/jobs/alfred/berlin/downlaod/download_actions.sh\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_dynamics_reproduction_requeue.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_dynamics_reproduction_requeue_2.0.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_lam_reproduction.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_tokenizer_repoduction.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_tokenizer_repoduction_requeue.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_tokenizer_repoduction_requeue_2.0.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/sample.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/sample_og_150k.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/sample_og_175k.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/sample_og_200k.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/sample_og_spawner.sh\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/spawner.sh\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/spawner_sample.sh\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/requeue/\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/requeue/og_coinrun_dynamics_reproduction.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/requeue/og_coinrun_lam_reproduction.sbatch\r\nslurm/jobs/alfred/berlin/jafar_og_reproduction/requeue/og_coinrun_tokenizer_repoduction.sbatch\r\nslurm/jobs/alfred/berlin/minecraft/\r\nslurm/jobs/alfred/berlin/minecraft/gt_act_sampling/\r\nslurm/jobs/alfred/berlin/minecraft/gt_act_sampling/sample_gt_act_maskgit.sbatch\r\nslurm/jobs/alfred/berlin/minecraft/minecraft_sampling/\r\nslurm/jobs/alfred/berlin/minecraft/minecraft_sampling/atari_dyn_maskgit_causal.sh\r\nslurm/jobs/alfred/berlin/minecraft/minecraft_sampling/sample_causal.sbatch\r\nslurm/jobs/alfred/berlin/minecraft/minecraft_sampling/sample_dynamics-maskgit-8-node-darkness-filter-3423250.sbatch\r\nslurm/jobs/alfred/helmholtz_cluster/\r\nslurm/jobs/alfred/helmholtz_cluster/jafar_og_reproduction/\r\nslurm/jobs/alfred/helmholtz_cluster/jafar_og_reproduction/generate_dataset.sbatch\r\nslurm/jobs/alfred/helmholtz_cluster/jafar_og_reproduction/generate_dataset_10m.sbatch\r\nslurm/jobs/alfred/helmholtz_cluster/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch\r\nslurm/jobs/alfred/helmholtz_cluster/jafar_og_reproduction/og_coinrun_lam_reproduction.sbatch\r\nslurm/jobs/alfred/helmholtz_cluster/jafar_og_reproduction/og_coinrun_tokenizer_repoduction.sbatch\r\nslurm/jobs/alfred/horeka/\r\nslurm/jobs/alfred/horeka/atari_dynamics/\r\nslurm/jobs/alfred/horeka/atari_dynamics/atari_dyn_maskgit_causal.sh\r\n",,terminal_output +5404,14748445,"TERMINAL",0,0,"slurm/jobs/alfred/horeka/atari_dynamics/atari_dynamics.sbatch\r\nslurm/jobs/alfred/horeka/dyn_gt_actions_ablation_prepend/\r\nslurm/jobs/alfred/horeka/dyn_gt_actions_ablation_prepend/dyn_gt_actions_ablation.sbatch\r\nslurm/jobs/alfred/horeka/dyn_gt_actions_ablation_prepend/dyn_gt_actions_maskgit_causal.sh\r\nslurm/jobs/alfred/horeka/lr_tuning/tokenizer/\r\nslurm/jobs/alfred/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch\r\nslurm/jobs/alfred/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch\r\nslurm/jobs/alfred/horeka/lr_tuning/tokenizer/train_tokenizer_lr_general.sbatch\r\nslurm/jobs/alfred/horeka/misc/\r\nslurm/jobs/alfred/horeka/misc/preproc_mp4_to_npy_open_ai.sbatch\r\nslurm/jobs/alfred/horeka/misc/preproc_npy_to_tfrecord_open_ai.sbatch\r\nslurm/jobs/alfred/horeka/misc/train_dyn_knoms.sbatch\r\nslurm/jobs/alfred/horeka/misc/train_lam_knoms.sbatch\r\nslurm/jobs/alfred/horeka/misc/train_tokenizer_knoms.sbatch\r\nslurm/jobs/franz/\r\nslurm/jobs/franz/berlin/\r\nslurm/jobs/franz/berlin/atari/\r\nslurm/jobs/franz/berlin/atari/atari_data_gen.sh\r\nslurm/jobs/franz/berlin/coinrun/\r\nslurm/jobs/franz/berlin/coinrun/coinrun_dynamics/\r\nslurm/jobs/franz/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_fp32_adam_moments.sh\r\nslurm/jobs/franz/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_fp32_layernorm.sh\r\nslurm/jobs/franz/berlin/coinrun/coinrun_lam/\r\nslurm/jobs/franz/berlin/coinrun/coinrun_lam/lam_coinrun_nan_investigation_100k_to_107k.sbatch\r\nslurm/jobs/franz/berlin/coinrun/coinrun_lam/lam_coinrun_nan_investigation_z_loss.sbatch\r\nslurm/jobs/franz/berlin/coinrun/coinrun_tokenizer/\r\nslurm/jobs/franz/berlin/coinrun/coinrun_tokenizer/coinrun_tokenizer_repoduction.sbatch\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/linear_lr/\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/linear_lr/tester.sh\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/linear_lr/train_dynamics_16_nodes.sbatch\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/linear_lr/train_dynamics_1_nodes.sbatch\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/linear_lr/train_dynamics_2_nodes.sbatch\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/linear_lr/train_dynamics_4_nodes.sbatch\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/linear_lr/train_dynamics_8_nodes.sbatch\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/sqrt_lr/\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/sqrt_lr/tester.sh\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/sqrt_lr/train_dynamics_16_nodes.sbatch\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/sqrt_lr/train_dynamics_1_nodes.sbatch\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/sqrt_lr/train_dynamics_2_nodes.sbatch\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/sqrt_lr/train_dynamics_4_nodes.sbatch\r\nslurm/jobs/franz/horeka/batchsize_scaling/dynamics_cotraining_new_arch/sqrt_lr/train_dynamics_8_nodes.sbatch\r\nslurm/jobs/mihir/horeka/\r\nslurm/jobs/mihir/horeka/sbatch_holiday.sh\r\nslurm/jobs/mihir/horeka/action_space_scaling/co_training/\r\nslurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/solo_lam/\r\nslurm/jobs/mihir/horeka/action_space_scaling/solo_lam/train_lam_actionspace_10.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/solo_lam/train_lam_actionspace_12.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/solo_lam/train_lam_actionspace_20.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/solo_lam/train_lam_actionspace_50.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/solo_lam/train_lam_actionspace_6.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/solo_lam/train_lam_actionspace_8.sbatch\r\nslurm/jobs/mihir/horeka/atari/\r\nslurm/jobs/mihir/horeka/atari/sample_causal.sbatch\r\nslurm/jobs/mihir/horeka/atari/sample_maskgit.sbatch\r\nslurm/jobs/mihir/horeka/atari/train_dynamics_causal.sbatch\r\nslurm/jobs/mihir/horeka/atari/train_dynamics_maskgit.sbatch\r\nslurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics/sqrt_lr/\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics/sqrt_lr/tester.sh\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics/sqrt_lr/train_dynamics_16_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics/sqrt_lr/train_dynamics_1_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics/sqrt_lr/train_dynamics_2_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics/sqrt_lr/train_dynamics_32_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics/sqrt_lr/train_dynamics_4_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics/sqrt_lr/train_dynamics_8_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/tester.sh\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/lam/linear_lr/\r\nslurm/jobs/mihir/horeka/batchsize_scaling/lam/linear_lr/train_lam_16_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/lam/linear_lr/train_lam_1_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/lam/linear_lr/train_lam_2_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/lam/linear_lr/train_lam_4_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/lam/linear_lr/train_lam_8_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/lam/sqrt_lr/\r\nslurm/jobs/mihir/horeka/batchsize_scaling/lam/sqrt_lr/train_lam_16_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/lam/sqrt_lr/train_lam_1_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/lam/sqrt_lr/train_lam_2_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/lam/sqrt_lr/train_lam_4_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/lam/sqrt_lr/train_lam_8_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/const_lr/\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/const_lr/train_tokenizer_16_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/const_lr/train_tokenizer_1_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/const_lr/train_tokenizer_2_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/const_lr/train_tokenizer_32_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/const_lr/train_tokenizer_4_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/const_lr/train_tokenizer_8_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/linear_lr/\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/linear_lr/tester.sh\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/linear_lr/train_tokenizer_16_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/linear_lr/train_tokenizer_1_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/linear_lr/train_tokenizer_2_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/linear_lr/train_tokenizer_32_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/linear_lr/train_tokenizer_4_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/linear_lr/train_tokenizer_8_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/tester.sh\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_16_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_1_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_2_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_32_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_4_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_8_nodes.sbatch\r\nslurm/jobs/mihir/horeka/breakout/\r\nslurm/jobs/mihir/horeka/breakout/sample_maskgit-gt-actions.sbatch\r\nslurm/jobs/mihir/horeka/breakout/sample_maskgit.sbatch\r\nslurm/jobs/mihir/horeka/breakout/big_runs/\r\nslurm/jobs/mihir/horeka/breakout/big_runs/train_dyn_single_gpu.sh\r\nslurm/jobs/mihir/horeka/breakout/big_runs/train_dyn_single_gpu_gt_actions.sh\r\nslurm/jobs/mihir/horeka/breakout/big_runs/train_dyn_single_gpu_gt_actions_smaller_lr.sh\r\nslurm/jobs/mihir/horeka/breakout/big_runs/train_dyn_single_gpu_smaller_lr.sh\r\nslurm/jobs/mihir/horeka/breakout/default_runs/\r\n",,terminal_output +5405,14748563,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh\r\nslurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh\r\nslurm/jobs/mihir/horeka/breakout/default_runs/train_lam_as_15.sh\r\nslurm/jobs/mihir/horeka/breakout/default_runs/train_lam_default.sh\r\nslurm/jobs/mihir/horeka/breakout/default_runs/train_tokenizer_default.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/full-prec/\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/full-prec/train_dyn_single_gpu.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/full-prec/train_dyn_single_gpu_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/full-prec/train_dyn_single_gpu_gt_actions.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/full-prec/train_dyn_single_gpu_gt_actions_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/full-prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/full-prec/train_dyn_single_gpu_smaller_lr_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu-nan-debug.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_smaller_lr_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/train_dyn_single_gpu.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/train_dyn_single_gpu_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/train_dyn_single_gpu_gt_actions.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/train_dyn_single_gpu_gt_actions_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/train_dyn_single_gpu_gt_actions_smaller_lr_50k.sh\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/z-loss/train_dyn_single_gpu_smaller_lr_50k.sh\r\nslurm/jobs/mihir/horeka/causal_big_runs/\r\nslurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_2_nodes.sbatch\r\nslurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes.sbatch\r\nslurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_dev.sh\r\nslurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch\r\nslurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark_req.sbatch\r\nslurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_dev.sbatch\r\nslurm/jobs/mihir/horeka/coinrun/\r\nslurm/jobs/mihir/horeka/coinrun/sample_causal.sbatch\r\nslurm/jobs/mihir/horeka/coinrun/sample_maskgit.sbatch\r\nslurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh\r\nslurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu_gt_actions.sh\r\nslurm/jobs/mihir/horeka/coinrun/train_dynamics_causal.sbatch\r\nslurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch\r\nslurm/jobs/mihir/horeka/coinrun/train_lam_single_gpu.sh\r\nslurm/jobs/mihir/horeka/coinrun/train_tokenizer_lr_1e-4.sbatch\r\nslurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh\r\nslurm/jobs/mihir/horeka/coinrun/default_runs/\r\nslurm/jobs/mihir/horeka/coinrun/default_runs/train_lam_as_15.sh\r\nslurm/jobs/mihir/horeka/coinrun/default_runs/train_lam_default.sh\r\nslurm/jobs/mihir/horeka/coinrun/default_runs/train_tokenizer_default.sh\r\nslurm/jobs/mihir/horeka/jafar_og_reproduction/\r\nslurm/jobs/mihir/horeka/jafar_og_reproduction/generate_dataset.sbatch\r\nslurm/jobs/mihir/horeka/jafar_og_reproduction/generate_dataset_10m.sbatch\r\nslurm/jobs/mihir/horeka/jafar_og_reproduction/generate_dataset_10m_gt_actions.sbatch\r\nslurm/jobs/mihir/horeka/jafar_og_reproduction/generate_dataset_10m_gt_actions_train_val_test.sbatch\r\nslurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch\r\nslurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_lam_reproduction.sbatch\r\nslurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_tokenizer_repoduction.sbatch\r\nslurm/jobs/mihir/horeka/lam/\r\nslurm/jobs/mihir/horeka/lam/coinrun/\r\nslurm/jobs/mihir/horeka/lam/coinrun/coinrun_lam_133M.sbatch\r\nslurm/jobs/mihir/horeka/lam/coinrun/coinrun_lam_311.sbatch\r\nslurm/jobs/mihir/horeka/lam/coinrun/coinrun_lam_400M.sbatch\r\nslurm/jobs/mihir/horeka/lam/coinrun/coinrun_lam_as_128.sbatch\r\nslurm/jobs/mihir/horeka/lam/coinrun/coinrun_lam_as_16.sbatch\r\nslurm/jobs/mihir/horeka/lam/coinrun/coinrun_lam_as_32.sbatch\r\nslurm/jobs/mihir/horeka/lam/coinrun/coinrun_lam_as_64.sbatch\r\nslurm/jobs/mihir/horeka/lam/coinrun/coinrun_lam_base.sbatch\r\nslurm/jobs/mihir/horeka/lam/coinrun/tmp\r\nslurm/jobs/mihir/horeka/lam/minecraft/\r\nslurm/jobs/mihir/horeka/lam/minecraft/train_lam_minecraft_1node_dev.sbatch\r\nslurm/jobs/mihir/horeka/lam/minecraft/train_lam_minecraft_8node-darkness-filter-133M.sbatch\r\nslurm/jobs/mihir/horeka/lam/minecraft/train_lam_minecraft_8node-darkness-filter-311M.sbatch\r\nslurm/jobs/mihir/horeka/lam/minecraft/train_lam_minecraft_8node-darkness-filter-400M.sbatch\r\nslurm/jobs/mihir/horeka/lam/minecraft/train_lam_minecraft_8node-darkness-filter-base.sbatch\r\nslurm/jobs/mihir/horeka/lam/minecraft/train_lam_minecraft_8node.sbatch\r\nslurm/jobs/mihir/horeka/lamap/\r\nslurm/jobs/mihir/horeka/lamap/coinrun_lam_base.sbatch\r\nslurm/jobs/mihir/horeka/lamap/coinrun_lam_base_dev.sh\r\nslurm/jobs/mihir/horeka/lr_tuning/dynamics/\r\nslurm/jobs/mihir/horeka/lr_tuning/dynamics/train_dyn_default_const.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/dynamics/train_dyn_default_cos.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/dynamics/train_dyn_default_wsd.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes_dev.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_dev.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-6_8nodes.sbatch\r\nslurm/jobs/mihir/horeka/mask_prob_fix/\r\nslurm/jobs/mihir/horeka/mask_prob_fix/train_dynamics_2_nodes.sbatch\r\nslurm/jobs/mihir/horeka/mask_prob_fix/train_dynamics_2_nodes_small_model.sbatch\r\nslurm/jobs/mihir/horeka/mask_prob_fix/train_dynamics_8_nodes.sbatch\r\nslurm/jobs/mihir/horeka/mask_prob_fix/train_dynamics_8_nodes_dev.sbatch\r\nslurm/jobs/mihir/horeka/mask_prob_fix/train_dynamics_8_nodes_dev_requeue.sbatch\r\nslurm/jobs/mihir/horeka/mask_prob_fix/train_dynamics_8_nodes_small_model.sbatch\r\nslurm/jobs/mihir/horeka/maskgit_big_runs/\r\nslurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes.sbatch\r\nslurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch\r\nslurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark_req.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/\r\nslurm/jobs/mihir/horeka/modelsize_scaling/runner.sh\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/B_train_dyn_3.5.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/C_train_dyn_6M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/D_train_dyn_12M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/E_train_dyn_18M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/lam/\r\nslurm/jobs/mihir/horeka/modelsize_scaling/lam/tester.sh\r\nslurm/jobs/mihir/horeka/modelsize_scaling/lam/train_lam_38M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/\r\nslurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/tester.sh\r\nslurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/train_tokenizer_127M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/train_tokenizer_227M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/train_tokenizer_37M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/train_tokenizer_74M.sbatch\r\nslurm/jobs/mihir/horeka/preprocessing/\r\nslurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh\r\nslurm/jobs/mihir/horeka/preprocessing/coinrun_chunked.sh\r\nslurm/jobs/mihir/horeka/preprocessing/preprocess_atari.sbatch\r\nslurm/jobs/mihir/horeka/preprocessing/preprocess_minecraft_multi_chunk.sbatch\r\nslurm/jobs/mihir/horeka/preprocessing/preprocess_minecraft_single_chunk.sbatch\r\nslurm/utils/\r\nslurm/utils/exclude.txt\r\nslurm/utils/alfred/\r\nslurm/utils/alfred/scp_scripts/\r\nslurm/utils/alfred/scp_scripts/copy_to_berlin.sh\r\nslurm/utils/alfred/scp_scripts/copy_to_local.sh\r\nslurm/utils/mihir/\r\nslurm/utils/mihir/weekend-job-requeuer.sh\r\nslurm/utils/mihir/weekend-job-starter.sh\r\ntests/\r\ntests/test_dataloader.py\r\nutils/\r\n",,terminal_output +5406,14750433,"TERMINAL",0,0,"\r\nsent 1,571,238 bytes received 13,076 bytes 452,661.14 bytes/sec\r\ntotal size is 28,033,824 speedup is 17.69\r\n]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +5407,14750943,"TERMINAL",0,0,"r",,terminal_output +5408,14751062,"TERMINAL",0,0,"u",,terminal_output +5409,14751195,"TERMINAL",0,0,"n",,terminal_output +5410,14751330,"TERMINAL",0,0,"n",,terminal_output +5411,14751471,"TERMINAL",0,0,"er",,terminal_output +5412,14751601,"TERMINAL",0,0,"-",,terminal_output +5413,14752084,"TERMINAL",0,0,"3",,terminal_output +5414,14752381,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_3[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_3]$ ",,terminal_output +5415,14752807,"TERMINAL",0,0,"s",,terminal_output +5416,14752960,"TERMINAL",0,0,"b",,terminal_output +5417,14753022,"TERMINAL",0,0,"a",,terminal_output +5418,14753130,"TERMINAL",0,0,"t",,terminal_output +5419,14753255,"TERMINAL",0,0,"c",,terminal_output +5420,14753318,"TERMINAL",0,0,"h",,terminal_output +5421,14753454,"TERMINAL",0,0," ",,terminal_output +5422,14753750,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5423,14754328,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh\r\n[?2004l\rSubmitted batch job 3527920\r\n]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_3[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_3]$ ",,terminal_output +5424,14755671,"TERMINAL",0,0,"de",,terminal_output +5425,14756000,"TERMINAL",0,0,"v",,terminal_output +5426,14756242,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +5427,14756649,"TERMINAL",0,0,"qu",,terminal_output +5428,14756763,"TERMINAL",0,0,"e",,terminal_output +5429,14756869,"TERMINAL",0,0,"u",,terminal_output +5430,14756930,"TERMINAL",0,0,"e",,terminal_output +5431,14757113,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: squeue --mehkn0402.localdomain: Sun Sep 28 17:33:21 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3527920 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3527904 accelerat train_dy tum_cte0 R\t6:42\t 1 hkn04293527905 accelerat train_dy tum_cte0 R\t6:42\t 1 hkn05343527891 accelerat train_dy tum_cte0 R10:43\t 1 hkn07303527866 dev_accel interact tum_cte0 R44:24\t 1 hkn0402",,terminal_output +5432,14758202,"TERMINAL",0,0,"23345",,terminal_output +5433,14759123,"TERMINAL",0,0,"34456",,terminal_output +5434,14760147,"TERMINAL",0,0,"45567",,terminal_output +5435,14761156,"TERMINAL",0,0,"56678",,terminal_output +5436,14762177,"TERMINAL",0,0,"67789",,terminal_output +5437,14763198,"TERMINAL",0,0,"788930",,terminal_output +5438,14764241,"TERMINAL",0,0,"899501",,terminal_output +5439,14765266,"TERMINAL",0,0,"9505012",,terminal_output +5440,14766293,"TERMINAL",0,0,"301123",,terminal_output +5441,14767264,"TERMINAL",0,0,"12234",,terminal_output +5442,14768281,"TERMINAL",0,0,"23345",,terminal_output +5443,14769297,"TERMINAL",0,0,"34456",,terminal_output +5444,14770387,"TERMINAL",0,0,"46678",,terminal_output +5445,14771410,"TERMINAL",0,0,"67789",,terminal_output +5446,14772407,"TERMINAL",0,0,"788940",,terminal_output +5447,14773457,"TERMINAL",0,0,"8991:001",,terminal_output +5448,14774393,"TERMINAL",0,0,"97:007:0012",,terminal_output +5449,14775410,"TERMINAL",0,0,"401123",,terminal_output +5450,14776531,"TERMINAL",0,0,"12234",,terminal_output +5451,14777601,"TERMINAL",0,0,"23345",,terminal_output +5452,14778465,"TERMINAL",0,0,"34456",,terminal_output +5453,14779500,"TERMINAL",0,0,"45567",,terminal_output +5454,14780528,"TERMINAL",0,0,"56678",,terminal_output +5455,14781549,"TERMINAL",0,0,"67789",,terminal_output +5456,14782574,"TERMINAL",0,0,"788950",,terminal_output +5457,14783599,"TERMINAL",0,0,"899101",,terminal_output +5458,14784568,"TERMINAL",0,0,"9101012",,terminal_output +5459,14785651,"TERMINAL",0,0,"501123",,terminal_output +5460,14786669,"TERMINAL",0,0,"12234",,terminal_output +5461,14787693,"TERMINAL",0,0,"23345",,terminal_output +5462,14788666,"TERMINAL",0,0,"34456",,terminal_output +5463,14789677,"TERMINAL",0,0,"45567",,terminal_output +5464,14790701,"TERMINAL",0,0,"56678",,terminal_output +5465,14791715,"TERMINAL",0,0,"67789",,terminal_output +5466,14792815,"TERMINAL",0,0,"78895:00",,terminal_output +5467,14793751,"TERMINAL",0,0,"899201",,terminal_output +5468,14794860,"TERMINAL",0,0,"9202012",,terminal_output +5469,14795798,"TERMINAL",0,0,"4:001123",,terminal_output +5470,14796910,"TERMINAL",0,0,"12234",,terminal_output +5471,14797821,"TERMINAL",0,0,"23345",,terminal_output +5472,14798857,"TERMINAL",0,0,"34456",,terminal_output +5473,14799880,"TERMINAL",0,0,"45567",,terminal_output +5474,14800878,"TERMINAL",0,0,"56678",,terminal_output +5475,14801896,"TERMINAL",0,0,"67789",,terminal_output +5476,14802940,"TERMINAL",0,0,"788910",,terminal_output +5477,14803931,"TERMINAL",0,0,"899301",,terminal_output +5478,14804951,"TERMINAL",0,0,"9303012",,terminal_output +5479,14805968,"TERMINAL",0,0,"101123",,terminal_output +5480,14806988,"TERMINAL",0,0,"12234",,terminal_output +5481,14808001,"TERMINAL",0,0,"23345",,terminal_output +5482,14809019,"TERMINAL",0,0,"34456",,terminal_output +5483,14810036,"TERMINAL",0,0,"45567",,terminal_output +5484,14811052,"TERMINAL",0,0,"56678",,terminal_output +5485,14812071,"TERMINAL",0,0,"67789",,terminal_output +5486,14813093,"TERMINAL",0,0,"788920",,terminal_output +5487,14814117,"TERMINAL",0,0,"899401",,terminal_output +5488,14815134,"TERMINAL",0,0,"9404012",,terminal_output +5489,14816160,"TERMINAL",0,0,"201123",,terminal_output +5490,14817157,"TERMINAL",0,0,"12234",,terminal_output +5491,14818174,"TERMINAL",0,0,"23345",,terminal_output +5492,14819232,"TERMINAL",0,0,"34456",,terminal_output +5493,14820213,"TERMINAL",0,0,"45567",,terminal_output +5494,14821279,"TERMINAL",0,0,"56678",,terminal_output +5495,14822256,"TERMINAL",0,0,"67789",,terminal_output +5496,14823329,"TERMINAL",0,0,"788930",,terminal_output +5497,14824287,"TERMINAL",0,0,"899501",,terminal_output +5498,14825377,"TERMINAL",0,0,"9515123",,terminal_output +5499,14826401,"TERMINAL",0,0,"312234",,terminal_output +5500,14827341,"TERMINAL",0,0,"23345",,terminal_output +5501,14828359,"TERMINAL",0,0,"34456",,terminal_output +5502,14829476,"TERMINAL",0,0,"45567",,terminal_output +5503,14830497,"TERMINAL",0,0,"56678",,terminal_output +5504,14831527,"TERMINAL",0,0,"67789",,terminal_output +5505,14832434,"TERMINAL",0,0,"788940",,terminal_output +5506,14833469,"TERMINAL",0,0,"8992:001",,terminal_output +5507,14834491,"TERMINAL",0,0,"98:008:0012",,terminal_output +5508,14835516,"TERMINAL",0,0,"40 Rhkn05161123",,terminal_output +5509,14836538,"TERMINAL",0,0,"112234",,terminal_output +5510,14837514,"TERMINAL",0,0,"223345",,terminal_output +5511,14838584,"TERMINAL",0,0,"334456",,terminal_output +5512,14839553,"TERMINAL",0,0,"445567",,terminal_output +5513,14840656,"TERMINAL",0,0,"556678",,terminal_output +5514,14841658,"TERMINAL",0,0,"667789",,terminal_output +5515,14842607,"TERMINAL",0,0,"7788950",,terminal_output +5516,14843644,"TERMINAL",0,0,"8899101",,terminal_output +5517,14844639,"TERMINAL",0,0,"99101012",,terminal_output +5518,14845657,"TERMINAL",0,0,"50101123",,terminal_output +5519,14846688,"TERMINAL",0,0,"112234",,terminal_output +5520,14847793,"TERMINAL",0,0,"223345",,terminal_output +5521,14848722,"TERMINAL",0,0,"334456",,terminal_output +5522,14849741,"TERMINAL",0,0,"445567",,terminal_output +5523,14850754,"TERMINAL",0,0,"556678",,terminal_output +5524,14851797,"TERMINAL",0,0,"667789",,terminal_output +5525,14852821,"TERMINAL",0,0,"778896:00",,terminal_output +5526,14853845,"TERMINAL",0,0,"8899201",,terminal_output +5527,14854866,"TERMINAL",0,0,"99202012",,terminal_output +5528,14855864,"TERMINAL",0,0,"5:00201123",,terminal_output +5529,14856880,"TERMINAL",0,0,"112234",,terminal_output +5530,14857893,"TERMINAL",0,0,"223345",,terminal_output +5531,14858912,"TERMINAL",0,0,"334456",,terminal_output +5532,14859927,"TERMINAL",0,0,"445567",,terminal_output +5533,14860945,"TERMINAL",0,0,"556678",,terminal_output +5534,14861964,"TERMINAL",0,0,"667789",,terminal_output +5535,14862984,"TERMINAL",0,0,"7788910",,terminal_output +5536,14864000,"TERMINAL",0,0,"8899301",,terminal_output +5537,14865019,"TERMINAL",0,0,"99303012",,terminal_output +5538,14866038,"TERMINAL",0,0,"10301123",,terminal_output +5539,14867157,"TERMINAL",0,0,"112234",,terminal_output +5540,14868079,"TERMINAL",0,0,"223345",,terminal_output +5541,14869108,"TERMINAL",0,0,"334456",,terminal_output +5542,14870126,"TERMINAL",0,0,"445567",,terminal_output +5543,14871151,"TERMINAL",0,0,"556678",,terminal_output +5544,14872175,"TERMINAL",0,0,"667789",,terminal_output +5545,14873161,"TERMINAL",0,0,"7788920",,terminal_output +5546,14874221,"TERMINAL",0,0,"8899401",,terminal_output +5547,14875245,"TERMINAL",0,0,"99404012",,terminal_output +5548,14876219,"TERMINAL",0,0,"20401123",,terminal_output +5549,14877232,"TERMINAL",0,0,"112234",,terminal_output +5550,14878307,"TERMINAL",0,0,"223345",,terminal_output +5551,14879267,"TERMINAL",0,0,"334456",,terminal_output +5552,14880365,"TERMINAL",0,0,"445567",,terminal_output +5553,14881302,"TERMINAL",0,0,"556678",,terminal_output +5554,14882412,"TERMINAL",0,0,"7788930",,terminal_output +5555,14883436,"TERMINAL",0,0,"8899501",,terminal_output +5556,14884460,"TERMINAL",0,0,"99505012",,terminal_output +5557,14884523,"TERMINAL",0,0,"bash",,terminal_focus +5558,14885383,"TERMINAL",0,0,"30501123",,terminal_output +5559,14885690,"jasmine/train_dynamics.py",0,0,"",python,tab +5560,14886387,"TERMINAL",0,0,"112234",,terminal_output +5561,14887037,"jasmine/train_dynamics.py",21923,0,"",python,selection_mouse +5562,14887400,"TERMINAL",0,0,"223345",,terminal_output +5563,14887689,"jasmine/train_dynamics.py",22125,0,"",python,selection_mouse +5564,14888146,"jasmine/train_dynamics.py",22003,0,"",python,selection_mouse +5565,14888417,"TERMINAL",0,0,"334456",,terminal_output +5566,14889480,"TERMINAL",0,0,"445567",,terminal_output +5567,14890052,"jasmine/genie.py",0,0,"",python,tab +5568,14890453,"TERMINAL",0,0,"556678",,terminal_output +5569,14890844,"jasmine/models/dynamics.py",0,0,"",python,tab +5570,14891468,"TERMINAL",0,0,"667789",,terminal_output +5571,14891841,"jasmine/models/dynamics.py",3923,0,"",python,selection_mouse +5572,14892488,"TERMINAL",0,0,"7788940",,terminal_output +5573,14893504,"TERMINAL",0,0,"88993:001",,terminal_output +5574,14894569,"TERMINAL",0,0,"999:009:0012",,terminal_output +5575,14895653,"TERMINAL",0,0,"401:001123",,terminal_output +5576,14896301,"jasmine/models/dynamics.py",2856,0,"",python,selection_mouse +5577,14896559,"TERMINAL",0,0,"112234",,terminal_output +5578,14896915,"jasmine/models/dynamics.py",2949,0,"",python,selection_mouse +5579,14897441,"jasmine/models/dynamics.py",2922,0,"",python,selection_mouse +5580,14897588,"TERMINAL",0,0,"223345",,terminal_output +5581,14898102,"jasmine/models/dynamics.py",2856,0,"",python,selection_mouse +5582,14898594,"TERMINAL",0,0,"334456",,terminal_output +5583,14898655,"jasmine/models/dynamics.py",2821,0,"",python,selection_mouse +5584,14899204,"jasmine/models/dynamics.py",2928,0,"",python,selection_mouse +5585,14899610,"TERMINAL",0,0,"445567",,terminal_output +5586,14900653,"TERMINAL",0,0,"556678",,terminal_output +5587,14901651,"TERMINAL",0,0,"667789",,terminal_output +5588,14902669,"TERMINAL",0,0,"7788950",,terminal_output +5589,14903686,"TERMINAL",0,0,"8899101",,terminal_output +5590,14904722,"TERMINAL",0,0,"99101012",,terminal_output +5591,14905723,"TERMINAL",0,0,"50101123",,terminal_output +5592,14906794,"jasmine/models/dynamics.py",3532,0,"",python,selection_mouse +5593,14906805,"TERMINAL",0,0,"112234",,terminal_output +5594,14907755,"TERMINAL",0,0,"223345",,terminal_output +5595,14908806,"TERMINAL",0,0,"334456",,terminal_output +5596,14909870,"TERMINAL",0,0,"445567",,terminal_output +5597,14910807,"TERMINAL",0,0,"556678",,terminal_output +5598,14911839,"TERMINAL",0,0,"667789",,terminal_output +5599,14912845,"TERMINAL",0,0,"778897:00",,terminal_output +5600,14913865,"TERMINAL",0,0,"8899201",,terminal_output +5601,14914881,"TERMINAL",0,0,"99202012",,terminal_output +5602,14915904,"TERMINAL",0,0,"6:00201123",,terminal_output +5603,14916939,"TERMINAL",0,0,"112234",,terminal_output +5604,14917952,"TERMINAL",0,0,"223345",,terminal_output +5605,14918972,"TERMINAL",0,0,"334456",,terminal_output +5606,14919988,"TERMINAL",0,0,"445567",,terminal_output +5607,14921005,"TERMINAL",0,0,"556678",,terminal_output +5608,14922063,"TERMINAL",0,0,"667789",,terminal_output +5609,14923040,"TERMINAL",0,0,"7788910",,terminal_output +5610,14924058,"TERMINAL",0,0,"8899301",,terminal_output +5611,14925117,"TERMINAL",0,0,"99303012",,terminal_output +5612,14926095,"TERMINAL",0,0,"10301123",,terminal_output +5613,14927108,"TERMINAL",0,0,"112234",,terminal_output +5614,14928124,"TERMINAL",0,0,"223345",,terminal_output +5615,14929146,"TERMINAL",0,0,"334456",,terminal_output +5616,14930157,"TERMINAL",0,0,"445567",,terminal_output +5617,14931183,"TERMINAL",0,0,"556678",,terminal_output +5618,14932194,"TERMINAL",0,0,"667789",,terminal_output +5619,14933307,"TERMINAL",0,0,"7788920",,terminal_output +5620,14934332,"TERMINAL",0,0,"8899401",,terminal_output +5621,14935253,"TERMINAL",0,0,"99404012",,terminal_output +5622,14936277,"TERMINAL",0,0,"20401123",,terminal_output +5623,14937280,"TERMINAL",0,0,"112234",,terminal_output +5624,14938325,"TERMINAL",0,0,"223345",,terminal_output +5625,14939352,"TERMINAL",0,0,"345567",,terminal_output +5626,14940381,"TERMINAL",0,0,"556678",,terminal_output +5627,14941349,"TERMINAL",0,0,"667789",,terminal_output +5628,14942454,"TERMINAL",0,0,"7788930",,terminal_output +5629,14943445,"TERMINAL",0,0,"8899501",,terminal_output +5630,14944469,"TERMINAL",0,0,"99505012",,terminal_output +5631,14945412,"TERMINAL",0,0,"30501123",,terminal_output +5632,14946518,"TERMINAL",0,0,"112234",,terminal_output +5633,14947544,"TERMINAL",0,0,"223345",,terminal_output +5634,14948466,"TERMINAL",0,0,"334456",,terminal_output +5635,14949524,"TERMINAL",0,0,"445567",,terminal_output +5636,14950502,"TERMINAL",0,0,"556678",,terminal_output +5637,14951541,"TERMINAL",0,0,"667789",,terminal_output +5638,14952534,"TERMINAL",0,0,"7788940",,terminal_output +5639,14953580,"TERMINAL",0,0,"88994:001",,terminal_output +5640,14954572,"TERMINAL",0,0,"9910:0010:0012",,terminal_output +5641,14955599,"TERMINAL",0,0,"402:001123",,terminal_output +5642,14956656,"TERMINAL",0,0,"112234",,terminal_output +5643,14957712,"TERMINAL",0,0,"223345",,terminal_output +5644,14958692,"TERMINAL",0,0,"334456",,terminal_output +5645,14959676,"TERMINAL",0,0,"445567",,terminal_output +5646,14960710,"TERMINAL",0,0,"556678",,terminal_output +5647,14961776,"TERMINAL",0,0,"667789",,terminal_output +5648,14962728,"TERMINAL",0,0,"7788950",,terminal_output +5649,14963777,"TERMINAL",0,0,"8899101",,terminal_output +5650,14964785,"TERMINAL",0,0,"99101012",,terminal_output +5651,14965661,"jasmine/models/dynamics.py",3532,0," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n\n",python,content +5652,14965778,"TERMINAL",0,0,"50101123",,terminal_output +5653,14965857,"jasmine/models/dynamics.py",3592,0," jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n",python,content +5654,14966134,"jasmine/models/dynamics.py",3646,0," jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n",python,content +5655,14966436,"jasmine/models/dynamics.py",3716,0," jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n",python,content +5656,14966600,"jasmine/models/dynamics.py",3794,0," jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n",python,content +5657,14966804,"jasmine/models/dynamics.py",3858,0," jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n",python,content +5658,14966818,"TERMINAL",0,0,"112234",,terminal_output +5659,14966987,"jasmine/models/dynamics.py",3920,0," jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n",python,content +5660,14967242,"jasmine/models/dynamics.py",3974,0," jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)\n",python,content +5661,14967246,"jasmine/models/dynamics.py",4036,1,"",python,content +5662,14967813,"TERMINAL",0,0,"223345",,terminal_output +5663,14968838,"TERMINAL",0,0,"334456",,terminal_output +5664,14969569,"jasmine/models/dynamics.py",4007,0,"",python,selection_mouse +5665,14969631,"jasmine/models/dynamics.py",4006,0,"",python,selection_command +5666,14969855,"TERMINAL",0,0,"445567",,terminal_output +5667,14970445,"jasmine/models/dynamics.py",4036,0,"",python,selection_command +5668,14970648,"jasmine/models/dynamics.py",4069,0,"",python,selection_command +5669,14970864,"TERMINAL",0,0,"556678",,terminal_output +5670,14971105,"jasmine/models/dynamics.py",4135,0,"\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,content +5671,14971107,"jasmine/models/dynamics.py",4144,0,"",python,selection_command +5672,14971792,"jasmine/models/dynamics.py",4145,0,"",python,selection_command +5673,14971921,"TERMINAL",0,0,"667789",,terminal_output +5674,14972276,"jasmine/models/dynamics.py",4146,0,"",python,selection_command +5675,14972301,"jasmine/models/dynamics.py",4147,0,"",python,selection_command +5676,14972384,"jasmine/models/dynamics.py",4148,0,"",python,selection_command +5677,14972396,"jasmine/models/dynamics.py",4149,0,"",python,selection_command +5678,14972405,"jasmine/models/dynamics.py",4150,0,"",python,selection_command +5679,14972432,"jasmine/models/dynamics.py",4151,0,"",python,selection_command +5680,14972494,"jasmine/models/dynamics.py",4152,0,"",python,selection_command +5681,14972495,"jasmine/models/dynamics.py",4153,0,"",python,selection_command +5682,14972518,"jasmine/models/dynamics.py",4154,0,"",python,selection_command +5683,14972550,"jasmine/models/dynamics.py",4155,0,"",python,selection_command +5684,14972581,"jasmine/models/dynamics.py",4156,0,"",python,selection_command +5685,14972608,"jasmine/models/dynamics.py",4157,0,"",python,selection_command +5686,14972634,"jasmine/models/dynamics.py",4158,0,"",python,selection_command +5687,14972671,"jasmine/models/dynamics.py",4159,0,"",python,selection_command +5688,14972698,"jasmine/models/dynamics.py",4160,0,"",python,selection_command +5689,14972722,"jasmine/models/dynamics.py",4161,0,"",python,selection_command +5690,14972784,"jasmine/models/dynamics.py",4162,0,"",python,selection_command +5691,14972785,"jasmine/models/dynamics.py",4163,0,"",python,selection_command +5692,14972871,"jasmine/models/dynamics.py",4164,0,"",python,selection_command +5693,14972923,"TERMINAL",0,0,"778898:00",,terminal_output +5694,14973361,"jasmine/models/dynamics.py",4165,0,"",python,selection_command +5695,14973417,"jasmine/models/dynamics.py",4166,0,"",python,selection_command +5696,14973417,"jasmine/models/dynamics.py",4167,0,"",python,selection_command +5697,14973470,"jasmine/models/dynamics.py",4168,0,"",python,selection_command +5698,14973478,"jasmine/models/dynamics.py",4169,0,"",python,selection_command +5699,14973531,"jasmine/models/dynamics.py",4170,0,"",python,selection_command +5700,14973537,"jasmine/models/dynamics.py",4171,0,"",python,selection_command +5701,14973614,"jasmine/models/dynamics.py",4172,0,"",python,selection_command +5702,14973615,"jasmine/models/dynamics.py",4173,0,"",python,selection_command +5703,14973628,"jasmine/models/dynamics.py",4174,0,"",python,selection_command +5704,14973683,"jasmine/models/dynamics.py",4175,0,"",python,selection_command +5705,14973691,"jasmine/models/dynamics.py",4176,0,"",python,selection_command +5706,14973747,"jasmine/models/dynamics.py",4177,0,"",python,selection_command +5707,14973757,"jasmine/models/dynamics.py",4178,0,"",python,selection_command +5708,14973805,"jasmine/models/dynamics.py",4179,0,"",python,selection_command +5709,14973818,"jasmine/models/dynamics.py",4180,0,"",python,selection_command +5710,14973867,"jasmine/models/dynamics.py",4181,0,"",python,selection_command +5711,14973879,"jasmine/models/dynamics.py",4182,0,"",python,selection_command +5712,14973934,"TERMINAL",0,0,"8899201",,terminal_output +5713,14974372,"jasmine/models/dynamics.py",4182,14,"",python,content +5714,14974717,"jasmine/models/dynamics.py",4182,0,"n",python,content +5715,14974718,"jasmine/models/dynamics.py",4183,0,"",python,selection_keyboard +5716,14974868,"jasmine/models/dynamics.py",4183,0,"o",python,content +5717,14974870,"jasmine/models/dynamics.py",4184,0,"",python,selection_keyboard +5718,14974942,"jasmine/models/dynamics.py",4184,0,"i",python,content +5719,14974944,"jasmine/models/dynamics.py",4185,0,"",python,selection_keyboard +5720,14974944,"TERMINAL",0,0,"99202012",,terminal_output +5721,14975004,"jasmine/models/dynamics.py",4185,0,"s",python,content +5722,14975005,"jasmine/models/dynamics.py",4186,0,"",python,selection_keyboard +5723,14975956,"TERMINAL",0,0,"7:00201123",,terminal_output +5724,14976975,"TERMINAL",0,0,"112234",,terminal_output +5725,14977995,"TERMINAL",0,0,"223345",,terminal_output +5726,14978966,"jasmine/models/dynamics.py",4182,4,"noise_augmented_vid_embed_BTNM",python,content +5727,14979018,"TERMINAL",0,0,"334456",,terminal_output +5728,14980043,"TERMINAL",0,0,"445567",,terminal_output +5729,14980269,"jasmine/models/dynamics.py",4185,0,"",python,selection_mouse +5730,14980416,"jasmine/models/dynamics.py",4182,30,"noise_augmented_vid_embed_BTNM",python,selection_mouse +5731,14981129,"TERMINAL",0,0,"556678",,terminal_output +5732,14982107,"TERMINAL",0,0,"667789",,terminal_output +5733,14982495,"jasmine/models/dynamics.py",4169,0,"",python,selection_mouse +5734,14982649,"jasmine/models/dynamics.py",4161,14,"vid_embed_BTNM",python,selection_mouse +5735,14982993,"jasmine/models/dynamics.py",4161,14,"",python,content +5736,14983088,"TERMINAL",0,0,"7788910",,terminal_output +5737,14983233,"jasmine/models/dynamics.py",4161,0,"noise_augmented_vid_embed_BTNM",python,content +5738,14984097,"jasmine/models/dynamics.py",4230,0,"",python,selection_mouse +5739,14984133,"TERMINAL",0,0,"8899301",,terminal_output +5740,14985133,"TERMINAL",0,0,"99303012",,terminal_output +5741,14985364,"jasmine/models/dynamics.py",3531,0,"",python,selection_mouse +5742,14985775,"jasmine/models/dynamics.py",3531,0,"\n ",python,content +5743,14985932,"jasmine/models/dynamics.py",3532,8,"",python,content +5744,14986159,"TERMINAL",0,0,"10301123",,terminal_output +5745,14987183,"TERMINAL",0,0,"112234",,terminal_output +5746,14988198,"TERMINAL",0,0,"223345",,terminal_output +5747,14989203,"TERMINAL",0,0,"334456",,terminal_output +5748,14990034,"jasmine/models/dynamics.py",3015,0,"",python,selection_mouse +5749,14990214,"TERMINAL",0,0,"445567",,terminal_output +5750,14990658,"jasmine/models/dynamics.py",2956,0,"",python,selection_mouse +5751,14991266,"TERMINAL",0,0,"556678",,terminal_output +5752,14991460,"jasmine/models/dynamics.py",2950,0,"",python,selection_mouse +5753,14992261,"jasmine/models/dynamics.py",2973,0,"",python,selection_mouse +5754,14992289,"TERMINAL",0,0,"667789",,terminal_output +5755,14992393,"jasmine/models/dynamics.py",2968,18,"noise_bucket_idx_B",python,selection_mouse +5756,14993050,"jasmine/models/dynamics.py",2958,0,"",python,selection_mouse +5757,14993272,"TERMINAL",0,0,"7788920",,terminal_output +5758,14993582,"jasmine/models/dynamics.py",2963,0,"",python,selection_mouse +5759,14993696,"jasmine/models/dynamics.py",2963,4,"clip",python,selection_mouse +5760,14994338,"TERMINAL",0,0,"8899401",,terminal_output +5761,14995304,"TERMINAL",0,0,"940414123",,terminal_output +5762,14996386,"TERMINAL",0,0,"2112234",,terminal_output +5763,14997342,"TERMINAL",0,0,"223345",,terminal_output +5764,14998357,"TERMINAL",0,0,"334456",,terminal_output +5765,14999460,"TERMINAL",0,0,"445567",,terminal_output +5766,15000392,"TERMINAL",0,0,"556678",,terminal_output +5767,15001505,"TERMINAL",0,0,"667789",,terminal_output +5768,15002222,"TERMINAL",0,0,"srun",,terminal_focus +5769,15002425,"TERMINAL",0,0,"7788930",,terminal_output +5770,15003443,"TERMINAL",0,0,"8899501",,terminal_output +5771,15003871,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +5772,15008266,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",0,0,"",shellscript,tab +5773,15008371,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +5774,15011247,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1961,0,"",shellscript,selection_mouse +5775,15012012,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1962,0,"",shellscript,selection_command +5776,15012448,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1958,4,"",shellscript,content +5777,15012557,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1957,1,"",shellscript,content +5778,15012811,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1951,6,"",shellscript,content +5779,15013258,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1951,0,"p",shellscript,content +5780,15013259,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1952,0,"",shellscript,selection_keyboard +5781,15013600,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1952,0,"r",shellscript,content +5782,15013601,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1953,0,"",shellscript,selection_keyboard +5783,15013675,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1953,0,"i",shellscript,content +5784,15013675,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1954,0,"",shellscript,selection_keyboard +5785,15013778,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1954,0,"n",shellscript,content +5786,15013779,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1955,0,"",shellscript,selection_keyboard +5787,15013868,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1955,0,"t",shellscript,content +5788,15013869,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1956,0,"",shellscript,selection_keyboard +5789,15013993,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1956,0,"-",shellscript,content +5790,15013994,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1957,0,"",shellscript,selection_keyboard +5791,15014208,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1957,0,"e",shellscript,content +5792,15014209,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1958,0,"",shellscript,selection_keyboard +5793,15014366,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1958,0,"v",shellscript,content +5794,15014369,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1959,0,"",shellscript,selection_keyboard +5795,15014510,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1959,0,"e",shellscript,content +5796,15014511,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1960,0,"",shellscript,selection_keyboard +5797,15014621,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1960,0,"r",shellscript,content +5798,15014622,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1961,0,"",shellscript,selection_keyboard +5799,15014835,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1961,0,"y",shellscript,content +5800,15014836,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1962,0,"",shellscript,selection_keyboard +5801,15015092,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1962,0,"t",shellscript,content +5802,15015093,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1963,0,"",shellscript,selection_keyboard +5803,15015431,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1963,0,"h",shellscript,content +5804,15015432,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1964,0,"",shellscript,selection_keyboard +5805,15015540,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1964,0,"i",shellscript,content +5806,15015541,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1965,0,"",shellscript,selection_keyboard +5807,15015631,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1965,0,"n",shellscript,content +5808,15015632,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1966,0,"",shellscript,selection_keyboard +5809,15015687,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1966,0,"g",shellscript,content +5810,15015688,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",1967,0,"",shellscript,selection_keyboard +5811,15016111,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2045,0,"",shellscript,selection_command +5812,15016820,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2043,0,"",shellscript,selection_command +5813,15017160,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2039,4,"",shellscript,content +5814,15017307,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2038,1,"",shellscript,content +5815,15017663,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2032,6,"",shellscript,content +5816,15018028,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2032,0,"p",shellscript,content +5817,15018028,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2033,0,"",shellscript,selection_keyboard +5818,15018178,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2033,0,"r",shellscript,content +5819,15018179,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2034,0,"",shellscript,selection_keyboard +5820,15018275,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2034,0,"i",shellscript,content +5821,15018276,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2035,0,"",shellscript,selection_keyboard +5822,15018351,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2035,0,"n",shellscript,content +5823,15018353,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2036,0,"",shellscript,selection_keyboard +5824,15018439,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2036,0,"t",shellscript,content +5825,15018439,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2037,0,"",shellscript,selection_keyboard +5826,15018587,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2037,0,"-",shellscript,content +5827,15018588,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2038,0,"",shellscript,selection_keyboard +5828,15018782,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2038,0,"e",shellscript,content +5829,15018783,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2039,0,"",shellscript,selection_keyboard +5830,15018937,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2039,0,"v",shellscript,content +5831,15018938,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2040,0,"",shellscript,selection_keyboard +5832,15019079,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2040,0,"e",shellscript,content +5833,15019080,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2041,0,"",shellscript,selection_keyboard +5834,15019192,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2041,0,"r",shellscript,content +5835,15019193,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2042,0,"",shellscript,selection_keyboard +5836,15019407,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2042,0,"y",shellscript,content +5837,15019409,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2043,0,"",shellscript,selection_keyboard +5838,15019618,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2043,0,"t",shellscript,content +5839,15019619,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2044,0,"",shellscript,selection_keyboard +5840,15019776,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2044,0,"i",shellscript,content +5841,15019777,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2045,0,"",shellscript,selection_keyboard +5842,15020107,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2044,1,"",shellscript,content +5843,15020290,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2044,0,"h",shellscript,content +5844,15020290,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2045,0,"",shellscript,selection_keyboard +5845,15020444,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2045,0,"i",shellscript,content +5846,15020444,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2046,0,"",shellscript,selection_keyboard +5847,15020501,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2046,0,"n",shellscript,content +5848,15020502,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2047,0,"",shellscript,selection_keyboard +5849,15020652,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2047,0,"g",shellscript,content +5850,15020652,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",2048,0,"",shellscript,selection_keyboard +5851,15033046,"TERMINAL",0,0,"s",,terminal_output +5852,15033710,"TERMINAL",0,0,"",,terminal_output +5853,15034786,"TERMINAL",0,0,"s",,terminal_output +5854,15034990,"TERMINAL",0,0,"y",,terminal_output +5855,15035052,"TERMINAL",0,0,"n",,terminal_output +5856,15035240,"TERMINAL",0,0,"c",,terminal_output +5857,15035437,"TERMINAL",0,0,"-",,terminal_output +5858,15035602,"TERMINAL",0,0,"r",,terminal_output +5859,15035682,"TERMINAL",0,0,"u",,terminal_output +5860,15035816,"TERMINAL",0,0,"n",,terminal_output +5861,15035898,"TERMINAL",0,0,"n",,terminal_output +5862,15035960,"TERMINAL",0,0,"e",,terminal_output +5863,15036020,"TERMINAL",0,0,"r",,terminal_output +5864,15036178,"TERMINAL",0,0,"- \r",,terminal_output +5865,15036239,"TERMINAL",0,0,"2",,terminal_output +5866,15036572,"TERMINAL",0,0,"\r\n[?2004l\rsending incremental file list\r\n",,terminal_output +5867,15036679,"TERMINAL",0,0,"jasmine/train_dynamics.py\r\njasmine/models/dynamics.py\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/\r\nslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh\r\n",,terminal_output +5868,15036755,"TERMINAL",0,0,"\r\nsent 76,161 bytes received 279 bytes 152,880.00 bytes/sec\r\ntotal size is 28,034,433 speedup is 366.75\r\n]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +5869,15037159,"TERMINAL",0,0,"r",,terminal_output +5870,15037305,"TERMINAL",0,0,"un",,terminal_output +5871,15037463,"TERMINAL",0,0,"n",,terminal_output +5872,15037525,"TERMINAL",0,0,"e",,terminal_output +5873,15037593,"TERMINAL",0,0,"r",,terminal_output +5874,15037697,"TERMINAL",0,0,"-",,terminal_output +5875,15037843,"TERMINAL",0,0,"2",,terminal_output +5876,15038214,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_2[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +5877,15038800,"TERMINAL",0,0,"s",,terminal_output +5878,15038960,"TERMINAL",0,0,"b",,terminal_output +5879,15039077,"TERMINAL",0,0,"a",,terminal_output +5880,15039137,"TERMINAL",0,0,"t",,terminal_output +5881,15039267,"TERMINAL",0,0,"c \r",,terminal_output +5882,15039329,"TERMINAL",0,0,"h",,terminal_output +5883,15039456,"TERMINAL",0,0," ",,terminal_output +5884,15039794,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh",,terminal_output +5885,15040700,"TERMINAL",0,0,"\rslurm/jobs/mihir/horeka/breakout/noise_schedule_runs/mixed_prec/train_dyn_single_gpu_50k.sh\r\n[?2004l\rSubmitted batch job 3527927\r\n]0;tum_cte0515@hkn0402:~/Projects/jasmine_jobs_2[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine_jobs_2]$ ",,terminal_output +5886,15042491,"TERMINAL",0,0,"d",,terminal_output +5887,15042974,"TERMINAL",0,0,"e",,terminal_output +5888,15043089,"TERMINAL",0,0,"v",,terminal_output +5889,15043156,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +5890,15043626,"TERMINAL",0,0,"q",,terminal_output +5891,15043706,"TERMINAL",0,0,"u",,terminal_output +5892,15043766,"TERMINAL",0,0,"e",,terminal_output +5893,15043827,"TERMINAL",0,0,"u",,terminal_output +5894,15043887,"TERMINAL",0,0,"e",,terminal_output +5895,15044063,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hhkn0402.localdomain: Sun Sep 28 17:38:08 2025JOBID PARTITION NAME USER STTIME NODES NODELIST(REASON)3527927 accelerat train_dy tum_cte0 PD0:001 (Priority)3527920 accelerat train_dy tum_cte0 R3:281 hkn05163527904 accelerat train_dy tum_cte0 R11:291 hkn04293527905 accelerat train_dy tum_cte0 R11:291 hkn05343527891 accelerat train_dy tum_cte0 R15:301 hkn07303527866 dev_accel interact tum_cte0 R49:111 hkn0402",,terminal_output +5896,15045060,"TERMINAL",0,0,"99303012",,terminal_output +5897,15045661,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn0402.localdomain: Sun Sep 28 17:38:10 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3527927 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3527920 accelerat train_dy tum_cte0 R\t3:30\t 1 hkn05163527904 accelerat train_dy tum_cte0 R11:31\t 1 hkn04293527905 accelerat train_dy tum_cte0 R11:31\t 1 hkn05343527891 accelerat train_dy tum_cte0 R15:32\t 1 hkn07303527866 dev_accel interact tum_cte0 R49:13\t 1 hkn0402",,terminal_output +5898,15046676,"TERMINAL",0,0,"112234",,terminal_output +5899,15047792,"TERMINAL",0,0,"223345",,terminal_output +5900,15048711,"TERMINAL",0,0,"334456",,terminal_output +5901,15049723,"TERMINAL",0,0,"445567",,terminal_output +5902,15050740,"TERMINAL",0,0,"556678",,terminal_output +5903,15051759,"TERMINAL",0,0,"667789",,terminal_output +5904,15052775,"TERMINAL",0,0,"7788920",,terminal_output +5905,15053831,"TERMINAL",0,0,"8899401",,terminal_output +5906,15054805,"TERMINAL",0,0,"99404012",,terminal_output +5907,15055879,"TERMINAL",0,0,"20401123",,terminal_output +5908,15056836,"TERMINAL",0,0,"112234",,terminal_output +5909,15057927,"TERMINAL",0,0,"223345",,terminal_output +5910,15058953,"TERMINAL",0,0,"334456",,terminal_output +5911,15059979,"TERMINAL",0,0,"445567",,terminal_output +5912,15060910,"TERMINAL",0,0,"556678",,terminal_output +5913,15061928,"TERMINAL",0,0,"667789",,terminal_output +5914,15062948,"TERMINAL",0,0,"7788930",,terminal_output +5915,15063961,"TERMINAL",0,0,"8899501",,terminal_output +5916,15064976,"TERMINAL",0,0,"99505012",,terminal_output +5917,15065992,"TERMINAL",0,0,"30501123",,terminal_output +5918,15067043,"TERMINAL",0,0,"112234",,terminal_output +5919,15068030,"TERMINAL",0,0,"223345",,terminal_output +5920,15069048,"TERMINAL",0,0,"334456",,terminal_output +5921,15070116,"TERMINAL",0,0,"445567",,terminal_output +5922,15071084,"TERMINAL",0,0,"556678",,terminal_output +5923,15072102,"TERMINAL",0,0,"667789",,terminal_output +5924,15073122,"TERMINAL",0,0,"7788940",,terminal_output +5925,15074211,"TERMINAL",0,0,"88996:001",,terminal_output +5926,15075236,"TERMINAL",0,0,"992:002:0012",,terminal_output +5927,15076370,"TERMINAL",0,0,"40 Rhkn05134:001123",,terminal_output +5928,15077322,"TERMINAL",0,0,"1223345",,terminal_output +5929,15078410,"TERMINAL",0,0,"3334456",,terminal_output +5930,15079340,"TERMINAL",0,0,"4445567",,terminal_output +5931,15080458,"TERMINAL",0,0,"5556678",,terminal_output +5932,15081373,"TERMINAL",0,0,"6667789",,terminal_output +5933,15082387,"TERMINAL",0,0,"77788950",,terminal_output +5934,15083405,"TERMINAL",0,0,"88899101",,terminal_output +5935,15084424,"TERMINAL",0,0,"999101012",,terminal_output +5936,15085440,"TERMINAL",0,0,"5010101123",,terminal_output +5937,15086499,"TERMINAL",0,0,"1112234",,terminal_output +5938,15087525,"TERMINAL",0,0,"2223345",,terminal_output +5939,15088547,"TERMINAL",0,0,"3334456",,terminal_output +5940,15089529,"TERMINAL",0,0,"4445567",,terminal_output +5941,15090530,"TERMINAL",0,0,"5556678",,terminal_output +5942,15091824,"TERMINAL",0,0,"6667789",,terminal_output +5943,15092832,"TERMINAL",0,0,"77788950:00",,terminal_output +5944,15093807,"TERMINAL",0,0,"88899201",,terminal_output +5945,15094825,"TERMINAL",0,0,"999202012",,terminal_output +5946,15095922,"TERMINAL",0,0,"9:0020201123",,terminal_output +5947,15096880,"TERMINAL",0,0,"1112234",,terminal_output +5948,15097872,"TERMINAL",0,0,"2223345",,terminal_output +5949,15098892,"TERMINAL",0,0,"3334456",,terminal_output +5950,15099906,"TERMINAL",0,0,"4445567",,terminal_output +5951,15100924,"TERMINAL",0,0,"5556678",,terminal_output +5952,15101962,"TERMINAL",0,0,"6667789",,terminal_output +5953,15102986,"TERMINAL",0,0,"77788910",,terminal_output +5954,15103975,"TERMINAL",0,0,"88899301",,terminal_output +5955,15104988,"TERMINAL",0,0,"999303012",,terminal_output +5956,15106004,"TERMINAL",0,0,"1030301123",,terminal_output +5957,15107021,"TERMINAL",0,0,"1112234",,terminal_output +5958,15108042,"TERMINAL",0,0,"2223345",,terminal_output +5959,15109061,"TERMINAL",0,0,"3334456",,terminal_output +5960,15110074,"TERMINAL",0,0,"4445567",,terminal_output +5961,15111092,"TERMINAL",0,0,"5556678",,terminal_output +5962,15112110,"TERMINAL",0,0,"6667789",,terminal_output +5963,15113127,"TERMINAL",0,0,"77788920",,terminal_output +5964,15114250,"TERMINAL",0,0,"88899401",,terminal_output +5965,15115171,"TERMINAL",0,0,"999404012",,terminal_output +5966,15116193,"TERMINAL",0,0,"2040401123",,terminal_output +5967,15117197,"TERMINAL",0,0,"1112234",,terminal_output +5968,15118241,"TERMINAL",0,0,"2223345",,terminal_output +5969,15119265,"TERMINAL",0,0,"3334456",,terminal_output +5970,15120289,"TERMINAL",0,0,"4445567",,terminal_output +5971,15121313,"TERMINAL",0,0,"5556678",,terminal_output +5972,15122287,"TERMINAL",0,0,"6667789",,terminal_output +5973,15123363,"TERMINAL",0,0,"788995031",,terminal_output +5974,15124386,"TERMINAL",0,0,"999505012",,terminal_output +5975,15125413,"TERMINAL",0,0,"3050501123",,terminal_output +5976,15126436,"TERMINAL",0,0,"1112234",,terminal_output +5977,15127382,"TERMINAL",0,0,"2223345",,terminal_output +5978,15128485,"TERMINAL",0,0,"3334456",,terminal_output +5979,15129507,"TERMINAL",0,0,"4445567",,terminal_output +5980,15130530,"TERMINAL",0,0,"5556678",,terminal_output +5981,15131459,"TERMINAL",0,0,"6667789",,terminal_output +5982,15132476,"TERMINAL",0,0,"77788940",,terminal_output +5983,15133500,"TERMINAL",0,0,"888997:001",,terminal_output +5984,15134523,"TERMINAL",0,0,"9993:003:0012",,terminal_output +5985,15135549,"TERMINAL",0,0,"401:005:001123",,terminal_output +5986,15136572,"TERMINAL",0,0,"1112234",,terminal_output +5987,15137600,"TERMINAL",0,0,"2223345",,terminal_output +5988,15138567,"TERMINAL",0,0,"3334456",,terminal_output +5989,15139590,"TERMINAL",0,0,"4445567",,terminal_output +5990,15140606,"TERMINAL",0,0,"5556678",,terminal_output +5991,15141624,"TERMINAL",0,0,"6667789",,terminal_output +5992,15142643,"TERMINAL",0,0,"77788950",,terminal_output +5993,15143739,"TERMINAL",0,0,"88899101",,terminal_output +5994,15144678,"TERMINAL",0,0,"999101012",,terminal_output +5995,15145696,"TERMINAL",0,0,"5010101123",,terminal_output +5996,15146715,"TERMINAL",0,0,"1112234",,terminal_output +5997,15147741,"TERMINAL",0,0,"2223345",,terminal_output +5998,15148755,"TERMINAL",0,0,"3334456",,terminal_output +5999,15149777,"TERMINAL",0,0,"4445567",,terminal_output +6000,15150795,"TERMINAL",0,0,"5556678",,terminal_output +6001,15151807,"TERMINAL",0,0,"6667789",,terminal_output +6002,15152857,"TERMINAL",0,0,"7778891:00",,terminal_output +6003,15153842,"TERMINAL",0,0,"88899201",,terminal_output +6004,15154862,"TERMINAL",0,0,"999202012",,terminal_output +6005,15155931,"TERMINAL",0,0,"40:0020201123",,terminal_output +6006,15156927,"TERMINAL",0,0,"1112234",,terminal_output +6007,15157973,"TERMINAL",0,0,"2223345",,terminal_output +6008,15158954,"TERMINAL",0,0,"3334456",,terminal_output +6009,15159971,"TERMINAL",0,0,"4445567",,terminal_output +6010,15160992,"TERMINAL",0,0,"5556678",,terminal_output +6011,15162009,"TERMINAL",0,0,"6667789",,terminal_output +6012,15163029,"TERMINAL",0,0,"77788910",,terminal_output +6013,15164045,"TERMINAL",0,0,"88899301",,terminal_output +6014,15165065,"TERMINAL",0,0,"999303012",,terminal_output +6015,15166083,"TERMINAL",0,0,"1030301123",,terminal_output +6016,15167106,"TERMINAL",0,0,"1112234",,terminal_output +6017,15168119,"TERMINAL",0,0,"2223345",,terminal_output +6018,15169239,"TERMINAL",0,0,"3334456",,terminal_output +6019,15170158,"TERMINAL",0,0,"4445567",,terminal_output +6020,15171174,"TERMINAL",0,0,"5556678",,terminal_output +6021,15172209,"TERMINAL",0,0,"6667789",,terminal_output +6022,15173230,"TERMINAL",0,0,"77788920",,terminal_output +6023,15174254,"TERMINAL",0,0,"88899401",,terminal_output +6024,15175279,"TERMINAL",0,0,"999404012",,terminal_output +6025,15176303,"TERMINAL",0,0,"2040401123",,terminal_output +6026,15177283,"TERMINAL",0,0,"1112234",,terminal_output +6027,15178350,"TERMINAL",0,0,"2334456",,terminal_output +6028,15179374,"TERMINAL",0,0,"4445567",,terminal_output +6029,15180398,"TERMINAL",0,0,"5556678",,terminal_output +6030,15181422,"TERMINAL",0,0,"6667789",,terminal_output +6031,15182447,"TERMINAL",0,0,"77788930",,terminal_output +6032,15183472,"TERMINAL",0,0,"88899501",,terminal_output +6033,15184495,"TERMINAL",0,0,"999505012",,terminal_output +6034,15185425,"TERMINAL",0,0,"3050501123",,terminal_output +6035,15186442,"TERMINAL",0,0,"1112234",,terminal_output +6036,15187463,"TERMINAL",0,0,"2223345",,terminal_output +6037,15188492,"TERMINAL",0,0,"3334456",,terminal_output +6038,15189498,"TERMINAL",0,0,"4445567",,terminal_output +6039,15190538,"TERMINAL",0,0,"5556678",,terminal_output +6040,15191564,"TERMINAL",0,0,"6667789",,terminal_output +6041,15192585,"TERMINAL",0,0,"77788940",,terminal_output +6042,15193566,"TERMINAL",0,0,"888998:001",,terminal_output +6043,15194586,"TERMINAL",0,0,"9994:004:0012",,terminal_output +6044,15195656,"TERMINAL",0,0,"402:006:001123",,terminal_output +6045,15196689,"TERMINAL",0,0,"1112234",,terminal_output +6046,15197707,"TERMINAL",0,0,"2223345",,terminal_output +6047,15198730,"TERMINAL",0,0,"3334456",,terminal_output +6048,15199755,"TERMINAL",0,0,"4445567",,terminal_output +6049,15200692,"TERMINAL",0,0,"5556678",,terminal_output +6050,15201708,"TERMINAL",0,0,"6667789",,terminal_output +6051,15202839,"TERMINAL",0,0,"77788950",,terminal_output +6052,15203850,"TERMINAL",0,0,"88899101",,terminal_output +6053,15204764,"TERMINAL",0,0,"999101012",,terminal_output +6054,15205783,"TERMINAL",0,0,"5010101123",,terminal_output +6055,15206821,"TERMINAL",0,0,"1112234",,terminal_output +6056,15207844,"TERMINAL",0,0,"2223345",,terminal_output +6057,15208841,"TERMINAL",0,0,"3334456",,terminal_output +6058,15209892,"TERMINAL",0,0,"4445567",,terminal_output +6059,15210924,"TERMINAL",0,0,"5556678",,terminal_output +6060,15211922,"TERMINAL",0,0,"6667789",,terminal_output +6061,15212963,"TERMINAL",0,0,"7778892:00",,terminal_output +6062,15213951,"TERMINAL",0,0,"88899201",,terminal_output +6063,15215018,"TERMINAL",0,0,"999202012",,terminal_output +6064,15215971,"TERMINAL",0,0,"1:0020201123",,terminal_output +6065,15216345,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +6066,15293472,"TERMINAL",0,0,"p",,terminal_output +6067,15293533,"TERMINAL",0,0,"y",,terminal_output +6068,15293717,"TERMINAL",0,0,"t",,terminal_output +6069,15293775,"TERMINAL",0,0,"h",,terminal_output +6070,15293944,"TERMINAL",0,0,"o",,terminal_output +6071,15294062,"TERMINAL",0,0,"n",,terminal_output +6072,15294125,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +6073,15294182,"TERMINAL",0,0,"Python 3.12.1 (main, Aug 20 2025, 00:00:00) [GCC 11.4.1 20231218 (Red Hat 11.4.1-4)] on linux\r\nType ""help"", ""copyright"", ""credits"" or ""license"" for more information.\r\n>>> ",,terminal_output +6074,15294997,"TERMINAL",0,0,"i",,terminal_output +6075,15295085,"TERMINAL",0,0,"m",,terminal_output +6076,15295288,"TERMINAL",0,0,"po",,terminal_output +6077,15295457,"TERMINAL",0,0,"r",,terminal_output +6078,15295688,"TERMINAL",0,0,"t ",,terminal_output +6079,15295967,"TERMINAL",0,0,"ja",,terminal_output +6080,15296252,"TERMINAL",0,0,"x",,terminal_output +6081,15296400,"TERMINAL",0,0,"\r\n",,terminal_output +6082,15296824,"TERMINAL",0,0,">>> ",,terminal_output +6083,15301647,"TERMINAL",0,0,"jax.numpy.asarray(0.69999999).astype(jax.numpy.float32)/0.7",,terminal_output +6084,15301983,"TERMINAL",0,0,"\r\n",,terminal_output +6085,15302287,"TERMINAL",0,0,"ERROR:2025-09-28 17:42:26,903:jax._src.xla_bridge:487: Jax plugin configuration error: Exception when calling jax_plugins.xla_cuda12.initialize()\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax_plugins/xla_cuda12/__init__.py"", line 201, in _version_check\r\n version = get_version()\r\n ^^^^^^^^^^^^^\r\nRuntimeError: jaxlib/cuda/versions_helpers.cc:81: operation cusparseGetProperty(MAJOR_VERSION, &major) failed: The cuSPARSE library was not found.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 485, in discover_pjrt_plugins\r\n plugin_module.initialize()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax_plugins/xla_cuda12/__init__.py"", line 328, in initialize\r\n _check_cuda_versions(raise_on_first_error=True)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax_plugins/xla_cuda12/__init__.py"", line 266, in _check_cuda_versions\r\n _version_check(""cuSPARSE"", cuda_versions.cusparse_get_version,\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax_plugins/xla_cuda12/__init__.py"", line 205, in _version_check\r\n raise RuntimeError(err_msg) from e\r\nRuntimeError: Unable to load cuSPARSE. Is it installed?\r\nWARNING:2025-09-28 17:42:26,913:jax._src.xla_bridge:864: An NVIDIA GPU may be present on this machine, but a CUDA-enabled jaxlib is not installed. Falling back to cpu.\r\nArray(1., dtype=float32)\r\n>>> ",,terminal_output +6086,15311031,"TERMINAL",0,0,"jax.numpy.asarray(0.69999999).astype(jax.numpy.float32)/0.7",,terminal_output +6087,15311834,"TERMINAL",0,0,"",,terminal_output +6088,15311971,"TERMINAL",0,0,"",,terminal_output +6089,15312167,"TERMINAL",0,0,"",,terminal_output +6090,15312323,"TERMINAL",0,0,"",,terminal_output +6091,15312437,"TERMINAL",0,0,"",,terminal_output +6092,15312640,"TERMINAL",0,0,"",,terminal_output +6093,15312881,"TERMINAL",0,0,"",,terminal_output +6094,15313280,"TERMINAL",0,0,"",,terminal_output +6095,15313775,"TERMINAL",0,0,").astype(jax.numpy.float32)/0.7",,terminal_output +6096,15313934,"TERMINAL",0,0,"\r\nArray(0.99999994, dtype=float32)\r\n>>> ",,terminal_output +6097,15320177,"TERMINAL",0,0,"jax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7",,terminal_output +6098,15321257,"TERMINAL",0,0,"",,terminal_output +6099,15322005,"TERMINAL",0,0,"",,terminal_output +6100,15322714,"TERMINAL",0,0,"jax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6101,15324747,"TERMINAL",0,0,"ajax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6102,15325024,"TERMINAL",0,0,"xjax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6103,15325260,"TERMINAL",0,0,"-jax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6104,15325849,"TERMINAL",0,0,"jax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6105,15326100,"TERMINAL",0,0,".jax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6106,15326363,"TERMINAL",0,0,"njax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6107,15326504,"TERMINAL",0,0,"ujax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6108,15326690,"TERMINAL",0,0,"mjax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6109,15326860,"TERMINAL",0,0,"pjax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6110,15326979,"TERMINAL",0,0,"yjax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6111,15327095,"TERMINAL",0,0,".jax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6112,15327554,"TERMINAL",0,0,"fljax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6113,15327723,"TERMINAL",0,0,"ojax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6114,15327854,"TERMINAL",0,0,"ojax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6115,15327915,"TERMINAL",0,0,"rjax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6116,15328450,"TERMINAL",0,0,"(jax.numpy.asarray(0.6999999).astype(jax.numpy.float32)/0.7\r",,terminal_output +6117,15328924,"TERMINAL",0,0,"",,terminal_output +6118,15329658,"TERMINAL",0,0,"",,terminal_output +6119,15330419,"TERMINAL",0,0,")",,terminal_output +6120,15330654,"TERMINAL",0,0,"\r\nArray(0., dtype=float32)\r\n>>> ",,terminal_output +6121,15405208,"TERMINAL",0,0,"\r\n",,terminal_output +6122,15405291,"TERMINAL",0,0,"]0;tum_cte0515@hkn0402:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0402 jasmine]$ ",,terminal_output +6123,15703369,"TERMINAL",0,0,"salloc: Job 3527866 has exceeded its time limit and its allocation has been revoked.\nslurmstepd: error: *** STEP 3527866.interactive ON hkn0402 CANCELLED AT 2025-09-28T17:49:08 DUE TO TIME LIMIT ***\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\n",,terminal_output +6124,15733026,"TERMINAL",0,0,"srun: error: hkn0402: task 0: Killed\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +6125,16211852,"jasmine/train_dynamics.py",0,0,"",python,tab +6126,16212535,"jasmine/models/dynamics.py",0,0,"",python,tab +6127,16214166,"jasmine/models/dynamics.py",3566,0,"",python,selection_mouse +6128,16214465,"jasmine/models/dynamics.py",3558,13,"noise_level_B",python,selection_mouse +6129,16219443,"jasmine/models/dynamics.py",4036,0,"",python,selection_mouse +6130,16219446,"jasmine/models/dynamics.py",4035,0,"",python,selection_command +6131,16222304,"jasmine/models/dynamics.py",4030,0,"",python,selection_mouse +6132,16223266,"jasmine/models/dynamics.py",4036,0,"",python,selection_mouse +6133,16223267,"jasmine/models/dynamics.py",4035,0,"",python,selection_command +6134,16223391,"jasmine/models/dynamics.py",4035,1,")",python,selection_mouse +6135,16223392,"jasmine/models/dynamics.py",4036,0,"",python,selection_command +6136,16223456,"jasmine/models/dynamics.py",4031,5,"BTNM)",python,selection_mouse +6137,16223472,"jasmine/models/dynamics.py",3958,78,"{}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6138,16223472,"jasmine/models/dynamics.py",3953,83,"ise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6139,16223494,"jasmine/models/dynamics.py",3949,87,"t_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6140,16223545,"jasmine/models/dynamics.py",3948,88,"rt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6141,16223546,"jasmine/models/dynamics.py",3884,152,"sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6142,16223547,"jasmine/models/dynamics.py",3882,154,"(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6143,16223559,"jasmine/models/dynamics.py",3880,156,"nt(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6144,16223576,"jasmine/models/dynamics.py",3879,157,"int(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6145,16223633,"jasmine/models/dynamics.py",3814,222,"rint(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6146,16223633,"jasmine/models/dynamics.py",3813,223,"print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6147,16223634,"jasmine/models/dynamics.py",3811,225,"g.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6148,16223655,"jasmine/models/dynamics.py",3731,305,"bug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6149,16223674,"jasmine/models/dynamics.py",3730,306,"ebug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6150,16223733,"jasmine/models/dynamics.py",3729,307,"debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6151,16223734,"jasmine/models/dynamics.py",3658,378,".debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6152,16223734,"jasmine/models/dynamics.py",3656,380,"ax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6153,16223755,"jasmine/models/dynamics.py",3655,381,"jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6154,16223813,"jasmine/models/dynamics.py",3654,382," jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6155,16223874,"jasmine/models/dynamics.py",3653,383," jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6156,16223894,"jasmine/models/dynamics.py",3599,437," jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6157,16223907,"jasmine/models/dynamics.py",3598,438," jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6158,16223924,"jasmine/models/dynamics.py",3597,439," jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6159,16223982,"jasmine/models/dynamics.py",3596,440," jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6160,16223988,"jasmine/models/dynamics.py",3535,501," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6161,16224025,"jasmine/models/dynamics.py",3534,502," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6162,16225165,"jasmine/models/dynamics.py",3534,0,"",python,selection_mouse +6163,16225166,"jasmine/models/dynamics.py",3533,8," ",python,selection_mouse +6164,16225400,"jasmine/models/dynamics.py",3533,62," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n ",python,selection_mouse +6165,16225430,"jasmine/models/dynamics.py",3533,63," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n ",python,selection_mouse +6166,16225492,"jasmine/models/dynamics.py",3533,64," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n ",python,selection_mouse +6167,16225493,"jasmine/models/dynamics.py",3533,65," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n ",python,selection_mouse +6168,16225502,"jasmine/models/dynamics.py",3533,66," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n ",python,selection_mouse +6169,16225517,"jasmine/models/dynamics.py",3533,121," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n ",python,selection_mouse +6170,16225535,"jasmine/models/dynamics.py",3533,122," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n ",python,selection_mouse +6171,16225586,"jasmine/models/dynamics.py",3533,125," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax",python,selection_mouse +6172,16225586,"jasmine/models/dynamics.py",3533,131," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug",python,selection_mouse +6173,16225601,"jasmine/models/dynamics.py",3533,201," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug",python,selection_mouse +6174,16225623,"jasmine/models/dynamics.py",3533,202," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.",python,selection_mouse +6175,16225676,"jasmine/models/dynamics.py",3533,207," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print",python,selection_mouse +6176,16225676,"jasmine/models/dynamics.py",3533,208," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(",python,selection_mouse +6177,16225676,"jasmine/models/dynamics.py",3533,231," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M",python,selection_mouse +6178,16225685,"jasmine/models/dynamics.py",3533,262," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n",python,selection_mouse +6179,16225700,"jasmine/models/dynamics.py",3533,302," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise",python,selection_mouse +6180,16225783,"jasmine/models/dynamics.py",3533,366," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus:",python,selection_mouse +6181,16225803,"jasmine/models/dynamics.py",3533,368," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {",python,selection_mouse +6182,16225833,"jasmine/models/dynamics.py",3533,369," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}",python,selection_mouse +6183,16225886,"jasmine/models/dynamics.py",3533,370," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}""",python,selection_mouse +6184,16225887,"jasmine/models/dynamics.py",3533,371," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"",",python,selection_mouse +6185,16225887,"jasmine/models/dynamics.py",3533,372," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", ",python,selection_mouse +6186,16225946,"jasmine/models/dynamics.py",3533,386," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus",python,selection_mouse +6187,16226216,"jasmine/models/dynamics.py",3533,440," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise",python,selection_mouse +6188,16226247,"jasmine/models/dynamics.py",3533,441," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)",python,selection_mouse +6189,16226734,"jasmine/models/dynamics.py",3533,503," jax.debug.print(""noise_level_B: {}"", noise_level_B)\n jax.debug.print(""noise_BTNM: {}"", noise_BTNM)\n jax.debug.print(""noise_bucket_idx_B: {}"", noise_bucket_idx_B)\n jax.debug.print(""noise_level_embed_B11M: {}"", noise_level_embed_B11M)\n jax.debug.print(""one_minus_noise: {}"", one_minus_noise)\n jax.debug.print(""sqrt_one_minus: {}"", sqrt_one_minus)\n jax.debug.print(""sqrt_noise: {}"", sqrt_noise)\n jax.debug.print(""vid_embed_BTNM: {}"", vid_embed_BTNM)",python,selection_mouse +6190,16295655,"jasmine/models/dynamics.py",4272,0,"",python,selection_mouse +6191,16296247,"jasmine/models/dynamics.py",3907,0,"",python,selection_mouse +6192,16296733,"jasmine/models/dynamics.py",4023,0,"",python,selection_mouse diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-04d29a91-f2bd-4b62-89da-678008b090861755160613254-2025_08_14-10.37.33.283/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-04d29a91-f2bd-4b62-89da-678008b090861755160613254-2025_08_14-10.37.33.283/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..111db95a0c01a6b88ebeffe20e6433f7d44ebce5 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-04d29a91-f2bd-4b62-89da-678008b090861755160613254-2025_08_14-10.37.33.283/source.csv @@ -0,0 +1,10125 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +1,4,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining/%x_%j.log\n#SBATCH --job-name=train_dynamics_causal_8_node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/holiday/causal/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\n# tokenizer with the new structure supporting larger ffn_dim\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401\n\nenv | grep SLURM\n\n# export PYTHONUNBUFFERED=1\n\nsrun python train_dynamics.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --dyna_type=causal \\n --num_latent_actions=100 \\n --darkness_threshold=50 \\n --max_lr=8e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-causal-8-node-darkness-filter-$slurm_job_id \\n --tags dynamics causal 8-node post-launch-main darkness-filter \\n --entity instant-uv \\n --project jafar \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16 \\n --dyna_ffn_dim=4096 \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +2,723,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:37:33 AM [info] Activating crowd-code\n10:37:33 AM [info] Recording started\n10:37:33 AM [info] Initializing git provider using file system watchers...\n10:37:33 AM [info] Git repository found\n10:37:33 AM [info] Git provider initialized successfully\n10:37:33 AM [info] Initial git state: [object Object]\n",Log,tab +3,1451,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +4,2714,"TERMINAL",0,0,"queue",,terminal_command +5,2778,"TERMINAL",0,0,"]633;E;2025-08-14 10:37:35 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 10:37:35 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-16:37:34\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 20:48:39\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]",,terminal_output +6,3845,"TERMINAL",0,0,"6641",,terminal_output +7,4886,"TERMINAL",0,0,"872",,terminal_output +8,5937,"TERMINAL",0,0,"983",,terminal_output +9,6973,"TERMINAL",0,0,"4094",,terminal_output +10,7378,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 10:37:40 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-16:37:39\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 20:48:44\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]",,terminal_output +11,8474,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 10:37:41 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-16:37:40\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 20:48:45\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]",,terminal_output +12,8602,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 10:37:41 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-16:37:40\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 20:48:45\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]",,terminal_output +13,9315,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +14,10786,"TERMINAL",0,0,"logs",,terminal_command +15,12854,"TERMINAL",0,0,"ls",,terminal_command +16,12896,"TERMINAL",0,0,"]633;E;2025-08-14 10:37:46 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +17,13251,"TERMINAL",0,0,"atari train_lam_action_space_scaling_50_3329804.log\r\nbig_run train_lam_action_space_scaling_50_3331286.log\r\nbig-runs train_lam_action_space_scaling_6_3318549.log\r\ncausal train_lam_action_space_scaling_6_3320178.log\r\ncoinrun train_lam_action_space_scaling_6_3321528.log\r\nmaskgit train_lam_action_space_scaling_6_3329790.log\r\nmaskgit-maskprob-fix train_lam_action_space_scaling_6_3329805.log\r\npreprocess train_lam_action_space_scaling_6_3331287.log\r\ntrain_dyn_causal_180M_3372931.log train_lam_action_space_scaling_8_3318550.log\r\ntrain_dyn_causal_180M_3372963.log train_lam_action_space_scaling_8_3329791.log\r\ntrain_dyn_causal_180M_3372969.log train_lam_action_space_scaling_8_3329806.log\r\ntrain_dyn_causal_180M_3373107.log train_lam_action_space_scaling_8_3331288.log\r\ntrain_dyn_causal_255M_3372932.log train_lam_minecraft_overfit_sample_3309655.log\r\ntrain_dyn_causal_255M_3372970.log train_lam_model_size_scaling_38M_3317098.log\r\ntrain_dyn_causal_255M_3373108.log train_lam_model_size_scaling_38M_3317115.log\r\ntrain_dyn_causal_356M_3372934.log train_lam_model_size_scaling_38M_3317231.log\r\ntrain_dyn_causal_356M_3372971.log train_tokenizer_batch_size_scaling_16_node_3321526.log\r\ntrain_dyn_causal_356M_3373109.log train_tokenizer_batch_size_scaling_1_node_3318551.log\r\ntrain_dyn_causal_500M_3372936.log train_tokenizer_batch_size_scaling_2_node_3318552.log\r\ntrain_dyn_causal_500M_3372972.log train_tokenizer_batch_size_scaling_2_node_3330806.log\r\ntrain_dyn_causal_500M_3373110.log train_tokenizer_batch_size_scaling_2_node_3330848.log\r\ntrain_dyn_new_arch-bugfixed-spatial-shift_3359343.log train_tokenizer_batch_size_scaling_2_node_3331282.log\r\ntrain_dyn_new_arch-bugfixed-temporal-shift_3359349.log train_tokenizer_batch_size_scaling_4_node_3318553.log\r\ntrain_dyn_yolorun_3333026.log train_tokenizer_batch_size_scaling_4_node_3320175.log\r\ntrain_dyn_yolorun_3333448.log train_tokenizer_batch_size_scaling_4_node_3321524.log\r\ntrain_dyn_yolorun_3335345.log train_tokenizer_batch_size_scaling_8_node_3320176.log\r\ntrain_dyn_yolorun_3335362.log train_tokenizer_batch_size_scaling_8_node_3321525.log\r\ntrain_dyn_yolorun_3348592.log train_tokenizer_minecraft_overfit_sample_3309656.log\r\ntrain_dyn_yolorun_new_arch_3351743.log train_tokenizer_model_size_scaling_127M_3317233.log\r\ntrain_dyn_yolorun_new_arch_3352103.log train_tokenizer_model_size_scaling_127M_3318554.log\r\ntrain_dyn_yolorun_new_arch_3352115.log train_tokenizer_model_size_scaling_140M_3313562.log\r\ntrain_dyn_yolorun_new_arch_3358457.log train_tokenizer_model_size_scaling_140M_3316019.log\r\ntrain_lam_action_space_scaling_10_3320179.log train_tokenizer_model_size_scaling_200M_3313563.log\r\ntrain_lam_action_space_scaling_10_3321529.log train_tokenizer_model_size_scaling_200M_3316020.log\r\ntrain_lam_action_space_scaling_10_3329786.log train_tokenizer_model_size_scaling_227M_3317234.log\r\ntrain_lam_action_space_scaling_10_3329801.log train_tokenizer_model_size_scaling_227M_3318555.log\r\ntrain_lam_action_space_scaling_10_3331283.log train_tokenizer_model_size_scaling_227M_3320173.log\r\ntrain_lam_action_space_scaling_12_3318546.log train_tokenizer_model_size_scaling_227M_3321523.log\r\ntrain_lam_action_space_scaling_12_3320177.log train_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_lam_action_space_scaling_12_3321527.log train_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_lam_action_space_scaling_12_3329787.log train_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_lam_action_space_scaling_12_3329802.log train_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_lam_action_space_scaling_12_3331284.log train_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_lam_action_space_scaling_20_3318547.log train_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_lam_action_space_scaling_20_3329788.log train_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_lam_action_space_scaling_20_3329803.log train_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_lam_action_space_scaling_20_3331285.log train_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_lam_action_space_scaling_50_3320180.log train_tokenizer_model_size_scaling_80M_3316026.log\r\ntrain_lam_action_space_scaling_50_3329789.log yoloruns\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +18,23366,"TERMINAL",0,0,"cd causal/",,terminal_command +19,23742,"TERMINAL",0,0,"ls",,terminal_command +20,30822,"TERMINAL",0,0,"cd dynamics-cotraining/",,terminal_command +21,31118,"TERMINAL",0,0,"ls",,terminal_command +22,31166,"TERMINAL",0,0,"]633;E;2025-08-14 10:38:04 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +23,31264,"TERMINAL",0,0,"train_dynamics_causal_2_node_3373407.log train_dynamics_causal_8_node_3388140.log train_dynamics_causal_8_node_3412397.log\r\ntrain_dynamics_causal_2_node_3373407.log_bak train_dynamics_causal_8_node_3389928.log train_dynamics_causal_8_node_3412399.log\r\ntrain_dynamics_causal_2_node_3388135.log train_dynamics_causal_8_node_3390458.log train_dynamics_causal_8_node_3414710.log\r\ntrain_dynamics_causal_2_node_3388147.log train_dynamics_causal_8_node_3393060.log train_dynamics_causal_8_node_3415128.log\r\ntrain_dynamics_causal_2_node_3389801.log train_dynamics_causal_8_node_3393061.log train_dynamics_causal_8_node_3415137.log\r\ntrain_dynamics_causal_2_node_3393065.log train_dynamics_causal_8_node_3393066.log train_dynamics_causal_8_node_3417223.log\r\ntrain_dynamics_causal_2_node_dev_3416774.log train_dynamics_causal_8_node_3412343.log train_dynamics_causal_8_node_3417224.log\r\ntrain_dynamics_causal_2_node_dev_3416819.log train_dynamics_causal_8_node_3412349.log train_dynamics_causal_8_node_3418831.log\r\ntrain_dynamics_causal_8_node_3373408.log train_dynamics_causal_8_node_3412356.log train_dynamics_causal_8_node_3418832.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining]633;D;0",,terminal_output +24,38596,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining/train_dynamics_causal_8_node_3418831.log",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining/%x_%j.log\n#SBATCH --job-name=train_dynamics_causal_8_node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/holiday/causal/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\n# tokenizer with the new structure supporting larger ffn_dim\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401\n\nenv | grep SLURM\n\n# export PYTHONUNBUFFERED=1\n\nsrun python train_dynamics.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --num_latent_actions=100 \\n --init_lr=0 \\n --dyna_type=causal \\n --max_lr=8e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-causal-8-node-$slurm_job_id \\n --tags dynamics causal 8-node post-launch-main \\n --entity instant-uv \\n --project jafar \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16 \\n --dyna_ffn_dim=4096 \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n/var/spool/slurmd/job3418831/slurm_script: line 42: .venv/bin/activate: No such file or directory\nSLURM_JOB_USER=tum_cte0515\nSLURM_TASKS_PER_NODE=4(x8)\nSLURM_JOB_UID=999226\nSLURM_TASK_PID=1596326\nSLURM_JOB_GPUS=0,1,2,3\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs\nSLURMD_NODENAME=hkn0415\nSLURM_JOB_START_TIME=1755085369\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1755258169\nSLURM_CPUS_ON_NODE=24\nSLURM_JOB_CPUS_PER_NODE=24(x8)\nSLURM_GPUS_ON_NODE=4\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=accelerated\nSLURM_TRES_PER_TASK=cpu=5\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=8\nSLURM_JOBID=3418831\nSLURM_JOB_QOS=normal\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=5\nSLURM_NTASKS=32\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e12.hkn0415\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_NODELIST=hkn[0415,0423-0424,0531,0605,0729,0735,0814]\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=32\nSLURM_NNODES=8\nSLURM_SUBMIT_HOST=hkn1990.localdomain\nSLURM_JOB_ID=3418831\nSLURM_NODEID=0\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_dynamics_causal_8_node\nSLURM_NTASKS_PER_NODE=4\nSLURM_JOB_GID=502226\nSLURM_JOB_NODELIST=hkn[0415,0423-0424,0531,0605,0729,0735,0814]\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\n2025-08-13 13:43:44.591197: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\n2025-08-13 13:43:44.591342: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\n2025-08-13 13:43:44.591430: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\n2025-08-13 13:43:44.591432: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: Getting local topologies failed: Error 1: GetKeyValue() timed out with key: cuda:local_topology/cuda/12 and duration: 2m\n\nError 2: GetKeyValue() timed out with key: cuda:local_topology/cuda/13 and duration: 2m\n\nError 3: GetKeyValue() timed out with key: cuda:local_topology/cuda/14 and duration: 2m\n\nError 4: GetKeyValue() timed out with key: cuda:local_topology/cuda/15 and duration: 2m\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': INTERNAL: Getting local topologies failed: Error 1: GetKeyValue() timed out with key: cuda:local_topology/cuda/12 and duration: 2m\n\nError 2: GetKeyValue() timed out with key: cuda:local_topology/cuda/13 and duration: 2m\n\nError 3: GetKeyValue() timed out with key: cuda:local_topology/cuda/14 and duration: 2m\n\nError 4: GetKeyValue() timed out with key: cuda:local_topology/cuda/15 and duration: 2m (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n2025-08-13 13:48:45.162543: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1798] Shutdown barrier in coordination service has failed:\nDEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\nThis suggests that the workers are out of sync. Either at least one worker (a) crashed early due to program error or scheduler events (e.g. preemption, eviction), (b) was too fast in its execution, or (c) too slow / hanging. Check the logs (both the program and scheduler events) for an earlier error to identify the root cause.\n2025-08-13 13:48:45.162669: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1839] Use error polling to propagate the following error to all tasks: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.162928: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.162977: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.163142: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.163205: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.163118: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.163466: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.163359: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\n2025-08-13 13:48:45.163442: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.163466: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.163406: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.163597: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.163655: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.163817: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\n2025-08-13 13:48:45.163509: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.163397: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.163605: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.163511: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.163582: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n2025-08-13 13:48:45.163608: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.163815: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\n2025-08-13 13:48:45.164136: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.163687: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\n2025-08-13 13:48:45.164140: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\n2025-08-13 13:48:45.164197: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.163644: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.163729: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.164303: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.164300: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.163148: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.175424: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.175375: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.175765: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.178589: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.175833: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.175902: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.179447: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.179456: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.176050: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.179966: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::13736296003622477263::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:14. Some timed out task names:\n/job:jax_worker/replica:0/task:18\n/job:jax_worker/replica:0/task:6\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::13736296003622477263']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 13:48:45.352090: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.352127: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755085725.351949699"",""description"":""Error received from peer ipv4:10.0.1.47:64207"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 13:48:45.352201: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.352180: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.352202: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755085725.352087203"",""description"":""Error received from peer ipv4:10.0.1.47:64207"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 13:48:45.352237: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.352271: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755085725.351927525"",""description"":""Error received from peer ipv4:10.0.1.47:64207"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 13:48:45.352243: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.352276: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n2025-08-13 13:48:45.352055: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.352078: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755085725.352132883"",""description"":""Error received from peer ipv4:10.0.1.47:64207"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n:{""created"":""@1755085725.351952405"",""description"":""Error received from peer ipv4:10.0.1.47:64207"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 13:48:45.352214: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.352283: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n2025-08-13 13:48:45.352298: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755085725.351994661"",""description"":""Error received from peer ipv4:10.0.1.47:64207"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 13:48:45.352195: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.352228: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755085725.351969875"",""description"":""Error received from peer ipv4:10.0.1.47:64207"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n:{""created"":""@1755085725.352052510"",""description"":""Error received from peer ipv4:10.0.1.47:64207"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 13:48:45.352284: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.352350: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755085725.352023698"",""description"":""Error received from peer ipv4:10.0.1.47:64207"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 13:48:45.352343: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.352406: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755085725.352099318"",""description"":""Error received from peer ipv4:10.0.1.47:64207"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 13:48:45.352353: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 13:48:45.352426: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755085725.352099423"",""description"":""Error received from peer ipv4:10.0.1.47:64207"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\nsrun: error: hkn0531: tasks 12-15: Aborted (core dumped)\nsrun: error: hkn0424: tasks 9,11: Aborted (core dumped)\nsrun: error: hkn0415: tasks 0,2: Aborted (core dumped)\nsrun: error: hkn0735: task 25: Aborted (core dumped)\nsrun: error: hkn0729: tasks 21,23: Aborted (core dumped)\nsrun: error: hkn0605: tasks 18-19: Aborted (core dumped)\nsrun: error: hkn0423: tasks 4,6: Aborted (core dumped)\nsrun: error: hkn0814: tasks 30-31: Aborted (core dumped)\nsrun: error: hkn0415: task 1: Aborted (core dumped)\nsrun: error: hkn0729: task 20: Aborted (core dumped)\nsrun: error: hkn0605: task 16: Aborted (core dumped)\nsrun: error: hkn0423: task 5: Aborted (core dumped)\nsrun: error: hkn0814: task 28: Aborted (core dumped)\nsrun: error: hkn0735: tasks 24,27: Aborted (core dumped)\nsrun: error: hkn0424: task 10: Aborted (core dumped)\nsrun: error: hkn0415: task 3: Aborted (core dumped)\nsrun: error: hkn0729: task 22: Aborted (core dumped)\nsrun: error: hkn0605: task 17: Aborted (core dumped)\nsrun: error: hkn0423: task 7: Aborted (core dumped)\nsrun: error: hkn0814: task 29: Aborted (core dumped)\nsrun: error: hkn0424: task 8: Aborted (core dumped)\nsrun: error: hkn0735: task 26: Aborted (core dumped)\n\n============================= JOB FEEDBACK =============================\n\nJob ID: 3418831\nCluster: hk\nUser/Group: tum_cte0515/hk-project-p0023960\nAccount: hk-project-p0023960\nState: FAILED (exit code 134)\nPartition: accelerated\nNodes: 8\nCores per node: 24\nNodelist: hkn[0415,0423-0424,0531,0605,0729,0735,0814]\nCPU Utilized: 00:02:35\nCPU Efficiency: 0.23% of 19:02:24 core-walltime\nJob Wall-clock time: 00:05:57\nStarttime: Wed Aug 13 13:42:49 2025\nEndtime: Wed Aug 13 13:48:46 2025\nMemory Utilized: 11.96 GB (estimated maximum)\nMemory Efficiency: 0.00% of 0.00 MB (0.00 MB/node)\nEnergy Consumed: 4294164 Joule / 1192.82333333333 Watthours\nAverage node power draw: 12028.4705882353 Watt\n",log,tab +25,39365,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining/train_dynamics_causal_8_node_3418831.log",746,0,"",log,selection_mouse +26,39384,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining/train_dynamics_causal_8_node_3418831.log",745,0,"",log,selection_command +27,39869,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining/train_dynamics_causal_8_node_3418831.log",61841,0,"",log,selection_command +28,82828,"TERMINAL",0,0,"cd ..",,terminal_command +29,82833,"TERMINAL",0,0,"]633;E;2025-08-14 10:38:56 cd ..;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal]633;D;0",,terminal_output +30,83954,"TERMINAL",0,0,"cd ..",,terminal_command +31,84396,"TERMINAL",0,0,"l",,terminal_command +32,84418,"TERMINAL",0,0,"]633;E;2025-08-14 10:38:57 l;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Cbash: l: command not found...\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;127",,terminal_output +33,85599,"TERMINAL",0,0,"ls",,terminal_command +34,85637,"TERMINAL",0,0,"]633;E;2025-08-14 10:38:58 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +35,85704,"TERMINAL",0,0,"atari train_lam_action_space_scaling_50_3329804.log\r\nbig_run train_lam_action_space_scaling_50_3331286.log\r\nbig-runs train_lam_action_space_scaling_6_3318549.log\r\ncausal train_lam_action_space_scaling_6_3320178.log\r\ncoinrun train_lam_action_space_scaling_6_3321528.log\r\nmaskgit train_lam_action_space_scaling_6_3329790.log\r\nmaskgit-maskprob-fix train_lam_action_space_scaling_6_3329805.log\r\npreprocess train_lam_action_space_scaling_6_3331287.log\r\ntrain_dyn_causal_180M_3372931.log train_lam_action_space_scaling_8_3318550.log\r\ntrain_dyn_causal_180M_3372963.log train_lam_action_space_scaling_8_3329791.log\r\ntrain_dyn_causal_180M_3372969.log train_lam_action_space_scaling_8_3329806.log\r\ntrain_dyn_causal_180M_3373107.log train_lam_action_space_scaling_8_3331288.log\r\ntrain_dyn_causal_255M_3372932.log train_lam_minecraft_overfit_sample_3309655.log\r\ntrain_dyn_causal_255M_3372970.log train_lam_model_size_scaling_38M_3317098.log\r\ntrain_dyn_causal_255M_3373108.log train_lam_model_size_scaling_38M_3317115.log\r\ntrain_dyn_causal_356M_3372934.log train_lam_model_size_scaling_38M_3317231.log\r\ntrain_dyn_causal_356M_3372971.log train_tokenizer_batch_size_scaling_16_node_3321526.log\r\ntrain_dyn_causal_356M_3373109.log train_tokenizer_batch_size_scaling_1_node_3318551.log\r\ntrain_dyn_causal_500M_3372936.log train_tokenizer_batch_size_scaling_2_node_3318552.log\r\ntrain_dyn_causal_500M_3372972.log train_tokenizer_batch_size_scaling_2_node_3330806.log\r\ntrain_dyn_causal_500M_3373110.log train_tokenizer_batch_size_scaling_2_node_3330848.log\r\ntrain_dyn_new_arch-bugfixed-spatial-shift_3359343.log train_tokenizer_batch_size_scaling_2_node_3331282.log\r\ntrain_dyn_new_arch-bugfixed-temporal-shift_3359349.log train_tokenizer_batch_size_scaling_4_node_3318553.log\r\ntrain_dyn_yolorun_3333026.log train_tokenizer_batch_size_scaling_4_node_3320175.log\r\ntrain_dyn_yolorun_3333448.log train_tokenizer_batch_size_scaling_4_node_3321524.log\r\ntrain_dyn_yolorun_3335345.log train_tokenizer_batch_size_scaling_8_node_3320176.log\r\ntrain_dyn_yolorun_3335362.log train_tokenizer_batch_size_scaling_8_node_3321525.log\r\ntrain_dyn_yolorun_3348592.log train_tokenizer_minecraft_overfit_sample_3309656.log\r\ntrain_dyn_yolorun_new_arch_3351743.log train_tokenizer_model_size_scaling_127M_3317233.log\r\ntrain_dyn_yolorun_new_arch_3352103.log train_tokenizer_model_size_scaling_127M_3318554.log\r\ntrain_dyn_yolorun_new_arch_3352115.log train_tokenizer_model_size_scaling_140M_3313562.log\r\ntrain_dyn_yolorun_new_arch_3358457.log train_tokenizer_model_size_scaling_140M_3316019.log\r\ntrain_lam_action_space_scaling_10_3320179.log train_tokenizer_model_size_scaling_200M_3313563.log\r\ntrain_lam_action_space_scaling_10_3321529.log train_tokenizer_model_size_scaling_200M_3316020.log\r\ntrain_lam_action_space_scaling_10_3329786.log train_tokenizer_model_size_scaling_227M_3317234.log\r\ntrain_lam_action_space_scaling_10_3329801.log train_tokenizer_model_size_scaling_227M_3318555.log\r\ntrain_lam_action_space_scaling_10_3331283.log train_tokenizer_model_size_scaling_227M_3320173.log\r\ntrain_lam_action_space_scaling_12_3318546.log train_tokenizer_model_size_scaling_227M_3321523.log\r\ntrain_lam_action_space_scaling_12_3320177.log train_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_lam_action_space_scaling_12_3321527.log train_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_lam_action_space_scaling_12_3329787.log train_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_lam_action_space_scaling_12_3329802.log train_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_lam_action_space_scaling_12_3331284.log train_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_lam_action_space_scaling_20_3318547.log train_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_lam_action_space_scaling_20_3329788.log train_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_lam_action_space_scaling_20_3329803.log train_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_lam_action_space_scaling_20_3331285.log train_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_lam_action_space_scaling_50_3320180.log train_tokenizer_model_size_scaling_80M_3316026.log\r\ntrain_lam_action_space_scaling_50_3329789.log yoloruns\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +36,87243,"TERMINAL",0,0,"cd maskgit",,terminal_command +37,87564,"TERMINAL",0,0,"ls",,terminal_command +38,87571,"TERMINAL",0,0,"]633;E;2025-08-14 10:39:00 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Cdynamics-cotraining\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit]633;D;0",,terminal_output +39,89811,"TERMINAL",0,0,"cd dynamics-cotraining/",,terminal_command +40,90204,"TERMINAL",0,0,"ls",,terminal_command +41,90234,"TERMINAL",0,0,"]633;E;2025-08-14 10:39:03 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Ctrain_dynamics_maskgit_8_node_3412350.log train_dynamics_maskgit_8_node_3417225.log train_dynamics_maskgit_8_node_3418833.log\r\ntrain_dynamics_maskgit_8_node_3412354.log train_dynamics_maskgit_8_node_3417226.log train_dynamics_maskgit_8_node_3418834.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;0",,terminal_output +42,94109,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418833.log",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --job-name=train_dynamics_maskgit_8_node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/holiday/maskgit/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --dyna_type=maskgit \\n --num_latent_actions=100 \\n --max_lr=8e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-maskgit-8-node-$slurm_job_id \\n --tags dynamics maskgit 8-node post-launch-main \\n --entity instant-uv \\n --project jafar \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16 \\n --dyna_ffn_dim=4096 \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n/var/spool/slurmd/job3418833/slurm_script: line 42: .venv/bin/activate: No such file or directory\nSLURM_JOB_USER=tum_cte0515\nSLURM_TASKS_PER_NODE=4(x8)\nSLURM_JOB_UID=999226\nSLURM_TASK_PID=2787129\nSLURM_JOB_GPUS=0,1,2,3\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs\nSLURMD_NODENAME=hkn0531\nSLURM_JOB_START_TIME=1755086164\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1755258964\nSLURM_CPUS_ON_NODE=24\nSLURM_JOB_CPUS_PER_NODE=24(x8)\nSLURM_GPUS_ON_NODE=4\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=accelerated\nSLURM_TRES_PER_TASK=cpu=5\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=8\nSLURM_JOBID=3418833\nSLURM_JOB_QOS=normal\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=5\nSLURM_NTASKS=32\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0531\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_NODELIST=hkn[0531,0605,0608,0621,0728,0812,0814,0821]\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=32\nSLURM_NNODES=8\nSLURM_SUBMIT_HOST=hkn1990.localdomain\nSLURM_JOB_ID=3418833\nSLURM_NODEID=0\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_dynamics_maskgit_8_node\nSLURM_NTASKS_PER_NODE=4\nSLURM_JOB_GID=502226\nSLURM_JOB_NODELIST=hkn[0531,0605,0608,0621,0728,0812,0814,0821]\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\n2025-08-13 13:56:59.039557: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\n2025-08-13 13:56:59.039649: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\n2025-08-13 13:56:59.039599: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\n2025-08-13 13:56:59.039730: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n2025-08-13 14:02:00.026193: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1798] Shutdown barrier in coordination service has failed:\nDEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\nThis suggests that the workers are out of sync. Either at least one worker (a) crashed early due to program error or scheduler events (e.g. preemption, eviction), (b) was too fast in its execution, or (c) too slow / hanging. Check the logs (both the program and scheduler events) for an earlier error to identify the root cause.\n2025-08-13 14:02:00.026235: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1839] Use error polling to propagate the following error to all tasks: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.026303: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.026375: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.026488: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.026484: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.026766: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.026781: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.026779: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.026952: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.026920: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n2025-08-13 14:02:00.026781: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.026833: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.026908: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.026745: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.027030: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.026932: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.026956: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.026986: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.026997: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\n2025-08-13 14:02:00.027172: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.027213: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.027389: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.027415: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\n2025-08-13 14:02:00.027093: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.027004: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.027102: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.027012: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.027091: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.027193: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\n2025-08-13 14:02:00.026992: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.027045: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.027109: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\n2025-08-13 14:02:00.027062: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.027017: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.027112: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\n2025-08-13 14:02:00.027032: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.027690: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\n2025-08-13 14:02:00.027421: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\n2025-08-13 14:02:00.027240: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.027532: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.027612: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.027657: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\n2025-08-13 14:02:00.027624: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.027483: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.027561: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.027595: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\n2025-08-13 14:02:00.027623: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.027612: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.027563: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.064663: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.064634: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.064853: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::1026922694718633181::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:28\n/job:jax_worker/replica:0/task:9\n [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::1026922694718633181'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:02:00.106695: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.106720: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n2025-08-13 14:02:00.106625: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.106647: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755086520.106534982"",""description"":""Error received from peer ipv4:10.0.1.99:64209"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 14:02:00.106864: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.106889: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755086520.106772094"",""description"":""Error received from peer ipv4:10.0.1.99:64209"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n:{""created"":""@1755086520.106600545"",""description"":""Error received from peer ipv4:10.0.1.99:64209"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 14:02:00.106886: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.106913: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n2025-08-13 14:02:00.107058: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.107082: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755086520.106784864"",""description"":""Error received from peer ipv4:10.0.1.99:64209"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n:{""created"":""@1755086520.106957136"",""description"":""Error received from peer ipv4:10.0.1.99:64209"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 14:02:00.107117: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:02:00.107178: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755086520.106878160"",""description"":""Error received from peer ipv4:10.0.1.99:64209"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\nsrun: error: hkn0531: tasks 0-3: Aborted (core dumped)\nsrun: error: hkn0605: tasks 4,6: Aborted (core dumped)\nsrun: error: hkn0728: tasks 18-19: Aborted (core dumped)\nsrun: error: hkn0608: tasks 8-9: Aborted (core dumped)\nsrun: error: hkn0621: tasks 12,14: Aborted (core dumped)\nsrun: error: hkn0814: tasks 25-26: Aborted (core dumped)\nsrun: error: hkn0812: tasks 21,23: Aborted (core dumped)\nsrun: error: hkn0821: tasks 28,31: Aborted (core dumped)\nsrun: error: hkn0605: task 7: Aborted (core dumped)\nsrun: error: hkn0728: task 16: Aborted (core dumped)\nsrun: error: hkn0621: task 15: Aborted (core dumped)\nsrun: error: hkn0608: task 11: Aborted (core dumped)\nsrun: error: hkn0812: task 20: Aborted (core dumped)\nsrun: error: hkn0814: task 27: Aborted (core dumped)\nsrun: error: hkn0821: task 30: Aborted (core dumped)\nsrun: error: hkn0605: task 5: Aborted (core dumped)\nsrun: error: hkn0728: task 17: Aborted (core dumped)\nsrun: error: hkn0608: task 10: Aborted (core dumped)\nsrun: error: hkn0621: task 13: Aborted (core dumped)\nsrun: error: hkn0812: task 22: Aborted (core dumped)\nsrun: error: hkn0814: task 24: Aborted (core dumped)\nsrun: error: hkn0821: task 29: Aborted (core dumped)\n\n============================= JOB FEEDBACK =============================\n\nJob ID: 3418833\nCluster: hk\nUser/Group: tum_cte0515/hk-project-p0023960\nAccount: hk-project-p0023960\nState: FAILED (exit code 134)\nPartition: accelerated\nNodes: 8\nCores per node: 24\nNodelist: hkn[0531,0605,0608,0621,0728,0812,0814,0821]\nCPU Utilized: 00:02:20\nCPU Efficiency: 0.20% of 19:02:24 core-walltime\nJob Wall-clock time: 00:05:57\nStarttime: Wed Aug 13 13:56:04 2025\nEndtime: Wed Aug 13 14:02:01 2025\nMemory Utilized: 11.92 GB (estimated maximum)\nMemory Efficiency: 0.00% of 0.00 MB (0.00 MB/node)\nEnergy Consumed: 2591063 Joule / 719.739722222222 Watthours\nAverage node power draw: 7257.87955182073 Watt\n",log,tab +43,102885,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --job-name=train_dynamics_maskgit_8_node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/holiday/maskgit/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --darkness_threshold=50 \\n --dyna_type=maskgit \\n --num_latent_actions=100 \\n --max_lr=8e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-maskgit-8-node-darkness-filter-$slurm_job_id \\n --tags dynamics maskgit 8-node post-launch-main darkness-filter \\n --entity instant-uv \\n --project jafar \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16 \\n --dyna_ffn_dim=4096 \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n/var/spool/slurmd/job3418834/slurm_script: line 42: .venv/bin/activate: No such file or directory\nSLURM_JOB_USER=tum_cte0515\nSLURM_TASKS_PER_NODE=4(x8)\nSLURM_JOB_UID=999226\nSLURM_TASK_PID=2792684\nSLURM_JOB_GPUS=0,1,2,3\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs\nSLURMD_NODENAME=hkn0531\nSLURM_JOB_START_TIME=1755086531\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1755259331\nSLURM_CPUS_ON_NODE=24\nSLURM_JOB_CPUS_PER_NODE=24(x8)\nSLURM_GPUS_ON_NODE=4\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=accelerated\nSLURM_TRES_PER_TASK=cpu=5\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=8\nSLURM_JOBID=3418834\nSLURM_JOB_QOS=normal\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=5\nSLURM_NTASKS=32\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0531\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_NODELIST=hkn[0531,0605,0608,0621,0728,0812,0814,0817]\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=32\nSLURM_NNODES=8\nSLURM_SUBMIT_HOST=hkn1990.localdomain\nSLURM_JOB_ID=3418834\nSLURM_NODEID=0\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_dynamics_maskgit_8_node\nSLURM_NTASKS_PER_NODE=4\nSLURM_JOB_GID=502226\nSLURM_JOB_NODELIST=hkn[0531,0605,0608,0621,0728,0812,0814,0817]\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\n2025-08-13 14:03:12.754648: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n2025-08-13 14:03:12.754945: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\n2025-08-13 14:03:12.754961: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n2025-08-13 14:03:12.755393: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n2025-08-13 14:08:13.373538: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1798] Shutdown barrier in coordination service has failed:\nDEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\nThis suggests that the workers are out of sync. Either at least one worker (a) crashed early due to program error or scheduler events (e.g. preemption, eviction), (b) was too fast in its execution, or (c) too slow / hanging. Check the logs (both the program and scheduler events) for an earlier error to identify the root cause.\n2025-08-13 14:08:13.373564: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1839] Use error polling to propagate the following error to all tasks: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.373830: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.373806: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.373972: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374005: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.373979: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374003: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374236: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374055: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374175: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374101: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374282: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\n2025-08-13 14:08:13.374175: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374096: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374298: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n2025-08-13 14:08:13.374440: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374446: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374299: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374317: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374276: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374215: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374160: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374372: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374510: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374639: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374520: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374599: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374375: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374697: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\n2025-08-13 14:08:13.374871: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\n2025-08-13 14:08:13.374620: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374414: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374600: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374669: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374823: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\n2025-08-13 14:08:13.374758: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374535: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.374784: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374805: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\n2025-08-13 14:08:13.374945: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\n2025-08-13 14:08:13.374790: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374890: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374801: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374866: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374963: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\n2025-08-13 14:08:13.374985: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374968: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\n2025-08-13 14:08:13.375000: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374968: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.374969: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.375006: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.375018: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.375020: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::7350480505452186648::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:2. Some timed out task names:\n/job:jax_worker/replica:0/task:29\n/job:jax_worker/replica:0/task:17\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::7350480505452186648']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-13 14:08:13.453060: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.453091: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n2025-08-13 14:08:13.453144: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.453164: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755086893.452946699"",""description"":""Error received from peer ipv4:10.0.1.99:64210"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n:{""created"":""@1755086893.453054591"",""description"":""Error received from peer ipv4:10.0.1.99:64210"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 14:08:13.453150: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.453220: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n2025-08-13 14:08:13.453124: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.453152: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755086893.452915579"",""description"":""Error received from peer ipv4:10.0.1.99:64210"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n:{""created"":""@1755086893.453028863"",""description"":""Error received from peer ipv4:10.0.1.99:64210"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-08-13 14:08:13.453301: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-13 14:08:13.453359: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1755086893.453059547"",""description"":""Error received from peer ipv4:10.0.1.99:64210"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\nsrun: error: hkn0531: tasks 0-3: Aborted (core dumped)\nsrun: error: hkn0814: tasks 25,27: Aborted (core dumped)\nsrun: error: hkn0728: tasks 16,18: Aborted (core dumped)\nsrun: error: hkn0605: tasks 4-5: Aborted (core dumped)\nsrun: error: hkn0608: tasks 8,10: Aborted (core dumped)\nsrun: error: hkn0817: tasks 29-30: Aborted (core dumped)\nsrun: error: hkn0812: tasks 20,22: Aborted (core dumped)\nsrun: error: hkn0621: tasks 13,15: Aborted (core dumped)\nsrun: error: hkn0728: task 19: Aborted (core dumped)\nsrun: error: hkn0814: task 24: Aborted (core dumped)\nsrun: error: hkn0605: task 7: Aborted (core dumped)\nsrun: error: hkn0608: task 11: Aborted (core dumped)\nsrun: error: hkn0817: task 31: Aborted (core dumped)\nsrun: error: hkn0621: task 14: Aborted (core dumped)\nsrun: error: hkn0812: task 23: Aborted (core dumped)\nsrun: error: hkn0814: task 26: Aborted (core dumped)\nsrun: error: hkn0728: task 17: Aborted (core dumped)\nsrun: error: hkn0608: task 9: Aborted (core dumped)\nsrun: error: hkn0605: task 6: Aborted (core dumped)\nsrun: error: hkn0817: task 28: Aborted (core dumped)\nsrun: error: hkn0812: task 21: Aborted (core dumped)\nsrun: error: hkn0621: task 12: Aborted (core dumped)\n\n============================= JOB FEEDBACK =============================\n\nJob ID: 3418834\nCluster: hk\nUser/Group: tum_cte0515/hk-project-p0023960\nAccount: hk-project-p0023960\nState: FAILED (exit code 134)\nPartition: accelerated\nNodes: 8\nCores per node: 24\nNodelist: hkn[0531,0605,0608,0621,0728,0812,0814,0817]\nCPU Utilized: 00:02:03\nCPU Efficiency: 0.18% of 19:21:36 core-walltime\nJob Wall-clock time: 00:06:03\nStarttime: Wed Aug 13 14:02:11 2025\nEndtime: Wed Aug 13 14:08:14 2025\nMemory Utilized: 11.91 GB (estimated maximum)\nMemory Efficiency: 0.00% of 0.00 MB (0.00 MB/node)\nEnergy Consumed: 1494485 Joule / 415.134722222222 Watthours\nAverage node power draw: 4117.03856749311 Watt\n",log,tab +44,325138,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8721,0,"",log,selection_mouse +45,325277,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,15,"XlaRuntimeError",log,selection_mouse +46,325559,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,17,"XlaRuntimeError: ",log,selection_mouse +47,325564,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,36,"XlaRuntimeError: FAILED_PRECONDITION",log,selection_mouse +48,325599,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8664,69,"plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError",log,selection_mouse +49,325708,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8675,58,", options, distributed_client)\njaxlib._jax.XlaRuntimeError",log,selection_mouse +50,325709,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8677,56,"options, distributed_client)\njaxlib._jax.XlaRuntimeError",log,selection_mouse +51,325710,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,27,"jaxlib._jax.XlaRuntimeError",log,selection_mouse +52,325711,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,38,"XlaRuntimeError: FAILED_PRECONDITION: ",log,selection_mouse +53,325729,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,40,"XlaRuntimeError: FAILED_PRECONDITION: No",log,selection_mouse +54,325776,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,41,"XlaRuntimeError: FAILED_PRECONDITION: No ",log,selection_mouse +55,325777,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,48,"XlaRuntimeError: FAILED_PRECONDITION: No visible",log,selection_mouse +56,325884,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,52,"XlaRuntimeError: FAILED_PRECONDITION: No visible GPU",log,selection_mouse +57,325917,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,53,"XlaRuntimeError: FAILED_PRECONDITION: No visible GPU ",log,selection_mouse +58,325953,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,60,"XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +59,326032,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,61,"XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +60,326457,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8779,0,"",log,selection_mouse +61,326474,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8778,0,"",log,selection_command +62,326630,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8778,2,".\n",log,selection_mouse +63,326678,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8779,1,"\n",log,selection_command +64,326749,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8759,20,"visible GPU devices.",log,selection_mouse +65,326749,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8756,23,"No visible GPU devices.",log,selection_mouse +66,326785,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8735,44,"FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +67,326906,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8734,45," FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +68,326944,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8733,46,": FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +69,326984,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,61,"XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +70,327190,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8717,62,".XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +71,327258,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8713,66,"_jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +72,327876,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8710,0,"",log,selection_mouse +73,328024,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,6,"jaxlib",log,selection_mouse +74,328236,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,7,"jaxlib.",log,selection_mouse +75,328271,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,11,"jaxlib._jax",log,selection_mouse +76,328271,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,12,"jaxlib._jax.",log,selection_mouse +77,328292,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,27,"jaxlib._jax.XlaRuntimeError",log,selection_mouse +78,328328,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8647,65,"get_c_api_client(plugin_name, options, distributed_client)\njaxlib",log,selection_mouse +79,328428,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,48,"jaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION",log,selection_mouse +80,328656,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,49,"jaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION:",log,selection_mouse +81,328692,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,50,"jaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: ",log,selection_mouse +82,328693,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,74,"jaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +83,328934,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,60,"jaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible",log,selection_mouse +84,328992,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,61,"jaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible ",log,selection_mouse +85,329025,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,64,"jaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU",log,selection_mouse +86,329645,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8776,0,"",log,selection_mouse +87,329789,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8771,7,"devices",log,selection_mouse +88,329941,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8767,11,"GPU devices",log,selection_mouse +89,329961,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8759,19,"visible GPU devices",log,selection_mouse +90,329977,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8677,101,"options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +91,329997,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8664,114,"plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +92,330032,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8647,131,"get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +93,330110,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,60,"XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +94,330286,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8771,9,"devices.\n",log,selection_mouse +95,330428,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8713,65,"_jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +96,330537,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8712,66,"._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +97,330614,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8706,72,"jaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +98,330810,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8633,145," return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +99,330847,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8512,266,"File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +100,330884,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,303,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +101,331356,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8477,0,"",log,selection_mouse +102,331357,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,9,"Traceback",log,selection_mouse +103,331513,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,35,"Traceback (most recent call last):\n",log,selection_mouse +104,331687,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,156,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n",log,selection_mouse +105,331712,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,231,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\n",log,selection_mouse +106,331793,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,305,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +107,331961,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,306,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\n",log,selection_mouse +108,332188,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,305,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +109,332809,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8777,0,"",log,selection_mouse +110,332810,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8771,7,"devices",log,selection_mouse +111,333032,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8767,11,"GPU devices",log,selection_mouse +112,333033,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8759,19,"visible GPU devices",log,selection_mouse +113,333033,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8735,43,"FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +114,333034,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8664,114,"plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +115,333049,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8647,131,"get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +116,333125,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8646,132,".get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +117,333160,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8642,136,"_xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +118,333238,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8641,137," _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +119,333266,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8519,259,"hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +120,333267,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8517,261,"""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +121,333286,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8516,262," ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +122,333287,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8512,266,"File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +123,333365,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8511,267," File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +124,333393,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8510,268," File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +125,333473,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,303,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +126,334476,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8477,0,"",log,selection_mouse +127,334477,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,9,"Traceback",log,selection_mouse +128,334673,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,43,"Traceback (most recent call last):\n File """,log,selection_mouse +129,334801,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,48,"Traceback (most recent call last):\n File ""/hkfs",log,selection_mouse +130,334802,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,61,"Traceback (most recent call last):\n File ""/hkfs/home/project",log,selection_mouse +131,334802,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,62,"Traceback (most recent call last):\n File ""/hkfs/home/project/",log,selection_mouse +132,334803,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,72,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project",log,selection_mouse +133,334803,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,200,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name",log,selection_mouse +134,334837,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,201,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name,",log,selection_mouse +135,334875,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,209,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options",log,selection_mouse +136,334909,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,281,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: ",log,selection_mouse +137,334942,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,283,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No",log,selection_mouse +138,334945,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,284,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No ",log,selection_mouse +139,334980,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,291,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible",log,selection_mouse +140,335057,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,305,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +141,335233,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,303,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +142,336020,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8777,0,"",log,selection_mouse +143,336021,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8771,7,"devices",log,selection_mouse +144,336222,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8767,11,"GPU devices",log,selection_mouse +145,336261,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8759,19,"visible GPU devices",log,selection_mouse +146,336296,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8756,22,"No visible GPU devices",log,selection_mouse +147,336297,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8754,24,": No visible GPU devices",log,selection_mouse +148,336332,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8735,43,"FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +149,336332,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8675,103,", options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +150,336368,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8664,114,"plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +151,336452,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8647,131,"get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +152,336457,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8537,241,"hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +153,336480,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8529,249,"project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +154,336513,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8528,250,"/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +155,336546,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8524,254,"home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +156,336583,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8523,255,"/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +157,336584,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8486,292,"most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +158,336650,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8485,293,"(most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +159,336684,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8484,294," (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +160,336685,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,303,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +161,337249,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8480,0,"",log,selection_mouse +162,337250,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,9,"Traceback",log,selection_mouse +163,337448,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,48,"Traceback (most recent call last):\n File ""/hkfs",log,selection_mouse +164,337484,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,53,"Traceback (most recent call last):\n File ""/hkfs/home",log,selection_mouse +165,337484,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,61,"Traceback (most recent call last):\n File ""/hkfs/home/project",log,selection_mouse +166,337519,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,62,"Traceback (most recent call last):\n File ""/hkfs/home/project/",log,selection_mouse +167,337519,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,72,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project",log,selection_mouse +168,337555,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,200,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name",log,selection_mouse +169,337591,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,209,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options",log,selection_mouse +170,337627,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,211,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, ",log,selection_mouse +171,337668,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,229,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client",log,selection_mouse +172,337750,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,295,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU",log,selection_mouse +173,337766,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,296,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU ",log,selection_mouse +174,337767,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,303,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices",log,selection_mouse +175,337903,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,304,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +176,338242,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8779,0,"",log,selection_mouse +177,338275,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8778,0,"",log,selection_command +178,338420,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8779,0,"",log,selection_mouse +179,338425,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8778,0,"",log,selection_command +180,338588,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8778,2,".\n",log,selection_mouse +181,338604,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8779,1,"\n",log,selection_command +182,338658,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8759,20,"visible GPU devices.",log,selection_mouse +183,338692,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8756,23,"No visible GPU devices.",log,selection_mouse +184,338693,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8754,25,": No visible GPU devices.",log,selection_mouse +185,338725,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8735,44,"FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +186,338762,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8664,115,"plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +187,338799,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8537,242,"hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +188,338833,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8529,250,"project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +189,338867,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8528,251,"/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +190,338867,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8524,255,"home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +191,338902,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8523,256,"/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +192,338936,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8519,260,"hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +193,339085,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8518,261,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +194,339119,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8517,262,"""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +195,339193,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8516,263," ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +196,339201,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,304,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.",log,selection_mouse +197,339835,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8477,0,"",log,selection_mouse +198,339835,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,9,"Traceback",log,selection_mouse +199,340037,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,44,"Traceback (most recent call last):\n File ""/",log,selection_mouse +200,340073,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,53,"Traceback (most recent call last):\n File ""/hkfs/home",log,selection_mouse +201,340074,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,188,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client",log,selection_mouse +202,340107,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,200,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name",log,selection_mouse +203,340139,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,209,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options",log,selection_mouse +204,340172,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,284,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No ",log,selection_mouse +205,340209,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,291,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible",log,selection_mouse +206,340290,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,305,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +207,340829,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8780,0,"",log,selection_mouse +208,341222,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8735,45,"FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +209,341255,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8734,46," FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +210,341255,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8718,62,"XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +211,341347,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8646,134,".get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +212,341375,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8642,138,"_xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +213,341454,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8641,139," _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +214,341455,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8635,145,"return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +215,341467,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8518,262,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +216,341478,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8517,263,"""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +217,341507,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8516,264," ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +218,341541,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8512,268,"File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +219,341851,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,305,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +220,342271,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8478,0,"",log,selection_mouse +221,342271,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,9,"Traceback",log,selection_mouse +222,342415,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,35,"Traceback (most recent call last):\n",log,selection_mouse +223,342574,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,156,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n",log,selection_mouse +224,342608,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,231,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\n",log,selection_mouse +225,342683,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,305,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +226,342764,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,306,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\n",log,selection_mouse +227,343180,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8780,0,"",log,selection_mouse +228,343504,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8735,45,"FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +229,343554,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8647,133,"get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +230,343555,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8490,290," recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +231,343555,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8474,306,"\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +232,343581,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8406,374,"During handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +233,343814,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8474,306,"\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +234,343874,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",8475,305,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_dynamics.py"", line 158, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n",log,selection_mouse +235,362884,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",3737,0,"",log,selection_mouse +236,363025,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",3735,3,"hkn",log,selection_mouse +237,755342,"TERMINAL",0,0,"queue",,terminal_command +238,755417,"TERMINAL",0,0,"]633;E;2025-08-14 10:50:08 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 10:50:08 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-16:50:07\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 21:01:12\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]",,terminal_output +239,756521,"TERMINAL",0,0,"983",,terminal_output +240,757543,"TERMINAL",0,0,"1094[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;0",,terminal_output +241,758653,"TERMINAL",0,0,"fqueue",,terminal_command +242,758717,"TERMINAL",0,0,"]633;E;2025-08-14 10:50:11 fqueue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue -o ""%.10i %.16P %.30j %.8u %.8T %.10M %.9l %.6D %R""hkn1993.localdomain: Thu Aug 14 10:50:11 2025JOBIDPARTITIONNAME USER STATE\t TIME TIME_LIMI NODES NODELIST(REASON)3415713\taccelerated train_dynamics_maskgit_1_node tum_cte0 PENDING\t 0:00 2-00:00:00\t1 (Priority)3412401\tacceleratedtrain_tokenizer_1e-4 tum_cte0 RUNNING 1-16:50:10 2-00:00:00\t8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832\taccelerated train_dynamics_causal_8_node tum_cte0 RUNNING 21:01:15 2-00:00:00\t8 hkn[0407,0415,0420,0422-0424,07[?7l2[?7h",,terminal_output +243,759753,"TERMINAL",0,0,"216",,terminal_output +244,760819,"TERMINAL",0,0,"327",,terminal_output +245,761826,"TERMINAL",0,0,"449",,terminal_output +246,762888,"TERMINAL",0,0,"6520",,terminal_output +247,764002,"TERMINAL",0,0,"761",,terminal_output +248,764937,"TERMINAL",0,0,"872",,terminal_output +249,765973,"TERMINAL",0,0,"983",,terminal_output +250,767068,"TERMINAL",0,0,"2094",,terminal_output +251,768090,"TERMINAL",0,0,"1205",,terminal_output +252,769082,"TERMINAL",0,0,"216",,terminal_output +253,770141,"TERMINAL",0,0,"327",,terminal_output +254,770570,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;0",,terminal_output +255,1040801,"TERMINAL",0,0,"queue",,terminal_command +256,1040857,"TERMINAL",0,0,"]633;E;2025-08-14 10:54:53 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 10:54:53 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-16:54:53\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 21:05:58\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]",,terminal_output +257,1041913,"TERMINAL",0,0,"549",,terminal_output +258,1042989,"TERMINAL",0,0,"656:00",,terminal_output +259,1044028,"TERMINAL",0,0,"761",,terminal_output +260,1045079,"TERMINAL",0,0,"872",,terminal_output +261,1046133,"TERMINAL",0,0,"983",,terminal_output +262,1047236,"TERMINAL",0,0,"5:0094",,terminal_output +263,1048269,"TERMINAL",0,0,"15:005",,terminal_output +264,1049388,"TERMINAL",0,0,"216",,terminal_output +265,1050337,"TERMINAL",0,0,"327",,terminal_output +266,1051536,"TERMINAL",0,0,"438",,terminal_output +267,1052433,"TERMINAL",0,0,"549",,terminal_output +268,1053590,"TERMINAL",0,0,"6510",,terminal_output +269,1054610,"TERMINAL",0,0,"761",,terminal_output +270,1055636,"TERMINAL",0,0,"872",,terminal_output +271,1056657,"TERMINAL",0,0,"983",,terminal_output +272,1057682,"TERMINAL",0,0,"1094",,terminal_output +273,1058722,"TERMINAL",0,0,"1105",,terminal_output +274,1059835,"TERMINAL",0,0,"216",,terminal_output +275,1060815,"TERMINAL",0,0,"338",,terminal_output +276,1061973,"TERMINAL",0,0,"549",,terminal_output +277,1062960,"TERMINAL",0,0,"6520",,terminal_output +278,1063972,"TERMINAL",0,0,"761",,terminal_output +279,1065028,"TERMINAL",0,0,"872",,terminal_output +280,1066078,"TERMINAL",0,0,"983",,terminal_output +281,1067123,"TERMINAL",0,0,"2094",,terminal_output +282,1068240,"TERMINAL",0,0,"1205",,terminal_output +283,1069253,"TERMINAL",0,0,"216",,terminal_output +284,1070319,"TERMINAL",0,0,"327",,terminal_output +285,1071402,"TERMINAL",0,0,"438",,terminal_output +286,1072426,"TERMINAL",0,0,"549",,terminal_output +287,1073438,"TERMINAL",0,0,"6530",,terminal_output +288,1074490,"TERMINAL",0,0,"761",,terminal_output +289,1075686,"TERMINAL",0,0,"872",,terminal_output +290,1076606,"TERMINAL",0,0,"983",,terminal_output +291,1077750,"TERMINAL",0,0,"3094",,terminal_output +292,1078776,"TERMINAL",0,0,"1305",,terminal_output +293,1079805,"TERMINAL",0,0,"216",,terminal_output +294,1080824,"TERMINAL",0,0,"327",,terminal_output +295,1081850,"TERMINAL",0,0,"449",,terminal_output +296,1082923,"TERMINAL",0,0,"6540",,terminal_output +297,1083939,"TERMINAL",0,0,"761",,terminal_output +298,1084989,"TERMINAL",0,0,"872",,terminal_output +299,1086043,"TERMINAL",0,0,"983",,terminal_output +300,1087085,"TERMINAL",0,0,"4094",,terminal_output +301,1088139,"TERMINAL",0,0,"1405",,terminal_output +302,1089222,"TERMINAL",0,0,"216",,terminal_output +303,1089622,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;0",,terminal_output +304,1093455,"TERMINAL",0,0,"dev",,terminal_command +305,1093981,"TERMINAL",0,0,"ls",,terminal_command +306,1094045,"TERMINAL",0,0,"]633;E;2025-08-14 10:55:47 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +307,1094173,"TERMINAL",0,0,"data frames LICENSE overfit_dir.zip requirements.txt tests wandb\r\ndata_atari generate_dataset.py log.log __pycache__ sample.py train_dynamics.py\r\ndebug genie.py logs README.md samples train_lam.py\r\nframe-knoms.png gifs models read_tf_record.py scripts_cremers train_tokenizer.py\r\nframe.png input_pipeline overfit_dir requirements-franz.txt slurm utils\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +308,1096844,"TERMINAL",0,0,"cd slurm/",,terminal_command +309,1097099,"TERMINAL",0,0,"ls",,terminal_command +310,1097143,"TERMINAL",0,0,"]633;E;2025-08-14 10:55:50 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +311,1097237,"TERMINAL",0,0,"common dev jobs README.md templates utils\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar/slurm]633;D;0",,terminal_output +312,1103680,"TERMINAL",0,0,"git status",,terminal_command +313,1103721,"TERMINAL",0,0,"]633;E;2025-08-14 10:55:56 git statu;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Cgit: 'statu' is not a git command. See 'git --help'.\r\n\r\nThe most similar commands are\r\n\tstatus\r\n\tstage\r\n\tstash\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar/slurm]633;D;1",,terminal_output +314,1108123,"TERMINAL",0,0,"git status",,terminal_command +315,1108164,"TERMINAL",0,0,"]633;E;2025-08-14 10:56:01 git status;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +316,1109072,"TERMINAL",0,0,"On branch main\r\nYour branch is up to date with 'origin/main'.\r\n\r\nChanges not staged for commit:\r\n (use ""git add ..."" to update what will be committed)\r\n (use ""git restore ..."" to discard changes in working directory)\r\n\tmodified: jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes.sbatch\r\n\tmodified: jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch\r\n\tmodified: jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes.sbatch\r\n\tmodified: jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch\r\n\r\nUntracked files:\r\n (use ""git add ..."" to include in what will be committed)\r\n\tutils/alfred/sqrt_lr_scaling.py\r\n\r\nno changes added to commit (use ""git add"" and/or ""git commit -a"")\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar/slurm]633;D;0",,terminal_output +317,1137989,"TERMINAL",0,0,"git pull",,terminal_command +318,1137998,"TERMINAL",0,0,"]633;E;2025-08-14 10:56:31 git pull;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +319,1139744,"TERMINAL",0,0,"remote: Enumerating objects: 53, done.\r\nremote: Counting objects: 1% (1/53)\rremote: Counting objects: 3% (2/53)\rremote: Counting objects: 5% (3/53)\rremote: Counting objects: 7% (4/53)\rremote: Counting objects: 9% (5/53)\rremote: Counting objects: 11% (6/53)\rremote: Counting objects: 13% (7/53)\rremote: Counting objects: 15% (8/53)\rremote: Counting objects: 16% (9/53)\rremote: Counting objects: 18% (10/53)\rremote: Counting objects: 20% (11/53)\rremote: Counting objects: 22% (12/53)\rremote: Counting objects: 24% (13/53)\rremote: Counting objects: 26% (14/53)\rremote: Counting objects: 28% (15/53)\rremote: Counting objects: 30% (16/53)\rremote: Counting objects: 32% (17/53)\rremote: Counting objects: 33% (18/53)\rremote: Counting objects: 35% (19/53)\rremote: Counting objects: 37% (20/53)\rremote: Counting objects: 39% (21/53)\rremote: Counting objects: 41% (22/53)\rremote: Counting objects: 43% (23/53)\rremote: Counting objects: 45% (24/53)\rremote: Counting objects: 47% (25/53)\rremote: Counting objects: 49% (26/53)\rremote: Counting objects: 50% (27/53)\rremote: Counting objects: 52% (28/53)\rremote: Counting objects: 54% (29/53)\rremote: Counting objects: 56% (30/53)\rremote: Counting objects: 58% (31/53)\rremote: Counting objects: 60% (32/53)\rremote: Counting objects: 62% (33/53)\rremote: Counting objects: 64% (34/53)\rremote: Counting objects: 66% (35/53)\rremote: Counting objects: 67% (36/53)\rremote: Counting objects: 69% (37/53)\rremote: Counting objects: 71% (38/53)\rremote: Counting objects: 73% (39/53)\rremote: Counting objects: 75% (40/53)\rremote: Counting objects: 77% (41/53)\rremote: Counting objects: 79% (42/53)\rremote: Counting objects: 81% (43/53)\rremote: Counting objects: 83% (44/53)\rremote: Counting objects: 84% (45/53)\rremote: Counting objects: 86% (46/53)\rremote: Counting objects: 88% (47/53)\rremote: Counting objects: 90% (48/53)\r",,terminal_output +320,1139842,"TERMINAL",0,0,"remote: Counting objects: 92% (49/53)\rremote: Counting objects: 94% (50/53)\rremote: Counting objects: 96% (51/53)\rremote: Counting objects: 98% (52/53)\rremote: Counting objects: 100% (53/53)\rremote: Counting objects: 100% (53/53), done.\r\nremote: Compressing objects: 3% (1/30)\rremote: Compressing objects: 6% (2/30)\rremote: Compressing objects: 10% (3/30)\rremote: Compressing objects: 13% (4/30)\rremote: Compressing objects: 16% (5/30)\rremote: Compressing objects: 20% (6/30)\rremote: Compressing objects: 23% (7/30)\rremote: Compressing objects: 26% (8/30)\rremote: Compressing objects: 30% (9/30)\rremote: Compressing objects: 33% (10/30)\rremote: Compressing objects: 36% (11/30)\rremote: Compressing objects: 40% (12/30)\rremote: Compressing objects: 43% (13/30)\rremote: Compressing objects: 46% (14/30)\rremote: Compressing objects: 50% (15/30)\rremote: Compressing objects: 53% (16/30)\rremote: Compressing objects: 56% (17/30)\rremote: Compressing objects: 60% (18/30)\rremote: Compressing objects: 63% (19/30)\rremote: Compressing objects: 66% (20/30)\rremote: Compressing objects: 70% (21/30)\rremote: Compressing objects: 73% (22/30)\rremote: Compressing objects: 76% (23/30)\rremote: Compressing objects: 80% (24/30)\rremote: Compressing objects: 83% (25/30)\rremote: Compressing objects: 86% (26/30)\rremote: Compressing objects: 90% (27/30)\rremote: Compressing objects: 93% (28/30)\rremote: Compressing objects: 96% (29/30)\rremote: Compressing objects: 100% (30/30)\rremote: Compressing objects: 100% (30/30), done.\r\nremote: Total 42 (delta 15), reused 35 (delta 10), pack-reused 0 (from 0)\r\n",,terminal_output +321,1139948,"TERMINAL",0,0,"Unpacking objects: 2% (1/42)\rUnpacking objects: 4% (2/42)\r",,terminal_output +322,1140038,"TERMINAL",0,0,"Unpacking objects: 7% (3/42)\rUnpacking objects: 9% (4/42)\rUnpacking objects: 11% (5/42)\rUnpacking objects: 14% (6/42)\rUnpacking objects: 16% (7/42)\r",,terminal_output +323,1140358,"TERMINAL",0,0,"Unpacking objects: 19% (8/42)\rUnpacking objects: 21% (9/42)\rUnpacking objects: 23% (10/42)\rUnpacking objects: 26% (11/42)\rUnpacking objects: 28% (12/42)\rUnpacking objects: 30% (13/42)\rUnpacking objects: 33% (14/42)\rUnpacking objects: 35% (15/42)\rUnpacking objects: 38% (16/42)\rUnpacking objects: 40% (17/42)\rUnpacking objects: 42% (18/42)\rUnpacking objects: 45% (19/42)\rUnpacking objects: 47% (20/42)\rUnpacking objects: 50% (21/42)\rUnpacking objects: 52% (22/42)\rUnpacking objects: 54% (23/42)\rUnpacking objects: 57% (24/42)\rUnpacking objects: 59% (25/42)\rUnpacking objects: 61% (26/42)\rUnpacking objects: 64% (27/42)\rUnpacking objects: 66% (28/42)\rUnpacking objects: 69% (29/42), 5.36 KiB | 10.00 KiB/s\rUnpacking objects: 71% (30/42), 5.36 KiB | 10.00 KiB/s\rUnpacking objects: 73% (31/42), 5.36 KiB | 10.00 KiB/s\r",,terminal_output +324,1140456,"TERMINAL",0,0,"Unpacking objects: 76% (32/42), 5.36 KiB | 10.00 KiB/s\rUnpacking objects: 78% (33/42), 5.36 KiB | 10.00 KiB/s\rUnpacking objects: 80% (34/42), 5.36 KiB | 10.00 KiB/s\rUnpacking objects: 83% (35/42), 5.36 KiB | 10.00 KiB/s\rUnpacking objects: 85% (36/42), 5.36 KiB | 10.00 KiB/s\rUnpacking objects: 88% (37/42), 5.36 KiB | 10.00 KiB/s\rUnpacking objects: 90% (38/42), 5.36 KiB | 10.00 KiB/s\r",,terminal_output +325,1140586,"TERMINAL",0,0,"Unpacking objects: 92% (39/42), 5.36 KiB | 10.00 KiB/s\rUnpacking objects: 95% (40/42), 5.36 KiB | 10.00 KiB/s\rUnpacking objects: 97% (41/42), 5.36 KiB | 10.00 KiB/s\rUnpacking objects: 100% (42/42), 5.36 KiB | 10.00 KiB/s\rUnpacking objects: 100% (42/42), 6.88 KiB | 10.00 KiB/s, done.\r\n",,terminal_output +326,1141139,"TERMINAL",0,0,"From github.com:p-doom/slurm\r\n 5ba7db4..d92a560 main -> origin/main\r\n",,terminal_output +327,1141301,"TERMINAL",0,0,"Updating 5ba7db4..d92a560\r\n",,terminal_output +328,1141477,"TERMINAL",0,0,"Fast-forward\r\n",,terminal_output +329,1141522,"TERMINAL",0,0," dev/alfred/horeka/input_pipeline_ws/actions/download_actions.sbatch | 18 +++++++\r\n dev/alfred/horeka/input_pipeline_ws/actions/download_actions_all.sh | 17 +++++++\r\n dev/alfred/horeka/jobs_cur/atari/sample_causal.sbatch | 35 +++++++++++++\r\n dev/alfred/horeka/jobs_cur/atari/sample_maskgit.sbatch | 35 +++++++++++++\r\n dev/alfred/horeka/jobs_cur/atari/train_dynamics_causal.sbatch | 83 ++++++++++++++++++++++++++++++\r\n dev/alfred/horeka/jobs_cur/atari/train_dynamics_maskgit.sbatch | 80 +++++++++++++++++++++++++++++\r\n dev/alfred/horeka/jobs_cur/atari/train_tokenizer_lr_1e-4.sbatch | 74 +++++++++++++++++++++++++++\r\n .../dyn_gt_actions_ablation_prepend/dyn_gt_actions_ablation_single_gpu copy.sbatch | 114 ++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/atari_dynamics/atari_dyn_maskgit_causal.sh | 13 +++++\r\n jobs/alfred/berlin/atari_dynamics/atari_dynamics.sbatch | 80 +++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/{ => downlaod}/download_actions.sbatch | 0\r\n jobs/alfred/berlin/{ => downlaod}/download_actions.sh | 0\r\n jobs/alfred/horeka/atari_dynamics/atari_dyn_maskgit_causal.sh | 13 +++++\r\n jobs/alfred/horeka/atari_dynamics/atari_dynamics.sbatch | 83 ++++++++++++++++++++++++++++++\r\n utils/alfred/scp_scripts/copy_to_berlin.sh | 7 +++\r\n 15 files changed, 652 insertions(+)\r\n create mode 100644 dev/alfred/horeka/input_pipeline_ws/actions/download_actions.sbatch\r\n create mode 100644 dev/alfred/horeka/input_pipeline_ws/actions/download_actions_all.sh\r\n create mode 100644 dev/alfred/horeka/jobs_cur/atari/sample_causal.sbatch\r\n create mode 100644 dev/alfred/horeka/jobs_cur/atari/sample_maskgit.sbatch\r\n create mode 100644 dev/alfred/horeka/jobs_cur/atari/train_dynamics_causal.sbatch\r\n create mode 100644 dev/alfred/horeka/jobs_cur/atari/train_dynamics_maskgit.sbatch\r\n create mode 100644 dev/alfred/horeka/jobs_cur/atari/train_tokenizer_lr_1e-4.sbatch\r\n create mode 100644 dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/dyn_gt_actions_ablation_single_gpu copy.sbatch\r\n create mode 100644 jobs/alfred/berlin/atari_dynamics/atari_dyn_maskgit_causal.sh\r\n create mode 100644 jobs/alfred/berlin/atari_dynamics/atari_dynamics.sbatch\r\n rename jobs/alfred/berlin/{ => downlaod}/download_actions.sbatch (100%)\r\n rename jobs/alfred/berlin/{ => downlaod}/download_actions.sh (100%)\r\n create mode 100644 jobs/alfred/horeka/atari_dynamics/atari_dyn_maskgit_causal.sh\r\n create mode 100644 jobs/alfred/horeka/atari_dynamics/atari_dynamics.sbatch\r\n create mode 100644 utils/alfred/scp_scripts/copy_to_berlin.sh\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar/slurm]633;D;0",,terminal_output +330,1231382,"TERMINAL",0,0,"git commit -am ""update scripts""",,terminal_command +331,1231437,"TERMINAL",0,0,"]633;E;2025-08-14 10:58:04 git commit -am ""update scripts"";7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +332,1231678,"TERMINAL",0,0,"[main d700975] update scripts\r\n 4 files changed, 4 insertions(+), 4 deletions(-)\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar/slurm]633;D;0",,terminal_output +333,1233815,"TERMINAL",0,0,"git push",,terminal_command +334,1233888,"TERMINAL",0,0,"]633;E;2025-08-14 10:58:06 git push;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +335,1235090,"TERMINAL",0,0,"Enumerating objects: 19, done.\r\nCounting objects: 5% (1/19)\rCounting objects: 10% (2/19)\rCounting objects: 15% (3/19)\rCounting objects: 21% (4/19)\rCounting objects: 26% (5/19)\rCounting objects: 31% (6/19)\rCounting objects: 36% (7/19)\rCounting objects: 42% (8/19)\rCounting objects: 47% (9/19)\rCounting objects: 52% (10/19)\rCounting objects: 57% (11/19)\rCounting objects: 63% (12/19)\rCounting objects: 68% (13/19)\rCounting objects: 73% (14/19)\rCounting objects: 78% (15/19)\rCounting objects: 84% (16/19)\rCounting objects: 89% (17/19)\rCounting objects: 94% (18/19)\rCounting objects: 100% (19/19)\rCounting objects: 100% (19/19), done.\r\nDelta compression using up to 152 threads\r\nCompressing objects: 9% (1/11)\rCompressing objects: 18% (2/11)\rCompressing objects: 27% (3/11)\rCompressing objects: 36% (4/11)\rCompressing objects: 45% (5/11)\rCompressing objects: 54% (6/11)\rCompressing objects: 63% (7/11)\rCompressing objects: 72% (8/11)\rCompressing objects: 81% (9/11)\rCompressing objects: 90% (10/11)\rCompressing objects: 100% (11/11)\rCompressing objects: 100% (11/11), done.\r\nWriting objects: 9% (1/11)\rWriting objects: 18% (2/11)\rWriting objects: 27% (3/11)\rWriting objects: 36% (4/11)\rWriting objects: 45% (5/11)\rWriting objects: 54% (6/11)\rWriting objects: 63% (7/11)\rWriting objects: 72% (8/11)\rWriting objects: 81% (9/11)\rWriting objects: 90% (10/11)\rWriting objects: 100% (11/11)\rWriting objects: 100% (11/11), 938 bytes | 469.00 KiB/s, done.\r\nTotal 11 (delta 8), reused 0 (delta 0), pack-reused 0\r\n",,terminal_output +336,1235220,"TERMINAL",0,0,"remote: Resolving deltas: 0% (0/8)\rremote: Resolving deltas: 12% (1/8)\rremote: Resolving deltas: 25% (2/8)\rremote: Resolving deltas: 37% (3/8)\rremote: Resolving deltas: 50% (4/8)\rremote: Resolving deltas: 62% (5/8)\rremote: Resolving deltas: 75% (6/8)\rremote: Resolving deltas: 87% (7/8)\rremote: Resolving deltas: 100% (8/8)\rremote: Resolving deltas: 100% (8/8), completed with 6 local objects.\r\n",,terminal_output +337,1235369,"TERMINAL",0,0,"To github.com:p-doom/slurm.git\r\n d92a560..d700975 main -> main\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar/slurm]633;D;0",,terminal_output +338,1239124,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_dev.sh",0,0,"# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/holiday/causal/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\n# tokenizer with the new structure supporting larger ffn_dim\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3416521\n\nenv | grep SLURM\n\nexport PYTHONUNBUFFERED=1\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=24 \\n --init_lr=0 \\n --darkness_threshold=50 \\n --max_lr=8e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-causal-test-$slurm_job_id \\n --tags dynamics causal \\n --entity instant-uv \\n --project jafar \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16 \\n --dyna_ffn_dim=4096 \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,tab +339,1240177,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +340,3809848,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2095,0,"",shellscript,selection_mouse +341,3809884,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2094,0,"",shellscript,selection_command +342,3815250,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/atari/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/atari/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_atari/array_records\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n --image_height=84 \\n --image_width=64 \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=atari-tokenizer-1e-4-$slurm_job_id \\n --tags tokenizer atari 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n\n",shellscript,tab +343,3816390,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1720,0,"",shellscript,selection_mouse +344,3816965,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1793,0,"",shellscript,selection_mouse +345,3816972,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1792,0,"",shellscript,selection_command +346,3837141,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1751,0,"",shellscript,selection_mouse +347,3838491,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1751,1,"",shellscript,content +348,3838614,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1751,1,"",shellscript,content +349,3839909,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1751,0,"4",shellscript,content +350,3839910,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1752,0,"",shellscript,selection_keyboard +351,3840023,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1752,0,"8",shellscript,content +352,3840024,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1753,0,"",shellscript,selection_keyboard +353,3840247,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1752,0,"",shellscript,selection_command +354,3842206,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",38,0,"",shellscript,selection_mouse +355,3842220,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",37,0,"",shellscript,selection_command +356,3842434,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",38,0,"",shellscript,selection_command +357,3842607,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",37,1,"",shellscript,content +358,3842709,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",37,0,"1",shellscript,content +359,3842710,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",38,0,"",shellscript,selection_keyboard +360,3843118,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",37,0,"",shellscript,selection_command +361,3853599,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1788,0,"",shellscript,selection_mouse +362,3854535,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1787,1,"",shellscript,content +363,3854771,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1787,0,"3",shellscript,content +364,3854772,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1788,0,"",shellscript,selection_keyboard +365,3855780,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1790,0,"",shellscript,selection_command +366,3855963,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1791,0,"",shellscript,selection_command +367,3856416,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1790,1,"",shellscript,content +368,3856503,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1790,0,"5",shellscript,content +369,3856504,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1791,0,"",shellscript,selection_keyboard +370,3859471,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1903,0,"",shellscript,selection_mouse +371,3860821,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1902,1,"",shellscript,content +372,3860910,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1902,0,"3",shellscript,content +373,3860911,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1903,0,"",shellscript,selection_keyboard +374,3861109,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1904,0,"",shellscript,selection_command +375,3861325,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1905,0,"",shellscript,selection_command +376,3861519,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1906,0,"",shellscript,selection_command +377,3862033,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1905,1,"",shellscript,content +378,3862132,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1905,0,"5",shellscript,content +379,3862133,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1906,0,"",shellscript,selection_keyboard +380,3862341,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1954,0,"",shellscript,selection_command +381,3862840,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1953,1,"",shellscript,content +382,3862876,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1953,0,"5",shellscript,content +383,3862877,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1954,0,"",shellscript,selection_keyboard +384,3863095,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1953,0,"",shellscript,selection_command +385,3863299,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1952,0,"",shellscript,selection_command +386,3863469,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1951,0,"",shellscript,selection_command +387,3864473,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1950,1,"",shellscript,content +388,3864561,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1950,0,"3",shellscript,content +389,3864562,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1951,0,"",shellscript,selection_keyboard +390,3865725,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",2041,0,"",shellscript,selection_mouse +391,3866304,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",2036,0,"",shellscript,selection_mouse +392,3866868,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",2004,0,"",shellscript,selection_mouse +393,3867435,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1982,0,"",shellscript,selection_mouse +394,3867973,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1874,0,"",shellscript,selection_mouse +395,3868613,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1825,0,"",shellscript,selection_mouse +396,3868898,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1825,37,"\n --log_checkpoint_interval=1000 \",shellscript,selection_mouse +397,3869081,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1862,0,"",shellscript,selection_mouse +398,3869724,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1793,0,"",shellscript,selection_mouse +399,3870328,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1862,0,"",shellscript,selection_mouse +400,3870940,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",1825,0,"",shellscript,selection_mouse +401,4031145,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",122,0,"",shellscript,selection_mouse +402,4032436,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",38,0,"",shellscript,selection_mouse +403,4033032,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",90,0,"",shellscript,selection_mouse +404,4033540,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",169,0,"",shellscript,selection_mouse +405,4034057,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",456,0,"",shellscript,selection_mouse +406,4034548,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",508,0,"",shellscript,selection_mouse +407,4213525,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",543,0,"",shellscript,selection_mouse +408,4214091,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_1e-4.sbatch",704,0,"",shellscript,selection_mouse +409,4223818,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/atari/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/atari/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_atari/array_records\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n --image_height=84 \\n --image_width=64 \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=48 \\n --init_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=atari-tokenizer-3e-5-$slurm_job_id \\n --tags tokenizer atari 3e-5 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n\n",shellscript,tab +410,4240554,"TERMINAL",0,0,"queue",,terminal_command +411,4240662,"TERMINAL",0,0,"]633;E;2025-08-14 11:48:13 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 11:48:13 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-17:48:12\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 21:59:17\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]",,terminal_output +412,4241732,"TERMINAL",0,0,"438",,terminal_output +413,4242756,"TERMINAL",0,0,"549",,terminal_output +414,4243782,"TERMINAL",0,0,"6520",,terminal_output +415,4244907,"TERMINAL",0,0,"761",,terminal_output +416,4245933,"TERMINAL",0,0,"883",,terminal_output +417,4246930,"TERMINAL",0,0,"2094",,terminal_output +418,4247980,"TERMINAL",0,0,"1205",,terminal_output +419,4248325,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",489,0,"",shellscript,selection_mouse +420,4249208,"TERMINAL",0,0,"216",,terminal_output +421,4250233,"TERMINAL",0,0,"327",,terminal_output +422,4250740,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",435,0,"",shellscript,selection_mouse +423,4251164,"TERMINAL",0,0,"438",,terminal_output +424,4251885,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",434,1,"",shellscript,content +425,4251955,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",434,0,"3",shellscript,content +426,4251956,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",435,0,"",shellscript,selection_keyboard +427,4252188,"TERMINAL",0,0,"549",,terminal_output +428,4252388,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",436,0,"",shellscript,selection_command +429,4252548,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",437,0,"",shellscript,selection_command +430,4252713,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",438,0,"",shellscript,selection_command +431,4252998,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",437,1,"",shellscript,content +432,4253190,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",437,0,"5",shellscript,content +433,4253191,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",438,0,"",shellscript,selection_keyboard +434,4253225,"TERMINAL",0,0,"6530",,terminal_output +435,4253819,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",437,0,"",shellscript,selection_command +436,4254252,"TERMINAL",0,0,"761",,terminal_output +437,4254538,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",576,0,"",shellscript,selection_mouse +438,4254539,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",575,0,"",shellscript,selection_command +439,4255288,"TERMINAL",0,0,"872",,terminal_output +440,4256326,"TERMINAL",0,0,"983",,terminal_output +441,4257385,"TERMINAL",0,0,"3094",,terminal_output +442,4258414,"TERMINAL",0,0,"1305",,terminal_output +443,4259548,"TERMINAL",0,0,"216",,terminal_output +444,4260573,"TERMINAL",0,0,"327",,terminal_output +445,4261646,"TERMINAL",0,0,"438",,terminal_output +446,4262633,"TERMINAL",0,0,"549",,terminal_output +447,4263750,"TERMINAL",0,0,"6540",,terminal_output +448,4264632,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar/slurm]633;D;0",,terminal_output +449,4267871,"TERMINAL",0,0,"idle",,terminal_command +450,4267935,"TERMINAL",0,0,"]633;E;2025-08-14 11:48:41 idle;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CPartition dev_cpuonly : 12 nodes idle\r\nPartition cpuonly : 182 nodes idle\r\nPartition dev_accelerated : 0 nodes idle\r\nPartition accelerated : 33 nodes idle\r\nPartition dev_accelerated-h100 : 0 nodes idle\r\nPartition accelerated-h100 : 1 nodes idle\r\nPartition large : 6 nodes idle\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar/slurm]633;D;0",,terminal_output +451,4558798,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",456,0,"",shellscript,selection_mouse +452,4558834,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",455,0,"",shellscript,selection_command +453,4560453,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",415,0,"",shellscript,selection_command +454,4560714,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",438,0,"\n",shellscript,content +455,4560979,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",439,0,"--reservation=llmtum",shellscript,content +456,4562522,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",400,0,"",shellscript,selection_mouse +457,4562652,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",400,6,"SBATCH",shellscript,selection_mouse +458,4563460,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",450,0,"",shellscript,selection_mouse +459,4563833,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",449,0,"",shellscript,selection_command +460,4564077,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",409,0,"",shellscript,selection_command +461,4564613,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",408,0,"",shellscript,selection_command +462,4564768,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",407,0,"",shellscript,selection_command +463,4564963,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",406,0,"",shellscript,selection_command +464,4565613,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",405,0,"",shellscript,selection_command +465,4566537,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",400,0,"",shellscript,selection_command +466,4567024,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",440,0,"",shellscript,selection_command +467,4567436,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",439,0,"",shellscript,selection_command +468,4568409,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",439,0,"#",shellscript,content +469,4568411,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",440,0,"",shellscript,selection_keyboard +470,4569518,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",439,0,"",shellscript,selection_command +471,4569715,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",440,0,"SBATC",shellscript,content +472,4569742,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",444,0,"",shellscript,selection_command +473,4570697,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",445,0,"",shellscript,selection_command +474,4571363,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",445,0,"H",shellscript,content +475,4571364,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",446,0,"",shellscript,selection_keyboard +476,4571788,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",446,0," ",shellscript,content +477,4571789,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",447,0,"",shellscript,selection_keyboard +478,4572095,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",446,0,"",shellscript,selection_command +479,4661432,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",485,0,"",shellscript,selection_mouse +480,4661455,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",484,0,"",shellscript,selection_command +481,4661981,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",465,0,"",shellscript,selection_mouse +482,4683705,"TERMINAL",0,0,"cd ..",,terminal_command +483,4685431,"TERMINAL",0,0,"git branch",,terminal_command +484,4685544,"TERMINAL",0,0,"]633;E;2025-08-14 11:55:38 git branch;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1h=\r add-wandb-name-and-tags\r\n before-nnx\r\n causal-mem-reduce\r\n causal-spatiotemporal-kv-cache\r\n causal-st-transformer\r\n causal-transformer-dynamics-model\r\n causal-transformer-nnx-no-kv-cache\r\n convert-to-jax-array-in-iter\r\n correct-batched-sampling\r\n dev\r\n dont-let-tf-see-gpu\r\n* feat/darkness-filter\r\n:",,terminal_output +485,4686845,"TERMINAL",0,0,"\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +486,4698666,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-1e-4-$slurm_job_id \\n --tags tokenizer big-run 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n\n",shellscript,tab +487,4709173,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=256 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-8-nodes-$slurm_job_id \\n --tags tokenizer big-run 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid",shellscript,tab +488,4710536,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes copy.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=256 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-8-nodes-$slurm_job_id \\n --tags tokenizer big-run 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid",shellscript,tab +489,4719300,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=256 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-8-nodes-$slurm_job_id \\n --tags tokenizer big-run 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid",shellscript,tab +490,4720488,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",38,0,"",shellscript,selection_mouse +491,4720490,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",37,0,"",shellscript,selection_command +492,4720781,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",38,0,"",shellscript,selection_command +493,4720965,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",37,1,"",shellscript,content +494,4721284,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",37,0,"4",shellscript,content +495,4721285,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",38,0,"",shellscript,selection_keyboard +496,4721766,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",37,0,"",shellscript,selection_command +497,4726498,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1734,0,"",shellscript,selection_mouse +498,4727626,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1733,1,"",shellscript,content +499,4727759,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1732,1,"",shellscript,content +500,4728106,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1731,1,"",shellscript,content +501,4733650,"TERMINAL",0,0,"python",,terminal_command +502,4733715,"TERMINAL",0,0,"]633;E;2025-08-14 11:56:26 python;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CPython 3.9.18 (main, Jun 27 2025, 00:00:00) \r\n[GCC 11.4.1 20231218 (Red Hat 11.4.1-4)] on linux\r\nType ""help"", ""copyright"", ""credits"" or ""license"" for more information.\r\n",,terminal_output +503,4733784,"TERMINAL",0,0,">>> ",,terminal_output +504,4752362,"TERMINAL",0,0,"^?",,terminal_output +505,4752800,"TERMINAL",0,0,"[?25l?[?25h",,terminal_output +506,4752929,"TERMINAL",0,0,"",,terminal_output +507,4753486,"TERMINAL",0,0,"384\r\n384\r\n>>> ",,terminal_output +508,4754869,"TERMINAL",0,0,"384\r\n384\r\n>>> ",,terminal_output +509,4756704,"TERMINAL",0,0,"3",,terminal_output +510,4757793,"TERMINAL",0,0,"[?25l8[?25h",,terminal_output +511,4757896,"TERMINAL",0,0,"[?25l4[?25h",,terminal_output +512,4758585,"TERMINAL",0,0,"[?25l/[?25h",,terminal_output +513,4758884,"TERMINAL",0,0,"[?25l2[?25h",,terminal_output +514,4758955,"TERMINAL",0,0,"\r\n192.0\r\n>>> ",,terminal_output +515,4761038,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1730,0,"",shellscript,selection_mouse +516,4762033,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1731,0,"",shellscript,selection_mouse +517,4762633,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1731,0,"1",shellscript,content +518,4762634,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1732,0,"",shellscript,selection_keyboard +519,4762769,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1732,0,"9",shellscript,content +520,4762770,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1733,0,"",shellscript,selection_keyboard +521,4762883,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1733,0,"2",shellscript,content +522,4762884,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1734,0,"",shellscript,selection_keyboard +523,4766449,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",0,0,"",shellscript,tab +524,4770566,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",1734,0,"",shellscript,selection_command +525,4771184,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",1733,1,"",shellscript,content +526,4771283,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",1732,1,"",shellscript,content +527,4771440,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",1731,1,"",shellscript,content +528,4772194,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",1731,0,"3",shellscript,content +529,4772195,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",1732,0,"",shellscript,selection_keyboard +530,4772998,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",1732,0,"8",shellscript,content +531,4772998,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",1733,0,"",shellscript,selection_keyboard +532,4773137,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",1733,0,"4",shellscript,content +533,4773138,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",1734,0,"",shellscript,selection_keyboard +534,4814230,"TERMINAL",0,0,"1",,terminal_output +535,4814501,"TERMINAL",0,0,"[?25l9[?25h",,terminal_output +536,4814611,"TERMINAL",0,0,"[?25l2[?25h",,terminal_output +537,4815272,"TERMINAL",0,0,"[?25l/[?25h[?25l2[?25h",,terminal_output +538,4815516,"TERMINAL",0,0,"\r\n96.0\r\n>>> ",,terminal_output +539,4821781,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",0,0,"",shellscript,tab +540,4824111,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes copy.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=4\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=192 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-8-nodes-$slurm_job_id \\n --tags tokenizer big-run 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid",shellscript,tab +541,4830145,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=4\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=192 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-8-nodes-$slurm_job_id \\n --tags tokenizer big-run 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid",shellscript,tab +542,4831142,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",38,0,"",shellscript,selection_mouse +543,4831143,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",37,0,"",shellscript,selection_command +544,4831834,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",38,0,"",shellscript,selection_command +545,4832023,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",37,1,"",shellscript,content +546,4832142,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",37,0,"2",shellscript,content +547,4832143,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",38,0,"",shellscript,selection_keyboard +548,4833022,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",37,0,"",shellscript,selection_command +549,4835791,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1733,0,"",shellscript,selection_mouse +550,4836470,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1734,0,"",shellscript,selection_command +551,4836899,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1733,1,"",shellscript,content +552,4837002,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1732,1,"",shellscript,content +553,4837309,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1731,1,"",shellscript,content +554,4838043,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1731,0,"9",shellscript,content +555,4838044,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1732,0,"",shellscript,selection_keyboard +556,4838228,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1732,0,"6",shellscript,content +557,4838229,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1733,0,"",shellscript,selection_keyboard +558,4842423,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",0,0,"",shellscript,tab +559,4843518,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",485,0,"",shellscript,selection_mouse +560,4843541,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",484,0,"",shellscript,selection_command +561,4843813,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",466,0,"",shellscript,selection_mouse +562,4848799,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",0,0,"",shellscript,tab +563,4852071,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",415,0,"",shellscript,selection_mouse +564,4853603,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",414,0,"",shellscript,selection_command +565,4854012,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",442,0,"\n#SBATCH --reservation=llmtum",shellscript,content +566,4854035,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",443,0,"",shellscript,selection_command +567,4855910,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",0,0,"",shellscript,tab +568,4856629,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",415,0,"",shellscript,selection_mouse +569,4857135,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",414,0,"",shellscript,selection_command +570,4857474,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",442,0,"\n#SBATCH --reservation=llmtum",shellscript,content +571,4857506,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",443,0,"",shellscript,selection_command +572,4882441,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",471,0,"",shellscript,selection_mouse +573,4882444,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",470,0,"",shellscript,selection_command +574,4883369,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",541,0,"",shellscript,selection_mouse +575,4884848,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",0,0,"",shellscript,tab +576,4887897,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",0,0,"",shellscript,tab +577,4889275,"TERMINAL",0,0,"\r\n",,terminal_output +578,4889332,"TERMINAL",0,0,"]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +579,4894460,"TERMINAL",0,0,"",,terminal_focus +580,4896722,"TERMINAL",0,0,"source /home/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/bin/activate",,terminal_command +581,4899107,"TERMINAL",0,0,"runner",,terminal_command +582,4900920,"TERMINAL",0,0,"sync-runner",,terminal_command +583,4901026,"TERMINAL",0,0,"]633;E;2025-08-14 11:59:14 sync-runner;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;Csending incremental file list\r\n",,terminal_output +584,4903181,"TERMINAL",0,0,"slurm/dev/alfred/horeka/\r\nslurm/dev/alfred/horeka/input_pipeline_ws/\r\nslurm/dev/alfred/horeka/input_pipeline_ws/actions/\r\nslurm/dev/alfred/horeka/input_pipeline_ws/actions/download_actions.sbatch\r\nslurm/dev/alfred/horeka/input_pipeline_ws/actions/download_actions_all.sh\r\nslurm/dev/alfred/horeka/jobs_cur/\r\nslurm/dev/alfred/horeka/jobs_cur/atari/\r\nslurm/dev/alfred/horeka/jobs_cur/atari/sample_causal.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/atari/sample_maskgit.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/atari/train_dynamics_causal.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/atari/train_dynamics_maskgit.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/atari/train_tokenizer_lr_1e-4.sbatch\r\nslurm/dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/\r\nslurm/dev/alfred/horeka/jobs_cur/dyn_gt_actions_ablation_prepend/dyn_gt_actions_ablation_single_gpu copy.sbatch\r\nslurm/jobs/alfred/\r\nslurm/jobs/alfred/berlin/\r\nslurm/jobs/alfred/berlin/atari_dynamics/\r\nslurm/jobs/alfred/berlin/atari_dynamics/atari_dyn_maskgit_causal.sh\r\nslurm/jobs/alfred/berlin/atari_dynamics/atari_dynamics.sbatch\r\nslurm/jobs/alfred/berlin/downlaod/\r\nslurm/jobs/alfred/berlin/downlaod/download_actions.sbatch\r\nslurm/jobs/alfred/berlin/downlaod/download_actions.sh\r\nslurm/jobs/alfred/horeka/\r\nslurm/jobs/alfred/horeka/atari_dynamics/\r\nslurm/jobs/alfred/horeka/atari_dynamics/atari_dyn_maskgit_causal.sh\r\nslurm/jobs/alfred/horeka/atari_dynamics/atari_dynamics.sbatch\r\nslurm/jobs/mihir/horeka/atari/\r\nslurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch\r\nslurm/utils/alfred/\r\nslurm/utils/alfred/scp_scripts/\r\nslurm/utils/alfred/scp_scripts/copy_to_berlin.sh\r\n",,terminal_output +585,4903560,"TERMINAL",0,0,"\r\nsent 52,243 bytes received 574 bytes 21,126.80 bytes/sec\r\ntotal size is 128,148,730 speedup is 2,426.28\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +586,4907524,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_8nodes.sbatch",0,0,"",shellscript,tab +587,4909337,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",0,0,"",shellscript,tab +588,4915400,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",0,0,"",shellscript,tab +589,4992724,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",,terminal_command +590,4992770,"TERMINAL",0,0,"]633;E;2025-08-14 12:00:45 sbatch slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;CSubmitted batch job 3421964\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +591,4994722,"TERMINAL",0,0,"queue",,terminal_command +592,4994792,"TERMINAL",0,0,"]633;E;2025-08-14 12:00:47 queue;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 12:00:47 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-18:00:46\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 22:11:51\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3421964 accelerat train_to tum_cte0 R\t0:01\t 1 hkn0720",,terminal_output +593,4995830,"TERMINAL",0,0,"8833",,terminal_output +594,4996874,"TERMINAL",0,0,"50944",,terminal_output +595,4997905,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +596,5003824,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",0,0,"",shellscript,tab +597,5008331,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",,terminal_command +598,5008408,"TERMINAL",0,0,"]633;E;2025-08-14 12:01:01 sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;CSubmitted batch job 3421973\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +599,5009784,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",0,0,"",shellscript,tab +600,5016350,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",,terminal_command +601,5016428,"TERMINAL",0,0,"]633;E;2025-08-14 12:01:09 sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;CSubmitted batch job 3421982\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +602,5017682,"TERMINAL",0,0,"queue",,terminal_command +603,5017800,"TERMINAL",0,0,"]633;E;2025-08-14 12:01:10 queue;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 12:01:10 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-18:01:09\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 22:12:14\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3421982 accelerat train_to tum_cte0 R\t0:00\t 4 hkn[0724,0726,0728,0731]3421973 accelerat train_to tum_cte0 R\t0:09\t 2 hkn[0722-0723]3421964 accelerat train_to tum_cte0 R\t0:24\t 1 hkn0720",,terminal_output +604,5018855,"TERMINAL",0,0,"11051105",,terminal_output +605,5019879,"TERMINAL",0,0,"227327",,terminal_output +606,5020889,"TERMINAL",0,0,"438438",,terminal_output +607,5021938,"TERMINAL",0,0,"549549",,terminal_output +608,5023054,"TERMINAL",0,0,"65206530",,terminal_output +609,5024026,"TERMINAL",0,0,"761761",,terminal_output +610,5025101,"TERMINAL",0,0,"872872",,terminal_output +611,5026122,"TERMINAL",0,0,"983983",,terminal_output +612,5027250,"TERMINAL",0,0,"20941094",,terminal_output +613,5028208,"TERMINAL",0,0,"12051205",,terminal_output +614,5029298,"TERMINAL",0,0,"216216",,terminal_output +615,5030528,"TERMINAL",0,0,"327327",,terminal_output +616,5031534,"TERMINAL",0,0,"438438",,terminal_output +617,5032781,"TERMINAL",0,0,"549549",,terminal_output +618,5033809,"TERMINAL",0,0,"65306540",,terminal_output +619,5034761,"TERMINAL",0,0,"761761",,terminal_output +620,5035716,"TERMINAL",0,0,"872872",,terminal_output +621,5036762,"TERMINAL",0,0,"983983",,terminal_output +622,5037809,"TERMINAL",0,0,"30942094",,terminal_output +623,5038855,"TERMINAL",0,0,"13162316",,terminal_output +624,5039901,"TERMINAL",0,0,"327327",,terminal_output +625,5040987,"TERMINAL",0,0,"438438",,terminal_output +626,5042211,"TERMINAL",0,0,"549549",,terminal_output +627,5043083,"TERMINAL",0,0,"65406550",,terminal_output +628,5044099,"TERMINAL",0,0,"761761",,terminal_output +629,5045169,"TERMINAL",0,0,"872872",,terminal_output +630,5046202,"TERMINAL",0,0,"983983",,terminal_output +631,5047323,"TERMINAL",0,0,"40943094",,terminal_output +632,5048344,"TERMINAL",0,0,"14051405",,terminal_output +633,5049368,"TERMINAL",0,0,"216216",,terminal_output +634,5050394,"TERMINAL",0,0,"327327",,terminal_output +635,5051526,"TERMINAL",0,0,"438438",,terminal_output +636,5052544,"TERMINAL",0,0,"549549",,terminal_output +637,5053668,"TERMINAL",0,0,"6550651:00",,terminal_output +638,5054704,"TERMINAL",0,0,"761761",,terminal_output +639,5055615,"TERMINAL",0,0,"872872",,terminal_output +640,5056743,"TERMINAL",0,0,"983983",,terminal_output +641,5057708,"TERMINAL",0,0,"50944094",,terminal_output +642,5058751,"TERMINAL",0,0,"15051505",,terminal_output +643,5059835,"TERMINAL",0,0,"216216",,terminal_output +644,5060839,"TERMINAL",0,0,"338438",,terminal_output +645,5061890,"TERMINAL",0,0,"549549",,terminal_output +646,5062938,"TERMINAL",0,0,"653:006510",,terminal_output +647,5063979,"TERMINAL",0,0,"761761",,terminal_output +648,5065028,"TERMINAL",0,0,"872872",,terminal_output +649,5066077,"TERMINAL",0,0,"983983",,terminal_output +650,5067199,"TERMINAL",0,0,"2:00945094",,terminal_output +651,5068169,"TERMINAL",0,0,"12:00511:005",,terminal_output +652,5069248,"TERMINAL",0,0,"216216",,terminal_output +653,5070256,"TERMINAL",0,0,"327327",,terminal_output +654,5071385,"TERMINAL",0,0,"438438",,terminal_output +655,5072512,"TERMINAL",0,0,"549549",,terminal_output +656,5073401,"TERMINAL",0,0,"65106520",,terminal_output +657,5074618,"TERMINAL",0,0,"761761",,terminal_output +658,5075582,"TERMINAL",0,0,"872872",,terminal_output +659,5076607,"TERMINAL",0,0,"983983",,terminal_output +660,5077634,"TERMINAL",0,0,"10941:0094",,terminal_output +661,5078859,"TERMINAL",0,0,"11051105",,terminal_output +662,5079678,"TERMINAL",0,0,"216216",,terminal_output +663,5080908,"TERMINAL",0,0,"327327",,terminal_output +664,5081936,"TERMINAL",0,0,"438438",,terminal_output +665,5082804,"TERMINAL",0,0,"549549",,terminal_output +666,5083856,"TERMINAL",0,0,"66217631",,terminal_output +667,5084903,"TERMINAL",0,0,"872872",,terminal_output +668,5085965,"TERMINAL",0,0,"983983",,terminal_output +669,5087063,"TERMINAL",0,0,"20941094",,terminal_output +670,5088052,"TERMINAL",0,0,"12051205",,terminal_output +671,5089202,"TERMINAL",0,0,"216216",,terminal_output +672,5090249,"TERMINAL",0,0,"327327",,terminal_output +673,5091299,"TERMINAL",0,0,"438438",,terminal_output +674,5092280,"TERMINAL",0,0,"549549",,terminal_output +675,5093301,"TERMINAL",0,0,"65306540",,terminal_output +676,5094327,"TERMINAL",0,0,"761761",,terminal_output +677,5095376,"TERMINAL",0,0,"872872",,terminal_output +678,5096518,"TERMINAL",0,0,"983983",,terminal_output +679,5097465,"TERMINAL",0,0,"30942094",,terminal_output +680,5098520,"TERMINAL",0,0,"13051305",,terminal_output +681,5099757,"TERMINAL",0,0,"216216",,terminal_output +682,5100673,"TERMINAL",0,0,"327327",,terminal_output +683,5101806,"TERMINAL",0,0,"438438",,terminal_output +684,5102799,"TERMINAL",0,0,"549549",,terminal_output +685,5103786,"TERMINAL",0,0,"65406550",,terminal_output +686,5104873,"TERMINAL",0,0,"761761",,terminal_output +687,5106099,"TERMINAL",0,0,"883983",,terminal_output +688,5107123,"TERMINAL",0,0,"40943094",,terminal_output +689,5108163,"TERMINAL",0,0,"14051405",,terminal_output +690,5109199,"TERMINAL",0,0,"216216",,terminal_output +691,5110374,"TERMINAL",0,0,"327327",,terminal_output +692,5111302,"TERMINAL",0,0,"438438",,terminal_output +693,5112455,"TERMINAL",0,0,"549549",,terminal_output +694,5113404,"TERMINAL",0,0,"6550652:00",,terminal_output +695,5114454,"TERMINAL",0,0,"761761",,terminal_output +696,5115520,"TERMINAL",0,0,"872872",,terminal_output +697,5116646,"TERMINAL",0,0,"983983",,terminal_output +698,5117605,"TERMINAL",0,0,"50944094",,terminal_output +699,5118652,"TERMINAL",0,0,"15051505",,terminal_output +700,5119721,"TERMINAL",0,0,"216216",,terminal_output +701,5120851,"TERMINAL",0,0,"327327",,terminal_output +702,5121802,"TERMINAL",0,0,"438438",,terminal_output +703,5122849,"TERMINAL",0,0,"554:006510",,terminal_output +704,5123933,"TERMINAL",0,0,"761761",,terminal_output +705,5124946,"TERMINAL",0,0,"872872",,terminal_output +706,5126076,"TERMINAL",0,0,"983983",,terminal_output +707,5127092,"TERMINAL",0,0,"3:00945094",,terminal_output +708,5128185,"TERMINAL",0,0,"13:00512:005",,terminal_output +709,5129138,"TERMINAL",0,0,"216216",,terminal_output +710,5130268,"TERMINAL",0,0,"327327",,terminal_output +711,5131230,"TERMINAL",0,0,"438438",,terminal_output +712,5132278,"TERMINAL",0,0,"549549",,terminal_output +713,5133326,"TERMINAL",0,0,"65106520",,terminal_output +714,5134375,"TERMINAL",0,0,"761761",,terminal_output +715,5135412,"TERMINAL",0,0,"872872",,terminal_output +716,5136509,"TERMINAL",0,0,"983983",,terminal_output +717,5137640,"TERMINAL",0,0,"10942:0094",,terminal_output +718,5138549,"TERMINAL",0,0,"11051105",,terminal_output +719,5139687,"TERMINAL",0,0,"216216",,terminal_output +720,5140709,"TERMINAL",0,0,"327327",,terminal_output +721,5141735,"TERMINAL",0,0,"438438",,terminal_output +722,5142791,"TERMINAL",0,0,"549549",,terminal_output +723,5143815,"TERMINAL",0,0,"65206530",,terminal_output +724,5144829,"TERMINAL",0,0,"772872",,terminal_output +725,5145887,"TERMINAL",0,0,"983983",,terminal_output +726,5146938,"TERMINAL",0,0,"20941094",,terminal_output +727,5147983,"TERMINAL",0,0,"12051205",,terminal_output +728,5149153,"TERMINAL",0,0,"216216",,terminal_output +729,5150077,"TERMINAL",0,0,"327327",,terminal_output +730,5151122,"TERMINAL",0,0,"438438",,terminal_output +731,5152165,"TERMINAL",0,0,"549549",,terminal_output +732,5153225,"TERMINAL",0,0,"65306540",,terminal_output +733,5154250,"TERMINAL",0,0,"761761",,terminal_output +734,5155307,"TERMINAL",0,0,"872872",,terminal_output +735,5156376,"TERMINAL",0,0,"983983",,terminal_output +736,5157393,"TERMINAL",0,0,"30942094",,terminal_output +737,5158532,"TERMINAL",0,0,"13051305",,terminal_output +738,5159552,"TERMINAL",0,0,"216216",,terminal_output +739,5160541,"TERMINAL",0,0,"327327",,terminal_output +740,5161577,"TERMINAL",0,0,"438438",,terminal_output +741,5162800,"TERMINAL",0,0,"549549",,terminal_output +742,5163788,"TERMINAL",0,0,"65406550",,terminal_output +743,5164717,"TERMINAL",0,0,"761761",,terminal_output +744,5165784,"TERMINAL",0,0,"872872",,terminal_output +745,5166815,"TERMINAL",0,0,"983983",,terminal_output +746,5167857,"TERMINAL",0,0,"4140531405",,terminal_output +747,5168906,"TERMINAL",0,0,"216216",,terminal_output +748,5169956,"TERMINAL",0,0,"327327",,terminal_output +749,5171000,"TERMINAL",0,0,"438438",,terminal_output +750,5172044,"TERMINAL",0,0,"549549",,terminal_output +751,5173111,"TERMINAL",0,0,"6550653:00",,terminal_output +752,5174198,"TERMINAL",0,0,"761761",,terminal_output +753,5175191,"TERMINAL",0,0,"872872",,terminal_output +754,5176214,"TERMINAL",0,0,"983983",,terminal_output +755,5177269,"TERMINAL",0,0,"50944094",,terminal_output +756,5178393,"TERMINAL",0,0,"15051505",,terminal_output +757,5179364,"TERMINAL",0,0,"216216",,terminal_output +758,5180421,"TERMINAL",0,0,"327327",,terminal_output +759,5181454,"TERMINAL",0,0,"438438",,terminal_output +760,5182697,"TERMINAL",0,0,"549549",,terminal_output +761,5183616,"TERMINAL",0,0,"655:006510",,terminal_output +762,5184731,"TERMINAL",0,0,"761761",,terminal_output +763,5185665,"TERMINAL",0,0,"872872",,terminal_output +764,5186680,"TERMINAL",0,0,"983983",,terminal_output +765,5187838,"TERMINAL",0,0,"4:00945094",,terminal_output +766,5188814,"TERMINAL",0,0,"14:00513:005",,terminal_output +767,5190068,"TERMINAL",0,0,"227327",,terminal_output +768,5191091,"TERMINAL",0,0,"438438",,terminal_output +769,5191927,"TERMINAL",0,0,"549549",,terminal_output +770,5192988,"TERMINAL",0,0,"65106520",,terminal_output +771,5194051,"TERMINAL",0,0,"761761",,terminal_output +772,5195087,"TERMINAL",0,0,"872872",,terminal_output +773,5196138,"TERMINAL",0,0,"983983",,terminal_output +774,5197188,"TERMINAL",0,0,"10943:0094",,terminal_output +775,5198268,"TERMINAL",0,0,"11051105",,terminal_output +776,5199284,"TERMINAL",0,0,"216216",,terminal_output +777,5200427,"TERMINAL",0,0,"327327",,terminal_output +778,5201371,"TERMINAL",0,0,"438438",,terminal_output +779,5202406,"TERMINAL",0,0,"549549",,terminal_output +780,5203446,"TERMINAL",0,0,"65206530",,terminal_output +781,5204609,"TERMINAL",0,0,"761761",,terminal_output +782,5205536,"TERMINAL",0,0,"872872",,terminal_output +783,5206586,"TERMINAL",0,0,"983983",,terminal_output +784,5207656,"TERMINAL",0,0,"20941094",,terminal_output +785,5208706,"TERMINAL",0,0,"12051205",,terminal_output +786,5209731,"TERMINAL",0,0,"216216",,terminal_output +787,5210903,"TERMINAL",0,0,"327327",,terminal_output +788,5211819,"TERMINAL",0,0,"438438",,terminal_output +789,5212844,"TERMINAL",0,0,"55306540",,terminal_output +790,5213931,"TERMINAL",0,0,"761761",,terminal_output +791,5214938,"TERMINAL",0,0,"872872",,terminal_output +792,5215975,"TERMINAL",0,0,"983983",,terminal_output +793,5217010,"TERMINAL",0,0,"30942094",,terminal_output +794,5218123,"TERMINAL",0,0,"13051305",,terminal_output +795,5219099,"TERMINAL",0,0,"216216",,terminal_output +796,5220141,"TERMINAL",0,0,"327327",,terminal_output +797,5221214,"TERMINAL",0,0,"438438",,terminal_output +798,5222302,"TERMINAL",0,0,"549549",,terminal_output +799,5223348,"TERMINAL",0,0,"65406550",,terminal_output +800,5224325,"TERMINAL",0,0,"761761",,terminal_output +801,5225396,"TERMINAL",0,0,"872872",,terminal_output +802,5226422,"TERMINAL",0,0,"983983",,terminal_output +803,5227486,"TERMINAL",0,0,"40943094",,terminal_output +804,5228491,"TERMINAL",0,0,"14051405",,terminal_output +805,5229536,"TERMINAL",0,0,"216216",,terminal_output +806,5230616,"TERMINAL",0,0,"327327",,terminal_output +807,5231617,"TERMINAL",0,0,"438438",,terminal_output +808,5232657,"TERMINAL",0,0,"549549",,terminal_output +809,5233698,"TERMINAL",0,0,"6550654:00",,terminal_output +810,5234850,"TERMINAL",0,0,"761761",,terminal_output +811,5235954,"TERMINAL",0,0,"872872",,terminal_output +812,5236874,"TERMINAL",0,0,"9944094",,terminal_output +813,5237887,"TERMINAL",0,0,"515051505",,terminal_output +814,5238931,"TERMINAL",0,0,"216216",,terminal_output +815,5239982,"TERMINAL",0,0,"327327",,terminal_output +816,5241018,"TERMINAL",0,0,"438438",,terminal_output +817,5242088,"TERMINAL",0,0,"549549",,terminal_output +818,5243104,"TERMINAL",0,0,"656:006510",,terminal_output +819,5244150,"TERMINAL",0,0,"761761",,terminal_output +820,5245264,"TERMINAL",0,0,"872872",,terminal_output +821,5246242,"TERMINAL",0,0,"983983",,terminal_output +822,5247309,"TERMINAL",0,0,"5:00945094",,terminal_output +823,5248337,"TERMINAL",0,0,"15:00514:005",,terminal_output +824,5249376,"TERMINAL",0,0,"216216",,terminal_output +825,5250491,"TERMINAL",0,0,"327327",,terminal_output +826,5251507,"TERMINAL",0,0,"438438",,terminal_output +827,5252538,"TERMINAL",0,0,"549549",,terminal_output +828,5253663,"TERMINAL",0,0,"65106520",,terminal_output +829,5254731,"TERMINAL",0,0,"761761",,terminal_output +830,5255808,"TERMINAL",0,0,"872872",,terminal_output +831,5256729,"TERMINAL",0,0,"983983",,terminal_output +832,5257866,"TERMINAL",0,0,"10944:0094",,terminal_output +833,5258787,"TERMINAL",0,0,"11051105",,terminal_output +834,5259910,"TERMINAL",0,0,"216216",,terminal_output +835,5260930,"TERMINAL",0,0,"338438",,terminal_output +836,5261893,"TERMINAL",0,0,"549549",,terminal_output +837,5262936,"TERMINAL",0,0,"65206530",,terminal_output +838,5263972,"TERMINAL",0,0,"761761",,terminal_output +839,5265012,"TERMINAL",0,0,"872872",,terminal_output +840,5266061,"TERMINAL",0,0,"983983",,terminal_output +841,5267174,"TERMINAL",0,0,"20941094",,terminal_output +842,5268150,"TERMINAL",0,0,"12051205",,terminal_output +843,5269223,"TERMINAL",0,0,"216216",,terminal_output +844,5270265,"TERMINAL",0,0,"327327",,terminal_output +845,5271274,"TERMINAL",0,0,"438438",,terminal_output +846,5272314,"TERMINAL",0,0,"549549",,terminal_output +847,5273422,"TERMINAL",0,0,"65306540",,terminal_output +848,5274447,"TERMINAL",0,0,"761761",,terminal_output +849,5275470,"TERMINAL",0,0,"872872",,terminal_output +850,5276492,"TERMINAL",0,0,"983983",,terminal_output +851,5277528,"TERMINAL",0,0,"30942094",,terminal_output +852,5278752,"TERMINAL",0,0,"13051305",,terminal_output +853,5279692,"TERMINAL",0,0,"216216",,terminal_output +854,5280797,"TERMINAL",0,0,"327327",,terminal_output +855,5281715,"TERMINAL",0,0,"438438",,terminal_output +856,5282825,"TERMINAL",0,0,"549549",,terminal_output +857,5283789,"TERMINAL",0,0,"65406550",,terminal_output +858,5284892,"TERMINAL",0,0,"772872",,terminal_output +859,5285917,"TERMINAL",0,0,"983983",,terminal_output +860,5286914,"TERMINAL",0,0,"40943094",,terminal_output +861,5287964,"TERMINAL",0,0,"14051405",,terminal_output +862,5289018,"TERMINAL",0,0,"216216",,terminal_output +863,5290064,"TERMINAL",0,0,"327327",,terminal_output +864,5291111,"TERMINAL",0,0,"438438",,terminal_output +865,5292152,"TERMINAL",0,0,"549549",,terminal_output +866,5293196,"TERMINAL",0,0,"6550655:00",,terminal_output +867,5294312,"TERMINAL",0,0,"761761",,terminal_output +868,5295338,"TERMINAL",0,0,"872872",,terminal_output +869,5296326,"TERMINAL",0,0,"983983",,terminal_output +870,5297373,"TERMINAL",0,0,"50944094",,terminal_output +871,5298512,"TERMINAL",0,0,"15051505",,terminal_output +872,5299466,"TERMINAL",0,0,"216216",,terminal_output +873,5300662,"TERMINAL",0,0,"327327",,terminal_output +874,5301584,"TERMINAL",0,0,"438438",,terminal_output +875,5302709,"TERMINAL",0,0,"549549",,terminal_output +876,5303634,"TERMINAL",0,0,"657:006510",,terminal_output +877,5304757,"TERMINAL",0,0,"761761",,terminal_output +878,5305783,"TERMINAL",0,0,"872872",,terminal_output +879,5306773,"TERMINAL",0,0,"983983",,terminal_output +880,5307805,"TERMINAL",0,0,"6:00945094",,terminal_output +881,5308886,"TERMINAL",0,0,"16:01625:016",,terminal_output +882,5309877,"TERMINAL",0,0,"327327",,terminal_output +883,5310923,"TERMINAL",0,0,"438438",,terminal_output +884,5312009,"TERMINAL",0,0,"549549",,terminal_output +885,5313012,"TERMINAL",0,0,"65106520",,terminal_output +886,5314078,"TERMINAL",0,0,"761761",,terminal_output +887,5315092,"TERMINAL",0,0,"872872",,terminal_output +888,5316146,"TERMINAL",0,0,"983983",,terminal_output +889,5317252,"TERMINAL",0,0,"10945:0094",,terminal_output +890,5318280,"TERMINAL",0,0,"11051105",,terminal_output +891,5319301,"TERMINAL",0,0,"216216",,terminal_output +892,5320327,"TERMINAL",0,0,"327327",,terminal_output +893,5321362,"TERMINAL",0,0,"438438",,terminal_output +894,5322540,"TERMINAL",0,0,"549549",,terminal_output +895,5323496,"TERMINAL",0,0,"65206530",,terminal_output +896,5324504,"TERMINAL",0,0,"761761",,terminal_output +897,5325647,"TERMINAL",0,0,"872872",,terminal_output +898,5326671,"TERMINAL",0,0,"983983",,terminal_output +899,5327693,"TERMINAL",0,0,"20941094",,terminal_output +900,5328722,"TERMINAL",0,0,"12051205",,terminal_output +901,5329742,"TERMINAL",0,0,"216216",,terminal_output +902,5330970,"TERMINAL",0,0,"327327",,terminal_output +903,5331877,"TERMINAL",0,0,"449549",,terminal_output +904,5332863,"TERMINAL",0,0,"65306540",,terminal_output +905,5333940,"TERMINAL",0,0,"761761",,terminal_output +906,5334955,"TERMINAL",0,0,"872872",,terminal_output +907,5336005,"TERMINAL",0,0,"983983",,terminal_output +908,5337055,"TERMINAL",0,0,"30942094",,terminal_output +909,5338140,"TERMINAL",0,0,"13051305",,terminal_output +910,5339160,"TERMINAL",0,0,"216216",,terminal_output +911,5340292,"TERMINAL",0,0,"327327",,terminal_output +912,5341254,"TERMINAL",0,0,"438438",,terminal_output +913,5342337,"TERMINAL",0,0,"549549",,terminal_output +914,5343361,"TERMINAL",0,0,"65406550",,terminal_output +915,5344386,"TERMINAL",0,0,"761761",,terminal_output +916,5345480,"TERMINAL",0,0,"872872",,terminal_output +917,5346870,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",0,0,"",shellscript,tab +918,5347037,"TERMINAL",0,0,"983983",,terminal_output +919,5347625,"TERMINAL",0,0,"40943094",,terminal_output +920,5348690,"TERMINAL",0,0,"14051405",,terminal_output +921,5349716,"TERMINAL",0,0,"216216",,terminal_output +922,5350537,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1797,0,"",shellscript,selection_mouse +923,5350765,"TERMINAL",0,0,"327327",,terminal_output +924,5351687,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1800,0,"",shellscript,selection_mouse +925,5351806,"TERMINAL",0,0,"438438",,terminal_output +926,5352715,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1797,0,"",shellscript,selection_mouse +927,5352893,"TERMINAL",0,0,"5550656:00",,terminal_output +928,5353305,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1801,0,"",shellscript,selection_mouse +929,5353923,"TERMINAL",0,0,"761761",,terminal_output +930,5353985,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1800,0,"",shellscript,selection_mouse +931,5354689,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1797,0,"",shellscript,selection_mouse +932,5354955,"TERMINAL",0,0,"872872",,terminal_output +933,5356089,"TERMINAL",0,0,"983983",,terminal_output +934,5357084,"TERMINAL",0,0,"50944094",,terminal_output +935,5358110,"TERMINAL",0,0,"15051505",,terminal_output +936,5359121,"TERMINAL",0,0,"216216",,terminal_output +937,5360171,"TERMINAL",0,0,"327327",,terminal_output +938,5361222,"TERMINAL",0,0,"438438",,terminal_output +939,5362305,"TERMINAL",0,0,"549549",,terminal_output +940,5363332,"TERMINAL",0,0,"658:006510",,terminal_output +941,5364522,"TERMINAL",0,0,"761761",,terminal_output +942,5365670,"TERMINAL",0,0,"872872",,terminal_output +943,5366450,"TERMINAL",0,0,"983983",,terminal_output +944,5367528,"TERMINAL",0,0,"7:00945094",,terminal_output +945,5368538,"TERMINAL",0,0,"17:00516:005",,terminal_output +946,5369586,"TERMINAL",0,0,"216216",,terminal_output +947,5370634,"TERMINAL",0,0,"327327",,terminal_output +948,5371686,"TERMINAL",0,0,"438438",,terminal_output +949,5372739,"TERMINAL",0,0,"549549",,terminal_output +950,5373778,"TERMINAL",0,0,"65106520",,terminal_output +951,5374820,"TERMINAL",0,0,"772872",,terminal_output +952,5375962,"TERMINAL",0,0,"983983",,terminal_output +953,5376920,"TERMINAL",0,0,"10946:0094",,terminal_output +954,5377955,"TERMINAL",0,0,"11051105",,terminal_output +955,5379003,"TERMINAL",0,0,"216216",,terminal_output +956,5380050,"TERMINAL",0,0,"327327",,terminal_output +957,5381099,"TERMINAL",0,0,"438438",,terminal_output +958,5382140,"TERMINAL",0,0,"549549",,terminal_output +959,5383186,"TERMINAL",0,0,"65206530",,terminal_output +960,5384320,"TERMINAL",0,0,"761761",,terminal_output +961,5385351,"TERMINAL",0,0,"872872",,terminal_output +962,5386367,"TERMINAL",0,0,"983983",,terminal_output +963,5387501,"TERMINAL",0,0,"20941094",,terminal_output +964,5388410,"TERMINAL",0,0,"12051205",,terminal_output +965,5389544,"TERMINAL",0,0,"216216",,terminal_output +966,5390566,"TERMINAL",0,0,"327327",,terminal_output +967,5391535,"TERMINAL",0,0,"438438",,terminal_output +968,5392596,"TERMINAL",0,0,"549549",,terminal_output +969,5393662,"TERMINAL",0,0,"65306540",,terminal_output +970,5394768,"TERMINAL",0,0,"761761",,terminal_output +971,5395726,"TERMINAL",0,0,"872872",,terminal_output +972,5396775,"TERMINAL",0,0,"983983",,terminal_output +973,5397812,"TERMINAL",0,0,"30942094",,terminal_output +974,5398894,"TERMINAL",0,0,"13162316",,terminal_output +975,5399928,"TERMINAL",0,0,"327327",,terminal_output +976,5400946,"TERMINAL",0,0,"438438",,terminal_output +977,5401993,"TERMINAL",0,0,"549549",,terminal_output +978,5403033,"TERMINAL",0,0,"65406550",,terminal_output +979,5404069,"TERMINAL",0,0,"761761",,terminal_output +980,5405137,"TERMINAL",0,0,"872872",,terminal_output +981,5406205,"TERMINAL",0,0,"983983",,terminal_output +982,5407222,"TERMINAL",0,0,"40943094",,terminal_output +983,5408261,"TERMINAL",0,0,"14051405",,terminal_output +984,5409309,"TERMINAL",0,0,"216216",,terminal_output +985,5410339,"TERMINAL",0,0,"327327",,terminal_output +986,5411393,"TERMINAL",0,0,"438438",,terminal_output +987,5412487,"TERMINAL",0,0,"549549",,terminal_output +988,5413488,"TERMINAL",0,0,"6550657:00",,terminal_output +989,5414734,"TERMINAL",0,0,"761761",,terminal_output +990,5415657,"TERMINAL",0,0,"872872",,terminal_output +991,5416684,"TERMINAL",0,0,"983983",,terminal_output +992,5417706,"TERMINAL",0,0,"50944094",,terminal_output +993,5419020,"TERMINAL",0,0,"15051505",,terminal_output +994,5419788,"TERMINAL",0,0,"216216",,terminal_output +995,5420881,"TERMINAL",0,0,"338438",,terminal_output +996,5421944,"TERMINAL",0,0,"549549",,terminal_output +997,5422965,"TERMINAL",0,0,"659:006510",,terminal_output +998,5423965,"TERMINAL",0,0,"761761",,terminal_output +999,5425006,"TERMINAL",0,0,"872872",,terminal_output +1000,5426056,"TERMINAL",0,0,"983983",,terminal_output +1001,5427131,"TERMINAL",0,0,"8:00945094",,terminal_output +1002,5428153,"TERMINAL",0,0,"18:00517:005",,terminal_output +1003,5429281,"TERMINAL",0,0,"216216",,terminal_output +1004,5430230,"TERMINAL",0,0,"327327",,terminal_output +1005,5431274,"TERMINAL",0,0,"438438",,terminal_output +1006,5432347,"TERMINAL",0,0,"549549",,terminal_output +1007,5433366,"TERMINAL",0,0,"65106520",,terminal_output +1008,5434602,"TERMINAL",0,0,"761761",,terminal_output +1009,5435522,"TERMINAL",0,0,"872872",,terminal_output +1010,5436548,"TERMINAL",0,0,"983983",,terminal_output +1011,5437569,"TERMINAL",0,0,"10947:0094",,terminal_output +1012,5438699,"TERMINAL",0,0,"11051105",,terminal_output +1013,5439721,"TERMINAL",0,0,"216216",,terminal_output +1014,5440748,"TERMINAL",0,0,"327327",,terminal_output +1015,5441771,"TERMINAL",0,0,"438438",,terminal_output +1016,5442794,"TERMINAL",0,0,"549549",,terminal_output +1017,5443816,"TERMINAL",0,0,"66217631",,terminal_output +1018,5444947,"TERMINAL",0,0,"872872",,terminal_output +1019,5446071,"TERMINAL",0,0,"983983",,terminal_output +1020,5447100,"TERMINAL",0,0,"20941094",,terminal_output +1021,5447977,"TERMINAL",0,0,"12051205",,terminal_output +1022,5449152,"TERMINAL",0,0,"216216",,terminal_output +1023,5450174,"TERMINAL",0,0,"327327",,terminal_output +1024,5451118,"TERMINAL",0,0,"438438",,terminal_output +1025,5452156,"TERMINAL",0,0,"549549",,terminal_output +1026,5453197,"TERMINAL",0,0,"65306540",,terminal_output +1027,5454264,"TERMINAL",0,0,"761761",,terminal_output +1028,5455284,"TERMINAL",0,0,"872872",,terminal_output +1029,5456324,"TERMINAL",0,0,"983983",,terminal_output +1030,5457436,"TERMINAL",0,0,"30942094",,terminal_output +1031,5458462,"TERMINAL",0,0,"13051305",,terminal_output +1032,5459585,"TERMINAL",0,0,"216216",,terminal_output +1033,5460612,"TERMINAL",0,0,"327327",,terminal_output +1034,5461537,"TERMINAL",0,0,"438438",,terminal_output +1035,5462566,"TERMINAL",0,0,"549549",,terminal_output +1036,5463684,"TERMINAL",0,0,"65406550",,terminal_output +1037,5464709,"TERMINAL",0,0,"761761",,terminal_output +1038,5465734,"TERMINAL",0,0,"872872",,terminal_output +1039,5466756,"TERMINAL",0,0,"983983",,terminal_output +1040,5467805,"TERMINAL",0,0,"40943094",,terminal_output +1041,5468948,"TERMINAL",0,0,"14162416",,terminal_output +1042,5469886,"TERMINAL",0,0,"327327",,terminal_output +1043,5470976,"TERMINAL",0,0,"438438",,terminal_output +1044,5471998,"TERMINAL",0,0,"549549",,terminal_output +1045,5473024,"TERMINAL",0,0,"6550658:00",,terminal_output +1046,5474069,"TERMINAL",0,0,"761761",,terminal_output +1047,5475109,"TERMINAL",0,0,"872872",,terminal_output +1048,5476153,"TERMINAL",0,0,"983983",,terminal_output +1049,5477196,"TERMINAL",0,0,"50944094",,terminal_output +1050,5478248,"TERMINAL",0,0,"15051505",,terminal_output +1051,5479294,"TERMINAL",0,0,"216216",,terminal_output +1052,5480485,"TERMINAL",0,0,"327327",,terminal_output +1053,5481392,"TERMINAL",0,0,"438438",,terminal_output +1054,5482438,"TERMINAL",0,0,"549549",,terminal_output +1055,5483478,"TERMINAL",0,0,"6520:006510",,terminal_output +1056,5484573,"TERMINAL",0,0,"761761",,terminal_output +1057,5485570,"TERMINAL",0,0,"872872",,terminal_output +1058,5486611,"TERMINAL",0,0,"983983",,terminal_output +1059,5487734,"TERMINAL",0,0,"9:00945094",,terminal_output +1060,5488773,"TERMINAL",0,0,"19:00518:005",,terminal_output +1061,5489735,"TERMINAL",0,0,"216216",,terminal_output +1062,5490779,"TERMINAL",0,0,"327327",,terminal_output +1063,5491828,"TERMINAL",0,0,"449549",,terminal_output +1064,5492882,"TERMINAL",0,0,"65106520",,terminal_output +1065,5493993,"TERMINAL",0,0,"761761",,terminal_output +1066,5495018,"TERMINAL",0,0,"872872",,terminal_output +1067,5496051,"TERMINAL",0,0,"983983",,terminal_output +1068,5497062,"TERMINAL",0,0,"10948:0094",,terminal_output +1069,5498114,"TERMINAL",0,0,"11051105",,terminal_output +1070,5499163,"TERMINAL",0,0,"216216",,terminal_output +1071,5500211,"TERMINAL",0,0,"327327",,terminal_output +1072,5501254,"TERMINAL",0,0,"438438",,terminal_output +1073,5502300,"TERMINAL",0,0,"549549",,terminal_output +1074,5503415,"TERMINAL",0,0,"65206530",,terminal_output +1075,5504405,"TERMINAL",0,0,"761761",,terminal_output +1076,5505440,"TERMINAL",0,0,"872872",,terminal_output +1077,5506510,"TERMINAL",0,0,"983983",,terminal_output +1078,5507580,"TERMINAL",0,0,"20941094",,terminal_output +1079,5508638,"TERMINAL",0,0,"12051205",,terminal_output +1080,5509631,"TERMINAL",0,0,"216216",,terminal_output +1081,5510688,"TERMINAL",0,0,"327327",,terminal_output +1082,5511825,"TERMINAL",0,0,"438438",,terminal_output +1083,5512801,"TERMINAL",0,0,"549549",,terminal_output +1084,5513807,"TERMINAL",0,0,"65306540",,terminal_output +1085,5514883,"TERMINAL",0,0,"772872",,terminal_output +1086,5515895,"TERMINAL",0,0,"983983",,terminal_output +1087,5516976,"TERMINAL",0,0,"30942094",,terminal_output +1088,5517991,"TERMINAL",0,0,"13051305",,terminal_output +1089,5519063,"TERMINAL",0,0,"216216",,terminal_output +1090,5520080,"TERMINAL",0,0,"327327",,terminal_output +1091,5521127,"TERMINAL",0,0,"438438",,terminal_output +1092,5522169,"TERMINAL",0,0,"549549",,terminal_output +1093,5523209,"TERMINAL",0,0,"65406550",,terminal_output +1094,5524303,"TERMINAL",0,0,"761761",,terminal_output +1095,5525332,"TERMINAL",0,0,"872872",,terminal_output +1096,5526356,"TERMINAL",0,0,"983983",,terminal_output +1097,5527478,"TERMINAL",0,0,"40943094",,terminal_output +1098,5528504,"TERMINAL",0,0,"14051405",,terminal_output +1099,5529504,"TERMINAL",0,0,"216216",,terminal_output +1100,5530544,"TERMINAL",0,0,"327327",,terminal_output +1101,5531678,"TERMINAL",0,0,"438438",,terminal_output +1102,5532676,"TERMINAL",0,0,"549549",,terminal_output +1103,5533728,"TERMINAL",0,0,"6550659:00",,terminal_output +1104,5534765,"TERMINAL",0,0,"761761",,terminal_output +1105,5535874,"TERMINAL",0,0,"872872",,terminal_output +1106,5536980,"TERMINAL",0,0,"50944094",,terminal_output +1107,5538032,"TERMINAL",0,0,"15051505",,terminal_output +1108,5539072,"TERMINAL",0,0,"216216",,terminal_output +1109,5540102,"TERMINAL",0,0,"327327",,terminal_output +1110,5541146,"TERMINAL",0,0,"438438",,terminal_output +1111,5542204,"TERMINAL",0,0,"549549",,terminal_output +1112,5543248,"TERMINAL",0,0,"651:006510",,terminal_output +1113,5544291,"TERMINAL",0,0,"761761",,terminal_output +1114,5545340,"TERMINAL",0,0,"872872",,terminal_output +1115,5546390,"TERMINAL",0,0,"983983",,terminal_output +1116,5547446,"TERMINAL",0,0,"10:00945094",,terminal_output +1117,5548674,"TERMINAL",0,0,"110:00519:005",,terminal_output +1118,5549601,"TERMINAL",0,0,"216216",,terminal_output +1119,5550620,"TERMINAL",0,0,"327327",,terminal_output +1120,5551751,"TERMINAL",0,0,"438438",,terminal_output +1121,5552773,"TERMINAL",0,0,"549549",,terminal_output +1122,5553757,"TERMINAL",0,0,"65106520",,terminal_output +1123,5554821,"TERMINAL",0,0,"761761",,terminal_output +1124,5555840,"TERMINAL",0,0,"883983",,terminal_output +1125,5556882,"TERMINAL",0,0,"10949:0094",,terminal_output +1126,5557982,"TERMINAL",0,0,"11051105",,terminal_output +1127,5559129,"TERMINAL",0,0,"216216",,terminal_output +1128,5560147,"TERMINAL",0,0,"327327",,terminal_output +1129,5561092,"TERMINAL",0,0,"438438",,terminal_output +1130,5562112,"TERMINAL",0,0,"549549",,terminal_output +1131,5562843,"TERMINAL",0,0,"bash",,terminal_focus +1132,5563164,"TERMINAL",0,0,"65206530",,terminal_output +1133,5564027,"TERMINAL",0,0,"watch",,terminal_focus +1134,5564228,"TERMINAL",0,0,"761761",,terminal_output +1135,5565484,"TERMINAL",0,0,"872872",,terminal_output +1136,5566389,"TERMINAL",0,0,"983983",,terminal_output +1137,5566782,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1138,5568892,"TERMINAL",0,0,"idling",,terminal_command +1139,5568975,"TERMINAL",0,0,"]633;E;2025-08-14 12:10:22 idling;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Aug 14 12:10:22 2025Partition dev_cpuonly: 12 nodes idle\rPartition cpuonly: 121 nodes idle\rPartition dev_accelerated:\t 0 nodes idle\rPartition accelerated: 21 nodes idle\rPartition dev_accelerated-h100 :\t 1 nodes idle\rPartition accelerated-h100:\t 1 nodes idle\rPartition large:\t 6 nodes idle",,terminal_output +1140,5570181,"TERMINAL",0,0,"3\t ",,terminal_output +1141,5571045,"TERMINAL",0,0,"4\t ",,terminal_output +1142,5572090,"TERMINAL",0,0,"5\t ",,terminal_output +1143,5573135,"TERMINAL",0,0,"6\t ",,terminal_output +1144,5574173,"TERMINAL",0,0,"7\t ",,terminal_output +1145,5575216,"TERMINAL",0,0,"8\t ",,terminal_output +1146,5576256,"TERMINAL",0,0,"9\t ",,terminal_output +1147,5577357,"TERMINAL",0,0,"30\t ",,terminal_output +1148,5578346,"TERMINAL",0,0,"1\t ",,terminal_output +1149,5579415,"TERMINAL",0,0,"2\t ",,terminal_output +1150,5580428,"TERMINAL",0,0,"3\t ",,terminal_output +1151,5581552,"TERMINAL",0,0,"4\t ",,terminal_output +1152,5582500,"TERMINAL",0,0,"5\t ",,terminal_output +1153,5583592,"TERMINAL",0,0,"6\t ",,terminal_output +1154,5584721,"TERMINAL",0,0,"7\t ",,terminal_output +1155,5585712,"TERMINAL",0,0,"8\t ",,terminal_output +1156,5586808,"TERMINAL",0,0,"99",,terminal_output +1157,5587793,"TERMINAL",0,0,"40\t ",,terminal_output +1158,5587987,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1159,5655969,"slurm/jobs/mihir/horeka/atari/train_dynamics_maskgit.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/atari/dynamics/maskgit/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/atari/dynamics/maskgit/%x_%j.log\n#SBATCH --job-name=train_dynamics_maskgit_1_node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_atari/array_records\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/atari/maskgit/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\n# @mihir TODO\ntokenizer_ckpt_dir=TODO\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --image_height=84 \\n --image_width=64 \\n --dyna_type=maskgit \\n --init_lr=0 \\n --max_lr=1e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=atari-dynamics-maskgit-2-node-$slurm_job_id \\n --tags atari dynamics maskgit 2-node \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +1160,5662631,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/maskgit/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/maskgit/%x_%j.log\n#SBATCH --job-name=train_dynamics_maskgit_1_node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/array_records\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/coinrun/maskgit/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3414046""\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=48 \\n --image_height=64 \\n --image_width=64 \\n --dyna_type=maskgit \\n --init_lr=0 \\n --max_lr=1e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=coinrun-dynamics-maskgit-1-node-$slurm_job_id \\n --tags coinrun dynamics maskgit 1-node \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +1161,5686753,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",,terminal_command +1162,5686784,"TERMINAL",0,0,"]633;E;2025-08-14 12:12:19 sbatch slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C",,terminal_output +1163,5686849,"TERMINAL",0,0,"Submitted batch job 3422007\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1164,5688066,"TERMINAL",0,0,"queue",,terminal_command +1165,5688157,"TERMINAL",0,0,"]633;E;2025-08-14 12:12:21 queue;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 12:12:21 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3422007 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-18:12:20\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 22:23:25\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3421982 accelerat train_to tum_cte0 R11:11\t 4 hkn[0724,0726,0728,0731]3421973 accelerat train_to tum_cte0 R11:20\t 2 hkn[0722-0723]3421964 accelerat train_to tum_cte0 R11:35\t 1 hkn0720",,terminal_output +1166,5689195,"TERMINAL",0,0,"216216",,terminal_output +1167,5690234,"TERMINAL",0,0,"327327",,terminal_output +1168,5691285,"TERMINAL",0,0,"438438",,terminal_output +1169,5692360,"TERMINAL",0,0,"549549",,terminal_output +1170,5693471,"TERMINAL",0,0,"65306540",,terminal_output +1171,5694496,"TERMINAL",0,0,"761761",,terminal_output +1172,5695422,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1173,5696901,"TERMINAL",0,0,"fqueue",,terminal_command +1174,5696968,"TERMINAL",0,0,"]633;E;2025-08-14 12:12:30 fqueue;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C[?1049h(B[?7hEvery 1.0s: squeue -o ""%.10i %.16P %.30j %.8u %.8T %.10M %.9l %.6D %R""hkn1993.localdomain: Thu Aug 14 12:12:30 2025JOBIDPARTITIONNAME USER STATE\t TIME TIME_LIMI NODES NODELIST(REASON)3415713\taccelerated train_dynamics_maskgit_1_node tum_cte0 PENDING\t 0:00 2-00:00:00\t1 (Priority)3422007\taccelerated train_dynamics_maskgit_1_node tum_cte0 PENDING\t 0:00 2-00:00:00\t1 (Priority)3412401\tacceleratedtrain_tokenizer_1e-4 tum_cte0 RUNNING 1-18:12:29 2-00:00:00\t8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832\taccelerated train_dynamics_causal_8_node tum_cte0 RUNNING 22:23:34 2-00:00:00\t8 hkn[0407,0415,0420,0422-0424,0729,0735]3421982\tacceleratedtrain_tokenizer_1e-4 tum_cte0 RUNNING\t 11:20 2-00:00:00\t4 hkn[0724,0726,0728,0731]3421973\tacceleratedtrain_tokenizer_1e-4 tum_cte0 RUNNING\t 11:29 2-00:00:00\t2 hkn[0722-0723]3421964\tacceleratedtrain_tokenizer_3e-5 tum_cte0 RUNNING\t 11:44 2-00:00:00\t1 hkn0720",,terminal_output +1175,5697956,"TERMINAL",0,0,"13051305",,terminal_output +1176,5699210,"TERMINAL",0,0,"216216",,terminal_output +1177,5700070,"TERMINAL",0,0,"327327",,terminal_output +1178,5701081,"TERMINAL",0,0,"438438",,terminal_output +1179,5702084,"TERMINAL",0,0,"549549",,terminal_output +1180,5703134,"TERMINAL",0,0,"65406550",,terminal_output +1181,5704152,"TERMINAL",0,0,"761761",,terminal_output +1182,5705188,"TERMINAL",0,0,"872872",,terminal_output +1183,5706277,"TERMINAL",0,0,"983983",,terminal_output +1184,5707260,"TERMINAL",0,0,"40943094",,terminal_output +1185,5708319,"TERMINAL",0,0,"14051405",,terminal_output +1186,5708862,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1187,5710829,"TERMINAL",0,0,"scancel 3422007",,terminal_command +1188,5710866,"TERMINAL",0,0,"]633;E;2025-08-14 12:12:43 scancel 3422007;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs",,terminal_output +1189,5714913,"TERMINAL",0,0,"fqueue",,terminal_command +1190,5714974,"TERMINAL",0,0,"]633;E;2025-08-14 12:12:48 fqueue;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C[?1049h(B[?7hEvery 1.0s: squeue -o ""%.10i %.16P %.30j %.8u %.8T %.10M %.9l %.6D %R""hkn1993.localdomain: Thu Aug 14 12:12:48 2025JOBIDPARTITIONNAME USER STATE\t TIME TIME_LIMI NODES NODELIST(REASON)3415713\taccelerated train_dynamics_maskgit_1_node tum_cte0 PENDING\t 0:00 2-00:00:00\t1 (Priority)3412401\tacceleratedtrain_tokenizer_1e-4 tum_cte0 RUNNING 1-18:12:47 2-00:00:00\t8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832\taccelerated train_dynamics_causal_8_node tum_cte0 RUNNING 22:23:52 2-00:00:00\t8 hkn[0407,0415,0420,0422-0424,0729,0735]3421982\tacceleratedtrain_tokenizer_1e-4 tum_cte0 RUNNING\t 11:38 2-00:00:00\t4 hkn[0724,0726,0728,0731]3421973\tacceleratedtrain_tokenizer_1e-4 tum_cte0 RUNNING\t 11:47 2-00:00:00\t2 hkn[0722-0723]3421964\tacceleratedtrain_tokenizer_3e-5 tum_cte0 RUNNING\t 12:02 2-00:00:00\t1 hkn0720",,terminal_output +1191,5715768,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1192,5719545,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",0,0,"",shellscript,tab +1193,5722165,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",483,0,"",shellscript,selection_mouse +1194,5722168,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",482,0,"",shellscript,selection_command +1195,5722649,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",444,0,"",shellscript,selection_mouse +1196,5723823,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",0,0,"",shellscript,tab +1197,5726517,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",0,0,"",shellscript,tab +1198,5727473,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",465,0,"\n#SBATCH --reservation=llmtum",shellscript,content +1199,5727496,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",466,0,"",shellscript,selection_command +1200,5730352,"TERMINAL",0,0,"sync-runner",,terminal_command +1201,5730436,"TERMINAL",0,0,"]633;E;2025-08-14 12:13:03 sync-runner;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;Csending incremental file list\r\n",,terminal_output +1202,5732208,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch\r\n",,terminal_output +1203,5732283,"TERMINAL",0,0,"\r\nsent 25,843 bytes received 184 bytes 10,410.80 bytes/sec\r\ntotal size is 128,148,759 speedup is 4,923.69\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1204,5736356,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",,terminal_command +1205,5736418,"TERMINAL",0,0,"]633;E;2025-08-14 12:13:09 sbatch slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;CSubmitted batch job 3422008\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1206,5737347,"TERMINAL",0,0,"queue",,terminal_command +1207,5737445,"TERMINAL",0,0,"]633;E;2025-08-14 12:13:10 queue;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 12:13:10 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-18:13:09\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 22:24:14\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R\t0:00\t 1 hkn07133421982 accelerat train_to tum_cte0 R12:00\t 4 hkn[0724,0726,0728,0731]3421973 accelerat train_to tum_cte0 R12:09\t 2 hkn[0722-0723]3421964 accelerat train_to tum_cte0 R12:24\t 1 hkn0720",,terminal_output +1208,5738424,"TERMINAL",0,0,"110511105",,terminal_output +1209,5739564,"TERMINAL",0,0,"2162216",,terminal_output +1210,5740519,"TERMINAL",0,0,"3273327",,terminal_output +1211,5741567,"TERMINAL",0,0,"4384438",,terminal_output +1212,5742604,"TERMINAL",0,0,"5495549",,terminal_output +1213,5743642,"TERMINAL",0,0,"652066530",,terminal_output +1214,5744754,"TERMINAL",0,0,"7617761",,terminal_output +1215,5745798,"TERMINAL",0,0,"8728872",,terminal_output +1216,5746772,"TERMINAL",0,0,"9839983",,terminal_output +1217,5747852,"TERMINAL",0,0,"2094101094",,terminal_output +1218,5748872,"TERMINAL",0,0,"221622216",,terminal_output +1219,5749996,"TERMINAL",0,0,"3273327",,terminal_output +1220,5750955,"TERMINAL",0,0,"4384438",,terminal_output +1221,5752043,"TERMINAL",0,0,"5495549",,terminal_output +1222,5753068,"TERMINAL",0,0,"653066540",,terminal_output +1223,5754125,"TERMINAL",0,0,"7617761",,terminal_output +1224,5755123,"TERMINAL",0,0,"8728872",,terminal_output +1225,5756142,"TERMINAL",0,0,"9839983",,terminal_output +1226,5757192,"TERMINAL",0,0,"3094202094",,terminal_output +1227,5758228,"TERMINAL",0,0,"130511305",,terminal_output +1228,5759288,"TERMINAL",0,0,"2162216",,terminal_output +1229,5760323,"TERMINAL",0,0,"3273327",,terminal_output +1230,5761488,"TERMINAL",0,0,"4384438",,terminal_output +1231,5762401,"TERMINAL",0,0,"5495549",,terminal_output +1232,5763514,"TERMINAL",0,0,"654066550",,terminal_output +1233,5764486,"TERMINAL",0,0,"7617761",,terminal_output +1234,5765560,"TERMINAL",0,0,"8728872",,terminal_output +1235,5766588,"TERMINAL",0,0,"9839983",,terminal_output +1236,5767625,"TERMINAL",0,0,"4094303094",,terminal_output +1237,5768711,"TERMINAL",0,0,"140511405",,terminal_output +1238,5769827,"TERMINAL",0,0,"2162216",,terminal_output +1239,5770886,"TERMINAL",0,0,"3273327",,terminal_output +1240,5771910,"TERMINAL",0,0,"4495549",,terminal_output +1241,5772881,"TERMINAL",0,0,"65506653:00",,terminal_output +1242,5773926,"TERMINAL",0,0,"7617761",,terminal_output +1243,5774979,"TERMINAL",0,0,"8728872",,terminal_output +1244,5776135,"TERMINAL",0,0,"9839983",,terminal_output +1245,5777136,"TERMINAL",0,0,"5094404094",,terminal_output +1246,5778130,"TERMINAL",0,0,"150511505",,terminal_output +1247,5779150,"TERMINAL",0,0,"2162216",,terminal_output +1248,5780203,"TERMINAL",0,0,"3273327",,terminal_output +1249,5781240,"TERMINAL",0,0,"4384438",,terminal_output +1250,5782277,"TERMINAL",0,0,"5495549",,terminal_output +1251,5783381,"TERMINAL",0,0,"655:0066510",,terminal_output +1252,5784371,"TERMINAL",0,0,"7617761",,terminal_output +1253,5785529,"TERMINAL",0,0,"8728872",,terminal_output +1254,5786486,"TERMINAL",0,0,"9839983",,terminal_output +1255,5787574,"TERMINAL",0,0,"4:0094505094",,terminal_output +1256,5788563,"TERMINAL",0,0,"14:005113:005",,terminal_output +1257,5789630,"TERMINAL",0,0,"2162216",,terminal_output +1258,5790978,"TERMINAL",0,0,"3273327",,terminal_output +1259,5791774,"TERMINAL",0,0,"4384438",,terminal_output +1260,5792799,"TERMINAL",0,0,"5495549",,terminal_output +1261,5793828,"TERMINAL",0,0,"651066520",,terminal_output +1262,5794837,"TERMINAL",0,0,"7728872",,terminal_output +1263,5795977,"TERMINAL",0,0,"9839983",,terminal_output +1264,5797133,"TERMINAL",0,0,"10941:003:0094",,terminal_output +1265,5798021,"TERMINAL",0,0,"110511105",,terminal_output +1266,5799147,"TERMINAL",0,0,"2162216",,terminal_output +1267,5800072,"TERMINAL",0,0,"3273327",,terminal_output +1268,5801195,"TERMINAL",0,0,"4384438",,terminal_output +1269,5802242,"TERMINAL",0,0,"5495549",,terminal_output +1270,5803196,"TERMINAL",0,0,"652066530",,terminal_output +1271,5804243,"TERMINAL",0,0,"7617761",,terminal_output +1272,5805284,"TERMINAL",0,0,"8728872",,terminal_output +1273,5806324,"TERMINAL",0,0,"9839983",,terminal_output +1274,5807368,"TERMINAL",0,0,"2094101094",,terminal_output +1275,5808463,"TERMINAL",0,0,"120511205",,terminal_output +1276,5809490,"TERMINAL",0,0,"2162216",,terminal_output +1277,5810514,"TERMINAL",0,0,"3273327",,terminal_output +1278,5811548,"TERMINAL",0,0,"4384438",,terminal_output +1279,5812664,"TERMINAL",0,0,"5495549",,terminal_output +1280,5813698,"TERMINAL",0,0,"653066540",,terminal_output +1281,5814716,"TERMINAL",0,0,"7617761",,terminal_output +1282,5815709,"TERMINAL",0,0,"8728872",,terminal_output +1283,5816752,"TERMINAL",0,0,"9839983",,terminal_output +1284,5817809,"TERMINAL",0,0,"3094202094",,terminal_output +1285,5818910,"TERMINAL",0,0,"131622316",,terminal_output +1286,5819935,"TERMINAL",0,0,"3273327",,terminal_output +1287,5820961,"TERMINAL",0,0,"4384438",,terminal_output +1288,5821982,"TERMINAL",0,0,"5495549",,terminal_output +1289,5823014,"TERMINAL",0,0,"654066550",,terminal_output +1290,5824162,"TERMINAL",0,0,"7617761",,terminal_output +1291,5825157,"TERMINAL",0,0,"8728872",,terminal_output +1292,5826146,"TERMINAL",0,0,"9839983",,terminal_output +1293,5827192,"TERMINAL",0,0,"4094303094",,terminal_output +1294,5828238,"TERMINAL",0,0,"140511405",,terminal_output +1295,5829293,"TERMINAL",0,0,"2162216",,terminal_output +1296,5830333,"TERMINAL",0,0,"3273327",,terminal_output +1297,5831371,"TERMINAL",0,0,"4384438",,terminal_output +1298,5832423,"TERMINAL",0,0,"5495549",,terminal_output +1299,5833554,"TERMINAL",0,0,"65506654:00",,terminal_output +1300,5834579,"TERMINAL",0,0,"7617761",,terminal_output +1301,5835603,"TERMINAL",0,0,"8728872",,terminal_output +1302,5836625,"TERMINAL",0,0,"9839983",,terminal_output +1303,5837650,"TERMINAL",0,0,"5094404094",,terminal_output +1304,5838778,"TERMINAL",0,0,"150511505",,terminal_output +1305,5839746,"TERMINAL",0,0,"2162216",,terminal_output +1306,5840928,"TERMINAL",0,0,"3273327",,terminal_output +1307,5841846,"TERMINAL",0,0,"4495549",,terminal_output +1308,5842861,"TERMINAL",0,0,"656:0066510",,terminal_output +1309,5844014,"TERMINAL",0,0,"7617761",,terminal_output +1310,5845028,"TERMINAL",0,0,"8728872",,terminal_output +1311,5846048,"TERMINAL",0,0,"9839983",,terminal_output +1312,5847064,"TERMINAL",0,0,"5:0094505094",,terminal_output +1313,5848096,"TERMINAL",0,0,"15:005114:005",,terminal_output +1314,5849173,"TERMINAL",0,0,"2162216",,terminal_output +1315,5850296,"TERMINAL",0,0,"3273327",,terminal_output +1316,5851241,"TERMINAL",0,0,"4384438",,terminal_output +1317,5852288,"TERMINAL",0,0,"5495549",,terminal_output +1318,5853366,"TERMINAL",0,0,"651066520",,terminal_output +1319,5854445,"TERMINAL",0,0,"7617761",,terminal_output +1320,5855456,"TERMINAL",0,0,"8728872",,terminal_output +1321,5856494,"TERMINAL",0,0,"9839983",,terminal_output +1322,5857515,"TERMINAL",0,0,"10942:004:0094",,terminal_output +1323,5858646,"TERMINAL",0,0,"110511105",,terminal_output +1324,5859673,"TERMINAL",0,0,"2162216",,terminal_output +1325,5860662,"TERMINAL",0,0,"3273327",,terminal_output +1326,5861717,"TERMINAL",0,0,"4384438",,terminal_output +1327,5862750,"TERMINAL",0,0,"5495549",,terminal_output +1328,5863790,"TERMINAL",0,0,"652066530",,terminal_output +1329,5864838,"TERMINAL",0,0,"7728872",,terminal_output +1330,5865918,"TERMINAL",0,0,"9839983",,terminal_output +1331,5866946,"TERMINAL",0,0,"2094101094",,terminal_output +1332,5867993,"TERMINAL",0,0,"120511205",,terminal_output +1333,5869191,"TERMINAL",0,0,"2162216",,terminal_output +1334,5870224,"TERMINAL",0,0,"3273327",,terminal_output +1335,5871138,"TERMINAL",0,0,"4384438",,terminal_output +1336,5872263,"TERMINAL",0,0,"5495549",,terminal_output +1337,5873219,"TERMINAL",0,0,"653066540",,terminal_output +1338,5874274,"TERMINAL",0,0,"7617761",,terminal_output +1339,5875327,"TERMINAL",0,0,"8728872",,terminal_output +1340,5876378,"TERMINAL",0,0,"9839983",,terminal_output +1341,5877486,"TERMINAL",0,0,"3094202094",,terminal_output +1342,5878478,"TERMINAL",0,0,"130511305",,terminal_output +1343,5879512,"TERMINAL",0,0,"2162216",,terminal_output +1344,5880563,"TERMINAL",0,0,"3273327",,terminal_output +1345,5881681,"TERMINAL",0,0,"4384438",,terminal_output +1346,5882710,"TERMINAL",0,0,"5495549",,terminal_output +1347,5883733,"TERMINAL",0,0,"654066550",,terminal_output +1348,5884963,"TERMINAL",0,0,"7617761",,terminal_output +1349,5885984,"TERMINAL",0,0,"8728872",,terminal_output +1350,5886836,"TERMINAL",0,0,"994303094",,terminal_output +1351,5887892,"TERMINAL",0,0,"4140511405",,terminal_output +1352,5888938,"TERMINAL",0,0,"2162216",,terminal_output +1353,5889993,"TERMINAL",0,0,"3273327",,terminal_output +1354,5891315,"TERMINAL",0,0,"4384438",,terminal_output +1355,5892126,"TERMINAL",0,0,"5495549",,terminal_output +1356,5893176,"TERMINAL",0,0,"65506655:00",,terminal_output +1357,5894279,"TERMINAL",0,0,"7617761",,terminal_output +1358,5895204,"TERMINAL",0,0,"8728872",,terminal_output +1359,5896255,"TERMINAL",0,0,"9839983",,terminal_output +1360,5897307,"TERMINAL",0,0,"5094404094",,terminal_output +1361,5898362,"TERMINAL",0,0,"150511505",,terminal_output +1362,5899406,"TERMINAL",0,0,"2162216",,terminal_output +1363,5900463,"TERMINAL",0,0,"3273327",,terminal_output +1364,5901504,"TERMINAL",0,0,"4384438",,terminal_output +1365,5902557,"TERMINAL",0,0,"5495549",,terminal_output +1366,5903609,"TERMINAL",0,0,"657:0066510",,terminal_output +1367,5904654,"TERMINAL",0,0,"7617761",,terminal_output +1368,5905709,"TERMINAL",0,0,"8728872",,terminal_output +1369,5906752,"TERMINAL",0,0,"9839983",,terminal_output +1370,5907795,"TERMINAL",0,0,"6:0094505094",,terminal_output +1371,5908918,"TERMINAL",0,0,"16:016225:016",,terminal_output +1372,5909901,"TERMINAL",0,0,"3273327",,terminal_output +1373,5910933,"TERMINAL",0,0,"4384438",,terminal_output +1374,5911993,"TERMINAL",0,0,"5495549",,terminal_output +1375,5913119,"TERMINAL",0,0,"651066520",,terminal_output +1376,5914076,"TERMINAL",0,0,"7617761",,terminal_output +1377,5915122,"TERMINAL",0,0,"8728872",,terminal_output +1378,5916173,"TERMINAL",0,0,"9839983",,terminal_output +1379,5917215,"TERMINAL",0,0,"10943:005:0094",,terminal_output +1380,5918266,"TERMINAL",0,0,"110511105",,terminal_output +1381,5919367,"TERMINAL",0,0,"2162216",,terminal_output +1382,5920365,"TERMINAL",0,0,"3273327",,terminal_output +1383,5921516,"TERMINAL",0,0,"4384438",,terminal_output +1384,5922539,"TERMINAL",0,0,"5495549",,terminal_output +1385,5923507,"TERMINAL",0,0,"652066530",,terminal_output +1386,5924587,"TERMINAL",0,0,"7617761",,terminal_output +1387,5925609,"TERMINAL",0,0,"8728872",,terminal_output +1388,5926660,"TERMINAL",0,0,"9839983",,terminal_output +1389,5927701,"TERMINAL",0,0,"2094101094",,terminal_output +1390,5928749,"TERMINAL",0,0,"120511205",,terminal_output +1391,5929810,"TERMINAL",0,0,"2162216",,terminal_output +1392,5930849,"TERMINAL",0,0,"3384438",,terminal_output +1393,5931887,"TERMINAL",0,0,"5495549",,terminal_output +1394,5932932,"TERMINAL",0,0,"653066540",,terminal_output +1395,5933978,"TERMINAL",0,0,"7617761",,terminal_output +1396,5935136,"TERMINAL",0,0,"8728872",,terminal_output +1397,5936070,"TERMINAL",0,0,"9839983",,terminal_output +1398,5937116,"TERMINAL",0,0,"3094202094",,terminal_output +1399,5938156,"TERMINAL",0,0,"130511305",,terminal_output +1400,5939235,"TERMINAL",0,0,"2162216",,terminal_output +1401,5940359,"TERMINAL",0,0,"3273327",,terminal_output +1402,5941291,"TERMINAL",0,0,"4384438",,terminal_output +1403,5942344,"TERMINAL",0,0,"5495549",,terminal_output +1404,5943389,"TERMINAL",0,0,"654066550",,terminal_output +1405,5944458,"TERMINAL",0,0,"7617761",,terminal_output +1406,5945483,"TERMINAL",0,0,"8728872",,terminal_output +1407,5946520,"TERMINAL",0,0,"9839983",,terminal_output +1408,5947577,"TERMINAL",0,0,"4094303094",,terminal_output +1409,5948622,"TERMINAL",0,0,"140511405",,terminal_output +1410,5949694,"TERMINAL",0,0,"2162216",,terminal_output +1411,5950704,"TERMINAL",0,0,"3273327",,terminal_output +1412,5951746,"TERMINAL",0,0,"4384438",,terminal_output +1413,5952795,"TERMINAL",0,0,"5495549",,terminal_output +1414,5953845,"TERMINAL",0,0,"66517766:01",,terminal_output +1415,5954885,"TERMINAL",0,0,"8728872",,terminal_output +1416,5955948,"TERMINAL",0,0,"9839983",,terminal_output +1417,5956970,"TERMINAL",0,0,"5094404094",,terminal_output +1418,5958074,"TERMINAL",0,0,"150511505",,terminal_output +1419,5959064,"TERMINAL",0,0,"2162216",,terminal_output +1420,5960118,"TERMINAL",0,0,"3273327",,terminal_output +1421,5961162,"TERMINAL",0,0,"4384438",,terminal_output +1422,5962242,"TERMINAL",0,0,"5495549",,terminal_output +1423,5963260,"TERMINAL",0,0,"658:0066510",,terminal_output +1424,5964304,"TERMINAL",0,0,"7617761",,terminal_output +1425,5965375,"TERMINAL",0,0,"8728872",,terminal_output +1426,5966430,"TERMINAL",0,0,"9839983",,terminal_output +1427,5967436,"TERMINAL",0,0,"7:0094505094",,terminal_output +1428,5968488,"TERMINAL",0,0,"17:005116:005",,terminal_output +1429,5969544,"TERMINAL",0,0,"2162216",,terminal_output +1430,5970667,"TERMINAL",0,0,"3273327",,terminal_output +1431,5971665,"TERMINAL",0,0,"4384438",,terminal_output +1432,5972670,"TERMINAL",0,0,"5495549",,terminal_output +1433,5973717,"TERMINAL",0,0,"651066520",,terminal_output +1434,5974870,"TERMINAL",0,0,"7617761",,terminal_output +1435,5975813,"TERMINAL",0,0,"8728872",,terminal_output +1436,5976854,"TERMINAL",0,0,"9944:006:0094",,terminal_output +1437,5977893,"TERMINAL",0,0,"1110511105",,terminal_output +1438,5979074,"TERMINAL",0,0,"2162216",,terminal_output +1439,5980101,"TERMINAL",0,0,"3273327",,terminal_output +1440,5981115,"TERMINAL",0,0,"4384438",,terminal_output +1441,5982138,"TERMINAL",0,0,"5495549",,terminal_output +1442,5983163,"TERMINAL",0,0,"652066530",,terminal_output +1443,5984174,"TERMINAL",0,0,"7617761",,terminal_output +1444,5985274,"TERMINAL",0,0,"8728872",,terminal_output +1445,5986337,"TERMINAL",0,0,"9839983",,terminal_output +1446,5987320,"TERMINAL",0,0,"2094101094",,terminal_output +1447,5988395,"TERMINAL",0,0,"120511205",,terminal_output +1448,5989410,"TERMINAL",0,0,"2162216",,terminal_output +1449,5990535,"TERMINAL",0,0,"3273327",,terminal_output +1450,5991559,"TERMINAL",0,0,"4384438",,terminal_output +1451,5992582,"TERMINAL",0,0,"5495549",,terminal_output +1452,5993606,"TERMINAL",0,0,"653066540",,terminal_output +1453,5994732,"TERMINAL",0,0,"7617761",,terminal_output +1454,5995707,"TERMINAL",0,0,"8728872",,terminal_output +1455,5996755,"TERMINAL",0,0,"9839983",,terminal_output +1456,5997808,"TERMINAL",0,0,"3094202094",,terminal_output +1457,5999034,"TERMINAL",0,0,"131622316",,terminal_output +1458,6000086,"TERMINAL",0,0,"3273327",,terminal_output +1459,6000978,"TERMINAL",0,0,"4384438",,terminal_output +1460,6001998,"TERMINAL",0,0,"5495549",,terminal_output +1461,6003130,"TERMINAL",0,0,"654066550",,terminal_output +1462,6004153,"TERMINAL",0,0,"7617761",,terminal_output +1463,6005289,"TERMINAL",0,0,"8728872",,terminal_output +1464,6006231,"TERMINAL",0,0,"9839983",,terminal_output +1465,6007329,"TERMINAL",0,0,"4094303094",,terminal_output +1466,6008297,"TERMINAL",0,0,"140511405",,terminal_output +1467,6009342,"TERMINAL",0,0,"2162216",,terminal_output +1468,6010388,"TERMINAL",0,0,"3273327",,terminal_output +1469,6011446,"TERMINAL",0,0,"4384438",,terminal_output +1470,6012553,"TERMINAL",0,0,"5495549",,terminal_output +1471,6013576,"TERMINAL",0,0,"65506657:00",,terminal_output +1472,6014590,"TERMINAL",0,0,"7617761",,terminal_output +1473,6015725,"TERMINAL",0,0,"8728872",,terminal_output +1474,6016768,"TERMINAL",0,0,"9839983",,terminal_output +1475,6017734,"TERMINAL",0,0,"5094404094",,terminal_output +1476,6018902,"TERMINAL",0,0,"150511505",,terminal_output +1477,6019823,"TERMINAL",0,0,"2273327",,terminal_output +1478,6020950,"TERMINAL",0,0,"4384438",,terminal_output +1479,6022076,"TERMINAL",0,0,"5495549",,terminal_output +1480,6023014,"TERMINAL",0,0,"659:0066510",,terminal_output +1481,6024026,"TERMINAL",0,0,"7617761",,terminal_output +1482,6025147,"TERMINAL",0,0,"8728872",,terminal_output +1483,6026171,"TERMINAL",0,0,"9839983",,terminal_output +1484,6027234,"TERMINAL",0,0,"8:0094505094",,terminal_output +1485,6028238,"TERMINAL",0,0,"18:005117:005",,terminal_output +1486,6029261,"TERMINAL",0,0,"2162216",,terminal_output +1487,6030305,"TERMINAL",0,0,"3273327",,terminal_output +1488,6031394,"TERMINAL",0,0,"4384438",,terminal_output +1489,6032401,"TERMINAL",0,0,"5495549",,terminal_output +1490,6033445,"TERMINAL",0,0,"651066520",,terminal_output +1491,6034565,"TERMINAL",0,0,"7617761",,terminal_output +1492,6035591,"TERMINAL",0,0,"8728872",,terminal_output +1493,6036577,"TERMINAL",0,0,"9839983",,terminal_output +1494,6037641,"TERMINAL",0,0,"10945:007:0094",,terminal_output +1495,6038767,"TERMINAL",0,0,"110511105",,terminal_output +1496,6039719,"TERMINAL",0,0,"2162216",,terminal_output +1497,6040813,"TERMINAL",0,0,"3273327",,terminal_output +1498,6041839,"TERMINAL",0,0,"4384438",,terminal_output +1499,6042855,"TERMINAL",0,0,"552066530",,terminal_output +1500,6044090,"TERMINAL",0,0,"7617761",,terminal_output +1501,6044950,"TERMINAL",0,0,"8728872",,terminal_output +1502,6046038,"TERMINAL",0,0,"9839983",,terminal_output +1503,6047047,"TERMINAL",0,0,"2094101094",,terminal_output +1504,6048107,"TERMINAL",0,0,"120511205",,terminal_output +1505,6049210,"TERMINAL",0,0,"2162216",,terminal_output +1506,6050268,"TERMINAL",0,0,"3273327",,terminal_output +1507,6051259,"TERMINAL",0,0,"4384438",,terminal_output +1508,6052283,"TERMINAL",0,0,"5495549",,terminal_output +1509,6053328,"TERMINAL",0,0,"653066540",,terminal_output +1510,6054372,"TERMINAL",0,0,"7617761",,terminal_output +1511,6055419,"TERMINAL",0,0,"8728872",,terminal_output +1512,6056466,"TERMINAL",0,0,"9839983",,terminal_output +1513,6057515,"TERMINAL",0,0,"3094202094",,terminal_output +1514,6058634,"TERMINAL",0,0,"130511305",,terminal_output +1515,6059608,"TERMINAL",0,0,"2162216",,terminal_output +1516,6060677,"TERMINAL",0,0,"3273327",,terminal_output +1517,6061702,"TERMINAL",0,0,"4384438",,terminal_output +1518,6062796,"TERMINAL",0,0,"5495549",,terminal_output +1519,6063955,"TERMINAL",0,0,"654066550",,terminal_output +1520,6064825,"TERMINAL",0,0,"7728872",,terminal_output +1521,6066006,"TERMINAL",0,0,"9839983",,terminal_output +1522,6067031,"TERMINAL",0,0,"4094303094",,terminal_output +1523,6068154,"TERMINAL",0,0,"140511405",,terminal_output +1524,6069281,"TERMINAL",0,0,"2162216",,terminal_output +1525,6070305,"TERMINAL",0,0,"3273327",,terminal_output +1526,6071189,"TERMINAL",0,0,"4384438",,terminal_output +1527,6072386,"TERMINAL",0,0,"5495549",,terminal_output +1528,6073284,"TERMINAL",0,0,"65506658:00",,terminal_output +1529,6074321,"TERMINAL",0,0,"7617761",,terminal_output +1530,6075373,"TERMINAL",0,0,"8728872",,terminal_output +1531,6076415,"TERMINAL",0,0,"9839983",,terminal_output +1532,6077457,"TERMINAL",0,0,"5094404094",,terminal_output +1533,6078602,"TERMINAL",0,0,"150511505",,terminal_output +1534,6079548,"TERMINAL",0,0,"2162216",,terminal_output +1535,6080647,"TERMINAL",0,0,"3273327",,terminal_output +1536,6081671,"TERMINAL",0,0,"4384438",,terminal_output +1537,6082680,"TERMINAL",0,0,"5495549",,terminal_output +1538,6083727,"TERMINAL",0,0,"6530:0066510",,terminal_output +1539,6084848,"TERMINAL",0,0,"7617761",,terminal_output +1540,6085816,"TERMINAL",0,0,"8839983",,terminal_output +1541,6086863,"TERMINAL",0,0,"9:0094505094",,terminal_output +1542,6087908,"TERMINAL",0,0,"19:005118:005",,terminal_output +1543,6088949,"TERMINAL",0,0,"2162216",,terminal_output +1544,6090004,"TERMINAL",0,0,"3273327",,terminal_output +1545,6091048,"TERMINAL",0,0,"4384438",,terminal_output +1546,6092220,"TERMINAL",0,0,"5495549",,terminal_output +1547,6093189,"TERMINAL",0,0,"651066520",,terminal_output +1548,6094376,"TERMINAL",0,0,"7617761",,terminal_output +1549,6095290,"TERMINAL",0,0,"8728872",,terminal_output +1550,6096312,"TERMINAL",0,0,"9839983",,terminal_output +1551,6097543,"TERMINAL",0,0,"10946:008:0094",,terminal_output +1552,6098411,"TERMINAL",0,0,"110511105",,terminal_output +1553,6099468,"TERMINAL",0,0,"2162216",,terminal_output +1554,6100501,"TERMINAL",0,0,"3273327",,terminal_output +1555,6101649,"TERMINAL",0,0,"4384438",,terminal_output +1556,6102597,"TERMINAL",0,0,"5495549",,terminal_output +1557,6103692,"TERMINAL",0,0,"652066530",,terminal_output +1558,6104714,"TERMINAL",0,0,"7617761",,terminal_output +1559,6105850,"TERMINAL",0,0,"8728872",,terminal_output +1560,6106786,"TERMINAL",0,0,"9839983",,terminal_output +1561,6107837,"TERMINAL",0,0,"202051111205",,terminal_output +1562,6108886,"TERMINAL",0,0,"2162216",,terminal_output +1563,6110032,"TERMINAL",0,0,"3273327",,terminal_output +1564,6111302,"TERMINAL",0,0,"4384438",,terminal_output +1565,6112085,"TERMINAL",0,0,"5495549",,terminal_output +1566,6113110,"TERMINAL",0,0,"653066540",,terminal_output +1567,6114124,"TERMINAL",0,0,"7617761",,terminal_output +1568,6115361,"TERMINAL",0,0,"8728872",,terminal_output +1569,6116218,"TERMINAL",0,0,"9839983",,terminal_output +1570,6117290,"TERMINAL",0,0,"3094202094",,terminal_output +1571,6118340,"TERMINAL",0,0,"130511305",,terminal_output +1572,6119357,"TERMINAL",0,0,"2162216",,terminal_output +1573,6120483,"TERMINAL",0,0,"3273327",,terminal_output +1574,6121447,"TERMINAL",0,0,"4384438",,terminal_output +1575,6122491,"TERMINAL",0,0,"5495549",,terminal_output +1576,6123554,"TERMINAL",0,0,"654066550",,terminal_output +1577,6124588,"TERMINAL",0,0,"7617761",,terminal_output +1578,6125704,"TERMINAL",0,0,"8728872",,terminal_output +1579,6126729,"TERMINAL",0,0,"9839983",,terminal_output +1580,6127790,"TERMINAL",0,0,"4094303094",,terminal_output +1581,6128774,"TERMINAL",0,0,"140511405",,terminal_output +1582,6129816,"TERMINAL",0,0,"2162216",,terminal_output +1583,6130927,"TERMINAL",0,0,"4384438",,terminal_output +1584,6131948,"TERMINAL",0,0,"5495549",,terminal_output +1585,6132973,"TERMINAL",0,0,"65506659:00",,terminal_output +1586,6133992,"TERMINAL",0,0,"7617761",,terminal_output +1587,6135042,"TERMINAL",0,0,"8728872",,terminal_output +1588,6136272,"TERMINAL",0,0,"9839983",,terminal_output +1589,6137172,"TERMINAL",0,0,"5094404094",,terminal_output +1590,6138239,"TERMINAL",0,0,"150511505",,terminal_output +1591,6139239,"TERMINAL",0,0,"2162216",,terminal_output +1592,6140330,"TERMINAL",0,0,"3273327",,terminal_output +1593,6141366,"TERMINAL",0,0,"4384438",,terminal_output +1594,6142410,"TERMINAL",0,0,"5495549",,terminal_output +1595,6143520,"TERMINAL",0,0,"651:0066510",,terminal_output +1596,6144499,"TERMINAL",0,0,"7617761",,terminal_output +1597,6145553,"TERMINAL",0,0,"8728872",,terminal_output +1598,6146606,"TERMINAL",0,0,"9839983",,terminal_output +1599,6147720,"TERMINAL",0,0,"20:0094505094",,terminal_output +1600,6148694,"TERMINAL",0,0,"120:005119:005",,terminal_output +1601,6149788,"TERMINAL",0,0,"2162216",,terminal_output +1602,6150784,"TERMINAL",0,0,"3273327",,terminal_output +1603,6152022,"TERMINAL",0,0,"4495549",,terminal_output +1604,6152873,"TERMINAL",0,0,"651066520",,terminal_output +1605,6153968,"TERMINAL",0,0,"7617761",,terminal_output +1606,6154990,"TERMINAL",0,0,"8728872",,terminal_output +1607,6156116,"TERMINAL",0,0,"9839983",,terminal_output +1608,6157140,"TERMINAL",0,0,"10947:009:0094",,terminal_output +1609,6158164,"TERMINAL",0,0,"110511105",,terminal_output +1610,6159296,"TERMINAL",0,0,"2162216",,terminal_output +1611,6160212,"TERMINAL",0,0,"3273327",,terminal_output +1612,6161232,"TERMINAL",0,0,"4384438",,terminal_output +1613,6162305,"TERMINAL",0,0,"5495549",,terminal_output +1614,6163388,"TERMINAL",0,0,"652066530",,terminal_output +1615,6164362,"TERMINAL",0,0,"7617761",,terminal_output +1616,6165422,"TERMINAL",0,0,"8728872",,terminal_output +1617,6166453,"TERMINAL",0,0,"9839983",,terminal_output +1618,6167498,"TERMINAL",0,0,"2094101094",,terminal_output +1619,6168611,"TERMINAL",0,0,"120511205",,terminal_output +1620,6169584,"TERMINAL",0,0,"2162216",,terminal_output +1621,6170622,"TERMINAL",0,0,"3273327",,terminal_output +1622,6171684,"TERMINAL",0,0,"4384438",,terminal_output +1623,6172797,"TERMINAL",0,0,"5495549",,terminal_output +1624,6173747,"TERMINAL",0,0,"653066540",,terminal_output +1625,6174856,"TERMINAL",0,0,"7617761",,terminal_output +1626,6175826,"TERMINAL",0,0,"8839983",,terminal_output +1627,6176872,"TERMINAL",0,0,"3094202094",,terminal_output +1628,6177910,"TERMINAL",0,0,"130511305",,terminal_output +1629,6179165,"TERMINAL",0,0,"2162216",,terminal_output +1630,6180188,"TERMINAL",0,0,"3273327",,terminal_output +1631,6181068,"TERMINAL",0,0,"4384438",,terminal_output +1632,6182104,"TERMINAL",0,0,"5495549",,terminal_output +1633,6183192,"TERMINAL",0,0,"654066550",,terminal_output +1634,6184309,"TERMINAL",0,0,"7617761",,terminal_output +1635,6185322,"TERMINAL",0,0,"8728872",,terminal_output +1636,6186532,"TERMINAL",0,0,"9839983",,terminal_output +1637,6187334,"TERMINAL",0,0,"4094303094",,terminal_output +1638,6188392,"TERMINAL",0,0,"140511405",,terminal_output +1639,6189443,"TERMINAL",0,0,"2162216",,terminal_output +1640,6190492,"TERMINAL",0,0,"3273327",,terminal_output +1641,6191648,"TERMINAL",0,0,"4384438",,terminal_output +1642,6192592,"TERMINAL",0,0,"5495549",,terminal_output +1643,6193697,"TERMINAL",0,0,"655066520:00",,terminal_output +1644,6194824,"TERMINAL",0,0,"7617761",,terminal_output +1645,6195745,"TERMINAL",0,0,"8728872",,terminal_output +1646,6196769,"TERMINAL",0,0,"9839983",,terminal_output +1647,6197822,"TERMINAL",0,0,"505054141505",,terminal_output +1648,6198873,"TERMINAL",0,0,"2162216",,terminal_output +1649,6199918,"TERMINAL",0,0,"3273327",,terminal_output +1650,6201030,"TERMINAL",0,0,"4384438",,terminal_output +1651,6202094,"TERMINAL",0,0,"5495549",,terminal_output +1652,6203065,"TERMINAL",0,0,"652:0066510",,terminal_output +1653,6204132,"TERMINAL",0,0,"7617761",,terminal_output +1654,6205203,"TERMINAL",0,0,"8728872",,terminal_output +1655,6206253,"TERMINAL",0,0,"9839983",,terminal_output +1656,6207324,"TERMINAL",0,0,"1:0094505094",,terminal_output +1657,6208346,"TERMINAL",0,0,"11:0051120:005",,terminal_output +1658,6209363,"TERMINAL",0,0,"2162216",,terminal_output +1659,6210493,"TERMINAL",0,0,"3273327",,terminal_output +1660,6211440,"TERMINAL",0,0,"4384438",,terminal_output +1661,6212484,"TERMINAL",0,0,"5495549",,terminal_output +1662,6213565,"TERMINAL",0,0,"651066520",,terminal_output +1663,6214590,"TERMINAL",0,0,"7617761",,terminal_output +1664,6215616,"TERMINAL",0,0,"8728872",,terminal_output +1665,6216739,"TERMINAL",0,0,"9839983",,terminal_output +1666,6217763,"TERMINAL",0,0,"10948:0020:0094",,terminal_output +1667,6218766,"TERMINAL",0,0,"110511105",,terminal_output +1668,6219807,"TERMINAL",0,0,"2162216",,terminal_output +1669,6220855,"TERMINAL",0,0,"3384438",,terminal_output +1670,6221892,"TERMINAL",0,0,"5495549",,terminal_output +1671,6222987,"TERMINAL",0,0,"652066530",,terminal_output +1672,6224008,"TERMINAL",0,0,"7617761",,terminal_output +1673,6225238,"TERMINAL",0,0,"8728872",,terminal_output +1674,6226172,"TERMINAL",0,0,"9839983",,terminal_output +1675,6227184,"TERMINAL",0,0,"2094101094",,terminal_output +1676,6228208,"TERMINAL",0,0,"120511205",,terminal_output +1677,6229231,"TERMINAL",0,0,"2162216",,terminal_output +1678,6230362,"TERMINAL",0,0,"3273327",,terminal_output +1679,6231385,"TERMINAL",0,0,"4384438",,terminal_output +1680,6232364,"TERMINAL",0,0,"5495549",,terminal_output +1681,6233414,"TERMINAL",0,0,"653066540",,terminal_output +1682,6234470,"TERMINAL",0,0,"7617761",,terminal_output +1683,6235518,"TERMINAL",0,0,"8728872",,terminal_output +1684,6236554,"TERMINAL",0,0,"9839983",,terminal_output +1685,6237631,"TERMINAL",0,0,"3094202094",,terminal_output +1686,6238660,"TERMINAL",0,0,"130511305",,terminal_output +1687,6239711,"TERMINAL",0,0,"2162216",,terminal_output +1688,6240778,"TERMINAL",0,0,"3273327",,terminal_output +1689,6241825,"TERMINAL",0,0,"4384438",,terminal_output +1690,6242860,"TERMINAL",0,0,"554066550",,terminal_output +1691,6243975,"TERMINAL",0,0,"7617761",,terminal_output +1692,6244998,"TERMINAL",0,0,"8728872",,terminal_output +1693,6246025,"TERMINAL",0,0,"9839983",,terminal_output +1694,6247053,"TERMINAL",0,0,"4094303094",,terminal_output +1695,6248098,"TERMINAL",0,0,"140511405",,terminal_output +1696,6249142,"TERMINAL",0,0,"2162216",,terminal_output +1697,6250222,"TERMINAL",0,0,"3273327",,terminal_output +1698,6251289,"TERMINAL",0,0,"4384438",,terminal_output +1699,6252275,"TERMINAL",0,0,"5495549",,terminal_output +1700,6253363,"TERMINAL",0,0,"65506651:00",,terminal_output +1701,6254424,"TERMINAL",0,0,"7617761",,terminal_output +1702,6255549,"TERMINAL",0,0,"8728872",,terminal_output +1703,6256467,"TERMINAL",0,0,"9839983",,terminal_output +1704,6257530,"TERMINAL",0,0,"5094404094",,terminal_output +1705,6258620,"TERMINAL",0,0,"150511505",,terminal_output +1706,6259645,"TERMINAL",0,0,"2162216",,terminal_output +1707,6260668,"TERMINAL",0,0,"3273327",,terminal_output +1708,6261694,"TERMINAL",0,0,"4384438",,terminal_output +1709,6262799,"TERMINAL",0,0,"5495549",,terminal_output +1710,6263842,"TERMINAL",0,0,"653:0066510",,terminal_output +1711,6264976,"TERMINAL",0,0,"7728872",,terminal_output +1712,6265875,"TERMINAL",0,0,"9839983",,terminal_output +1713,6267014,"TERMINAL",0,0,"2:0094505094",,terminal_output +1714,6268042,"TERMINAL",0,0,"12:005111:005",,terminal_output +1715,6269022,"TERMINAL",0,0,"2162216",,terminal_output +1716,6270104,"TERMINAL",0,0,"3273327",,terminal_output +1717,6271112,"TERMINAL",0,0,"4384438",,terminal_output +1718,6272170,"TERMINAL",0,0,"5495549",,terminal_output +1719,6273206,"TERMINAL",0,0,"651066520",,terminal_output +1720,6274255,"TERMINAL",0,0,"7617761",,terminal_output +1721,6275351,"TERMINAL",0,0,"8728872",,terminal_output +1722,6276438,"TERMINAL",0,0,"9839983",,terminal_output +1723,6277403,"TERMINAL",0,0,"10949:001:0094",,terminal_output +1724,6278440,"TERMINAL",0,0,"110511105",,terminal_output +1725,6279507,"TERMINAL",0,0,"2162216",,terminal_output +1726,6280544,"TERMINAL",0,0,"3273327",,terminal_output +1727,6281662,"TERMINAL",0,0,"4384438",,terminal_output +1728,6282687,"TERMINAL",0,0,"5495549",,terminal_output +1729,6283654,"TERMINAL",0,0,"652066530",,terminal_output +1730,6284693,"TERMINAL",0,0,"7617761",,terminal_output +1731,6285756,"TERMINAL",0,0,"8728872",,terminal_output +1732,6286793,"TERMINAL",0,0,"9839983",,terminal_output +1733,6287844,"TERMINAL",0,0,"202051111205",,terminal_output +1734,6288932,"TERMINAL",0,0,"2162216",,terminal_output +1735,6289958,"TERMINAL",0,0,"3273327",,terminal_output +1736,6290977,"TERMINAL",0,0,"4384438",,terminal_output +1737,6292029,"TERMINAL",0,0,"5495549",,terminal_output +1738,6293082,"TERMINAL",0,0,"653066540",,terminal_output +1739,6294258,"TERMINAL",0,0,"7617761",,terminal_output +1740,6295282,"TERMINAL",0,0,"8728872",,terminal_output +1741,6296304,"TERMINAL",0,0,"9839983",,terminal_output +1742,6297270,"TERMINAL",0,0,"3094202094",,terminal_output +1743,6298386,"TERMINAL",0,0,"130511305",,terminal_output +1744,6299379,"TERMINAL",0,0,"2162216",,terminal_output +1745,6300381,"TERMINAL",0,0,"3273327",,terminal_output +1746,6301430,"TERMINAL",0,0,"4384438",,terminal_output +1747,6302477,"TERMINAL",0,0,"5495549",,terminal_output +1748,6303528,"TERMINAL",0,0,"654066550",,terminal_output +1749,6304559,"TERMINAL",0,0,"7617761",,terminal_output +1750,6305601,"TERMINAL",0,0,"8728872",,terminal_output +1751,6306644,"TERMINAL",0,0,"9839983",,terminal_output +1752,6307680,"TERMINAL",0,0,"4094303094",,terminal_output +1753,6308736,"TERMINAL",0,0,"140511405",,terminal_output +1754,6309766,"TERMINAL",0,0,"2162216",,terminal_output +1755,6310810,"TERMINAL",0,0,"3273327",,terminal_output +1756,6311857,"TERMINAL",0,0,"5495549",,terminal_output +1757,6312902,"TERMINAL",0,0,"65506652:00",,terminal_output +1758,6313945,"TERMINAL",0,0,"7617761",,terminal_output +1759,6315042,"TERMINAL",0,0,"8728872",,terminal_output +1760,6316066,"TERMINAL",0,0,"9839983",,terminal_output +1761,6317092,"TERMINAL",0,0,"5094404094",,terminal_output +1762,6318216,"TERMINAL",0,0,"150511505",,terminal_output +1763,6319183,"TERMINAL",0,0,"2162216",,terminal_output +1764,6320256,"TERMINAL",0,0,"3273327",,terminal_output +1765,6321320,"TERMINAL",0,0,"4384438",,terminal_output +1766,6322326,"TERMINAL",0,0,"5495549",,terminal_output +1767,6322637,"TERMINAL",0,0,"bash",,terminal_focus +1768,6323233,"TERMINAL",0,0,"watch",,terminal_focus +1769,6323369,"TERMINAL",0,0,"654:0066510",,terminal_output +1770,6323617,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1771,6324569,"TERMINAL",0,0,"idling",,terminal_command +1772,6324649,"TERMINAL",0,0,"]633;E;2025-08-14 12:22:57 idling;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Aug 14 12:22:57 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 121 nodes idle\rPartition dev_accelerated:\t 0 nodes idle\rPartition accelerated: 19 nodes idle\rPartition dev_accelerated-h100 :\t 1 nodes idle\rPartition accelerated-h100:\t 1 nodes idle\rPartition large:\t 6 nodes idle",,terminal_output +1773,6325689,"TERMINAL",0,0,"8\t ",,terminal_output +1774,6326739,"TERMINAL",0,0,"9\t ",,terminal_output +1775,6327801,"TERMINAL",0,0,"3:00\t ",,terminal_output +1776,6328906,"TERMINAL",0,0,"1\t ",,terminal_output +1777,6329865,"TERMINAL",0,0,"3\t ",,terminal_output +1778,6331125,"TERMINAL",0,0,"4\t ",,terminal_output +1779,6332041,"TERMINAL",0,0,"5\t ",,terminal_output +1780,6333167,"TERMINAL",0,0,"6\t ",,terminal_output +1781,6334039,"TERMINAL",0,0,"7\t ",,terminal_output +1782,6335112,"TERMINAL",0,0,"8\t ",,terminal_output +1783,6336145,"TERMINAL",0,0,"9\t ",,terminal_output +1784,6337211,"TERMINAL",0,0,"10\t ",,terminal_output +1785,6338208,"TERMINAL",0,0,"1\t ",,terminal_output +1786,6339261,"TERMINAL",0,0,"2\t ",,terminal_output +1787,6340335,"TERMINAL",0,0,"3\t ",,terminal_output +1788,6341352,"TERMINAL",0,0,"4\t ",,terminal_output +1789,6342390,"TERMINAL",0,0,"5\t ",,terminal_output +1790,6343434,"TERMINAL",0,0,"6\t ",,terminal_output +1791,6344639,"TERMINAL",0,0,"7\t ",,terminal_output +1792,6345544,"TERMINAL",0,0,"8\t ",,terminal_output +1793,6346558,"TERMINAL",0,0,"9\t ",,terminal_output +1794,6347613,"TERMINAL",0,0,"20\t ",,terminal_output +1795,6348737,"TERMINAL",0,0,"1\t ",,terminal_output +1796,6349758,"TERMINAL",0,0,"2\t ",,terminal_output +1797,6350760,"TERMINAL",0,0,"3\t ",,terminal_output +1798,6351809,"TERMINAL",0,0,"48",,terminal_output +1799,6352826,"TERMINAL",0,0,"5\t ",,terminal_output +1800,6354058,"TERMINAL",0,0,"7\t ",,terminal_output +1801,6355082,"TERMINAL",0,0,"8\t ",,terminal_output +1802,6355951,"TERMINAL",0,0,"9\t ",,terminal_output +1803,6357030,"TERMINAL",0,0,"30\t ",,terminal_output +1804,6358031,"TERMINAL",0,0,"1\t ",,terminal_output +1805,6359178,"TERMINAL",0,0,"2\t ",,terminal_output +1806,6360201,"TERMINAL",0,0,"3\t ",,terminal_output +1807,6361226,"TERMINAL",0,0,"4\t ",,terminal_output +1808,6362248,"TERMINAL",0,0,"5\t ",,terminal_output +1809,6363233,"TERMINAL",0,0,"6\t ",,terminal_output +1810,6364299,"TERMINAL",0,0,"7\t ",,terminal_output +1811,6365526,"TERMINAL",0,0,"8\t ",,terminal_output +1812,6366392,"TERMINAL",0,0,"9\t ",,terminal_output +1813,6367577,"TERMINAL",0,0,"40\t ",,terminal_output +1814,6368497,"TERMINAL",0,0,"1\t ",,terminal_output +1815,6369488,"TERMINAL",0,0,"2\t ",,terminal_output +1816,6370542,"TERMINAL",0,0,"3\t ",,terminal_output +1817,6371566,"TERMINAL",0,0,"4\t ",,terminal_output +1818,6372609,"TERMINAL",0,0,"5\t ",,terminal_output +1819,6373652,"TERMINAL",0,0,"6\t ",,terminal_output +1820,6374690,"TERMINAL",0,0,"7\t ",,terminal_output +1821,6375768,"TERMINAL",0,0,"8\t ",,terminal_output +1822,6376772,"TERMINAL",0,0,"9\t ",,terminal_output +1823,6377810,"TERMINAL",0,0,"509",,terminal_output +1824,6378849,"TERMINAL",0,0,"2\t ",,terminal_output +1825,6379893,"TERMINAL",0,0,"3\t ",,terminal_output +1826,6380935,"TERMINAL",0,0,"4\t ",,terminal_output +1827,6381966,"TERMINAL",0,0,"5\t ",,terminal_output +1828,6383138,"TERMINAL",0,0,"6\t ",,terminal_output +1829,6384162,"TERMINAL",0,0,"7\t ",,terminal_output +1830,6385102,"TERMINAL",0,0,"8\t ",,terminal_output +1831,6386215,"TERMINAL",0,0,"9\t ",,terminal_output +1832,6387183,"TERMINAL",0,0,"4:00\t ",,terminal_output +1833,6388219,"TERMINAL",0,0,"1\t ",,terminal_output +1834,6389287,"TERMINAL",0,0,"2\t ",,terminal_output +1835,6390410,"TERMINAL",0,0,"3\t ",,terminal_output +1836,6391342,"TERMINAL",0,0,"4\t ",,terminal_output +1837,6392459,"TERMINAL",0,0,"5\t ",,terminal_output +1838,6393447,"TERMINAL",0,0,"6\t ",,terminal_output +1839,6394515,"TERMINAL",0,0,"7\t ",,terminal_output +1840,6395499,"TERMINAL",0,0,"8\t ",,terminal_output +1841,6396542,"TERMINAL",0,0,"9\t ",,terminal_output +1842,6397583,"TERMINAL",0,0,"10\t ",,terminal_output +1843,6398707,"TERMINAL",0,0,"1\t ",,terminal_output +1844,6399730,"TERMINAL",0,0,"2\t ",,terminal_output +1845,6400755,"TERMINAL",0,0,"3\t ",,terminal_output +1846,6401775,"TERMINAL",0,0,"4\t ",,terminal_output +1847,6402799,"TERMINAL",0,0,"5\t ",,terminal_output +1848,6403928,"TERMINAL",0,0,"6\t ",,terminal_output +1849,6405055,"TERMINAL",0,0,"8\t ",,terminal_output +1850,6405980,"TERMINAL",0,0,"9\t ",,terminal_output +1851,6407004,"TERMINAL",0,0,"20\t ",,terminal_output +1852,6408024,"TERMINAL",0,0,"1\t ",,terminal_output +1853,6409080,"TERMINAL",0,0,"2\t ",,terminal_output +1854,6409512,"TERMINAL",0,0,"bash",,terminal_focus +1855,6410084,"TERMINAL",0,0,"3\t ",,terminal_output +1856,6410957,"TERMINAL",0,0,"watch",,terminal_focus +1857,6411119,"TERMINAL",0,0,"4\t ",,terminal_output +1858,6412158,"TERMINAL",0,0,"5\t ",,terminal_output +1859,6412690,"TERMINAL",0,0,"bash",,terminal_focus +1860,6413227,"TERMINAL",0,0,"6\t ",,terminal_output +1861,6413355,"TERMINAL",0,0,"watch",,terminal_focus +1862,6413905,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1863,6426861,"TERMINAL",0,0,"scontrol show reservation",,terminal_command +1864,6426936,"TERMINAL",0,0,"]633;E;2025-08-14 12:24:40 scontrol show reservation;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C",,terminal_output +1865,6427009,"TERMINAL",0,0,"ReservationName=llmtum StartTime=2025-08-14T12:00:00 EndTime=2025-09-30T23:59:00 Duration=47-11:59:00\r\n Nodes=hkn[0521-0528,0530-0534,0536,0701-0704,0708,0710-0715,0720,0722-0724,0726,0728,0731] NodeCnt=32 CoreCnt=2432 Features=(null) PartitionName=accelerated Flags=\r\n TRES=cpu=4864\r\n Users=(null) Groups=(null) Accounts=hk-project-p0023960 Licenses=(null) State=ACTIVE BurstBuffer=(null)\r\n MaxStartDelay=(null)\r\n\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1866,6635883,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",0,0,"",shellscript,tab +1867,6636702,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",450,0,"",shellscript,selection_mouse +1868,6636844,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",449,11,"reservation",shellscript,selection_mouse +1869,6636982,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",439,29,"#SBATCH --reservation=llmtum\n",shellscript,selection_mouse +1870,8890509,"TERMINAL",0,0,"bash",,terminal_focus +1871,8892237,"TERMINAL",0,0,"idling",,terminal_command +1872,8892312,"TERMINAL",0,0,"]633;E;2025-08-14 13:05:45 idling;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Aug 14 13:05:45 2025Partition dev_cpuonly: 12 nodes idle\rPartition cpuonly: 212 nodes idle\rPartition dev_accelerated:\t 0 nodes idle\rPartition accelerated: 13 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 1 nodes idle\rPartition large:\t 6 nodes idle",,terminal_output +1873,8893352,"TERMINAL",0,0,"6\t ",,terminal_output +1874,8894346,"TERMINAL",0,0,"721",,terminal_output +1875,8895447,"TERMINAL",0,0,"8\t ",,terminal_output +1876,8896472,"TERMINAL",0,0,"9\t ",,terminal_output +1877,8897509,"TERMINAL",0,0,"50\t ",,terminal_output +1878,8898519,"TERMINAL",0,0,"1\t ",,terminal_output +1879,8899543,"TERMINAL",0,0,"2\t ",,terminal_output +1880,8900776,"TERMINAL",0,0,"3\t ",,terminal_output +1881,8901731,"TERMINAL",0,0,"4\t ",,terminal_output +1882,8902782,"TERMINAL",0,0,"5\t ",,terminal_output +1883,8903739,"TERMINAL",0,0,"6\t ",,terminal_output +1884,8904727,"TERMINAL",0,0,"7\t ",,terminal_output +1885,8905773,"TERMINAL",0,0,"8\t ",,terminal_output +1886,8906854,"TERMINAL",0,0,"9\t ",,terminal_output +1887,8907851,"TERMINAL",0,0,"6:00\t ",,terminal_output +1888,8908421,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",571,0,"",shellscript,selection_mouse +1889,8908884,"TERMINAL",0,0,"2\t ",,terminal_output +1890,8908928,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",467,0,"",shellscript,selection_mouse +1891,8908962,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",466,0,"",shellscript,selection_command +1892,8909922,"TERMINAL",0,0,"3\t ",,terminal_output +1893,8910960,"TERMINAL",0,0,"4\t ",,terminal_output +1894,8911998,"TERMINAL",0,0,"5\t ",,terminal_output +1895,8913061,"TERMINAL",0,0,"6\t ",,terminal_output +1896,8914083,"TERMINAL",0,0,"7\t ",,terminal_output +1897,8915209,"TERMINAL",0,0,"8\t ",,terminal_output +1898,8916153,"TERMINAL",0,0,"9\t ",,terminal_output +1899,8917223,"TERMINAL",0,0,"10\t ",,terminal_output +1900,8918383,"TERMINAL",0,0,"1\t ",,terminal_output +1901,8919305,"TERMINAL",0,0,"2\t ",,terminal_output +1902,8920330,"TERMINAL",0,0,"3\t ",,terminal_output +1903,8921350,"TERMINAL",0,0,"4\t ",,terminal_output +1904,8922415,"TERMINAL",0,0,"5\t ",,terminal_output +1905,8923506,"TERMINAL",0,0,"6\t ",,terminal_output +1906,8924473,"TERMINAL",0,0,"7\t ",,terminal_output +1907,8925562,"TERMINAL",0,0,"8\t ",,terminal_output +1908,8926548,"TERMINAL",0,0,"9\t ",,terminal_output +1909,8927647,"TERMINAL",0,0,"201",,terminal_output +1910,8928632,"TERMINAL",0,0,"1\t ",,terminal_output +1911,8929665,"TERMINAL",0,0,"2\t ",,terminal_output +1912,8930758,"TERMINAL",0,0,"3\t ",,terminal_output +1913,8931796,"TERMINAL",0,0,"4\t ",,terminal_output +1914,8932788,"TERMINAL",0,0,"5\t ",,terminal_output +1915,8933952,"TERMINAL",0,0,"6\t ",,terminal_output +1916,8934861,"TERMINAL",0,0,"8\t ",,terminal_output +1917,8935902,"TERMINAL",0,0,"9\t ",,terminal_output +1918,8936943,"TERMINAL",0,0,"30\t ",,terminal_output +1919,8938046,"TERMINAL",0,0,"1\t ",,terminal_output +1920,8939067,"TERMINAL",0,0,"2\t ",,terminal_output +1921,8940058,"TERMINAL",0,0,"3\t ",,terminal_output +1922,8941118,"TERMINAL",0,0,"4\t ",,terminal_output +1923,8942245,"TERMINAL",0,0,"5\t ",,terminal_output +1924,8943269,"TERMINAL",0,0,"6\t ",,terminal_output +1925,8944292,"TERMINAL",0,0,"7\t ",,terminal_output +1926,8945314,"TERMINAL",0,0,"8\t ",,terminal_output +1927,8946447,"TERMINAL",0,0,"9\t ",,terminal_output +1928,8947472,"TERMINAL",0,0,"400",,terminal_output +1929,8948493,"TERMINAL",0,0,"1\t ",,terminal_output +1930,8949435,"TERMINAL",0,0,"25",,terminal_output +1931,8950480,"TERMINAL",0,0,"3\t ",,terminal_output +1932,8951514,"TERMINAL",0,0,"4\t ",,terminal_output +1933,8952585,"TERMINAL",0,0,"5\t ",,terminal_output +1934,8953615,"TERMINAL",0,0,"6\t ",,terminal_output +1935,8954764,"TERMINAL",0,0,"7\t ",,terminal_output +1936,8955837,"TERMINAL",0,0,"8\t ",,terminal_output +1937,8956791,"TERMINAL",0,0,"9\t ",,terminal_output +1938,8957782,"TERMINAL",0,0,"50\t ",,terminal_output +1939,8958805,"TERMINAL",0,0,"1\t ",,terminal_output +1940,8959830,"TERMINAL",0,0,"2\t ",,terminal_output +1941,8960890,"TERMINAL",0,0,"43",,terminal_output +1942,8961910,"TERMINAL",0,0,"52",,terminal_output +1943,8962951,"TERMINAL",0,0,"6\t ",,terminal_output +1944,8963987,"TERMINAL",0,0,"7\t ",,terminal_output +1945,8965031,"TERMINAL",0,0,"8\t ",,terminal_output +1946,8966061,"TERMINAL",0,0,"9\t ",,terminal_output +1947,8967125,"TERMINAL",0,0,"7:003",,terminal_output +1948,8968150,"TERMINAL",0,0,"1\t ",,terminal_output +1949,8969227,"TERMINAL",0,0,"2\t ",,terminal_output +1950,8970216,"TERMINAL",0,0,"3\t ",,terminal_output +1951,8971327,"TERMINAL",0,0,"4\t ",,terminal_output +1952,8972315,"TERMINAL",0,0,"52",,terminal_output +1953,8973351,"TERMINAL",0,0,"6\t ",,terminal_output +1954,8974398,"TERMINAL",0,0,"7\t ",,terminal_output +1955,8975523,"TERMINAL",0,0,"8\t ",,terminal_output +1956,8976494,"TERMINAL",0,0,"9\t ",,terminal_output +1957,8977569,"TERMINAL",0,0,"10\t ",,terminal_output +1958,8978531,"TERMINAL",0,0,"1\t ",,terminal_output +1959,8979566,"TERMINAL",0,0,"2\t ",,terminal_output +1960,8980643,"TERMINAL",0,0,"3\t ",,terminal_output +1961,8981667,"TERMINAL",0,0,"4\t ",,terminal_output +1962,8982813,"TERMINAL",0,0,"53",,terminal_output +1963,8983819,"TERMINAL",0,0,"6\t ",,terminal_output +1964,8984840,"TERMINAL",0,0,"7\t ",,terminal_output +1965,8985803,"TERMINAL",0,0,"8\t ",,terminal_output +1966,8986843,"TERMINAL",0,0,"9\t ",,terminal_output +1967,8987882,"TERMINAL",0,0,"21\t ",,terminal_output +1968,8988921,"TERMINAL",0,0,"2\t ",,terminal_output +1969,8989966,"TERMINAL",0,0,"3\t ",,terminal_output +1970,8991006,"TERMINAL",0,0,"4\t ",,terminal_output +1971,8992045,"TERMINAL",0,0,"54",,terminal_output +1972,8993094,"TERMINAL",0,0,"6\t ",,terminal_output +1973,8994124,"TERMINAL",0,0,"71",,terminal_output +1974,8995191,"TERMINAL",0,0,"8\t ",,terminal_output +1975,8996275,"TERMINAL",0,0,"9\t ",,terminal_output +1976,8997345,"TERMINAL",0,0,"30\t ",,terminal_output +1977,8998278,"TERMINAL",0,0,"1\t ",,terminal_output +1978,8999320,"TERMINAL",0,0,"2\t ",,terminal_output +1979,9000406,"TERMINAL",0,0,"3\t ",,terminal_output +1980,9001572,"TERMINAL",0,0,"4\t ",,terminal_output +1981,9002455,"TERMINAL",0,0,"5\t ",,terminal_output +1982,9003582,"TERMINAL",0,0,"6\t ",,terminal_output +1983,9004604,"TERMINAL",0,0,"7\t ",,terminal_output +1984,9005629,"TERMINAL",0,0,"8\t ",,terminal_output +1985,9006630,"TERMINAL",0,0,"9\t ",,terminal_output +1986,9007644,"TERMINAL",0,0,"40\t ",,terminal_output +1987,9008689,"TERMINAL",0,0,"1\t ",,terminal_output +1988,9009729,"TERMINAL",0,0,"2\t ",,terminal_output +1989,9010987,"TERMINAL",0,0,"3\t ",,terminal_output +1990,9011873,"TERMINAL",0,0,"4\t ",,terminal_output +1991,9012848,"TERMINAL",0,0,"6\t ",,terminal_output +1992,9014024,"TERMINAL",0,0,"7\t ",,terminal_output +1993,9014992,"TERMINAL",0,0,"8\t ",,terminal_output +1994,9016074,"TERMINAL",0,0,"9\t ",,terminal_output +1995,9017014,"TERMINAL",0,0,"50\t ",,terminal_output +1996,9018046,"TERMINAL",0,0,"1\t ",,terminal_output +1997,9019018,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +1998,9022701,"TERMINAL",0,0,"git branch",,terminal_command +1999,9022781,"TERMINAL",0,0,"]633;E;2025-08-14 13:07:55 git branch;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1h=\r",,terminal_output +2000,9022982,"TERMINAL",0,0," add-wandb-name-and-tags\r\n before-nnx\r\n causal-mem-reduce\r\n causal-spatiotemporal-kv-cache\r\n causal-st-transformer\r\n causal-transformer-dynamics-model\r\n causal-transformer-nnx-no-kv-cache\r\n convert-to-jax-array-in-iter\r\n correct-batched-sampling\r\n dev\r\n dont-let-tf-see-gpu\r\n* feat/darkness-filter\r\n:",,terminal_output +2001,9027893,"TERMINAL",0,0,"\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +2002,9029576,"TERMINAL",0,0,"git diff",,terminal_command +2003,9029609,"TERMINAL",0,0,"]633;E;2025-08-14 13:08:02 git diff;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1h=\r\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +2004,9089088,"TERMINAL",0,0,"git status",,terminal_command +2005,9089163,"TERMINAL",0,0,"]633;E;2025-08-14 13:09:02 git status;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +2006,9089238,"TERMINAL",0,0,"On branch feat/darkness-filter\r\nYour branch is up to date with 'origin/feat/darkness-filter'.\r\n\r\nUntracked files:\r\n (use ""git add ..."" to include in what will be committed)\r\n\tdata/\r\n\tdata_atari/\r\n\tlog.log\r\n\tlogs/\r\n\toverfit_dir.zip\r\n\tread_tf_record.py\r\n\trequirements-franz.txt\r\n\tsamples/\r\n\tscripts_cremers/\r\n\tslurm/\r\n\tutils/visualizer.py\r\n\r\nnothing added to commit but untracked files present (use ""git add"" to track)\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +2007,9098530,"TERMINAL",0,0,"git push",,terminal_command +2008,9098603,"TERMINAL",0,0,"]633;E;2025-08-14 13:09:11 git push;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +2009,9100120,"TERMINAL",0,0,"Everything up-to-date\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +2010,9155931,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_command +2011,9155961,"TERMINAL",0,0,"]633;E;2025-08-14 13:10:09 source .venv/bin/activate;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +2012,9157222,"TERMINAL",0,0,"bash",,terminal_focus +2013,9158374,"TERMINAL",0,0,"bash",,terminal_focus +2014,9161202,"TERMINAL",0,0,"git branch",,terminal_command +2015,9161256,"TERMINAL",0,0,"]633;E;2025-08-14 13:10:14 git branch;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1h=\r add-wandb-name-and-tags\r\n before-nnx\r\n causal-mem-reduce\r\n causal-spatiotemporal-kv-cache\r\n causal-st-transformer\r\n causal-transformer-dynamics-model\r\n causal-transformer-nnx-no-kv-cache\r\n convert-to-jax-array-in-iter\r\n correct-batched-sampling\r\n dev\r\n dont-let-tf-see-gpu\r\n* feat/darkness-filter\r\n:",,terminal_output +2016,9162865,"TERMINAL",0,0,"\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +2017,9164276,"TERMINAL",0,0,"bash",,terminal_focus +2018,9166817,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",0,0,"",shellscript,tab +2019,9175193,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",394,0,"",shellscript,selection_mouse +2020,9175782,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",465,0,"",shellscript,selection_mouse +2021,9175782,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",464,0,"",shellscript,selection_command +2022,9175944,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",436,29,"train_dynamics_maskgit_1_node",shellscript,selection_mouse +2023,9175945,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",437,28,"rain_dynamics_maskgit_1_node",shellscript,selection_command +2024,9176403,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",387,0,"",shellscript,selection_mouse +2025,9176564,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",382,7,"coinrun",shellscript,selection_mouse +2026,9176726,"slurm/jobs/mihir/horeka/coinrun/train_dynamics_maskgit.sbatch",294,123,"#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/maskgit/%x_%j.log\n",shellscript,selection_mouse +2027,9177808,"TERMINAL",0,0,"bash",,terminal_focus +2028,9183438,"TERMINAL",0,0,"logs",,terminal_command +2029,9183957,"TERMINAL",0,0,"ls",,terminal_command +2030,9184003,"TERMINAL",0,0,"]633;E;2025-08-14 13:10:37 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +2031,9184357,"TERMINAL",0,0,"atari train_lam_action_space_scaling_50_3329804.log\r\nbig_run train_lam_action_space_scaling_50_3331286.log\r\nbig-runs train_lam_action_space_scaling_6_3318549.log\r\ncausal train_lam_action_space_scaling_6_3320178.log\r\ncoinrun train_lam_action_space_scaling_6_3321528.log\r\nmaskgit train_lam_action_space_scaling_6_3329790.log\r\nmaskgit-maskprob-fix train_lam_action_space_scaling_6_3329805.log\r\npreprocess train_lam_action_space_scaling_6_3331287.log\r\ntrain_dyn_causal_180M_3372931.log train_lam_action_space_scaling_8_3318550.log\r\ntrain_dyn_causal_180M_3372963.log train_lam_action_space_scaling_8_3329791.log\r\ntrain_dyn_causal_180M_3372969.log train_lam_action_space_scaling_8_3329806.log\r\ntrain_dyn_causal_180M_3373107.log train_lam_action_space_scaling_8_3331288.log\r\ntrain_dyn_causal_255M_3372932.log train_lam_minecraft_overfit_sample_3309655.log\r\ntrain_dyn_causal_255M_3372970.log train_lam_model_size_scaling_38M_3317098.log\r\ntrain_dyn_causal_255M_3373108.log train_lam_model_size_scaling_38M_3317115.log\r\ntrain_dyn_causal_356M_3372934.log train_lam_model_size_scaling_38M_3317231.log\r\ntrain_dyn_causal_356M_3372971.log train_tokenizer_batch_size_scaling_16_node_3321526.log\r\ntrain_dyn_causal_356M_3373109.log train_tokenizer_batch_size_scaling_1_node_3318551.log\r\ntrain_dyn_causal_500M_3372936.log train_tokenizer_batch_size_scaling_2_node_3318552.log\r\ntrain_dyn_causal_500M_3372972.log train_tokenizer_batch_size_scaling_2_node_3330806.log\r\ntrain_dyn_causal_500M_3373110.log train_tokenizer_batch_size_scaling_2_node_3330848.log\r\ntrain_dyn_new_arch-bugfixed-spatial-shift_3359343.log train_tokenizer_batch_size_scaling_2_node_3331282.log\r\ntrain_dyn_new_arch-bugfixed-temporal-shift_3359349.log train_tokenizer_batch_size_scaling_4_node_3318553.log\r\ntrain_dyn_yolorun_3333026.log train_tokenizer_batch_size_scaling_4_node_3320175.log\r\ntrain_dyn_yolorun_3333448.log train_tokenizer_batch_size_scaling_4_node_3321524.log\r\ntrain_dyn_yolorun_3335345.log train_tokenizer_batch_size_scaling_8_node_3320176.log\r\ntrain_dyn_yolorun_3335362.log train_tokenizer_batch_size_scaling_8_node_3321525.log\r\ntrain_dyn_yolorun_3348592.log train_tokenizer_minecraft_overfit_sample_3309656.log\r\ntrain_dyn_yolorun_new_arch_3351743.log train_tokenizer_model_size_scaling_127M_3317233.log\r\ntrain_dyn_yolorun_new_arch_3352103.log train_tokenizer_model_size_scaling_127M_3318554.log\r\ntrain_dyn_yolorun_new_arch_3352115.log train_tokenizer_model_size_scaling_140M_3313562.log\r\ntrain_dyn_yolorun_new_arch_3358457.log train_tokenizer_model_size_scaling_140M_3316019.log\r\ntrain_lam_action_space_scaling_10_3320179.log train_tokenizer_model_size_scaling_200M_3313563.log\r\ntrain_lam_action_space_scaling_10_3321529.log train_tokenizer_model_size_scaling_200M_3316020.log\r\ntrain_lam_action_space_scaling_10_3329786.log train_tokenizer_model_size_scaling_227M_3317234.log\r\ntrain_lam_action_space_scaling_10_3329801.log train_tokenizer_model_size_scaling_227M_3318555.log\r\ntrain_lam_action_space_scaling_10_3331283.log train_tokenizer_model_size_scaling_227M_3320173.log\r\ntrain_lam_action_space_scaling_12_3318546.log train_tokenizer_model_size_scaling_227M_3321523.log\r\ntrain_lam_action_space_scaling_12_3320177.log train_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_lam_action_space_scaling_12_3321527.log train_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_lam_action_space_scaling_12_3329787.log train_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_lam_action_space_scaling_12_3329802.log train_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_lam_action_space_scaling_12_3331284.log train_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_lam_action_space_scaling_20_3318547.log train_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_lam_action_space_scaling_20_3329788.log train_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_lam_action_space_scaling_20_3329803.log train_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_lam_action_space_scaling_20_3331285.log train_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_lam_action_space_scaling_50_3320180.log train_tokenizer_model_size_scaling_80M_3316026.log\r\ntrain_lam_action_space_scaling_50_3329789.log yoloruns\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +2032,9197786,"TERMINAL",0,0,"cd causal/",,terminal_command +2033,9197822,"TERMINAL",0,0,"]633;E;2025-08-14 13:10:50 cd causal/;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal]633;D;0",,terminal_output +2034,9198086,"TERMINAL",0,0,"ls",,terminal_command +2035,9198097,"TERMINAL",0,0,"]633;E;2025-08-14 13:10:51 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Cdynamics-cotraining\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal]633;D;0",,terminal_output +2036,9199456,"TERMINAL",0,0,"cd dynamics-cotraining/",,terminal_command +2037,9199682,"TERMINAL",0,0,"ls",,terminal_command +2038,9199755,"TERMINAL",0,0,"]633;E;2025-08-14 13:10:52 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +2039,9199832,"TERMINAL",0,0,"train_dynamics_causal_2_node_3373407.log train_dynamics_causal_8_node_3388140.log train_dynamics_causal_8_node_3412397.log\r\ntrain_dynamics_causal_2_node_3373407.log_bak train_dynamics_causal_8_node_3389928.log train_dynamics_causal_8_node_3412399.log\r\ntrain_dynamics_causal_2_node_3388135.log train_dynamics_causal_8_node_3390458.log train_dynamics_causal_8_node_3414710.log\r\ntrain_dynamics_causal_2_node_3388147.log train_dynamics_causal_8_node_3393060.log train_dynamics_causal_8_node_3415128.log\r\ntrain_dynamics_causal_2_node_3389801.log train_dynamics_causal_8_node_3393061.log train_dynamics_causal_8_node_3415137.log\r\ntrain_dynamics_causal_2_node_3393065.log train_dynamics_causal_8_node_3393066.log train_dynamics_causal_8_node_3417223.log\r\ntrain_dynamics_causal_2_node_dev_3416774.log train_dynamics_causal_8_node_3412343.log train_dynamics_causal_8_node_3417224.log\r\ntrain_dynamics_causal_2_node_dev_3416819.log train_dynamics_causal_8_node_3412349.log train_dynamics_causal_8_node_3418831.log\r\ntrain_dynamics_causal_8_node_3373408.log train_dynamics_causal_8_node_3412356.log train_dynamics_causal_8_node_3418832.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining]633;D;0",,terminal_output +2040,9224374,"TERMINAL",0,0,"dev",,terminal_command +2041,9235196,"TERMINAL",0,0,"bash",,terminal_focus +2042,9247383,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +2043,9251074,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2115,0,"",shellscript,selection_mouse +2044,9251075,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2114,0,"",shellscript,selection_command +2045,9251967,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2060,0,"",shellscript,selection_mouse +2046,9253329,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2060,4,"",shellscript,content +2047,9254155,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2060,0,"7",shellscript,content +2048,9254156,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2061,0,"",shellscript,selection_keyboard +2049,9254786,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2061,0,"6",shellscript,content +2050,9254787,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2062,0,"",shellscript,selection_keyboard +2051,9255291,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2061,1,"",shellscript,content +2052,9255401,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2060,1,"",shellscript,content +2053,9255894,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2060,0,"6",shellscript,content +2054,9255895,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2061,0,"",shellscript,selection_keyboard +2055,9255976,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2061,0," ",shellscript,content +2056,9255977,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2062,0,"",shellscript,selection_keyboard +2057,9256229,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2061,0,"",shellscript,selection_command +2058,9256866,"TERMINAL",0,0,"bash",,terminal_focus +2059,9261936,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",,terminal_command +2060,9262012,"TERMINAL",0,0,"]633;E;2025-08-14 13:11:55 sbatch slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CSubmitted batch job 3422101\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +2061,9267138,"TERMINAL",0,0,"scancel 3422101",,terminal_command +2062,9271522,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",0,0,"",shellscript,tab +2063,9272769,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",460,0,"",shellscript,selection_mouse +2064,9275052,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +2065,9275934,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",481,0,"",shellscript,selection_mouse +2066,9276343,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",436,0,"",shellscript,selection_mouse +2067,9276894,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",468,0,"\n#SBATCH --reservation=llmtum",shellscript,content +2068,9276899,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",469,0,"",shellscript,selection_command +2069,9289519,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",,terminal_command +2070,9289542,"TERMINAL",0,0,"]633;E;2025-08-14 13:12:22 sbatch slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CSubmitted batch job 3422103\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +2071,9292215,"TERMINAL",0,0,"queue",,terminal_command +2072,9292279,"TERMINAL",0,0,"]633;E;2025-08-14 13:12:25 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 13:12:25 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-19:12:24\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 23:23:29\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422103 accelerat train_dy tum_cte0 R\t0:02\t 8 hkn[0521-0528]3422008 accelerat train_dy tum_cte0 R59:15\t 1 hkn07133421982 accelerat train_to tum_cte0 R 1:11:15\t 4 hkn[0724,0726,0728,0731]3421973 accelerat train_to tum_cte0 R 1:11:24\t 2 hkn[0722-0723]3421964 accelerat train_to tum_cte0 R 1:11:39\t 1 hkn0720",,terminal_output +2073,9293321,"TERMINAL",0,0,"6530366540",,terminal_output +2074,9294404,"TERMINAL",0,0,"76147761",,terminal_output +2075,9295447,"TERMINAL",0,0,"87258872",,terminal_output +2076,9296540,"TERMINAL",0,0,"98369983",,terminal_output +2077,9296677,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +2078,9298288,"TERMINAL",0,0,"logs",,terminal_command +2079,9298779,"TERMINAL",0,0,"ls",,terminal_command +2080,9298824,"TERMINAL",0,0,"]633;E;2025-08-14 13:12:31 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +2081,9299088,"TERMINAL",0,0,"atari train_lam_action_space_scaling_50_3329804.log\r\nbig_run train_lam_action_space_scaling_50_3331286.log\r\nbig-runs train_lam_action_space_scaling_6_3318549.log\r\ncausal train_lam_action_space_scaling_6_3320178.log\r\ncoinrun train_lam_action_space_scaling_6_3321528.log\r\nmaskgit train_lam_action_space_scaling_6_3329790.log\r\nmaskgit-maskprob-fix train_lam_action_space_scaling_6_3329805.log\r\npreprocess train_lam_action_space_scaling_6_3331287.log\r\ntrain_dyn_causal_180M_3372931.log train_lam_action_space_scaling_8_3318550.log\r\ntrain_dyn_causal_180M_3372963.log train_lam_action_space_scaling_8_3329791.log\r\ntrain_dyn_causal_180M_3372969.log train_lam_action_space_scaling_8_3329806.log\r\ntrain_dyn_causal_180M_3373107.log train_lam_action_space_scaling_8_3331288.log\r\ntrain_dyn_causal_255M_3372932.log train_lam_minecraft_overfit_sample_3309655.log\r\ntrain_dyn_causal_255M_3372970.log train_lam_model_size_scaling_38M_3317098.log\r\ntrain_dyn_causal_255M_3373108.log train_lam_model_size_scaling_38M_3317115.log\r\ntrain_dyn_causal_356M_3372934.log train_lam_model_size_scaling_38M_3317231.log\r\ntrain_dyn_causal_356M_3372971.log train_tokenizer_batch_size_scaling_16_node_3321526.log\r\ntrain_dyn_causal_356M_3373109.log train_tokenizer_batch_size_scaling_1_node_3318551.log\r\ntrain_dyn_causal_500M_3372936.log train_tokenizer_batch_size_scaling_2_node_3318552.log\r\ntrain_dyn_causal_500M_3372972.log train_tokenizer_batch_size_scaling_2_node_3330806.log\r\ntrain_dyn_causal_500M_3373110.log train_tokenizer_batch_size_scaling_2_node_3330848.log\r\ntrain_dyn_new_arch-bugfixed-spatial-shift_3359343.log train_tokenizer_batch_size_scaling_2_node_3331282.log\r\ntrain_dyn_new_arch-bugfixed-temporal-shift_3359349.log train_tokenizer_batch_size_scaling_4_node_3318553.log\r\ntrain_dyn_yolorun_3333026.log train_tokenizer_batch_size_scaling_4_node_3320175.log\r\ntrain_dyn_yolorun_3333448.log train_tokenizer_batch_size_scaling_4_node_3321524.log\r\ntrain_dyn_yolorun_3335345.log train_tokenizer_batch_size_scaling_8_node_3320176.log\r\ntrain_dyn_yolorun_3335362.log train_tokenizer_batch_size_scaling_8_node_3321525.log\r\ntrain_dyn_yolorun_3348592.log train_tokenizer_minecraft_overfit_sample_3309656.log\r\ntrain_dyn_yolorun_new_arch_3351743.log train_tokenizer_model_size_scaling_127M_3317233.log\r\ntrain_dyn_yolorun_new_arch_3352103.log train_tokenizer_model_size_scaling_127M_3318554.log\r\ntrain_dyn_yolorun_new_arch_3352115.log train_tokenizer_model_size_scaling_140M_3313562.log\r\ntrain_dyn_yolorun_new_arch_3358457.log train_tokenizer_model_size_scaling_140M_3316019.log\r\ntrain_lam_action_space_scaling_10_3320179.log train_tokenizer_model_size_scaling_200M_3313563.log\r\ntrain_lam_action_space_scaling_10_3321529.log train_tokenizer_model_size_scaling_200M_3316020.log\r\ntrain_lam_action_space_scaling_10_3329786.log train_tokenizer_model_size_scaling_227M_3317234.log\r\ntrain_lam_action_space_scaling_10_3329801.log train_tokenizer_model_size_scaling_227M_3318555.log\r\ntrain_lam_action_space_scaling_10_3331283.log train_tokenizer_model_size_scaling_227M_3320173.log\r\ntrain_lam_action_space_scaling_12_3318546.log train_tokenizer_model_size_scaling_227M_3321523.log\r\ntrain_lam_action_space_scaling_12_3320177.log train_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_lam_action_space_scaling_12_3321527.log train_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_lam_action_space_scaling_12_3329787.log train_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_lam_action_space_scaling_12_3329802.log train_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_lam_action_space_scaling_12_3331284.log train_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_lam_action_space_scaling_20_3318547.log train_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_lam_action_space_scaling_20_3329788.log train_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_lam_action_space_scaling_20_3329803.log train_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_lam_action_space_scaling_20_3331285.log train_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_lam_action_space_scaling_50_3320180.log train_tokenizer_model_size_scaling_80M_3316026.log\r\ntrain_lam_action_space_scaling_50_3329789.log yoloruns\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +2082,9303383,"TERMINAL",0,0,"cd causal/",,terminal_command +2083,9303414,"TERMINAL",0,0,"]633;E;2025-08-14 13:12:36 cd causal/;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal]633;D;0",,terminal_output +2084,9303742,"TERMINAL",0,0,"ls",,terminal_command +2085,9303820,"TERMINAL",0,0,"]633;E;2025-08-14 13:12:36 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Cdynamics-cotraining\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal]633;D;0",,terminal_output +2086,9305117,"TERMINAL",0,0,"cd dynamics-cotraining/",,terminal_command +2087,9305413,"TERMINAL",0,0,"ls",,terminal_command +2088,9310020,"TERMINAL",0,0,"ls",,terminal_command +2089,9310067,"TERMINAL",0,0,"]633;E;2025-08-14 13:12:43 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +2090,9310208,"TERMINAL",0,0,"train_dynamics_causal_2_node_3373407.log train_dynamics_causal_8_node_3388140.log train_dynamics_causal_8_node_3412397.log\r\ntrain_dynamics_causal_2_node_3373407.log_bak train_dynamics_causal_8_node_3389928.log train_dynamics_causal_8_node_3412399.log\r\ntrain_dynamics_causal_2_node_3388135.log train_dynamics_causal_8_node_3390458.log train_dynamics_causal_8_node_3414710.log\r\ntrain_dynamics_causal_2_node_3388147.log train_dynamics_causal_8_node_3393060.log train_dynamics_causal_8_node_3415128.log\r\ntrain_dynamics_causal_2_node_3389801.log train_dynamics_causal_8_node_3393061.log train_dynamics_causal_8_node_3415137.log\r\ntrain_dynamics_causal_2_node_3393065.log train_dynamics_causal_8_node_3393066.log train_dynamics_causal_8_node_3417223.log\r\ntrain_dynamics_causal_2_node_dev_3416774.log train_dynamics_causal_8_node_3412343.log train_dynamics_causal_8_node_3417224.log\r\ntrain_dynamics_causal_2_node_dev_3416819.log train_dynamics_causal_8_node_3412349.log train_dynamics_causal_8_node_3418831.log\r\ntrain_dynamics_causal_8_node_3373408.log train_dynamics_causal_8_node_3412356.log train_dynamics_causal_8_node_3418832.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining]633;D;0",,terminal_output +2091,9319827,"TERMINAL",0,0,"queue",,terminal_command +2092,9319912,"TERMINAL",0,0,"]633;E;2025-08-14 13:12:53 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 13:12:53 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-19:12:52\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 23:23:57\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422103 accelerat train_dy tum_cte0 R\t0:30\t 8 hkn[0521-0528]3422008 accelerat train_dy tum_cte0 R59:43\t 1 hkn07133421982 accelerat train_to tum_cte0 R 1:11:43\t 4 hkn[0724,0726,0728,0731]3421973 accelerat train_to tum_cte0 R 1:11:52\t 2 hkn[0722-0723]3421964 accelerat train_to tum_cte0 R 1:12:07\t 1 hkn0720",,terminal_output +2093,9320964,"TERMINAL",0,0,"43814438",,terminal_output +2094,9322069,"TERMINAL",0,0,"54925549",,terminal_output +2095,9323087,"TERMINAL",0,0,"654:00366510",,terminal_output +2096,9324099,"TERMINAL",0,0,"76147761",,terminal_output +2097,9324600,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",219,0,"",shellscript,selection_mouse +2098,9325122,"TERMINAL",0,0,"87258872",,terminal_output +2099,9326151,"TERMINAL",0,0,"bash",,terminal_focus +2100,9326165,"TERMINAL",0,0,"98369983",,terminal_output +2101,9327218,"TERMINAL",0,0,"3:00947505094",,terminal_output +2102,9328292,"TERMINAL",0,0,"13:0058112:005",,terminal_output +2103,9329343,"TERMINAL",0,0,"21692216",,terminal_output +2104,9329624,"TERMINAL",0,0,"logs",,terminal_command +2105,9329653,"TERMINAL",0,0,"]633;E;2025-08-14 13:13:02 logs;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +2106,9330045,"TERMINAL",0,0,"ls",,terminal_command +2107,9330083,"TERMINAL",0,0,"]633;E;2025-08-14 13:13:03 ls;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;Catari train_lam_action_space_scaling_50_3329804.log\r\nbig_run train_lam_action_space_scaling_50_3331286.log\r\nbig-runs train_lam_action_space_scaling_6_3318549.log\r\ncausal train_lam_action_space_scaling_6_3320178.log\r\ncoinrun train_lam_action_space_scaling_6_3321528.log\r\nmaskgit train_lam_action_space_scaling_6_3329790.log\r\nmaskgit-maskprob-fix train_lam_action_space_scaling_6_3329805.log\r\npreprocess train_lam_action_space_scaling_6_3331287.log\r\ntrain_dyn_causal_180M_3372931.log train_lam_action_space_scaling_8_3318550.log\r\ntrain_dyn_causal_180M_3372963.log train_lam_action_space_scaling_8_3329791.log\r\ntrain_dyn_causal_180M_3372969.log train_lam_action_space_scaling_8_3329806.log\r\ntrain_dyn_causal_180M_3373107.log train_lam_action_space_scaling_8_3331288.log\r\ntrain_dyn_causal_255M_3372932.log train_lam_minecraft_overfit_sample_3309655.log\r\ntrain_dyn_causal_255M_3372970.log train_lam_model_size_scaling_38M_3317098.log\r\ntrain_dyn_causal_255M_3373108.log train_lam_model_size_scaling_38M_3317115.log\r\ntrain_dyn_causal_356M_3372934.log train_lam_model_size_scaling_38M_3317231.log\r\ntrain_dyn_causal_356M_3372971.log train_tokenizer_batch_size_scaling_16_node_3321526.log\r\ntrain_dyn_causal_356M_3373109.log train_tokenizer_batch_size_scaling_1_node_3318551.log\r\ntrain_dyn_causal_500M_3372936.log train_tokenizer_batch_size_scaling_2_node_3318552.log\r\ntrain_dyn_causal_500M_3372972.log train_tokenizer_batch_size_scaling_2_node_3330806.log\r\ntrain_dyn_causal_500M_3373110.log train_tokenizer_batch_size_scaling_2_node_3330848.log\r\ntrain_dyn_new_arch-bugfixed-spatial-shift_3359343.log train_tokenizer_batch_size_scaling_2_node_3331282.log\r\ntrain_dyn_new_arch-bugfixed-temporal-shift_3359349.log train_tokenizer_batch_size_scaling_4_node_3318553.log\r\ntrain_dyn_yolorun_3333026.log train_tokenizer_batch_size_scaling_4_node_3320175.log\r\ntrain_dyn_yolorun_3333448.log train_tokenizer_batch_size_scaling_4_node_3321524.log\r\ntrain_dyn_yolorun_3335345.log train_tokenizer_batch_size_scaling_8_node_3320176.log\r\ntrain_dyn_yolorun_3335362.log train_tokenizer_batch_size_scaling_8_node_3321525.log\r\ntrain_dyn_yolorun_3348592.log train_tokenizer_minecraft_overfit_sample_3309656.log\r\ntrain_dyn_yolorun_new_arch_3351743.log train_tokenizer_model_size_scaling_127M_3317233.log\r\ntrain_dyn_yolorun_new_arch_3352103.log train_tokenizer_model_size_scaling_127M_3318554.log\r\ntrain_dyn_yolorun_new_arch_3352115.log train_tokenizer_model_size_scaling_140M_3313562.log\r\ntrain_dyn_yolorun_new_arch_3358457.log train_tokenizer_model_size_scaling_140M_3316019.log\r\ntrain_lam_action_space_scaling_10_3320179.log train_tokenizer_model_size_scaling_200M_3313563.log\r\ntrain_lam_action_space_scaling_10_3321529.log train_tokenizer_model_size_scaling_200M_3316020.log\r\ntrain_lam_action_space_scaling_10_3329786.log train_tokenizer_model_size_scaling_227M_3317234.log\r\ntrain_lam_action_space_scaling_10_3329801.log train_tokenizer_model_size_scaling_227M_3318555.log\r\ntrain_lam_action_space_scaling_10_3331283.log train_tokenizer_model_size_scaling_227M_3320173.log\r\ntrain_lam_action_space_scaling_12_3318546.log train_tokenizer_model_size_scaling_227M_3321523.log\r\ntrain_lam_action_space_scaling_12_3320177.log train_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_lam_action_space_scaling_12_3321527.log train_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_lam_action_space_scaling_12_3329787.log train_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_lam_action_space_scaling_12_3329802.log train_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_lam_action_space_scaling_12_3331284.log train_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_lam_action_space_scaling_20_3318547.log train_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_lam_action_space_scaling_20_3329788.log train_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_lam_action_space_scaling_20_3329803.log train_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_lam_action_space_scaling_20_3331285.log train_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_lam_action_space_scaling_50_3320180.log train_tokenizer_model_size_scaling_80M_3316026.log\r\ntrain_lam_action_space_scaling_50_3329789.log yoloruns\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +2108,9330445,"TERMINAL",0,0,"327403327",,terminal_output +2109,9331406,"TERMINAL",0,0,"43814438",,terminal_output +2110,9332470,"TERMINAL",0,0,"cd causal/",,terminal_command +2111,9332471,"TERMINAL",0,0,"54925549",,terminal_output +2112,9332669,"TERMINAL",0,0,"ls",,terminal_command +2113,9332674,"TERMINAL",0,0,"]633;E;2025-08-14 13:13:05 ls;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;Cdynamics-cotraining\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal]633;D;0",,terminal_output +2114,9333490,"TERMINAL",0,0,"6510366520",,terminal_output +2115,9334281,"TERMINAL",0,0,"cd dynamics-cotraining/",,terminal_command +2116,9334532,"TERMINAL",0,0,"76147761",,terminal_output +2117,9334666,"TERMINAL",0,0,"ls",,terminal_command +2118,9334703,"TERMINAL",0,0,"]633;E;2025-08-14 13:13:07 ls;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;Ctrain_dynamics_causal_2_node_3373407.log train_dynamics_causal_8_node_3389928.log train_dynamics_causal_8_node_3414710.log\r\ntrain_dynamics_causal_2_node_3373407.log_bak train_dynamics_causal_8_node_3390458.log train_dynamics_causal_8_node_3415128.log\r\ntrain_dynamics_causal_2_node_3388135.log train_dynamics_causal_8_node_3393060.log train_dynamics_causal_8_node_3415137.log\r\ntrain_dynamics_causal_2_node_3388147.log train_dynamics_causal_8_node_3393061.log train_dynamics_causal_8_node_3417223.log\r\ntrain_dynamics_causal_2_node_3389801.log train_dynamics_causal_8_node_3393066.log train_dynamics_causal_8_node_3417224.log\r\ntrain_dynamics_causal_2_node_3393065.log train_dynamics_causal_8_node_3412343.log train_dynamics_causal_8_node_3418831.log\r\ntrain_dynamics_causal_2_node_dev_3416774.log train_dynamics_causal_8_node_3412349.log train_dynamics_causal_8_node_3418832.log\r\ntrain_dynamics_causal_2_node_dev_3416819.log train_dynamics_causal_8_node_3412356.log train_dynamics_causal_8_node_3422103.log\r\ntrain_dynamics_causal_8_node_3373408.log train_dynamics_causal_8_node_3412397.log\r\ntrain_dynamics_causal_8_node_3388140.log train_dynamics_causal_8_node_3412399.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining]633;D;0",,terminal_output +2119,9335602,"TERMINAL",0,0,"87258872",,terminal_output +2120,9336616,"TERMINAL",0,0,"98369983",,terminal_output +2121,9337717,"TERMINAL",0,0,"109471:00:002:0094",,terminal_output +2122,9338700,"TERMINAL",0,0,"1105811105",,terminal_output +2123,9339868,"TERMINAL",0,0,"21692216",,terminal_output +2124,9340890,"TERMINAL",0,0,"327503327",,terminal_output +2125,9341885,"TERMINAL",0,0,"44925549",,terminal_output +2126,9342863,"TERMINAL",0,0,"6520366530",,terminal_output +2127,9343997,"TERMINAL",0,0,"76147761",,terminal_output +2128,9344999,"TERMINAL",0,0,"87258872",,terminal_output +2129,9346043,"TERMINAL",0,0,"98369983",,terminal_output +2130,9347136,"TERMINAL",0,0,"20947101094",,terminal_output +2131,9348161,"TERMINAL",0,0,"1205811205",,terminal_output +2132,9349111,"TERMINAL",0,0,"21692216",,terminal_output +2133,9350220,"TERMINAL",0,0,"3271:003327",,terminal_output +2134,9350474,"TERMINAL",0,0,"tail -f train_dynamics_causal_8_node_3422103.log",,terminal_command +2135,9350549,"TERMINAL",0,0,"]633;E;2025-08-14 13:13:23 tail -f train_dynamics_causal_8_node_3422103.log ;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;CSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0521-0528]\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +2136,9351204,"TERMINAL",0,0,"43814438",,terminal_output +2137,9352237,"TERMINAL",0,0,"54925549",,terminal_output +2138,9353277,"TERMINAL",0,0,"6530366540",,terminal_output +2139,9354317,"TERMINAL",0,0,"76147761",,terminal_output +2140,9355362,"TERMINAL",0,0,"87258872",,terminal_output +2141,9355530,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +2142,9356457,"TERMINAL",0,0,"98369983",,terminal_output +2143,9356757,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.19.11\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250814_131328-3422103\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run dynamics-causal-8-node-darkness-filter-3422103\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/3422103\r\n",,terminal_output +2144,9357453,"TERMINAL",0,0,"30947202094",,terminal_output +2145,9358496,"TERMINAL",0,0,"1305811305",,terminal_output +2146,9359733,"TERMINAL",0,0,"21692216",,terminal_output +2147,9360757,"TERMINAL",0,0,"327103327",,terminal_output +2148,9361627,"TERMINAL",0,0,"43814438",,terminal_output +2149,9362782,"TERMINAL",0,0,"54925549",,terminal_output +2150,9363831,"TERMINAL",0,0,"6540366550",,terminal_output +2151,9364803,"TERMINAL",0,0,"76147761",,terminal_output +2152,9365820,"TERMINAL",0,0,"87258872",,terminal_output +2153,9366852,"TERMINAL",0,0,"9947303094",,terminal_output +2154,9367927,"TERMINAL",0,0,"41405811405",,terminal_output +2155,9368946,"TERMINAL",0,0,"21692216",,terminal_output +2156,9370177,"TERMINAL",0,0,"327203327",,terminal_output +2157,9371041,"TERMINAL",0,0,"43814438",,terminal_output +2158,9372079,"TERMINAL",0,0,"54925549",,terminal_output +2159,9373128,"TERMINAL",0,0,"655036653:00",,terminal_output +2160,9374164,"TERMINAL",0,0,"76147761",,terminal_output +2161,9375208,"TERMINAL",0,0,"87258872",,terminal_output +2162,9376260,"TERMINAL",0,0,"98369983",,terminal_output +2163,9377304,"TERMINAL",0,0,"50947404094",,terminal_output +2164,9378344,"TERMINAL",0,0,"1505811505",,terminal_output +2165,9379392,"TERMINAL",0,0,"21692216",,terminal_output +2166,9380433,"TERMINAL",0,0,"327303327",,terminal_output +2167,9381485,"TERMINAL",0,0,"43814438",,terminal_output +2168,9382566,"TERMINAL",0,0,"54925549",,terminal_output +2169,9383591,"TERMINAL",0,0,"655:00366510",,terminal_output +2170,9384718,"TERMINAL",0,0,"76147761",,terminal_output +2171,9385676,"TERMINAL",0,0,"87258872",,terminal_output +2172,9386749,"TERMINAL",0,0,"98369983",,terminal_output +2173,9387753,"TERMINAL",0,0,"4:00947505094",,terminal_output +2174,9388812,"TERMINAL",0,0,"14:0058113:005",,terminal_output +2175,9389939,"TERMINAL",0,0,"227403327",,terminal_output +2176,9391096,"TERMINAL",0,0,"43814438",,terminal_output +2177,9392089,"TERMINAL",0,0,"54925549",,terminal_output +2178,9392959,"TERMINAL",0,0,"6510366520",,terminal_output +2179,9394009,"TERMINAL",0,0,"76147761",,terminal_output +2180,9395059,"TERMINAL",0,0,"87258872",,terminal_output +2181,9396132,"TERMINAL",0,0,"98369983",,terminal_output +2182,9397138,"TERMINAL",0,0,"109471:003:0094",,terminal_output +2183,9398181,"TERMINAL",0,0,"1105811105",,terminal_output +2184,9399222,"TERMINAL",0,0,"21692216",,terminal_output +2185,9400270,"TERMINAL",0,0,"327503327",,terminal_output +2186,9401316,"TERMINAL",0,0,"43814438",,terminal_output +2187,9402366,"TERMINAL",0,0,"54925549",,terminal_output +2188,9403412,"TERMINAL",0,0,"6520366530",,terminal_output +2189,9404481,"TERMINAL",0,0,"76147761",,terminal_output +2190,9405512,"TERMINAL",0,0,"87258872",,terminal_output +2191,9406558,"TERMINAL",0,0,"98369983",,terminal_output +2192,9407655,"TERMINAL",0,0,"20947101094",,terminal_output +2193,9408680,"TERMINAL",0,0,"1205811205",,terminal_output +2194,9409709,"TERMINAL",0,0,"21692216",,terminal_output +2195,9410760,"TERMINAL",0,0,"3272:003327",,terminal_output +2196,9411812,"TERMINAL",0,0,"43814438",,terminal_output +2197,9412845,"TERMINAL",0,0,"5530366540",,terminal_output +2198,9413900,"TERMINAL",0,0,"76147761",,terminal_output +2199,9414943,"TERMINAL",0,0,"87258872",,terminal_output +2200,9415159,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining]633;D;130",,terminal_output +2201,9416052,"TERMINAL",0,0,"98369983",,terminal_output +2202,9417038,"TERMINAL",0,0,"30947202094",,terminal_output +2203,9418116,"TERMINAL",0,0,"1305811305",,terminal_output +2204,9419331,"TERMINAL",0,0,"21692216",,terminal_output +2205,9420172,"TERMINAL",0,0,"327103327",,terminal_output +2206,9421219,"TERMINAL",0,0,"43814438",,terminal_output +2207,9422262,"TERMINAL",0,0,"54925549",,terminal_output +2208,9423311,"TERMINAL",0,0,"6540366550",,terminal_output +2209,9424450,"TERMINAL",0,0,"76147761",,terminal_output +2210,9425436,"TERMINAL",0,0,"87258872",,terminal_output +2211,9425542,"TERMINAL",0,0,"watch",,terminal_focus +2212,9426116,"TERMINAL",0,0,"bash",,terminal_focus +2213,9426500,"TERMINAL",0,0,"98369983",,terminal_output +2214,9427495,"TERMINAL",0,0,"40947303094",,terminal_output +2215,9427571,"TERMINAL",0,0,"ls",,terminal_command +2216,9427623,"TERMINAL",0,0,"]633;E;2025-08-14 13:14:40 ls;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C",,terminal_output +2217,9427757,"TERMINAL",0,0,"train_dynamics_causal_2_node_3373407.log train_dynamics_causal_8_node_3389928.log train_dynamics_causal_8_node_3414710.log\r\ntrain_dynamics_causal_2_node_3373407.log_bak train_dynamics_causal_8_node_3390458.log train_dynamics_causal_8_node_3415128.log\r\ntrain_dynamics_causal_2_node_3388135.log train_dynamics_causal_8_node_3393060.log train_dynamics_causal_8_node_3415137.log\r\ntrain_dynamics_causal_2_node_3388147.log train_dynamics_causal_8_node_3393061.log train_dynamics_causal_8_node_3417223.log\r\ntrain_dynamics_causal_2_node_3389801.log train_dynamics_causal_8_node_3393066.log train_dynamics_causal_8_node_3417224.log\r\ntrain_dynamics_causal_2_node_3393065.log train_dynamics_causal_8_node_3412343.log train_dynamics_causal_8_node_3418831.log\r\ntrain_dynamics_causal_2_node_dev_3416774.log train_dynamics_causal_8_node_3412349.log train_dynamics_causal_8_node_3418832.log\r\ntrain_dynamics_causal_2_node_dev_3416819.log train_dynamics_causal_8_node_3412356.log train_dynamics_causal_8_node_3422103.log\r\ntrain_dynamics_causal_8_node_3373408.log train_dynamics_causal_8_node_3412397.log\r\ntrain_dynamics_causal_8_node_3388140.log train_dynamics_causal_8_node_3412399.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining]633;D;0",,terminal_output +2218,9428525,"TERMINAL",0,0,"1405811405",,terminal_output +2219,9429573,"TERMINAL",0,0,"21692216",,terminal_output +2220,9430614,"TERMINAL",0,0,"327203327",,terminal_output +2221,9431651,"TERMINAL",0,0,"43814438",,terminal_output +2222,9432700,"TERMINAL",0,0,"54925549",,terminal_output +2223,9433742,"TERMINAL",0,0,"655036654:00",,terminal_output +2224,9434792,"TERMINAL",0,0,"76147761",,terminal_output +2225,9435826,"TERMINAL",0,0,"88369983",,terminal_output +2226,9436859,"TERMINAL",0,0,"50947404094",,terminal_output +2227,9437909,"TERMINAL",0,0,"1505811505",,terminal_output +2228,9439000,"TERMINAL",0,0,"21692216",,terminal_output +2229,9440014,"TERMINAL",0,0,"327303327",,terminal_output +2230,9441022,"TERMINAL",0,0,"43814438",,terminal_output +2231,9442067,"TERMINAL",0,0,"54925549",,terminal_output +2232,9443187,"TERMINAL",0,0,"656:00366510",,terminal_output +2233,9444211,"TERMINAL",0,0,"76147761",,terminal_output +2234,9445204,"TERMINAL",0,0,"87258872",,terminal_output +2235,9446262,"TERMINAL",0,0,"98369983",,terminal_output +2236,9447264,"TERMINAL",0,0,"5:00947505094",,terminal_output +2237,9448304,"TERMINAL",0,0,"15:0058114:005",,terminal_output +2238,9448440,"TERMINAL",0,0,"tail -f train_dynamics_causal_8_node_https://wandb.ai/instant-uv/jafar/runs/3422103^Cog ",,terminal_command +2239,9448477,"TERMINAL",0,0,"^C\r\n\r[?2004l\r[?2004h[?2004l\r\r\n]633;E;;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining]633;D",,terminal_output +2240,9449437,"TERMINAL",0,0,"21692216",,terminal_output +2241,9450460,"TERMINAL",0,0,"327403327",,terminal_output +2242,9451421,"TERMINAL",0,0,"43814438",,terminal_output +2243,9452461,"TERMINAL",0,0,"54925549",,terminal_output +2244,9453502,"TERMINAL",0,0,"6510366520",,terminal_output +2245,9454557,"TERMINAL",0,0,"76147761",,terminal_output +2246,9455598,"TERMINAL",0,0,"87258872",,terminal_output +2247,9456654,"TERMINAL",0,0,"98369983",,terminal_output +2248,9457192,"TERMINAL",0,0,"tail -f train_dynamics_causal_8_node_3422097.log",,terminal_command +2249,9457684,"TERMINAL",0,0,"109472:004:0094",,terminal_output +2250,9458783,"TERMINAL",0,0,"1105811105",,terminal_output +2251,9459735,"TERMINAL",0,0,"21692216",,terminal_output +2252,9460808,"TERMINAL",0,0,"327503327",,terminal_output +2253,9461830,"TERMINAL",0,0,"44925549",,terminal_output +2254,9462850,"TERMINAL",0,0,"6520366530",,terminal_output +2255,9464080,"TERMINAL",0,0,"76147761",,terminal_output +2256,9464928,"TERMINAL",0,0,"87258872",,terminal_output +2257,9466009,"TERMINAL",0,0,"98369983",,terminal_output +2258,9467152,"TERMINAL",0,0,"20947101094",,terminal_output +2259,9468073,"TERMINAL",0,0,"1205811205",,terminal_output +2260,9469404,"TERMINAL",0,0,"21692216",,terminal_output +2261,9470149,"TERMINAL",0,0,"3273:003327",,terminal_output +2262,9471220,"TERMINAL",0,0,"43814438",,terminal_output +2263,9472256,"TERMINAL",0,0,"54925549",,terminal_output +2264,9473301,"TERMINAL",0,0,"6530366540",,terminal_output +2265,9474372,"TERMINAL",0,0,"76147761",,terminal_output +2266,9475546,"TERMINAL",0,0,"87258872",,terminal_output +2267,9476468,"TERMINAL",0,0,"98369983",,terminal_output +2268,9477500,"TERMINAL",0,0,"30947202094",,terminal_output +2269,9478617,"TERMINAL",0,0,"1305811305",,terminal_output +2270,9479642,"TERMINAL",0,0,"21692216",,terminal_output +2271,9480669,"TERMINAL",0,0,"327103327",,terminal_output +2272,9481665,"TERMINAL",0,0,"43814438",,terminal_output +2273,9482741,"TERMINAL",0,0,"54925549",,terminal_output +2274,9483759,"TERMINAL",0,0,"6540366550",,terminal_output +2275,9484848,"TERMINAL",0,0,"76147761",,terminal_output +2276,9485995,"TERMINAL",0,0,"88369983",,terminal_output +2277,9486881,"TERMINAL",0,0,"40947303094",,terminal_output +2278,9487920,"TERMINAL",0,0,"1405811405",,terminal_output +2279,9489015,"TERMINAL",0,0,"21692216",,terminal_output +2280,9490098,"TERMINAL",0,0,"327203327",,terminal_output +2281,9491228,"TERMINAL",0,0,"43814438",,terminal_output +2282,9492089,"TERMINAL",0,0,"54925549",,terminal_output +2283,9493195,"TERMINAL",0,0,"655036655:00",,terminal_output +2284,9494190,"TERMINAL",0,0,"76147761",,terminal_output +2285,9495223,"TERMINAL",0,0,"87258872",,terminal_output +2286,9496274,"TERMINAL",0,0,"98369983",,terminal_output +2287,9497334,"TERMINAL",0,0,"50947404094",,terminal_output +2288,9498400,"TERMINAL",0,0,"1505811505",,terminal_output +2289,9499406,"TERMINAL",0,0,"21692216",,terminal_output +2290,9500454,"TERMINAL",0,0,"327303327",,terminal_output +2291,9501492,"TERMINAL",0,0,"43814438",,terminal_output +2292,9502581,"TERMINAL",0,0,"54925549",,terminal_output +2293,9503582,"TERMINAL",0,0,"657:00366510",,terminal_output +2294,9504631,"TERMINAL",0,0,"76147761",,terminal_output +2295,9505674,"TERMINAL",0,0,"87258872",,terminal_output +2296,9506740,"TERMINAL",0,0,"98369983",,terminal_output +2297,9507765,"TERMINAL",0,0,"6:00947505094",,terminal_output +2298,9508828,"TERMINAL",0,0,"16:0169225:016",,terminal_output +2299,9509955,"TERMINAL",0,0,"327403327",,terminal_output +2300,9510936,"TERMINAL",0,0,"43814438",,terminal_output +2301,9512002,"TERMINAL",0,0,"54925549",,terminal_output +2302,9513033,"TERMINAL",0,0,"6510366520",,terminal_output +2303,9514049,"TERMINAL",0,0,"76147761",,terminal_output +2304,9515282,"TERMINAL",0,0,"87258872",,terminal_output +2305,9516315,"TERMINAL",0,0,"98369983",,terminal_output +2306,9517327,"TERMINAL",0,0,"109473:005:0094",,terminal_output +2307,9518251,"TERMINAL",0,0,"1105811105",,terminal_output +2308,9519243,"TERMINAL",0,0,"21692216",,terminal_output +2309,9520294,"TERMINAL",0,0,"327503327",,terminal_output +2310,9521322,"TERMINAL",0,0,"43814438",,terminal_output +2311,9522358,"TERMINAL",0,0,"54925549",,terminal_output +2312,9523409,"TERMINAL",0,0,"6520366530",,terminal_output +2313,9524436,"TERMINAL",0,0,"76147761",,terminal_output +2314,9525519,"TERMINAL",0,0,"87258872",,terminal_output +2315,9526532,"TERMINAL",0,0,"98369983",,terminal_output +2316,9527586,"TERMINAL",0,0,"20947101094",,terminal_output +2317,9528691,"TERMINAL",0,0,"1205811205",,terminal_output +2318,9529716,"TERMINAL",0,0,"21692216",,terminal_output +2319,9530740,"TERMINAL",0,0,"3274:003327",,terminal_output +2320,9531868,"TERMINAL",0,0,"43814438",,terminal_output +2321,9532815,"TERMINAL",0,0,"5530366540",,terminal_output +2322,9533919,"TERMINAL",0,0,"76147761",,terminal_output +2323,9534907,"TERMINAL",0,0,"87258872",,terminal_output +2324,9535949,"TERMINAL",0,0,"98369983",,terminal_output +2325,9537092,"TERMINAL",0,0,"30947202094",,terminal_output +2326,9538113,"TERMINAL",0,0,"1305811305",,terminal_output +2327,9539178,"TERMINAL",0,0,"21692216",,terminal_output +2328,9540102,"TERMINAL",0,0,"327103327",,terminal_output +2329,9541190,"TERMINAL",0,0,"43814438",,terminal_output +2330,9542307,"TERMINAL",0,0,"54925549",,terminal_output +2331,9543239,"TERMINAL",0,0,"6540366550",,terminal_output +2332,9544288,"TERMINAL",0,0,"76147761",,terminal_output +2333,9545390,"TERMINAL",0,0,"87258872",,terminal_output +2334,9546370,"TERMINAL",0,0,"98369983",,terminal_output +2335,9547434,"TERMINAL",0,0,"40947303094",,terminal_output +2336,9548506,"TERMINAL",0,0,"1405811405",,terminal_output +2337,9549494,"TERMINAL",0,0,"21692216",,terminal_output +2338,9550521,"TERMINAL",0,0,"327203327",,terminal_output +2339,9551570,"TERMINAL",0,0,"43814438",,terminal_output +2340,9552598,"TERMINAL",0,0,"54925549",,terminal_output +2341,9553685,"TERMINAL",0,0,"655036656:00",,terminal_output +2342,9554705,"TERMINAL",0,0,"76147761",,terminal_output +2343,9555728,"TERMINAL",0,0,"87258872",,terminal_output +2344,9556764,"TERMINAL",0,0,"98369983",,terminal_output +2345,9557810,"TERMINAL",0,0,"50947404094",,terminal_output +2346,9558900,"TERMINAL",0,0,"1516922516",,terminal_output +2347,9559925,"TERMINAL",0,0,"327303327",,terminal_output +2348,9561207,"TERMINAL",0,0,"43814438",,terminal_output +2349,9562180,"TERMINAL",0,0,"54925549",,terminal_output +2350,9563060,"TERMINAL",0,0,"658:00366510",,terminal_output +2351,9564227,"TERMINAL",0,0,"76147761",,terminal_output +2352,9565251,"TERMINAL",0,0,"87258872",,terminal_output +2353,9566277,"TERMINAL",0,0,"98369983",,terminal_output +2354,9567240,"TERMINAL",0,0,"7:00947505094",,terminal_output +2355,9568288,"TERMINAL",0,0,"17:0058116:005",,terminal_output +2356,9569357,"TERMINAL",0,0,"21692216",,terminal_output +2357,9570377,"TERMINAL",0,0,"327403327",,terminal_output +2358,9571418,"TERMINAL",0,0,"43814438",,terminal_output +2359,9572521,"TERMINAL",0,0,"54925549",,terminal_output +2360,9573492,"TERMINAL",0,0,"6510366520",,terminal_output +2361,9574577,"TERMINAL",0,0,"76147761",,terminal_output +2362,9575576,"TERMINAL",0,0,"87258872",,terminal_output +2363,9576714,"TERMINAL",0,0,"98369983",,terminal_output +2364,9577659,"TERMINAL",0,0,"109474:006:0094",,terminal_output +2365,9578701,"TERMINAL",0,0,"1105811105",,terminal_output +2366,9579774,"TERMINAL",0,0,"21692216",,terminal_output +2367,9580788,"TERMINAL",0,0,"327503327",,terminal_output +2368,9581839,"TERMINAL",0,0,"44925549",,terminal_output +2369,9582867,"TERMINAL",0,0,"6520366530",,terminal_output +2370,9583992,"TERMINAL",0,0,"76147761",,terminal_output +2371,9584974,"TERMINAL",0,0,"87258872",,terminal_output +2372,9586146,"TERMINAL",0,0,"98369983",,terminal_output +2373,9587030,"TERMINAL",0,0,"20947101094",,terminal_output +2374,9588078,"TERMINAL",0,0,"1205811205",,terminal_output +2375,9589209,"TERMINAL",0,0,"21692216",,terminal_output +2376,9590235,"TERMINAL",0,0,"3275:003327",,terminal_output +2377,9591263,"TERMINAL",0,0,"43814438",,terminal_output +2378,9592282,"TERMINAL",0,0,"54925549",,terminal_output +2379,9593310,"TERMINAL",0,0,"6530366540",,terminal_output +2380,9594358,"TERMINAL",0,0,"76147761",,terminal_output +2381,9595399,"TERMINAL",0,0,"87258872",,terminal_output +2382,9596467,"TERMINAL",0,0,"98369983",,terminal_output +2383,9597476,"TERMINAL",0,0,"30947202094",,terminal_output +2384,9598528,"TERMINAL",0,0,"1305811305",,terminal_output +2385,9599701,"TERMINAL",0,0,"21692216",,terminal_output +2386,9600782,"TERMINAL",0,0,"327103327",,terminal_output +2387,9601703,"TERMINAL",0,0,"43814438",,terminal_output +2388,9602698,"TERMINAL",0,0,"54925549",,terminal_output +2389,9603761,"TERMINAL",0,0,"6540366550",,terminal_output +2390,9604876,"TERMINAL",0,0,"76147761",,terminal_output +2391,9605820,"TERMINAL",0,0,"88369983",,terminal_output +2392,9606862,"TERMINAL",0,0,"40947303094",,terminal_output +2393,9607904,"TERMINAL",0,0,"1405811405",,terminal_output +2394,9609082,"TERMINAL",0,0,"21692216",,terminal_output +2395,9610003,"TERMINAL",0,0,"327203327",,terminal_output +2396,9611032,"TERMINAL",0,0,"43814438",,terminal_output +2397,9612096,"TERMINAL",0,0,"54925549",,terminal_output +2398,9613147,"TERMINAL",0,0,"655036657:00",,terminal_output +2399,9614177,"TERMINAL",0,0,"76147761",,terminal_output +2400,9614815,"TERMINAL",0,0,"scancel 3422103",,terminal_command +2401,9614850,"TERMINAL",0,0,"]633;E;2025-08-14 13:17:47 scancel 3422103;101ae38c-fbbc-44ac-a671-e512ad144ba4]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining]633;D;0",,terminal_output +2402,9615227,"TERMINAL",0,0,"M83422103 accelerat train_dy tum_cte0 CG\t5:25\t 8 hkn[0521-0528]728872",,terminal_output +2403,9616255,"TERMINAL",0,0,"9839983",,terminal_output +2404,9617272,"TERMINAL",0,0,"5094404094",,terminal_output +2405,9618325,"TERMINAL",0,0,"150511505",,terminal_output +2406,9619372,"TERMINAL",0,0,"2162216",,terminal_output +2407,9620446,"TERMINAL",0,0,"3273327",,terminal_output +2408,9621470,"TERMINAL",0,0,"4384438",,terminal_output +2409,9622488,"TERMINAL",0,0,"5495549",,terminal_output +2410,9623543,"TERMINAL",0,0,"659:0066510",,terminal_output +2411,9624644,"TERMINAL",0,0,"7617761",,terminal_output +2412,9625667,"TERMINAL",0,0,"8728872",,terminal_output +2413,9626692,"TERMINAL",0,0,"9839983",,terminal_output +2414,9627776,"TERMINAL",0,0,"8:0094505094",,terminal_output +2415,9628945,"TERMINAL",0,0,"18:005117:005",,terminal_output +2416,9629973,"TERMINAL",0,0,"2162216",,terminal_output +2417,9630945,"TERMINAL",0,0,"3384438",,terminal_output +2418,9631912,"TERMINAL",0,0,"5495549",,terminal_output +2419,9633038,"TERMINAL",0,0,"651066520",,terminal_output +2420,9634070,"TERMINAL",0,0,"7617761",,terminal_output +2421,9635088,"TERMINAL",0,0,"8728872",,terminal_output +2422,9636059,"TERMINAL",0,0,"9839983",,terminal_output +2423,9637140,"TERMINAL",0,0,"10945:007:0094",,terminal_output +2424,9638147,"TERMINAL",0,0,"110511105",,terminal_output +2425,9639182,"TERMINAL",0,0,"2162216",,terminal_output +2426,9640326,"TERMINAL",0,0,"3273327",,terminal_output +2427,9641273,"TERMINAL",0,0,"4384438",,terminal_output +2428,9642318,"TERMINAL",0,0,"5495549",,terminal_output +2429,9643358,"TERMINAL",0,0,"652066530",,terminal_output +2430,9644392,"TERMINAL",0,0,"7617761",,terminal_output +2431,9645442,"TERMINAL",0,0,"8728872",,terminal_output +2432,9646495,"TERMINAL",0,0,"9839983",,terminal_output +2433,9647517,"TERMINAL",0,0,"2094101094",,terminal_output +2434,9648568,"TERMINAL",0,0,"120511205",,terminal_output +2435,9649595,"TERMINAL",0,0,"2162216",,terminal_output +2436,9650655,"TERMINAL",0,0,"3273327",,terminal_output +2437,9651738,"TERMINAL",0,0,"4384438",,terminal_output +2438,9652802,"TERMINAL",0,0,"5495549",,terminal_output +2439,9653783,"TERMINAL",0,0,"653066540",,terminal_output +2440,9654969,"TERMINAL",0,0,"7617761",,terminal_output +2441,9655877,"TERMINAL",0,0,"8839983",,terminal_output +2442,9656884,"TERMINAL",0,0,"3094202094",,terminal_output +2443,9657919,"TERMINAL",0,0,"130511305",,terminal_output +2444,9659058,"TERMINAL",0,0,"2162216",,terminal_output +2445,9659998,"TERMINAL",0,0,"3273327",,terminal_output +2446,9661080,"TERMINAL",0,0,"\r4384438",,terminal_output +2447,9662083,"TERMINAL",0,0,"5495549",,terminal_output +2448,9663370,"TERMINAL",0,0,"654066550",,terminal_output +2449,9664173,"TERMINAL",0,0,"7617761",,terminal_output +2450,9665299,"TERMINAL",0,0,"8728872",,terminal_output +2451,9666243,"TERMINAL",0,0,"9839983",,terminal_output +2452,9667279,"TERMINAL",0,0,"4094303094",,terminal_output +2453,9668320,"TERMINAL",0,0,"140511405",,terminal_output +2454,9669362,"TERMINAL",0,0,"2162216",,terminal_output +2455,9670400,"TERMINAL",0,0,"3273327",,terminal_output +2456,9671475,"TERMINAL",0,0,"4384438",,terminal_output +2457,9672480,"TERMINAL",0,0,"5495549",,terminal_output +2458,9673521,"TERMINAL",0,0,"65506658:00",,terminal_output +2459,9674561,"TERMINAL",0,0,"7617761",,terminal_output +2460,9675639,"TERMINAL",0,0,"8728872",,terminal_output +2461,9676660,"TERMINAL",0,0,"9839983",,terminal_output +2462,9677678,"TERMINAL",0,0,"5094404094",,terminal_output +2463,9678810,"TERMINAL",0,0,"150511505",,terminal_output +2464,9679752,"TERMINAL",0,0,"2162216",,terminal_output +2465,9680794,"TERMINAL",0,0,"3273327",,terminal_output +2466,9681887,"TERMINAL",0,0,"4495549",,terminal_output +2467,9682893,"TERMINAL",0,0,"6530:0066510",,terminal_output +2468,9684034,"TERMINAL",0,0,"7617761",,terminal_output +2469,9684972,"TERMINAL",0,0,"8728872",,terminal_output +2470,9686083,"TERMINAL",0,0,"9839983",,terminal_output +2471,9687109,"TERMINAL",0,0,"9:0094505094",,terminal_output +2472,9688092,"TERMINAL",0,0,"19:005118:005",,terminal_output +2473,9689144,"TERMINAL",0,0,"2162216",,terminal_output +2474,9690184,"TERMINAL",0,0,"3273327",,terminal_output +2475,9691308,"TERMINAL",0,0,"4384438",,terminal_output +2476,9692269,"TERMINAL",0,0,"5495549",,terminal_output +2477,9693315,"TERMINAL",0,0,"651066520",,terminal_output +2478,9694348,"TERMINAL",0,0,"7617761",,terminal_output +2479,9695387,"TERMINAL",0,0,"8728872",,terminal_output +2480,9696435,"TERMINAL",0,0,"9839983",,terminal_output +2481,9697510,"TERMINAL",0,0,"10946:008:0094",,terminal_output +2482,9698516,"TERMINAL",0,0,"110511105",,terminal_output +2483,9699566,"TERMINAL",0,0,"2162216",,terminal_output +2484,9700638,"TERMINAL",0,0,"3273327",,terminal_output +2485,9701638,"TERMINAL",0,0,"4384438",,terminal_output +2486,9702673,"TERMINAL",0,0,"5495549",,terminal_output +2487,9703802,"TERMINAL",0,0,"652066530",,terminal_output +2488,9704802,"TERMINAL",0,0,"7617761",,terminal_output +2489,9705795,"TERMINAL",0,0,"8728872",,terminal_output +2490,9706836,"TERMINAL",0,0,"994101094",,terminal_output +2491,9707872,"TERMINAL",0,0,"2120511205",,terminal_output +2492,9708920,"TERMINAL",0,0,"2162216",,terminal_output +2493,9709946,"TERMINAL",0,0,"3273327",,terminal_output +2494,9710998,"TERMINAL",0,0,"4384438",,terminal_output +2495,9712096,"TERMINAL",0,0,"5495549",,terminal_output +2496,9713119,"TERMINAL",0,0,"653066540",,terminal_output +2497,9714154,"TERMINAL",0,0,"7617761",,terminal_output +2498,9715164,"TERMINAL",0,0,"8728872",,terminal_output +2499,9716268,"TERMINAL",0,0,"9839983",,terminal_output +2500,9717249,"TERMINAL",0,0,"3094202094",,terminal_output +2501,9718397,"TERMINAL",0,0,"130511305",,terminal_output +2502,9719338,"TERMINAL",0,0,"2162216",,terminal_output +2503,9720383,"TERMINAL",0,0,"3273327",,terminal_output +2504,9721422,"TERMINAL",0,0,"4384438",,terminal_output +2505,9721467,"TERMINAL",0,0,"watch",,terminal_focus +2506,9722040,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/causal/dynamics-cotraining]633;D;0",,terminal_output +2507,9725293,"TERMINAL",0,0,"bash",,terminal_focus +2508,9774606,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",0,0,"",shellscript,tab +2509,9775918,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",0,0,"",shellscript,tab +2510,9792378,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1527,0,"",shellscript,selection_mouse +2511,9792948,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1517,0,"",shellscript,selection_mouse +2512,9793451,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1432,0,"",shellscript,selection_mouse +2513,9794032,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1511,0,"",shellscript,selection_mouse +2514,9794585,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1517,0,"",shellscript,selection_mouse +2515,9794735,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1516,3,"big",shellscript,selection_mouse +2516,9794930,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1516,4,"big-",shellscript,selection_mouse +2517,9794972,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1516,8,"big-runs",shellscript,selection_mouse +2518,9796542,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",1517,0,"",shellscript,selection_mouse +2519,12100126,"TERMINAL",0,0,"bash",,terminal_focus +2520,12101077,"TERMINAL",0,0,"id",,terminal_command +2521,12806247,"TERMINAL",0,0,"logs",,terminal_command +2522,12807752,"TERMINAL",0,0,"cd ..",,terminal_command +2523,12815406,"TERMINAL",0,0,"cd logs_franz/",,terminal_command +2524,12815675,"TERMINAL",0,0,"ls",,terminal_command +2525,12815727,"TERMINAL",0,0,"]633;E;2025-08-14 14:11:08 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +2526,12815923,"TERMINAL",0,0,"big-runs\r\ncausal\r\ndynamics-cotraining-modelsize-scaling\r\nmaskgit\r\ntokenizer-lr-scaling\r\ntrain_tokenizer_batch_size_scaling_linear_lr_scaling_16_nodes_3322407.log\r\ntrain_tokenizer_batch_size_scaling_linear_lr_scaling_16_nodes_3325607.log\r\ntrain_tokenizer_batch_size_scaling_linear_lr_scaling_2_nodes_3322404.log\r\ntrain_tokenizer_batch_size_scaling_linear_lr_scaling_2_nodes_3325603.log\r\ntrain_tokenizer_batch_size_scaling_linear_lr_scaling_4_nodes_3322405.log\r\ntrain_tokenizer_batch_size_scaling_linear_lr_scaling_4_nodes_3325605.log\r\ntrain_tokenizer_batch_size_scaling_linear_lr_scaling_8_nodes_3322406.log\r\ntrain_tokenizer_batch_size_scaling_linear_lr_scaling_8_nodes_3325606.log\r\ntrain_tokenizer_no_warmup_ablation_3322446.log\r\ntrain_tokenizer_no_warmup_ablation_3325608.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz]633;D;0",,terminal_output +2527,12819069,"TERMINAL",0,0,"cd causal/",,terminal_command +2528,12819480,"TERMINAL",0,0,"ls",,terminal_command +2529,12820910,"TERMINAL",0,0,"cd dynamics-cotraining/",,terminal_command +2530,12822029,"TERMINAL",0,0,"ls",,terminal_command +2531,12823981,"TERMINAL",0,0,"tail -f train_dynamics_causal_8_node_3422130.log",,terminal_command +2532,12824053,"TERMINAL",0,0,"]633;E;2025-08-14 14:11:17 tail -f train_dynamics_causal_8_node_3422130.log ;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CFiltering out sequence with average brightness 48.25839369270832, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.81418723394097, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 4.37693323784722, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.93325634809028, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.28341355642361, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.290651599826386, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.685171124565974, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 7.973923466145835, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.96465803776041, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.555535260416665, which is below the darkness threshold 50.0.\r\n",,terminal_output +2533,12825090,"TERMINAL",0,0,"Filtering out sequence with average brightness 6.248491104166666, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.026395559027776, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.32088695746528, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.287892904513892, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.699837139322916, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.35339623003471, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.515516414062496, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.779079039062502, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 31.77856176866319, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.801213352430556, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.445178874565972, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.22831657595486, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.642893111111114, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.82428636067709, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.826017976128472, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.369552199218745, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.205715572048604, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.80163310763888, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.21067757769097, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.756402393229166, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.90283489192707, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.064562754774306, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.96316022178819, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.351987908854174, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.47436232682293, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.904636282986104, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.30252202300348, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.50610748003473, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.83450543229167, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 3, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 43.50629163671874, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.25673835112847, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.17340608333334, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 34.061884940104164, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.26984409418403, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.782408529513887, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.534414733072918, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.83258071657986, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.201719534722223, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.418512315538194, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.291210371961807, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.396078554253478, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.273443664062498, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.447791945746527, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.783089046440985, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.316311935329859, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.161475852430556, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.72070090451389, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.54771968793403, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.68763006163195, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.59348041970486, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.49130341753471, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.410399775607639, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.73322909157986, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.389401227864575, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.299453903211806, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.42551810199654, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.461762073350695, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.428984342881945, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 8.41517641623264, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.258012031250004, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.84490517274306, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.395675162326405, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.939019926215266, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.32156702560764, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.05500592751736, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.14982335807292, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.28745239453125, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.03377534982639, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 1, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 38.291162500868055, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.500775280381944, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.11585116927084, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.341027131944436, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.53352068663194, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.105931896267364, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.851931579861116, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.783686277777777, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 14, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 3.430726105468751, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.391618353732646, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.483258674479174, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.90867199392361, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.89368974956597, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.308401263454858, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.57481587109373, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.550750368923605, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.28874605946181, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.313530924479169, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.177338001302086, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.77951588411458, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.298603133680555, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 4.929624919704861, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 34.7758957578125, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.7887266111111, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.57599053993056, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.004942078125001, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.866432825086806, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.99970993012153, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.484809298611108, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.990326822048612, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.76507261979167, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.50315575217014, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.640106145833336, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.475888818576388, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.887308973524306, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 5.163022049045138, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.16183431814236, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.22922659505208, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.61054293098958, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.001348415798603, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.836997713975695, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.84720079123263, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.436243806857643, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.55236843446179, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.661866783420137, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 2, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 5.745865147135416, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.16920592013887, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.739338965711806, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.64539661024304, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.69868717621527, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.61229421744791, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.408982079427076, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.878024863715275, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 4.087599994357639, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.782459053385416, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.26742317404514, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.066657614149303, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.102210330295144, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.68000962022569, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.108358296006934, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.01184064149306, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.07952100173611, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.23764219401041, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.273121290364582, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.67403715755209, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.159634221788195, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.859693625434026, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.267163098958335, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.97735686935763, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.05856661848958, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.21590727300348, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.729370394965265, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.069526227864582, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.73526423394098, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.858907507378472, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.11182256857639, which is below the darkness threshold 50.0.\r\n",,terminal_output +2534,12826139,"TERMINAL",0,0,"Filtering out sequence with average brightness 41.464187955295145, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.02931689626736, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.360279158420138, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.369152598090274, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.589123648003472, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.655651488281247, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.331612240885416, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.972827269097222, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.879556708767364, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.65433171397569, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 4.830970013888888, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.30589482986111, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.284072398871528, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.68896297743055, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.66199507335069, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.532190937934027, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.90413924479166, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.854636536892354, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.51112091102431, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.90547663975693, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 31.335604068576384, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.77122214453125, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.825450489149308, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.98555696961805, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.725477375000004, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.201561266927065, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.48539023394097, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 31.880347332031246, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.520327518229164, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.54553545833332, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.441437454427074, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.25730537890625, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.692825124565964, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.114622904947916, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.15385722309028, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.720363145833325, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 2, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 42.261345737413194, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.205406408854167, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.28746638802084, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.706706666232641, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.66873425911458, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.760344266059025, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.609183381076406, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.39644596310764, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.231705520399295, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.815763887586805, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.838116223958338, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.20830425130209, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 31.612449674045138, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.64177461154513, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 2.6250118437500003, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.652351098090275, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.236154825520833, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.732674315972222, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.76571675434028, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.09408901215278, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.613179714843746, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 4.137212364583333, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.480846973524304, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.922143979166663, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.552360985677087, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.33964979774306, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.98366849435763, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 1, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 42.19896424131946, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.28097957465277, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.31633569487847, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.227408264756946, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.91574588802084, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.008100246527775, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.670641799479167, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.22008243446181, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.730366986979167, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.35534505859375, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.85925859765625, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.549803424913193, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.906939588541654, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.471726490885413, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.98253778689236, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.01438754513889, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 6.185368413628472, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.60145767447917, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.11110933550347, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 6.471034640190972, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.89805972309027, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.59533146440973, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.08339677517361, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.241696391059023, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.71200667708333, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.3321095030382, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.191030071180556, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.631816058159721, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.175148952256944, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.51353421961806, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.14789116493056, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.62533254817709, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 6.170116545572916, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.80586637456596, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.00045060069444, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.97285501996528, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.87322573263887, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.405553090711805, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.41258210937499, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.1705804375, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.187679411024307, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.418184020399305, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.342379329861103, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.958920232638885, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.8499512404514, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.48266377777778, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.634783483506947, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.30427697699653, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.932166159288194, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.47815596701389, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.1499930264757, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.42777363107638, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.340278833333333, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.62191007595485, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.362136504340278, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.656163871527777, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.05662233637153, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.05428888888888, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.809410006510415, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 31.894950679687504, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.252755410156247, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.19600844618055, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.576112999565964, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.02128690104168, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.638448952256944, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.71198654036458, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.075061014322912, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.5968681484375, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.306010959201394, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.53528334027777, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.10780612065972, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.81797238715277, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.30913177126736, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.3419572313368, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.09825508420139, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.21295064366319, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.09468324956597, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.846888147569448, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.994928267361104, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.56066381163193, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.63902894878471, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.60075063628472, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.298270464843746, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.6677207938368, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.656851608940972, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.44361700217013, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.01214288064236, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.749617506076376, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.156503000868064, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.700470125434016, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.692577922309034, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.39059139019096, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.944281329427085, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.98361345312499, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.160125339409724, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.88992796657987, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.04279984027778, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.35145857118055, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.421225546006944, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.91267774782986, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 4.961705045138888, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 2.8816884709201394, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.59353025737847, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.814528867621526, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.700266996527777, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.792198299913192, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.6903745577257, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.18161632031251, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.671145737847226, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.81221030989583, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.980041447048606, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.09131488541667, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.40287151649305, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.277096176215277, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.947070907118055, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.08994145182291, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.293038946180555, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.391716476128476, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.729449763020833, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.963518907118054, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.76784844444445, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.53240935329861, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.994459208767367, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.574013672743053, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 8.741062166666666, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.988881016059025, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.19184354470487, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.19927317404513, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.343309812934024, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.72330868923611, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 1, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 32.185184884114584, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.69413179600695, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.24734658463541, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.79358963845487, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.850761896701385, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.34360246137154, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.71899516189236, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.432555463541675, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.210264149305555, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.53775068663194, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.44783652387153, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.963878419704862, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.25145430815973, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.86519064236111, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.735804779513884, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.322472456597225, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.741105429687508, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.57213467838541, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.844601434027794, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.257634926649303, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.814926668836804, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.93532238845486, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.714602090711807, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.851494995225694, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.99958494140625, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.6568950546875, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.45888567925347, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.252253840277776, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.36907496310765, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.463286263020834, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 14, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 34.204493317274306, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.224961368923594, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.96766257769097, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.94426772743055, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.308978797309027, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.23581079340278, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.14020086979166, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.915178693142366, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 1, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 24.836928993489593, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.46069146050348, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.223201267795135, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.69297598567709, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.54188554210069, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.28372880555556, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.88308756901043, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.61451967144097, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.35835799392361, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.337642720920142, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 1, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 41.409151208333334, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.406037435329864, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.80914919357639, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.79415559331597, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.30357367578125, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.03802509201388, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.667169706597225, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.482073514756944, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.695623681423605, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.18017578038195, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.07815861458333, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.55580262847222, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.946071781684026, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.152852060763891, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.03724315451389, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.62049548307292, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.550028684027776, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.214637049479162, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.13950512500001, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.879113871961803, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.897983416232641, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 3.1173806132812505, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.17875073828126, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.61300194140625, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.561536807291645, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 7.561142590711805, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.73062853689238, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.12580900260417, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.00336980815972, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.213181888020834, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.3500005625, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.97662650998263, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.280135458767361, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.642397835069442, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.310776618055554, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.83485321180554, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.975132686631946, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.12815167361112, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.344765733940974, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.7087069014757, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 1, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 41.07775581423611, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 15, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 47.44254116015625, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 31.810507785156243, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.39119336154514, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.84895600607638, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.23003100781251, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.59613746180556, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.287873690104156, which is below the darkness threshold 50.0.\r\n",,terminal_output +2535,12827134,"TERMINAL",0,0,"Filtering out sequence with average brightness 43.69283903168401, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 15, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 27.917037622395828, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.067997086805555, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.96044309505208, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 7.540668100260416, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.785180790364585, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.823761831597224, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.454952747395836, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.185998623263888, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.684065997829855, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.12983470529514, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.855540866753476, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.12479111631945, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.028692759548612, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.21180633072916, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.90460269314237, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.90338756944444, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.623259755642366, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.91172327647569, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.921175553819445, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.222967984375003, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 8, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 47.231410530381936, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.86746961197917, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.6776365486111, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.367244151909716, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.82563311631944, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.699530347656264, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.92555766406249, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 3.3143295755208344, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.81272714800346, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.594541773437495, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.249960812065975, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.86814599956596, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.002799543402773, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.08420196484374, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.22825359244791, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.950280488281244, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.465649786458334, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.34526884722222, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.5924828125, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.927245735243055, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.33458676822917, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.226825424479166, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.90795716883681, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.409973788194442, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.67603242664931, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.258384671874996, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.77652489019097, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.915616134982635, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.82448556553819, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.082291970052076, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 2, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 19.404355583333334, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.597213169270834, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.246115174913186, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.23935652690973, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.46171998828126, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.299551327690974, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.16669221875, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.90996709722223, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.06910895095486, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.54379095703124, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.2939763671875, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.59947224348959, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.5598106610243, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.079517940538192, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.87307643619791, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.28747654470487, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.07631629340278, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.949276658420146, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.00432015104167, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 3.5492821276041666, which is below the darkness threshold 50.0.\r\n",,terminal_output +2536,12829188,"TERMINAL",0,0,"Filtering out sequence with average brightness 47.198341880208325, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.40728582291666, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.340056402777776, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.36404093402778, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.675579679253467, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.49749498611111, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.786759471788194, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.022464996527773, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.37718409505208, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.494366769097216, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.81318576649305, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.960078757378472, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.55674429557292, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.61251941145834, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.87782646744791, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.31903804340277, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.632209563368047, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.04666593142361, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.57846514756944, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.715776291666664, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.87392485937499, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.89639959331597, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.464378066406248, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.613432802517362, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.25213565060765, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.99575514192709, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 7.040606244357639, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.980406286892354, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.94310283420139, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.83854703428819, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.880232010416666, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 31.190118578125, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.774855567274304, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.21080336241319, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.119482376302084, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.237615, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.454526124565973, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.63910823090278, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 4.867321037760415, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.77601594097221, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.749389063368056, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.85911107161458, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.37265406684028, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.13102692534722, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.82766136588541, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.967041765624984, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 6.80206761892361, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.521571290364584, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.396949013888896, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.604577486979146, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.617727667968758, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 14, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 22.07299042404514, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.08642860069443, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.63921443359376, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.564326149739586, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.47013460546875, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.608252359375005, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.525131890190973, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.18121266189236, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 11, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 35.667371157118055, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.838729708333332, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.527233295138892, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.40376301996527, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.632272440104167, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.950613430555553, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.452043534288194, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.918594010850693, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 15, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 35.20020058333334, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.76126238064235, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.2591979578993, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.493715072916665, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.27150613585069, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.15607400607639, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.192714589409725, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.346160309461803, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.620414561197904, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 34.67117838368055, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.10995779166667, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.57885933680555, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.498023140190966, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 1, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 20.69735179427084, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.25380546614583, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.50233857682292, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.52856073697917, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.21707877734375, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.35167131336805, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.06861614713542, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.72666005685764, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.00228797395833, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.047586053385412, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.888655702690976, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.10108218185764, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.532313350694434, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.33136517621527, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.60887864105902, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.703137481336807, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.59762666102431, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.657530521701396, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.9746550733507, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.555182551649295, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 14, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 13.38810968793403, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.121301209635412, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.04249641276042, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.596459095920135, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.095723200520833, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.66970105642362, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.87765283940971, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.40955750260417, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.666515397135417, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.01179643532987, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.615333272569444, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.35851682378473, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.90901723828126, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.34328444574651, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 8.337617865017362, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.25026827604167, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.602502460937494, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.86566910677083, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 7.414749640190972, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.963286688802086, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.992752450086805, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.578313046875001, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.04283973263889, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.162726562065973, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.09226792578125, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.359980235243057, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.770273491753464, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.321374582899306, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.02284484548611, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.894497451822915, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.778345146267359, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.80967586501737, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.182953991319444, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.925921832899306, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.28472446571182, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.625564571614582, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.18354208984375, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.54798894444444, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.527609464409718, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 5.510687426649303, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.84359200390624, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 1, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 32.99687792751736, which is below the darkness threshold 50.0.\r\n",,terminal_output +2537,12830132,"TERMINAL",0,0,"Filtering out sequence with average brightness 20.584517582899295, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.85491881380207, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.615683417968754, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.143158724392363, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.94644012369791, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 3.4838853832465277, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.21313230295139, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.16577651909722, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.88132032725694, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 34.88540337630208, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.54971527690973, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.577339532118053, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.62320235112846, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.43717969965277, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.99100119140624, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.40595642404514, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.085197149739585, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.599812453559025, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.23602431423611, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.65727039019096, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.744193267361112, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.396549269097214, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.37920105251737, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.88340945572917, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.44764171267361, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.05193628515625, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.49738452126736, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.712017695312497, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.51124050911458, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.46037941536458, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.26853993793402, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.72962086197916, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.70888270095486, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.908322227430554, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.766876286024306, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.73980721180554, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 4.247701313368054, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 14, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 22.44674719140625, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.785861887586806, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.94737023480904, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.1168075750868, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.21994015494791, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 10.18011365581597, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.2814726766493, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.044114148871515, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.984321318576384, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.078478967013886, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.24182104817708, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.405390644097226, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.648661205295141, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.70184042317708, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.753682626736115, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.23127126519097, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.96963414583333, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.82039700217013, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.6048155655382, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.00690611241319, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.57101773003472, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.671724417100695, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.5533767578125, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.960590548611115, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.34763674435764, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.14637792925347, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.065293465711804, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.546474480034718, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.009234914062493, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.458230740885416, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.286202267795126, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.899898362413204, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.30979557248264, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.96450533159722, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.260386103298607, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.677346032118045, which is below the darkness threshold 50.0.\r\n",,terminal_output +2538,12831144,"TERMINAL",0,0,"Filtering out episode with length 6, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 32.862843753906255, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.6795984874132, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.04458074609376, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.03164207769096, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.19243184244791, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.491503715277766, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.70997716319444, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 34.44411474175347, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.51261280078125, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.43365499479167, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.43792941796874, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.06409200911457, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.603385859809025, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.681302502604154, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.13043981032986, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.41378156206597, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.59997234114583, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.51647772699653, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.322956052517355, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.059334759548605, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.522987754774306, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 5.639856405381943, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.359573090711807, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.435467890624995, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.813368881510407, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.21610072699653, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.390818709201387, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 14, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 32.46940284852431, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.49886976779513, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.542988711371526, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.401129498263888, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.9952937282986, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.99812473871528, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.78518679947916, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.359289539062502, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.147173033854164, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 28.586465228732635, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.76773339105903, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.705631027777777, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 34.34850776996528, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 31.211462479600687, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.417603453993056, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.66003829861111, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.65604112456597, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.00930867404513, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.369178937065975, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.50743714105902, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.594955946180555, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 34.08040523828125, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.857080977864584, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.230154930121525, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.74555226996527, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.60679850130209, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.19186065842013, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.635319138020833, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.494882779079866, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.02789135546875, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.289346652777766, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.733238383246528, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.65975987413195, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.98584924348958, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.991792206597218, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.43363980555555, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 6, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 35.46088256814235, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.901566312499998, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.99567389409722, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.208291830295124, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.805345717013887, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.4354558485243, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.443111936197916, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.96171985720487, which is below the darkness threshold 50.0.\r\n",,terminal_output +2539,12832160,"TERMINAL",0,0,"Filtering out sequence with average brightness 45.56216610677084, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.700271710503472, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.305229101996527, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 3.317470993923611, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.59633501736111, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.11482972092013, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.57808961458333, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.96180032118055, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.93626566796875, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.80606769444444, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 34.5031797452257, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.61120937282986, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 31.283901809895834, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.080622509982632, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.607679651041664, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.03240911935765, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.853900456597213, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.543171285156248, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.07596157118056, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.24305379253473, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.75990732942708, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.92470306510416, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.737464522569447, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.533172161024304, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.28104738237848, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.57187957508681, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.28303004513889, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.434875995225692, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.62051286458332, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.736149002604165, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.433547449218754, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.88892702907986, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.29590201128471, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.18326280598959, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.085143552951397, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.49563035243056, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.344287523003466, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.285958331163195, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.690684492187497, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 1, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 20.1771911922743, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.59776680815973, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.795674034722214, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.211715393229163, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 21.258736057291667, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.41129976605903, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.04576054296875, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.01707182118056, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.896484003472228, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.56767216623265, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.46249483463542, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.489087004340274, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.13610442578125, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 7.291982520833333, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.769582384982634, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.28725516536459, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.12756687065973, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.228390414496516, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.72119899652776, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.56270264105902, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 8.796899195746526, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.470349585937484, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.16261827821181, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.01894919140624, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.733159609809025, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.07459774262153, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.731335756076405, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 20.736553951822913, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.892008654513885, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.73564039019097, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.27213774001736, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.067593857204855, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.235811047309028, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 12.818970408420137, which is below the darkness threshold 50.0.\r\n",,terminal_output +2540,12833144,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/causal/dynamics-cotraining]633;D;130",,terminal_output +2541,12834172,"TERMINAL",0,0,"cd ..",,terminal_command +2542,12834521,"TERMINAL",0,0,"ls",,terminal_command +2543,12834555,"TERMINAL",0,0,"]633;E;2025-08-14 14:11:27 l;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Cbash: l: command not found...\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/causal]633;D;127",,terminal_output +2544,12882919,"TERMINAL",0,0,"tail -f train_dynamics_causal_8_node_3422130.log",,terminal_command +2545,12885341,"TERMINAL",0,0,"cd ..",,terminal_command +2546,12886744,"TERMINAL",0,0,"cd maskgit/",,terminal_command +2547,12887027,"TERMINAL",0,0,"ls",,terminal_command +2548,12887056,"TERMINAL",0,0,"]633;E;2025-08-14 14:12:20 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Cdynamics-cotraining\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit]633;D;0",,terminal_output +2549,12888271,"TERMINAL",0,0,"cd dynamics-cotraining/",,terminal_command +2550,12891181,"TERMINAL",0,0,"ls",,terminal_command +2551,12891218,"TERMINAL",0,0,"]633;E;2025-08-14 14:12:24 ls;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Ctrain_dynamics_maskgit_8_node_3422132.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining]633;D;0",,terminal_output +2552,12893569,"TERMINAL",0,0,"tail -f train_dynamics_maskgit_8_node_3422132.log",,terminal_command +2553,12893641,"TERMINAL",0,0,"]633;E;2025-08-14 14:12:26 tail -f train_dynamics_maskgit_8_node_3422132.log ;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CNodelist: hkn[0530-0534,0536,0701,0715]\r\nCPU Utilized: 00:02:54\r\nCPU Efficiency: 0.25% of 19:02:24 core-walltime\r\nJob Wall-clock time: 00:05:57\r\nStarttime: Thu Aug 14 13:24:52 2025\r\nEndtime: Thu Aug 14 13:30:49 2025\r\nMemory Utilized: 12.67 GB (estimated maximum)\r\nMemory Efficiency: 0.00% of 0.00 MB (0.00 MB/node)\r\nEnergy Consumed: 1339227 Joule / 372.0075 Watthours\r\nAverage node power draw: 3751.33613445378 Watt\r\n",,terminal_output +2554,12896067,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining]633;D;130",,terminal_output +2555,12900284,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --job-name=train_dynamics_maskgit_8_node\n#SBATCH --requeue\n#SBATCH --reservation=llmtum\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/holiday/maskgit/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --darkness_threshold=50 \\n --dyna_type=maskgit \\n --num_latent_actions=6 \\n --max_lr=8e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-maskgit-8-node-darkness-filter-$slurm_job_id \\n --tags dynamics maskgit 8-node post-launch-main darkness-filter \\n --entity instant-uv \\n --project jafar \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16 \\n --dyna_ffn_dim=4096 \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\nSLURM_JOB_USER=tum_dbd0378\nSLURM_TASKS_PER_NODE=4(x8)\nSLURM_JOB_UID=996262\nSLURM_TASK_PID=17640\nSLURM_JOB_GPUS=0,1,2,3\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar\nSLURMD_NODENAME=hkn0530\nSLURM_JOB_START_TIME=1755170692\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1755343492\nSLURM_CPUS_ON_NODE=24\nSLURM_JOB_CPUS_PER_NODE=24(x8)\nSLURM_GPUS_ON_NODE=4\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=accelerated\nSLURM_TRES_PER_TASK=cpu=5\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=8\nSLURM_JOBID=3422132\nSLURM_JOB_QOS=normal\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=5\nSLURM_JOB_RESERVATION=llmtum\nSLURM_NTASKS=32\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0530\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_NODELIST=hkn[0530-0534,0536,0701,0715]\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=32\nSLURM_NNODES=8\nSLURM_SUBMIT_HOST=hkn1990.localdomain\nSLURM_JOB_ID=3422132\nSLURM_NODEID=0\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_dynamics_maskgit_8_node\nSLURM_NTASKS_PER_NODE=4\nSLURM_JOB_GID=502226\nSLURM_JOB_NODELIST=hkn[0530-0534,0536,0701,0715]\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\n2025-08-14 13:25:47.802222: W external/xla/xla/service/platform_util.cc:220] unable to create StreamExecutor for CUDA:3: CUDA error: Failed call to cuDeviceGet: CUDA_ERROR_INVALID_DEVICE: invalid device ordinal\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/train_dynamics.py"", line 158, in \n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': INTERNAL: no supported devices found for platform CUDA (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: Getting local topologies failed: Error 1: GetKeyValue() timed out with key: cuda:local_topology/cuda/19 and duration: 2m\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/train_dynamics.py"", line 158, in \n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': INTERNAL: Getting local topologies failed: Error 1: GetKeyValue() timed out with key: cuda:local_topology/cuda/19 and duration: 2m (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: DEADLINE_EXCEEDED: GetKeyValue() timed out with key: cuda:global_topology/cuda and duration: 5m\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/train_dynamics.py"", line 158, in \n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': DEADLINE_EXCEEDED: GetKeyValue() timed out with key: cuda:global_topology/cuda and duration: 5m (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: DEADLINE_EXCEEDED: GetKeyValue() timed out with key: cuda:global_topology/cuda and duration: 5m\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/train_dynamics.py"", line 158, in \n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': DEADLINE_EXCEEDED: GetKeyValue() timed out with key: cuda:global_topology/cuda and duration: 5m (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: DEADLINE_EXCEEDED: GetKeyValue() timed out with key: cuda:global_topology/cuda and duration: 5m\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/train_dynamics.py"", line 158, in \n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': DEADLINE_EXCEEDED: GetKeyValue() timed out with key: cuda:global_topology/cuda and duration: 5m (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n2025-08-14 13:30:48.404038: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1798] Shutdown barrier in coordination service has failed:\nDEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\nThis suggests that the workers are out of sync. Either at least one worker (a) crashed early due to program error or scheduler events (e.g. preemption, eviction), (b) was too fast in its execution, or (c) too slow / hanging. Check the logs (both the program and scheduler events) for an earlier error to identify the root cause.\n2025-08-14 13:30:48.404124: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1839] Use error polling to propagate the following error to all tasks: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.404702: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.404606: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.404610: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.404778: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405016: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.404842: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.404891: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\n2025-08-14 13:30:48.404932: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405085: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.404913: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405092: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.404916: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405030: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405001: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\n2025-08-14 13:30:48.405055: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405142: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405049: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405214: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405066: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405307: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405334: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\n2025-08-14 13:30:48.405487: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\n2025-08-14 13:30:48.405255: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405406: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405332: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\n2025-08-14 13:30:48.405365: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405622: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405408: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\n2025-08-14 13:30:48.405451: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405527: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405438: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405537: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405620: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405587: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405776: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405493: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\n2025-08-14 13:30:48.405770: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405462: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405773: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\n2025-08-14 13:30:48.405592: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405765: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\n2025-08-14 13:30:48.405437: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405935: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405994: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405866: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\n2025-08-14 13:30:48.406066: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.405889: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\n2025-08-14 13:30:48.406077: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.404956: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.404986: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405212: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.405510: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.421574: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.421570: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.421590: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.421569: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.512644: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.512567: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.512644: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.512749: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-08-14 13:30:48.513476: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.513541: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.513499: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-08-14 13:30:48.513545: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::3040337475261767296::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 5/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:19. Some timed out task names:\n/job:jax_worker/replica:0/task:20\n/job:jax_worker/replica:0/task:30\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::3040337475261767296']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nsrun: error: hkn0534: task 19: Aborted (core dumped)\nsrun: error: hkn0534: tasks 16,18: Aborted (core dumped)\nsrun: error: hkn0533: tasks 12,14: Aborted (core dumped)\nsrun: error: hkn0532: tasks 9,11: Aborted (core dumped)\nsrun: error: hkn0530: tasks 0,2: Aborted (core dumped)\nsrun: error: hkn0531: tasks 4,7: Aborted (core dumped)\nsrun: error: hkn0536: tasks 21-22: Aborted (core dumped)\nsrun: error: hkn0701: tasks 26-27: Aborted (core dumped)\nsrun: error: hkn0715: tasks 30-31: Aborted (core dumped)\nsrun: error: hkn0534: task 17: Aborted (core dumped)\nsrun: error: hkn0533: task 15: Aborted (core dumped)\nsrun: error: hkn0532: task 8: Aborted (core dumped)\nsrun: error: hkn0531: task 6: Aborted (core dumped)\nsrun: error: hkn0530: task 3: Aborted (core dumped)\nsrun: error: hkn0536: task 23: Aborted (core dumped)\nsrun: error: hkn0701: task 25: Aborted (core dumped)\nsrun: error: hkn0715: task 28: Aborted (core dumped)\nsrun: error: hkn0533: task 13: Aborted (core dumped)\nsrun: error: hkn0532: task 10: Aborted (core dumped)\nsrun: error: hkn0530: task 1: Aborted (core dumped)\nsrun: error: hkn0531: task 5: Aborted (core dumped)\nsrun: error: hkn0536: task 20: Aborted (core dumped)\nsrun: error: hkn0701: task 24: Aborted (core dumped)\nsrun: error: hkn0715: task 29: Aborted (core dumped)\n\n============================= JOB FEEDBACK =============================\n\nJob ID: 3422132\nCluster: hk\nUser/Group: tum_dbd0378/hk-project-p0023960\nAccount: hk-project-p0023960\nState: FAILED (exit code 134)\nPartition: accelerated\nNodes: 8\nCores per node: 24\nNodelist: hkn[0530-0534,0536,0701,0715]\nCPU Utilized: 00:02:54\nCPU Efficiency: 0.25% of 19:02:24 core-walltime\nJob Wall-clock time: 00:05:57\nStarttime: Thu Aug 14 13:24:52 2025\nEndtime: Thu Aug 14 13:30:49 2025\nMemory Utilized: 12.67 GB (estimated maximum)\nMemory Efficiency: 0.00% of 0.00 MB (0.00 MB/node)\nEnergy Consumed: 1339227 Joule / 372.0075 Watthours\nAverage node power draw: 3751.33613445378 Watt\n",log,tab +2556,12900733,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",489,0,"",log,selection_mouse +2557,12900767,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",488,0,"",log,selection_command +2558,12904425,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",562,0,"",log,selection_mouse +2559,12905011,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",460,0,"",log,selection_mouse +2560,12946943,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",6017,0,"",log,selection_mouse +2561,12947065,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",6016,12,"RuntimeError",log,selection_mouse +2562,12958605,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",6163,0,"",log,selection_mouse +2563,12958758,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",6161,6,"plugin",log,selection_mouse +2564,12958920,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",6016,209,"RuntimeError: Unable to initialize backend 'cuda': INTERNAL: no supported devices found for platform CUDA (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n",log,selection_mouse +2565,12960277,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",6080,0,"",log,selection_mouse +2566,12960277,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",6080,9,"supported",log,selection_mouse +2567,12960459,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",6016,209,"RuntimeError: Unable to initialize backend 'cuda': INTERNAL: no supported devices found for platform CUDA (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n",log,selection_mouse +2568,12968171,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",13857,0,"",log,selection_mouse +2569,12968883,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",13788,0,"",log,selection_mouse +2570,12969018,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",13780,11,"GetKeyValue",log,selection_mouse +2571,12969164,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",13732,125,"jaxlib._jax.XlaRuntimeError: DEADLINE_EXCEEDED: GetKeyValue() timed out with key: cuda:global_topology/cuda and duration: 5m\n",log,selection_mouse +2572,12987028,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4964,0,"",log,selection_mouse +2573,12987173,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,8,"INTERNAL",log,selection_mouse +2574,12987366,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,9,"INTERNAL:",log,selection_mouse +2575,12987386,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,12,"INTERNAL: no",log,selection_mouse +2576,12987403,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,22,"INTERNAL: no supported",log,selection_mouse +2577,12987481,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,23,"INTERNAL: no supported ",log,selection_mouse +2578,12987482,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,30,"INTERNAL: no supported devices",log,selection_mouse +2579,12987526,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4908,57,"distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL",log,selection_mouse +2580,12987561,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4928,37,"jaxlib._jax.XlaRuntimeError: INTERNAL",log,selection_mouse +2581,12987562,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,36,"INTERNAL: no supported devices found",log,selection_mouse +2582,12987596,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,37,"INTERNAL: no supported devices found ",log,selection_mouse +2583,12987597,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,40,"INTERNAL: no supported devices found for",log,selection_mouse +2584,12987630,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,41,"INTERNAL: no supported devices found for ",log,selection_mouse +2585,12987665,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,49,"INTERNAL: no supported devices found for platform",log,selection_mouse +2586,12987817,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,50,"INTERNAL: no supported devices found for platform ",log,selection_mouse +2587,12987855,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4957,54,"INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2588,12988372,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",5011,0,"",log,selection_mouse +2589,12988386,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",5010,0,"",log,selection_command +2590,12988827,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",5011,0,"",log,selection_mouse +2591,12988838,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",5010,0,"",log,selection_command +2592,12988990,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",5010,1,"A",log,selection_mouse +2593,12988996,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",5011,0,"",log,selection_command +2594,12989009,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",5009,2,"DA",log,selection_mouse +2595,12989025,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",5006,5," CUDA",log,selection_mouse +2596,12989048,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4999,12,"latform CUDA",log,selection_mouse +2597,12989060,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4993,18," for platform CUDA",log,selection_mouse +2598,12989083,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4912,99,"ributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2599,12989098,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4749,262,"_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2600,12989116,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4744,267,"0/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2601,12989130,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4695,316,"nt(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2602,12989160,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4544,467,"oject-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2603,12989165,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4540,471,"k-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2604,12989192,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4498,513,"ion.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2605,12989192,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4495,516,"ration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2606,12989229,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4340,671,"e/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2607,12989230,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4338,673,"ome/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2608,12989262,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4298,713," _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2609,12989262,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4297,714,"= _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2610,12989298,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4296,715," = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2611,12989338,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4295,716,"d = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2612,12989338,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4147,864,"hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2613,12989375,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4146,865,"/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2614,12989415,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4145,866,"""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2615,12989455,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4109,902,"ack (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2616,12989456,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4108,903,"back (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2617,12989490,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4107,904,"eback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2618,12989491,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4106,905,"ceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2619,12989492,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4105,906,"aceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2620,12989527,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4104,907,"raceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2621,12989528,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4103,908,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2622,12989658,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",3892,1119,"2025-08-14 13:25:47.802222: W external/xla/xla/service/platform_util.cc:220] unable to create StreamExecutor for CUDA:3: CUDA error: Failed call to cuDeviceGet: CUDA_ERROR_INVALID_DEVICE: invalid device ordinal\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_dbd0378/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA",log,selection_mouse +2623,12998021,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4322,0,"",log,selection_mouse +2624,12998058,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3422132.log",4321,0,"",log,selection_command +2625,14273636,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",0,0,"",shellscript,tab +2626,14274240,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",0,0,"",shellscript,tab +2627,14276107,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +2628,14276810,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",0,0,"",log,tab +2629,14278829,"sample.py",0,0,"from dataclasses import dataclass\nimport time\nimport os\nimport optax\n\nimport dm_pix as pix\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nimport numpy as np\nimport orbax.checkpoint as ocp\nfrom PIL import Image, ImageDraw\nimport tyro\nfrom flax import nnx\n\nfrom genie import Genie\nfrom utils.dataloader import get_dataloader\n\n\n@dataclass\nclass Args:\n # Experiment\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data/coinrun_episodes""\n checkpoint: str = """"\n # Sampling\n batch_size: int = 1\n maskgit_steps: int = 25\n temperature: float = 1.0\n sample_argmax: bool = True\n start_frame: int = 0\n # Tokenizer checkpoint\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n # LAM checkpoint\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n # Dynamics checkpoint\n dyna_type: str = ""maskgit""\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n # Additional parameters\n darkness_threshold: float = 0.0\n\n\nargs = tyro.cli(Args)\n\nif __name__ == ""__main__"":\n """"""\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n jax.distributed.initialize()\n\n rng = jax.random.key(args.seed)\n\n # --- Load Genie checkpoint ---\n rngs = nnx.Rngs(rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=False,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n # FIXME (f.srambical): implement spatiotemporal KV caching and set decode=True\n decode=False,\n rngs=rngs,\n )\n\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n dummy_tx = optax.adamw(\n learning_rate=optax.linear_schedule(0.0001, 0.0001, 10000),\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n dummy_optimizer = nnx.Optimizer(genie, dummy_tx)\n\n abstract_optimizer = nnx.eval_shape(lambda: dummy_optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(dummy_optimizer, restored_optimizer_state)\n\n # --- Define sampling function ---\n def _sampling_fn(model: Genie, batch: dict) -> jax.Array:\n """"""Runs Genie.sample with pre-defined generation hyper-parameters.""""""\n if args.dyna_type == ""maskgit"":\n return model.sample(\n batch,\n args.seq_len,\n args.maskgit_steps,\n args.temperature,\n args.sample_argmax,\n )\n elif args.dyna_type == ""causal"":\n return model.sample_causal(\n batch,\n args.seq_len,\n args.temperature,\n args.sample_argmax,\n )\n else:\n raise ValueError(f""Invalid dynamics type: {args.dyna_type}"")\n\n # --- Define autoregressive sampling loop ---\n def _autoreg_sample(genie, rng, video_batch_BSHWC, action_batch_E):\n input_video_BTHWC = video_batch_BSHWC[:, : args.start_frame]\n rng, _rng = jax.random.split(rng)\n batch = dict(videos=input_video_BTHWC, latent_actions=action_batch_E, rng=_rng)\n generated_vid_BSHWC = _sampling_fn(genie, batch)\n return generated_vid_BSHWC\n\n # --- Get video + latent actions ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n args.batch_size,\n args.image_height,\n args.image_width,\n args.image_channels,\n # We don't use workers in order to avoid grain shutdown issues (https://github.com/google/grain/issues/398)\n num_workers=0,\n prefetch_buffer_size=1,\n seed=args.seed,\n darkness_threshold=args.darkness_threshold,\n )\n dataloader = iter(dataloader)\n video_batch_BSHWC = next(dataloader)\n gt_video = jnp.asarray(video_batch_BSHWC, dtype=jnp.float32) / 255.0\n video_batch_BSHWC = gt_video.astype(args.dtype)\n # Get latent actions for all videos in the batch\n batch = dict(videos=video_batch_BSHWC)\n action_batch_E = genie.vq_encode(batch, training=False)\n\n # --- Sample + evaluate video ---\n recon_video_BSHWC = _autoreg_sample(genie, rng, video_batch_BSHWC, action_batch_E)\n recon_video_BSHWC = recon_video_BSHWC.astype(jnp.float32)\n gt = gt_video[:, : recon_video_BSHWC.shape[1]].clip(0, 1).reshape(-1, *gt_video.shape[2:])\n recon = recon_video_BSHWC.clip(0, 1).reshape(-1, *recon_video_BSHWC.shape[2:])\n ssim = jnp.asarray(\n pix.ssim(gt[:, args.start_frame:], recon[:, args.start_frame:])\n ).mean()\n print(f""SSIM: {ssim}"")\n\n # --- Construct video ---\n true_videos = (gt_video * 255).astype(np.uint8)\n pred_videos = (recon_video_BSHWC * 255).astype(np.uint8)\n video_comparison = np.zeros((2, *recon_video_BSHWC.shape), dtype=np.uint8)\n video_comparison[0] = true_videos[:, : args.seq_len]\n video_comparison[1] = pred_videos\n frames = einops.rearrange(video_comparison, ""n b t h w c -> t (b h) (n w) c"")\n\n # --- Save video ---\n imgs = [Image.fromarray(img) for img in frames]\n # Write actions on each frame, on each row (i.e., for each video in the batch, on the GT row)\n B, S, _, _, _ = video_batch_BSHWC.shape\n action_batch_BSm11 = jnp.reshape(action_batch_E, (B, S-1, 1))\n for t, img in enumerate(imgs[1:]):\n d = ImageDraw.Draw(img)\n for row in range(action_batch_BSm11.shape[0]):\n action = action_batch_BSm11[row, t, 0]\n y_offset = row * video_batch_BSHWC.shape[2] + 2\n d.text((2, y_offset), f""{action}"", fill=255)\n imgs[0].save(\n f""generation_{time.time()}.gif"",\n save_all=True,\n append_images=imgs[1:],\n duration=250,\n loop=0,\n )\n",python,tab +2630,14279643,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py",0,0,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n darkness_threshold: float = 0.0\n # Tokenizer\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_type: str = ""maskgit"" # supported options: maskgit, causal\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n wandb_id: str = """"\n\n\nargs = tyro.cli(Args)\n\n\ndef dynamics_loss_fn(\n model: Genie, inputs: dict\n) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n """"""Compute masked dynamics loss""""""\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n model.train()\n outputs = model(inputs, training=True)\n mask = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n gt = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]), size=args.num_latent_actions, fill_value=0\n )\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]), size=args.num_patch_latents, fill_value=0\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n psnr=psnr,\n ssim=ssim,\n codebook_usage_lam=codebook_usage_lam,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n\n@nnx.jit\ndef train_step(\n model: Genie, optimizer: nnx.Optimizer, inputs: dict\n) -> tuple[jax.Array, jax.Array, dict]:\n """"""Update state and compute metrics""""""\n\n def loss_fn(model: Genie) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n return dynamics_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(model)\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n decode=False,\n rngs=rngs,\n )\n\n _, params, _ = nnx.split(genie, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n optimizer = nnx.Optimizer(genie, tx)\n del genie\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n darkness_threshold=args.darkness_threshold,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n optimizer = restore_genie_components(optimizer, replicated_sharding, rng, args)\n # NOTE: We have to remove the (unused) tokenizer vq dropout due flax.nnx lazily initializing modules.\n # Specifically, the first dynamics model checkpoint will contain the vq dropout module,\n # but the first full restore will fail due to nnx not initializing the module when\n # dropout is set to 0.0.\n del optimizer.model.tokenizer.vq.drop\n\n # --- TRAIN LOOP ---\n dataloader = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in grain_iterator\n )\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng_mask = jax.random.split(rng, 2)\n inputs = dict(videos=videos, mask_rng=_rng_mask)\n loss, recon, metrics = train_step(optimizer.model, optimizer, inputs)\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n grain_iterator # type: ignore\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +2631,14283777,"TERMINAL",0,0,"",,terminal_focus +2632,14286008,"TERMINAL",0,0,"source /home/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/bin/activate",,terminal_command +2633,14286027,"TERMINAL",0,0,"]633;E;2025-08-14 14:35:39 source /home/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/bin/activate;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +2634,14287602,"TERMINAL",0,0,"queue",,terminal_command +2635,14287690,"TERMINAL",0,0,"]633;E;2025-08-14 14:35:40 queue;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 14:35:40 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-20:35:39\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-00:46:44\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 2:22:30\t 1 hkn07133421982 accelerat train_to tum_cte0 R 2:34:30\t 4 hkn[0724,0726,0728,0731]3421973 accelerat train_to tum_cte0 R 2:34:39\t 2 hkn[0722-0723]3421964 accelerat train_to tum_cte0 R 2:34:54\t 1 hkn0720",,terminal_output +2636,14288736,"TERMINAL",0,0,"140511405",,terminal_output +2637,14289774,"TERMINAL",0,0,"2162216",,terminal_output +2638,14290822,"TERMINAL",0,0,"3384438",,terminal_output +2639,14291871,"TERMINAL",0,0,"5495549",,terminal_output +2640,14292915,"TERMINAL",0,0,"65506655:00",,terminal_output +2641,14293966,"TERMINAL",0,0,"7617761",,terminal_output +2642,14295154,"TERMINAL",0,0,"8728872",,terminal_output +2643,14296072,"TERMINAL",0,0,"9839983",,terminal_output +2644,14297095,"TERMINAL",0,0,"5094404094",,terminal_output +2645,14298136,"TERMINAL",0,0,"150511505",,terminal_output +2646,14299177,"TERMINAL",0,0,"2162216",,terminal_output +2647,14300218,"TERMINAL",0,0,"3273327",,terminal_output +2648,14301287,"TERMINAL",0,0,"4384438",,terminal_output +2649,14302316,"TERMINAL",0,0,"5495549",,terminal_output +2650,14303442,"TERMINAL",0,0,"657:0066510",,terminal_output +2651,14304484,"TERMINAL",0,0,"7617761",,terminal_output +2652,14305423,"TERMINAL",0,0,"8728872",,terminal_output +2653,14306506,"TERMINAL",0,0,"9839983",,terminal_output +2654,14307540,"TERMINAL",0,0,"6:0094505094",,terminal_output +2655,14308564,"TERMINAL",0,0,"16:005115:005",,terminal_output +2656,14309694,"TERMINAL",0,0,"2162216",,terminal_output +2657,14310766,"TERMINAL",0,0,"3273327",,terminal_output +2658,14311852,"TERMINAL",0,0,"4384438",,terminal_output +2659,14312728,"TERMINAL",0,0,"5495549",,terminal_output +2660,14313775,"TERMINAL",0,0,"651066520",,terminal_output +2661,14314824,"TERMINAL",0,0,"7728872",,terminal_output +2662,14315873,"TERMINAL",0,0,"9839983",,terminal_output +2663,14316920,"TERMINAL",0,0,"10943:005:0094",,terminal_output +2664,14317972,"TERMINAL",0,0,"110511105",,terminal_output +2665,14319053,"TERMINAL",0,0,"2162216",,terminal_output +2666,14320153,"TERMINAL",0,0,"3273327",,terminal_output +2667,14321109,"TERMINAL",0,0,"4384438",,terminal_output +2668,14322157,"TERMINAL",0,0,"5495549",,terminal_output +2669,14323230,"TERMINAL",0,0,"652066530",,terminal_output +2670,14324334,"TERMINAL",0,0,"7617761",,terminal_output +2671,14325356,"TERMINAL",0,0,"8728872",,terminal_output +2672,14326381,"TERMINAL",0,0,"9839983",,terminal_output +2673,14327404,"TERMINAL",0,0,"2094101094",,terminal_output +2674,14328431,"TERMINAL",0,0,"120511205",,terminal_output +2675,14329554,"TERMINAL",0,0,"2162216",,terminal_output +2676,14330580,"TERMINAL",0,0,"3273327",,terminal_output +2677,14331604,"TERMINAL",0,0,"4384438",,terminal_output +2678,14332630,"TERMINAL",0,0,"5495549",,terminal_output +2679,14333639,"TERMINAL",0,0,"653066540",,terminal_output +2680,14334882,"TERMINAL",0,0,"7617761",,terminal_output +2681,14335741,"TERMINAL",0,0,"8728872",,terminal_output +2682,14336764,"TERMINAL",0,0,"9839983",,terminal_output +2683,14337838,"TERMINAL",0,0,"303052121305",,terminal_output +2684,14338977,"TERMINAL",0,0,"2162216",,terminal_output +2685,14339999,"TERMINAL",0,0,"3273327",,terminal_output +2686,14341024,"TERMINAL",0,0,"4384438",,terminal_output +2687,14342050,"TERMINAL",0,0,"5495549",,terminal_output +2688,14343077,"TERMINAL",0,0,"654066550",,terminal_output +2689,14344116,"TERMINAL",0,0,"7617761",,terminal_output +2690,14345206,"TERMINAL",0,0,"8728872",,terminal_output +2691,14346354,"TERMINAL",0,0,"9839983",,terminal_output +2692,14347257,"TERMINAL",0,0,"4094303094",,terminal_output +2693,14348281,"TERMINAL",0,0,"140511405",,terminal_output +2694,14349423,"TERMINAL",0,0,"2162216",,terminal_output +2695,14350446,"TERMINAL",0,0,"3273327",,terminal_output +2696,14351471,"TERMINAL",0,0,"4384438",,terminal_output +2697,14352500,"TERMINAL",0,0,"5495549",,terminal_output +2698,14353501,"TERMINAL",0,0,"65506656:00",,terminal_output +2699,14354550,"TERMINAL",0,0,"7617761",,terminal_output +2700,14355674,"TERMINAL",0,0,"8728872",,terminal_output +2701,14356697,"TERMINAL",0,0,"9839983",,terminal_output +2702,14357681,"TERMINAL",0,0,"5094404094",,terminal_output +2703,14358743,"TERMINAL",0,0,"150511505",,terminal_output +2704,14359770,"TERMINAL",0,0,"2162216",,terminal_output +2705,14360805,"TERMINAL",0,0,"3273327",,terminal_output +2706,14361958,"TERMINAL",0,0,"4495549",,terminal_output +2707,14362888,"TERMINAL",0,0,"658:0066510",,terminal_output +2708,14363935,"TERMINAL",0,0,"7617761",,terminal_output +2709,14365002,"TERMINAL",0,0,"8728872",,terminal_output +2710,14366030,"TERMINAL",0,0,"9839983",,terminal_output +2711,14367151,"TERMINAL",0,0,"7:0094505094",,terminal_output +2712,14368168,"TERMINAL",0,0,"17:005116:005",,terminal_output +2713,14369189,"TERMINAL",0,0,"2162216",,terminal_output +2714,14370245,"TERMINAL",0,0,"3273327",,terminal_output +2715,14371269,"TERMINAL",0,0,"4384438",,terminal_output +2716,14372315,"TERMINAL",0,0,"5495549",,terminal_output +2717,14373383,"TERMINAL",0,0,"651066520",,terminal_output +2718,14374409,"TERMINAL",0,0,"7617761",,terminal_output +2719,14375472,"TERMINAL",0,0,"8728872",,terminal_output +2720,14376557,"TERMINAL",0,0,"9839983",,terminal_output +2721,14377533,"TERMINAL",0,0,"10944:006:0094",,terminal_output +2722,14378709,"TERMINAL",0,0,"110511105",,terminal_output +2723,14379618,"TERMINAL",0,0,"2162216",,terminal_output +2724,14380757,"TERMINAL",0,0,"3273327",,terminal_output +2725,14381782,"TERMINAL",0,0,"4384438",,terminal_output +2726,14382744,"TERMINAL",0,0,"5495549",,terminal_output +2727,14383786,"TERMINAL",0,0,"652066530",,terminal_output +2728,14384854,"TERMINAL",0,0,"7728872",,terminal_output +2729,14385987,"TERMINAL",0,0,"9839983",,terminal_output +2730,14387003,"TERMINAL",0,0,"2094101094",,terminal_output +2731,14387971,"TERMINAL",0,0,"120511205",,terminal_output +2732,14389010,"TERMINAL",0,0,"2162216",,terminal_output +2733,14390176,"TERMINAL",0,0,"3273327",,terminal_output +2734,14391100,"TERMINAL",0,0,"4384438",,terminal_output +2735,14392141,"TERMINAL",0,0,"5495549",,terminal_output +2736,14393182,"TERMINAL",0,0,"653066540",,terminal_output +2737,14394232,"TERMINAL",0,0,"7617761",,terminal_output +2738,14395298,"TERMINAL",0,0,"8728872",,terminal_output +2739,14396327,"TERMINAL",0,0,"9839983",,terminal_output +2740,14397365,"TERMINAL",0,0,"3094202094",,terminal_output +2741,14398412,"TERMINAL",0,0,"130511305",,terminal_output +2742,14399498,"TERMINAL",0,0,"2162216",,terminal_output +2743,14400522,"TERMINAL",0,0,"3273327",,terminal_output +2744,14401647,"TERMINAL",0,0,"4384438",,terminal_output +2745,14402672,"TERMINAL",0,0,"5495549",,terminal_output +2746,14403942,"TERMINAL",0,0,"664177651",,terminal_output +2747,14405028,"TERMINAL",0,0,"8728872",,terminal_output +2748,14406081,"TERMINAL",0,0,"9839983",,terminal_output +2749,14407056,"TERMINAL",0,0,"4094303094",,terminal_output +2750,14408128,"TERMINAL",0,0,"140511405",,terminal_output +2751,14409156,"TERMINAL",0,0,"2162216",,terminal_output +2752,14410192,"TERMINAL",0,0,"3273327",,terminal_output +2753,14411273,"TERMINAL",0,0,"4384438",,terminal_output +2754,14412262,"TERMINAL",0,0,"5495549",,terminal_output +2755,14413320,"TERMINAL",0,0,"65506657:00",,terminal_output +2756,14414448,"TERMINAL",0,0,"7617761",,terminal_output +2757,14415470,"TERMINAL",0,0,"8728872",,terminal_output +2758,14416493,"TERMINAL",0,0,"9839983",,terminal_output +2759,14417520,"TERMINAL",0,0,"5094404094",,terminal_output +2760,14418542,"TERMINAL",0,0,"150511505",,terminal_output +2761,14419568,"TERMINAL",0,0,"2162216",,terminal_output +2762,14420898,"TERMINAL",0,0,"3273327",,terminal_output +2763,14421729,"TERMINAL",0,0,"4384438",,terminal_output +2764,14422688,"TERMINAL",0,0,"5495549",,terminal_output +2765,14423765,"TERMINAL",0,0,"659:0066510",,terminal_output +2766,14424768,"TERMINAL",0,0,"7617761",,terminal_output +2767,14425915,"TERMINAL",0,0,"8839983",,terminal_output +2768,14426856,"TERMINAL",0,0,"8:0094505094",,terminal_output +2769,14427897,"TERMINAL",0,0,"18:005117:005",,terminal_output +2770,14429128,"TERMINAL",0,0,"2162216",,terminal_output +2771,14430012,"TERMINAL",0,0,"3273327",,terminal_output +2772,14431139,"TERMINAL",0,0,"4384438",,terminal_output +2773,14432101,"TERMINAL",0,0,"5495549",,terminal_output +2774,14433186,"TERMINAL",0,0,"651066520",,terminal_output +2775,14434170,"TERMINAL",0,0,"7617761",,terminal_output +2776,14435190,"TERMINAL",0,0,"8728872",,terminal_output +2777,14436269,"TERMINAL",0,0,"9839983",,terminal_output +2778,14437285,"TERMINAL",0,0,"10945:007:0094",,terminal_output +2779,14438324,"TERMINAL",0,0,"110511105",,terminal_output +2780,14439432,"TERMINAL",0,0,"2162216",,terminal_output +2781,14440434,"TERMINAL",0,0,"3273327",,terminal_output +2782,14441479,"TERMINAL",0,0,"4384438",,terminal_output +2783,14442510,"TERMINAL",0,0,"5495549",,terminal_output +2784,14443631,"TERMINAL",0,0,"652066530",,terminal_output +2785,14444655,"TERMINAL",0,0,"7617761",,terminal_output +2786,14445678,"TERMINAL",0,0,"8728872",,terminal_output +2787,14446705,"TERMINAL",0,0,"9839983",,terminal_output +2788,14447728,"TERMINAL",0,0,"2094101094",,terminal_output +2789,14448755,"TERMINAL",0,0,"120511205",,terminal_output +2790,14449784,"TERMINAL",0,0,"2162216",,terminal_output +2791,14450900,"TERMINAL",0,0,"3384438",,terminal_output +2792,14452005,"TERMINAL",0,0,"5495549",,terminal_output +2793,14452922,"TERMINAL",0,0,"653066540",,terminal_output +2794,14453972,"TERMINAL",0,0,"7617761",,terminal_output +2795,14455133,"TERMINAL",0,0,"8728872",,terminal_output +2796,14456134,"TERMINAL",0,0,"9839983",,terminal_output +2797,14457192,"TERMINAL",0,0,"3094202094",,terminal_output +2798,14458227,"TERMINAL",0,0,"130511305",,terminal_output +2799,14459284,"TERMINAL",0,0,"2162216",,terminal_output +2800,14460425,"TERMINAL",0,0,"3273327",,terminal_output +2801,14461380,"TERMINAL",0,0,"4384438",,terminal_output +2802,14462473,"TERMINAL",0,0,"5495549",,terminal_output +2803,14463499,"TERMINAL",0,0,"654066550",,terminal_output +2804,14464524,"TERMINAL",0,0,"7617761",,terminal_output +2805,14465753,"TERMINAL",0,0,"8728872",,terminal_output +2806,14466671,"TERMINAL",0,0,"9839983",,terminal_output +2807,14467659,"TERMINAL",0,0,"4094303094",,terminal_output +2808,14468721,"TERMINAL",0,0,"140511405",,terminal_output +2809,14469738,"TERMINAL",0,0,"2162216",,terminal_output +2810,14470869,"TERMINAL",0,0,"3273327",,terminal_output +2811,14471898,"TERMINAL",0,0,"4495549",,terminal_output +2812,14472882,"TERMINAL",0,0,"65506658:00",,terminal_output +2813,14473943,"TERMINAL",0,0,"7617761",,terminal_output +2814,14474969,"TERMINAL",0,0,"8728872",,terminal_output +2815,14476129,"TERMINAL",0,0,"9839983",,terminal_output +2816,14477078,"TERMINAL",0,0,"5094404094",,terminal_output +2817,14478137,"TERMINAL",0,0,"150511505",,terminal_output +2818,14479144,"TERMINAL",0,0,"2162216",,terminal_output +2819,14480194,"TERMINAL",0,0,"3273327",,terminal_output +2820,14481316,"TERMINAL",0,0,"4384438",,terminal_output +2821,14482294,"TERMINAL",0,0,"5495549",,terminal_output +2822,14483340,"TERMINAL",0,0,"6550:0066510",,terminal_output +2823,14484489,"TERMINAL",0,0,"7617761",,terminal_output +2824,14485620,"TERMINAL",0,0,"8728872",,terminal_output +2825,14486622,"TERMINAL",0,0,"9839983",,terminal_output +2826,14487766,"TERMINAL",0,0,"9:0094505094",,terminal_output +2827,14488790,"TERMINAL",0,0,"19:005118:005",,terminal_output +2828,14489768,"TERMINAL",0,0,"2162216",,terminal_output +2829,14490818,"TERMINAL",0,0,"3384438",,terminal_output +2830,14491903,"TERMINAL",0,0,"5495549",,terminal_output +2831,14492894,"TERMINAL",0,0,"651066520",,terminal_output +2832,14493952,"TERMINAL",0,0,"7617761",,terminal_output +2833,14494987,"TERMINAL",0,0,"8728872",,terminal_output +2834,14496063,"TERMINAL",0,0,"9839983",,terminal_output +2835,14497117,"TERMINAL",0,0,"10946:008:0094",,terminal_output +2836,14498121,"TERMINAL",0,0,"110511105",,terminal_output +2837,14499153,"TERMINAL",0,0,"2162216",,terminal_output +2838,14500200,"TERMINAL",0,0,"3273327",,terminal_output +2839,14501245,"TERMINAL",0,0,"4384438",,terminal_output +2840,14502280,"TERMINAL",0,0,"5495549",,terminal_output +2841,14503340,"TERMINAL",0,0,"652066530",,terminal_output +2842,14504457,"TERMINAL",0,0,"7617761",,terminal_output +2843,14505426,"TERMINAL",0,0,"8728872",,terminal_output +2844,14506505,"TERMINAL",0,0,"9839983",,terminal_output +2845,14507505,"TERMINAL",0,0,"2094101094",,terminal_output +2846,14508553,"TERMINAL",0,0,"120511205",,terminal_output +2847,14509784,"TERMINAL",0,0,"2162216",,terminal_output +2848,14510807,"TERMINAL",0,0,"3273327",,terminal_output +2849,14511729,"TERMINAL",0,0,"4384438",,terminal_output +2850,14512756,"TERMINAL",0,0,"5495549",,terminal_output +2851,14513884,"TERMINAL",0,0,"653066540",,terminal_output +2852,14514905,"TERMINAL",0,0,"7617761",,terminal_output +2853,14515910,"TERMINAL",0,0,"8839983",,terminal_output +2854,14516903,"TERMINAL",0,0,"3094202094",,terminal_output +2855,14517945,"TERMINAL",0,0,"130511305",,terminal_output +2856,14519042,"TERMINAL",0,0,"2162216",,terminal_output +2857,14520129,"TERMINAL",0,0,"3273327",,terminal_output +2858,14521251,"TERMINAL",0,0,"4384438",,terminal_output +2859,14522278,"TERMINAL",0,0,"5495549",,terminal_output +2860,14523198,"TERMINAL",0,0,"654066550",,terminal_output +2861,14524225,"TERMINAL",0,0,"7617761",,terminal_output +2862,14525274,"TERMINAL",0,0,"8728872",,terminal_output +2863,14526318,"TERMINAL",0,0,"9839983",,terminal_output +2864,14527395,"TERMINAL",0,0,"4094303094",,terminal_output +2865,14528531,"TERMINAL",0,0,"140511405",,terminal_output +2866,14529548,"TERMINAL",0,0,"2162216",,terminal_output +2867,14530571,"TERMINAL",0,0,"3273327",,terminal_output +2868,14531595,"TERMINAL",0,0,"4384438",,terminal_output +2869,14532724,"TERMINAL",0,0,"5495549",,terminal_output +2870,14533691,"TERMINAL",0,0,"65506659:00",,terminal_output +2871,14534769,"TERMINAL",0,0,"7617761",,terminal_output +2872,14535894,"TERMINAL",0,0,"8728872",,terminal_output +2873,14536800,"TERMINAL",0,0,"9839983",,terminal_output +2874,14537829,"TERMINAL",0,0,"505054141505",,terminal_output +2875,14538968,"TERMINAL",0,0,"2162216",,terminal_output +2876,14539992,"TERMINAL",0,0,"3273327",,terminal_output +2877,14540947,"TERMINAL",0,0,"4384438",,terminal_output +2878,14541988,"TERMINAL",0,0,"5495549",,terminal_output +2879,14543061,"TERMINAL",0,0,"651:0066510",,terminal_output +2880,14544078,"TERMINAL",0,0,"7617761",,terminal_output +2881,14545252,"TERMINAL",0,0,"8728872",,terminal_output +2882,14546164,"TERMINAL",0,0,"9839983",,terminal_output +2883,14547202,"TERMINAL",0,0,"40:0094505094",,terminal_output +2884,14548246,"TERMINAL",0,0,"140:005119:005",,terminal_output +2885,14549313,"TERMINAL",0,0,"2162216",,terminal_output +2886,14550340,"TERMINAL",0,0,"3273327",,terminal_output +2887,14551494,"TERMINAL",0,0,"4384438",,terminal_output +2888,14552588,"TERMINAL",0,0,"5495549",,terminal_output +2889,14553609,"TERMINAL",0,0,"651066520",,terminal_output +2890,14554839,"TERMINAL",0,0,"7617761",,terminal_output +2891,14555718,"TERMINAL",0,0,"8728872",,terminal_output +2892,14556806,"TERMINAL",0,0,"9839983",,terminal_output +2893,14557828,"TERMINAL",0,0,"10947:009:0094",,terminal_output +2894,14558834,"TERMINAL",0,0,"110511105",,terminal_output +2895,14560063,"TERMINAL",0,0,"2273327",,terminal_output +2896,14560915,"TERMINAL",0,0,"4384438",,terminal_output +2897,14561965,"TERMINAL",0,0,"5495549",,terminal_output +2898,14563030,"TERMINAL",0,0,"652066530",,terminal_output +2899,14564071,"TERMINAL",0,0,"7617761",,terminal_output +2900,14565266,"TERMINAL",0,0,"8728872",,terminal_output +2901,14566268,"TERMINAL",0,0,"9839983",,terminal_output +2902,14567356,"TERMINAL",0,0,"2094101094",,terminal_output +2903,14568464,"TERMINAL",0,0,"120511205",,terminal_output +2904,14569483,"TERMINAL",0,0,"2162216",,terminal_output +2905,14570509,"TERMINAL",0,0,"3273327",,terminal_output +2906,14571550,"TERMINAL",0,0,"4384438",,terminal_output +2907,14572657,"TERMINAL",0,0,"5495549",,terminal_output +2908,14573682,"TERMINAL",0,0,"653066540",,terminal_output +2909,14574706,"TERMINAL",0,0,"7617761",,terminal_output +2910,14575766,"TERMINAL",0,0,"8728872",,terminal_output +2911,14576803,"TERMINAL",0,0,"9839983",,terminal_output +2912,14577856,"TERMINAL",0,0,"303052121305",,terminal_output +2913,14578958,"TERMINAL",0,0,"2162216",,terminal_output +2914,14580168,"TERMINAL",0,0,"3273327",,terminal_output +2915,14581051,"TERMINAL",0,0,"4384438",,terminal_output +2916,14582065,"TERMINAL",0,0,"5495549",,terminal_output +2917,14583205,"TERMINAL",0,0,"654066550",,terminal_output +2918,14584159,"TERMINAL",0,0,"7617761",,terminal_output +2919,14585210,"TERMINAL",0,0,"8728872",,terminal_output +2920,14586283,"TERMINAL",0,0,"9839983",,terminal_output +2921,14587313,"TERMINAL",0,0,"4094303094",,terminal_output +2922,14588361,"TERMINAL",0,0,"140511405",,terminal_output +2923,14589412,"TERMINAL",0,0,"2162216",,terminal_output +2924,14590475,"TERMINAL",0,0,"3273327",,terminal_output +2925,14591506,"TERMINAL",0,0,"4384438",,terminal_output +2926,14592553,"TERMINAL",0,0,"5495549",,terminal_output +2927,14593649,"TERMINAL",0,0,"655066540:00",,terminal_output +2928,14594651,"TERMINAL",0,0,"7617761",,terminal_output +2929,14595904,"TERMINAL",0,0,"8728872",,terminal_output +2930,14596741,"TERMINAL",0,0,"9839983",,terminal_output +2931,14597830,"TERMINAL",0,0,"5094404094",,terminal_output +2932,14598877,"TERMINAL",0,0,"151622516",,terminal_output +2933,14599898,"TERMINAL",0,0,"3273327",,terminal_output +2934,14601031,"TERMINAL",0,0,"4384438",,terminal_output +2935,14601980,"TERMINAL",0,0,"5495549",,terminal_output +2936,14603070,"TERMINAL",0,0,"652:0066510",,terminal_output +2937,14604097,"TERMINAL",0,0,"7617761",,terminal_output +2938,14605326,"TERMINAL",0,0,"8728872",,terminal_output +2939,14606349,"TERMINAL",0,0,"9839983",,terminal_output +2940,14607211,"TERMINAL",0,0,"1:0094505094",,terminal_output +2941,14608255,"TERMINAL",0,0,"11:0051140:005",,terminal_output +2942,14609303,"TERMINAL",0,0,"2162216",,terminal_output +2943,14610442,"TERMINAL",0,0,"3273327",,terminal_output +2944,14611390,"TERMINAL",0,0,"4384438",,terminal_output +2945,14612558,"TERMINAL",0,0,"5495549",,terminal_output +2946,14613514,"TERMINAL",0,0,"651066520",,terminal_output +2947,14614538,"TERMINAL",0,0,"7617761",,terminal_output +2948,14615611,"TERMINAL",0,0,"8728872",,terminal_output +2949,14616792,"TERMINAL",0,0,"9839983",,terminal_output +2950,14617817,"TERMINAL",0,0,"10948:0040:0094",,terminal_output +2951,14618710,"TERMINAL",0,0,"110511105",,terminal_output +2952,14619761,"TERMINAL",0,0,"2162216",,terminal_output +2953,14620798,"TERMINAL",0,0,"3273327",,terminal_output +2954,14621912,"TERMINAL",0,0,"4495549",,terminal_output +2955,14622890,"TERMINAL",0,0,"652066530",,terminal_output +2956,14624208,"TERMINAL",0,0,"7617761",,terminal_output +2957,14625084,"TERMINAL",0,0,"8728872",,terminal_output +2958,14626213,"TERMINAL",0,0,"9839983",,terminal_output +2959,14627168,"TERMINAL",0,0,"2094101094",,terminal_output +2960,14628159,"TERMINAL",0,0,"120511205",,terminal_output +2961,14629169,"TERMINAL",0,0,"2162216",,terminal_output +2962,14630311,"TERMINAL",0,0,"3273327",,terminal_output +2963,14631346,"TERMINAL",0,0,"4384438",,terminal_output +2964,14632285,"TERMINAL",0,0,"5495549",,terminal_output +2965,14633378,"TERMINAL",0,0,"653066540",,terminal_output +2966,14634365,"TERMINAL",0,0,"7617761",,terminal_output +2967,14635427,"TERMINAL",0,0,"8728872",,terminal_output +2968,14636452,"TERMINAL",0,0,"9839983",,terminal_output +2969,14637484,"TERMINAL",0,0,"3094202094",,terminal_output +2970,14638602,"TERMINAL",0,0,"130511305",,terminal_output +2971,14639626,"TERMINAL",0,0,"2162216",,terminal_output +2972,14640760,"TERMINAL",0,0,"3273327",,terminal_output +2973,14641780,"TERMINAL",0,0,"4384438",,terminal_output +2974,14642802,"TERMINAL",0,0,"5495549",,terminal_output +2975,14643824,"TERMINAL",0,0,"654066550",,terminal_output +2976,14644847,"TERMINAL",0,0,"7617761",,terminal_output +2977,14645832,"TERMINAL",0,0,"8839983",,terminal_output +2978,14646890,"TERMINAL",0,0,"4094303094",,terminal_output +2979,14647932,"TERMINAL",0,0,"140511405",,terminal_output +2980,14648978,"TERMINAL",0,0,"2162216",,terminal_output +2981,14650017,"TERMINAL",0,0,"3273327",,terminal_output +2982,14651055,"TERMINAL",0,0,"4384438",,terminal_output +2983,14652150,"TERMINAL",0,0,"5495549",,terminal_output +2984,14653177,"TERMINAL",0,0,"65506651:00",,terminal_output +2985,14654180,"TERMINAL",0,0,"7617761",,terminal_output +2986,14655295,"TERMINAL",0,0,"8728872",,terminal_output +2987,14656277,"TERMINAL",0,0,"9839983",,terminal_output +2988,14657326,"TERMINAL",0,0,"5094404094",,terminal_output +2989,14658370,"TERMINAL",0,0,"150511505",,terminal_output +2990,14659410,"TERMINAL",0,0,"2162216",,terminal_output +2991,14660450,"TERMINAL",0,0,"3273327",,terminal_output +2992,14661541,"TERMINAL",0,0,"4384438",,terminal_output +2993,14662538,"TERMINAL",0,0,"5495549",,terminal_output +2994,14663586,"TERMINAL",0,0,"653:0066510",,terminal_output +2995,14664719,"TERMINAL",0,0,"7617761",,terminal_output +2996,14665741,"TERMINAL",0,0,"8728872",,terminal_output +2997,14666717,"TERMINAL",0,0,"9839983",,terminal_output +2998,14667786,"TERMINAL",0,0,"2:0094505094",,terminal_output +2999,14668811,"TERMINAL",0,0,"12:005111:005",,terminal_output +3000,14670040,"TERMINAL",0,0,"2273327",,terminal_output +3001,14670961,"TERMINAL",0,0,"4384438",,terminal_output +3002,14671925,"TERMINAL",0,0,"5495549",,terminal_output +3003,14673021,"TERMINAL",0,0,"651066520",,terminal_output +3004,14674079,"TERMINAL",0,0,"7617761",,terminal_output +3005,14675062,"TERMINAL",0,0,"8728872",,terminal_output +3006,14676223,"TERMINAL",0,0,"9839983",,terminal_output +3007,14677211,"TERMINAL",0,0,"10949:001:0094",,terminal_output +3008,14678236,"TERMINAL",0,0,"110511105",,terminal_output +3009,14679260,"TERMINAL",0,0,"2162216",,terminal_output +3010,14680271,"TERMINAL",0,0,"3273327",,terminal_output +3011,14681317,"TERMINAL",0,0,"4384438",,terminal_output +3012,14682362,"TERMINAL",0,0,"5495549",,terminal_output +3013,14683454,"TERMINAL",0,0,"652066530",,terminal_output +3014,14684495,"TERMINAL",0,0,"7617761",,terminal_output +3015,14685493,"TERMINAL",0,0,"8728872",,terminal_output +3016,14686534,"TERMINAL",0,0,"9839983",,terminal_output +3017,14687656,"TERMINAL",0,0,"2094101094",,terminal_output +3018,14688680,"TERMINAL",0,0,"120511205",,terminal_output +3019,14689710,"TERMINAL",0,0,"2162216",,terminal_output +3020,14691032,"TERMINAL",0,0,"3273327",,terminal_output +3021,14691777,"TERMINAL",0,0,"4384438",,terminal_output +3022,14692830,"TERMINAL",0,0,"553066540",,terminal_output +3023,14694001,"TERMINAL",0,0,"7617761",,terminal_output +3024,14694929,"TERMINAL",0,0,"8728872",,terminal_output +3025,14695954,"TERMINAL",0,0,"9839983",,terminal_output +3026,14697004,"TERMINAL",0,0,"3094202094",,terminal_output +3027,14698243,"TERMINAL",0,0,"130511305",,terminal_output +3028,14699092,"TERMINAL",0,0,"2162216",,terminal_output +3029,14700357,"TERMINAL",0,0,"3273327",,terminal_output +3030,14701175,"TERMINAL",0,0,"4384438",,terminal_output +3031,14702213,"TERMINAL",0,0,"5495549",,terminal_output +3032,14703258,"TERMINAL",0,0,"654066550",,terminal_output +3033,14704302,"TERMINAL",0,0,"7617761",,terminal_output +3034,14705342,"TERMINAL",0,0,"8728872",,terminal_output +3035,14706497,"TERMINAL",0,0,"9839983",,terminal_output +3036,14707522,"TERMINAL",0,0,"4094303094",,terminal_output +3037,14708549,"TERMINAL",0,0,"140511405",,terminal_output +3038,14709568,"TERMINAL",0,0,"2162216",,terminal_output +3039,14710592,"TERMINAL",0,0,"3273327",,terminal_output +3040,14711608,"TERMINAL",0,0,"4384438",,terminal_output +3041,14712825,"TERMINAL",0,0,"5495549",,terminal_output +3042,14713766,"TERMINAL",0,0,"65506652:00",,terminal_output +3043,14714893,"TERMINAL",0,0,"7617761",,terminal_output +3044,14715920,"TERMINAL",0,0,"8728872",,terminal_output +3045,14716837,"TERMINAL",0,0,"994404094",,terminal_output +3046,14717889,"TERMINAL",0,0,"5150511505",,terminal_output +3047,14718991,"TERMINAL",0,0,"2162216",,terminal_output +3048,14719978,"TERMINAL",0,0,"3273327",,terminal_output +3049,14721036,"TERMINAL",0,0,"4384438",,terminal_output +3050,14722161,"TERMINAL",0,0,"5495549",,terminal_output +3051,14723104,"TERMINAL",0,0,"654:0066510",,terminal_output +3052,14724319,"TERMINAL",0,0,"7617761",,terminal_output +3053,14725198,"TERMINAL",0,0,"8728872",,terminal_output +3054,14726260,"TERMINAL",0,0,"9839983",,terminal_output +3055,14727386,"TERMINAL",0,0,"3:0094505094",,terminal_output +3056,14728324,"TERMINAL",0,0,"13:005112:005",,terminal_output +3057,14729374,"TERMINAL",0,0,"2162216",,terminal_output +3058,14730418,"TERMINAL",0,0,"3273327",,terminal_output +3059,14731483,"TERMINAL",0,0,"4384438",,terminal_output +3060,14732508,"TERMINAL",0,0,"5495549",,terminal_output +3061,14733631,"TERMINAL",0,0,"651066520",,terminal_output +3062,14734654,"TERMINAL",0,0,"7617761",,terminal_output +3063,14735680,"TERMINAL",0,0,"8728872",,terminal_output +3064,14736703,"TERMINAL",0,0,"9839983",,terminal_output +3065,14737795,"TERMINAL",0,0,"109430:002:0094",,terminal_output +3066,14738852,"TERMINAL",0,0,"110511105",,terminal_output +3067,14739879,"TERMINAL",0,0,"2273327",,terminal_output +3068,14740901,"TERMINAL",0,0,"4384438",,terminal_output +3069,14741923,"TERMINAL",0,0,"5495549",,terminal_output +3070,14742969,"TERMINAL",0,0,"652066530",,terminal_output +3071,14744020,"TERMINAL",0,0,"7617761",,terminal_output +3072,14745229,"TERMINAL",0,0,"8728872",,terminal_output +3073,14746124,"TERMINAL",0,0,"9839983",,terminal_output +3074,14747146,"TERMINAL",0,0,"2094101094",,terminal_output +3075,14748306,"TERMINAL",0,0,"120511205",,terminal_output +3076,14749303,"TERMINAL",0,0,"2162216",,terminal_output +3077,14750287,"TERMINAL",0,0,"3273327",,terminal_output +3078,14751351,"TERMINAL",0,0,"4384438",,terminal_output +3079,14752382,"TERMINAL",0,0,"5495549",,terminal_output +3080,14753429,"TERMINAL",0,0,"653066540",,terminal_output +3081,14754478,"TERMINAL",0,0,"7617761",,terminal_output +3082,14755543,"TERMINAL",0,0,"8728872",,terminal_output +3083,14756640,"TERMINAL",0,0,"9839983",,terminal_output +3084,14757612,"TERMINAL",0,0,"3094202094",,terminal_output +3085,14758656,"TERMINAL",0,0,"130511305",,terminal_output +3086,14759703,"TERMINAL",0,0,"2162216",,terminal_output +3087,14760776,"TERMINAL",0,0,"3273327",,terminal_output +3088,14761997,"TERMINAL",0,0,"4384438",,terminal_output +3089,14762844,"TERMINAL",0,0,"554066550",,terminal_output +3090,14763941,"TERMINAL",0,0,"7617761",,terminal_output +3091,14765228,"TERMINAL",0,0,"8728872",,terminal_output +3092,14766313,"TERMINAL",0,0,"9839983",,terminal_output +3093,14767272,"TERMINAL",0,0,"4094303094",,terminal_output +3094,14768341,"TERMINAL",0,0,"140511405",,terminal_output +3095,14769385,"TERMINAL",0,0,"2162216",,terminal_output +3096,14770429,"TERMINAL",0,0,"3273327",,terminal_output +3097,14771522,"TERMINAL",0,0,"4384438",,terminal_output +3098,14772545,"TERMINAL",0,0,"5495549",,terminal_output +3099,14773569,"TERMINAL",0,0,"65506653:00",,terminal_output +3100,14774693,"TERMINAL",0,0,"7617761",,terminal_output +3101,14775663,"TERMINAL",0,0,"8728872",,terminal_output +3102,14776741,"TERMINAL",0,0,"9839983",,terminal_output +3103,14777749,"TERMINAL",0,0,"5094404094",,terminal_output +3104,14778789,"TERMINAL",0,0,"150511505",,terminal_output +3105,14779837,"TERMINAL",0,0,"2273327",,terminal_output +3106,14780955,"TERMINAL",0,0,"4384438",,terminal_output +3107,14781929,"TERMINAL",0,0,"5495549",,terminal_output +3108,14782988,"TERMINAL",0,0,"655:0066510",,terminal_output +3109,14784054,"TERMINAL",0,0,"7617761",,terminal_output +3110,14785063,"TERMINAL",0,0,"8728872",,terminal_output +3111,14786103,"TERMINAL",0,0,"9839983",,terminal_output +3112,14787220,"TERMINAL",0,0,"4:0094505094",,terminal_output +3113,14788211,"TERMINAL",0,0,"14:005113:005",,terminal_output +3114,14789232,"TERMINAL",0,0,"2162216",,terminal_output +3115,14790269,"TERMINAL",0,0,"3273327",,terminal_output +3116,14791309,"TERMINAL",0,0,"4384438",,terminal_output +3117,14792347,"TERMINAL",0,0,"5495549",,terminal_output +3118,14793392,"TERMINAL",0,0,"651066520",,terminal_output +3119,14794457,"TERMINAL",0,0,"7617761",,terminal_output +3120,14795584,"TERMINAL",0,0,"8728872",,terminal_output +3121,14796539,"TERMINAL",0,0,"9839983",,terminal_output +3122,14797635,"TERMINAL",0,0,"10941:003:0094",,terminal_output +3123,14798658,"TERMINAL",0,0,"110511105",,terminal_output +3124,14799885,"TERMINAL",0,0,"2162216",,terminal_output +3125,14800768,"TERMINAL",0,0,"3273327",,terminal_output +3126,14801760,"TERMINAL",0,0,"4384438",,terminal_output +3127,14802801,"TERMINAL",0,0,"5495549",,terminal_output +3128,14803880,"TERMINAL",0,0,"662177631",,terminal_output +3129,14805004,"TERMINAL",0,0,"8728872",,terminal_output +3130,14806027,"TERMINAL",0,0,"9839983",,terminal_output +3131,14806972,"TERMINAL",0,0,"2094101094",,terminal_output +3132,14808075,"TERMINAL",0,0,"120511205",,terminal_output +3133,14809067,"TERMINAL",0,0,"2162216",,terminal_output +3134,14810226,"TERMINAL",0,0,"3273327",,terminal_output +3135,14811147,"TERMINAL",0,0,"4384438",,terminal_output +3136,14812274,"TERMINAL",0,0,"5495549",,terminal_output +3137,14813298,"TERMINAL",0,0,"653066540",,terminal_output +3138,14814273,"TERMINAL",0,0,"7617761",,terminal_output +3139,14815312,"TERMINAL",0,0,"8728872",,terminal_output +3140,14816381,"TERMINAL",0,0,"9839983",,terminal_output +3141,14817400,"TERMINAL",0,0,"3094202094",,terminal_output +3142,14818440,"TERMINAL",0,0,"130511305",,terminal_output +3143,14819547,"TERMINAL",0,0,"2162216",,terminal_output +3144,14820519,"TERMINAL",0,0,"3273327",,terminal_output +3145,14821568,"TERMINAL",0,0,"4384438",,terminal_output +3146,14822619,"TERMINAL",0,0,"5495549",,terminal_output +3147,14823853,"TERMINAL",0,0,"654066550",,terminal_output +3148,14824699,"TERMINAL",0,0,"7617761",,terminal_output +3149,14825792,"TERMINAL",0,0,"8728872",,terminal_output +3150,14826799,"TERMINAL",0,0,"9839983",,terminal_output +3151,14827831,"TERMINAL",0,0,"4094303094",,terminal_output +3152,14829255,"TERMINAL",0,0,"241622416",,terminal_output +3153,14829956,"TERMINAL",0,0,"3273327",,terminal_output +3154,14830994,"TERMINAL",0,0,"4384438",,terminal_output +3155,14832033,"TERMINAL",0,0,"5495549",,terminal_output +3156,14833079,"TERMINAL",0,0,"65506654:00",,terminal_output +3157,14834115,"TERMINAL",0,0,"7617761",,terminal_output +3158,14835360,"TERMINAL",0,0,"8728872",,terminal_output +3159,14836239,"TERMINAL",0,0,"9839983",,terminal_output +3160,14837261,"TERMINAL",0,0,"5094404094",,terminal_output +3161,14838285,"TERMINAL",0,0,"150511505",,terminal_output +3162,14839324,"TERMINAL",0,0,"2162216",,terminal_output +3163,14840374,"TERMINAL",0,0,"3273327",,terminal_output +3164,14841417,"TERMINAL",0,0,"4384438",,terminal_output +3165,14842483,"TERMINAL",0,0,"5495549",,terminal_output +3166,14843506,"TERMINAL",0,0,"656:0066510",,terminal_output +3167,14844553,"TERMINAL",0,0,"7617761",,terminal_output +3168,14845657,"TERMINAL",0,0,"8728872",,terminal_output +3169,14846642,"TERMINAL",0,0,"9839983",,terminal_output +3170,14847708,"TERMINAL",0,0,"5:0094505094",,terminal_output +3171,14848785,"TERMINAL",0,0,"15:005114:005",,terminal_output +3172,14849855,"TERMINAL",0,0,"2162216",,terminal_output +3173,14850888,"TERMINAL",0,0,"3384438",,terminal_output +3174,14851875,"TERMINAL",0,0,"5495549",,terminal_output +3175,14852908,"TERMINAL",0,0,"651066520",,terminal_output +3176,14853964,"TERMINAL",0,0,"7617761",,terminal_output +3177,14855050,"TERMINAL",0,0,"8728872",,terminal_output +3178,14856101,"TERMINAL",0,0,"9839983",,terminal_output +3179,14857099,"TERMINAL",0,0,"10942:004:0094",,terminal_output +3180,14858132,"TERMINAL",0,0,"110511105",,terminal_output +3181,14859173,"TERMINAL",0,0,"2162216",,terminal_output +3182,14860304,"TERMINAL",0,0,"3273327",,terminal_output +3183,14861298,"TERMINAL",0,0,"4384438",,terminal_output +3184,14862299,"TERMINAL",0,0,"5495549",,terminal_output +3185,14863335,"TERMINAL",0,0,"652066530",,terminal_output +3186,14864387,"TERMINAL",0,0,"7617761",,terminal_output +3187,14865430,"TERMINAL",0,0,"8728872",,terminal_output +3188,14866548,"TERMINAL",0,0,"9839983",,terminal_output +3189,14867571,"TERMINAL",0,0,"2094101094",,terminal_output +3190,14868603,"TERMINAL",0,0,"120511205",,terminal_output +3191,14869621,"TERMINAL",0,0,"2162216",,terminal_output +3192,14870642,"TERMINAL",0,0,"3273327",,terminal_output +3193,14871664,"TERMINAL",0,0,"4384438",,terminal_output +3194,14872794,"TERMINAL",0,0,"5495549",,terminal_output +3195,14873753,"TERMINAL",0,0,"653066540",,terminal_output +3196,14874841,"TERMINAL",0,0,"7617761",,terminal_output +3197,14875845,"TERMINAL",0,0,"8839983",,terminal_output +3198,14876881,"TERMINAL",0,0,"3094202094",,terminal_output +3199,14877923,"TERMINAL",0,0,"130511305",,terminal_output +3200,14879139,"TERMINAL",0,0,"2162216",,terminal_output +3201,14880066,"TERMINAL",0,0,"3273327",,terminal_output +3202,14881066,"TERMINAL",0,0,"4384438",,terminal_output +3203,14882108,"TERMINAL",0,0,"5495549",,terminal_output +3204,14883240,"TERMINAL",0,0,"654066550",,terminal_output +3205,14884192,"TERMINAL",0,0,"7617761",,terminal_output +3206,14885592,"TERMINAL",0,0,"8728872",,terminal_output +3207,14886620,"TERMINAL",0,0,"9839983",,terminal_output +3208,14887643,"TERMINAL",0,0,"4094303094",,terminal_output +3209,14888701,"TERMINAL",0,0,"140511405",,terminal_output +3210,14889714,"TERMINAL",0,0,"bash",,terminal_focus +3211,14889746,"TERMINAL",0,0,"2162216",,terminal_output +3212,14890819,"TERMINAL",0,0,"3273327",,terminal_output +3213,14891801,"TERMINAL",0,0,"4384438",,terminal_output +3214,14892902,"TERMINAL",0,0,"55506655:00",,terminal_output +3215,14893889,"TERMINAL",0,0,"7617761",,terminal_output +3216,14894037,"TERMINAL",0,0,"tail -f train_dynamics_maskgit_8_node_3422132.log",,terminal_command +3217,14894088,"TERMINAL",0,0,"]633;E;2025-08-14 14:45:47 tail -f train_dynamics_maskgit_8_node_3422132.log ;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CNodelist: hkn[0530-0534,0536,0701,0715]\r\nCPU Utilized: 00:02:54\r\nCPU Efficiency: 0.25% of 19:02:24 core-walltime\r\nJob Wall-clock time: 00:05:57\r\nStarttime: Thu Aug 14 13:24:52 2025\r\nEndtime: Thu Aug 14 13:30:49 2025\r\nMemory Utilized: 12.67 GB (estimated maximum)\r\nMemory Efficiency: 0.00% of 0.00 MB (0.00 MB/node)\r\nEnergy Consumed: 1339227 Joule / 372.0075 Watthours\r\nAverage node power draw: 3751.33613445378 Watt\r\n",,terminal_output +3218,14895025,"TERMINAL",0,0,"8728872",,terminal_output +3219,14895517,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining]633;D;130",,terminal_output +3220,14896038,"TERMINAL",0,0,"9839983",,terminal_output +3221,14897022,"TERMINAL",0,0,"5094404094",,terminal_output +3222,14898086,"TERMINAL",0,0,"150511505",,terminal_output +3223,14899113,"TERMINAL",0,0,"2162216",,terminal_output +3224,14900162,"TERMINAL",0,0,"3273327",,terminal_output +3225,14901263,"TERMINAL",0,0,"4384438",,terminal_output +3226,14902264,"TERMINAL",0,0,"5495549",,terminal_output +3227,14903292,"TERMINAL",0,0,"657:0066510",,terminal_output +3228,14904333,"TERMINAL",0,0,"7617761",,terminal_output +3229,14905376,"TERMINAL",0,0,"8728872",,terminal_output +3230,14906423,"TERMINAL",0,0,"9839983",,terminal_output +3231,14907458,"TERMINAL",0,0,"6:0094505094",,terminal_output +3232,14908531,"TERMINAL",0,0,"16:005115:005",,terminal_output +3233,14909556,"TERMINAL",0,0,"2162216",,terminal_output +3234,14910683,"TERMINAL",0,0,"3273327",,terminal_output +3235,14911631,"TERMINAL",0,0,"4384438",,terminal_output +3236,14912746,"TERMINAL",0,0,"5495549",,terminal_output +3237,14913715,"TERMINAL",0,0,"651066520",,terminal_output +3238,14914759,"TERMINAL",0,0,"7617761",,terminal_output +3239,14916009,"TERMINAL",0,0,"8728872",,terminal_output +3240,14916842,"TERMINAL",0,0,"9943:005:0094",,terminal_output +3241,14917887,"TERMINAL",0,0,"1110511105",,terminal_output +3242,14918954,"TERMINAL",0,0,"2162216",,terminal_output +3243,14920042,"TERMINAL",0,0,"3273327",,terminal_output +3244,14921010,"TERMINAL",0,0,"4384438",,terminal_output +3245,14922053,"TERMINAL",0,0,"5495549",,terminal_output +3246,14923174,"TERMINAL",0,0,"652066530",,terminal_output +3247,14924132,"TERMINAL",0,0,"7617761",,terminal_output +3248,14925220,"TERMINAL",0,0,"8728872",,terminal_output +3249,14926230,"TERMINAL",0,0,"9839983",,terminal_output +3250,14927282,"TERMINAL",0,0,"2094101094",,terminal_output +3251,14928333,"TERMINAL",0,0,"120511205",,terminal_output +3252,14929351,"TERMINAL",0,0,"2162216",,terminal_output +3253,14930398,"TERMINAL",0,0,"3273327",,terminal_output +3254,14931442,"TERMINAL",0,0,"4384438",,terminal_output +3255,14932490,"TERMINAL",0,0,"5495549",,terminal_output +3256,14933623,"TERMINAL",0,0,"653066540",,terminal_output +3257,14934568,"TERMINAL",0,0,"7617761",,terminal_output +3258,14935617,"TERMINAL",0,0,"8728872",,terminal_output +3259,14936654,"TERMINAL",0,0,"9839983",,terminal_output +3260,14937695,"TERMINAL",0,0,"3094202094",,terminal_output +3261,14938741,"TERMINAL",0,0,"130511305",,terminal_output +3262,14939793,"TERMINAL",0,0,"2162216",,terminal_output +3263,14940833,"TERMINAL",0,0,"3384438",,terminal_output +3264,14941914,"TERMINAL",0,0,"5495549",,terminal_output +3265,14942910,"TERMINAL",0,0,"654066550",,terminal_output +3266,14944106,"TERMINAL",0,0,"7617761",,terminal_output +3267,14945089,"TERMINAL",0,0,"8728872",,terminal_output +3268,14946117,"TERMINAL",0,0,"9839983",,terminal_output +3269,14947138,"TERMINAL",0,0,"4094303094",,terminal_output +3270,14948163,"TERMINAL",0,0,"140511405",,terminal_output +3271,14949286,"TERMINAL",0,0,"2162216",,terminal_output +3272,14950416,"TERMINAL",0,0,"3273327",,terminal_output +3273,14951321,"TERMINAL",0,0,"4384438",,terminal_output +3274,14952353,"TERMINAL",0,0,"5495549",,terminal_output +3275,14953384,"TERMINAL",0,0,"65506656:00",,terminal_output +3276,14954428,"TERMINAL",0,0,"7617761",,terminal_output +3277,14955474,"TERMINAL",0,0,"8728872",,terminal_output +3278,14955975,"TERMINAL",0,0,"watch",,terminal_focus +3279,14956525,"TERMINAL",0,0,"9839983",,terminal_output +3280,14956874,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +3281,14958617,"TERMINAL",0,0,"idling",,terminal_command +3282,14958717,"TERMINAL",0,0,"]633;E;2025-08-14 14:46:51 idling;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Aug 14 14:46:51 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 225 nodes idle\rPartition dev_accelerated:\t 0 nodes idle\rPartition accelerated: 10 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 1 nodes idle\rPartition large:\t 7 nodes idle",,terminal_output +3283,14959736,"TERMINAL",0,0,"2\t ",,terminal_output +3284,14960785,"TERMINAL",0,0,"3\t ",,terminal_output +3285,14961882,"TERMINAL",0,0,"4\t ",,terminal_output +3286,14962862,"TERMINAL",0,0,"6\t ",,terminal_output +3287,14963931,"TERMINAL",0,0,"7\t ",,terminal_output +3288,14964954,"TERMINAL",0,0,"8\t ",,terminal_output +3289,14966079,"TERMINAL",0,0,"98",,terminal_output +3290,14967032,"TERMINAL",0,0,"7:00\t ",,terminal_output +3291,14968128,"TERMINAL",0,0,"1\t ",,terminal_output +3292,14969115,"TERMINAL",0,0,"2\t ",,terminal_output +3293,14970298,"TERMINAL",0,0,"3\t ",,terminal_output +3294,14971205,"TERMINAL",0,0,"4\t ",,terminal_output +3295,14972328,"TERMINAL",0,0,"5\t ",,terminal_output +3296,14973351,"TERMINAL",0,0,"6\t ",,terminal_output +3297,14974323,"TERMINAL",0,0,"7\t ",,terminal_output +3298,14975366,"TERMINAL",0,0,"8\t ",,terminal_output +3299,14976406,"TERMINAL",0,0,"9\t ",,terminal_output +3300,14977447,"TERMINAL",0,0,"10\t ",,terminal_output +3301,14978488,"TERMINAL",0,0,"1\t ",,terminal_output +3302,14979603,"TERMINAL",0,0,"2\t ",,terminal_output +3303,14980706,"TERMINAL",0,0,"3\t ",,terminal_output +3304,14981658,"TERMINAL",0,0,"4\t ",,terminal_output +3305,14982652,"TERMINAL",0,0,"5\t ",,terminal_output +3306,14983687,"TERMINAL",0,0,"6\t ",,terminal_output +3307,14984818,"TERMINAL",0,0,"7\t ",,terminal_output +3308,14985843,"TERMINAL",0,0,"8\t ",,terminal_output +3309,14986811,"TERMINAL",0,0,"9\t ",,terminal_output +3310,14987882,"TERMINAL",0,0,"20\t ",,terminal_output +3311,14988915,"TERMINAL",0,0,"2\t ",,terminal_output +3312,14989931,"TERMINAL",0,0,"3\t ",,terminal_output +3313,14991069,"TERMINAL",0,0,"4\t ",,terminal_output +3314,14992009,"TERMINAL",0,0,"5\t ",,terminal_output +3315,14993057,"TERMINAL",0,0,"6\t ",,terminal_output +3316,14994088,"TERMINAL",0,0,"7\t ",,terminal_output +3317,14995137,"TERMINAL",0,0,"8\t ",,terminal_output +3318,14996175,"TERMINAL",0,0,"9\t ",,terminal_output +3319,14997216,"TERMINAL",0,0,"30\t ",,terminal_output +3320,14998448,"TERMINAL",0,0,"1\t ",,terminal_output +3321,14999316,"TERMINAL",0,0,"2\t ",,terminal_output +3322,15000355,"TERMINAL",0,0,"3\t ",,terminal_output +3323,15001371,"TERMINAL",0,0,"4\t ",,terminal_output +3324,15002422,"TERMINAL",0,0,"5\t ",,terminal_output +3325,15003448,"TERMINAL",0,0,"6\t ",,terminal_output +3326,15004491,"TERMINAL",0,0,"7\t ",,terminal_output +3327,15005535,"TERMINAL",0,0,"8\t ",,terminal_output +3328,15006632,"TERMINAL",0,0,"97 7",,terminal_output +3329,15007610,"TERMINAL",0,0,"40\t ",,terminal_output +3330,15008780,"TERMINAL",0,0,"1\t ",,terminal_output +3331,15009701,"TERMINAL",0,0,"2\t ",,terminal_output +3332,15010828,"TERMINAL",0,0,"3\t ",,terminal_output +3333,15011772,"TERMINAL",0,0,"4\t ",,terminal_output +3334,15012816,"TERMINAL",0,0,"5\t ",,terminal_output +3335,15013855,"TERMINAL",0,0,"7\t ",,terminal_output +3336,15014896,"TERMINAL",0,0,"8\t ",,terminal_output +3337,15015937,"TERMINAL",0,0,"9\t ",,terminal_output +3338,15016981,"TERMINAL",0,0,"50\t ",,terminal_output +3339,15018102,"TERMINAL",0,0,"1\t ",,terminal_output +3340,15019068,"TERMINAL",0,0,"2\t ",,terminal_output +3341,15020138,"TERMINAL",0,0,"3\t ",,terminal_output +3342,15021303,"TERMINAL",0,0,"4\t ",,terminal_output +3343,15022174,"TERMINAL",0,0,"5\t ",,terminal_output +3344,15023246,"TERMINAL",0,0,"6\t ",,terminal_output +3345,15024281,"TERMINAL",0,0,"7\t ",,terminal_output +3346,15025290,"TERMINAL",0,0,"8\t ",,terminal_output +3347,15025680,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +3348,15026869,"TERMINAL",0,0,"queue",,terminal_command +3349,15026929,"TERMINAL",0,0,"]633;E;2025-08-14 14:48:00 queue;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 14:48:00 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-20:47:59\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-00:59:04\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 2:34:50\t 1 hkn07133421982 accelerat train_to tum_cte0 R 2:46:50\t 4 hkn[0724,0726,0728,0731]3421973 accelerat train_to tum_cte0 R 2:46:59\t 2 hkn[0722-0723]3421964 accelerat train_to tum_cte0 R 2:47:14\t 1 hkn0720",,terminal_output +3350,15027969,"TERMINAL",0,0,"18:005117:005",,terminal_output +3351,15029089,"TERMINAL",0,0,"2162216",,terminal_output +3352,15030062,"TERMINAL",0,0,"3273327",,terminal_output +3353,15031165,"TERMINAL",0,0,"4384438",,terminal_output +3354,15032157,"TERMINAL",0,0,"5495549",,terminal_output +3355,15033286,"TERMINAL",0,0,"651066520",,terminal_output +3356,15034279,"TERMINAL",0,0,"7617761",,terminal_output +3357,15035310,"TERMINAL",0,0,"8728872",,terminal_output +3358,15036430,"TERMINAL",0,0,"9839983",,terminal_output +3359,15037469,"TERMINAL",0,0,"10945:007:0094",,terminal_output +3360,15038478,"TERMINAL",0,0,"110511105",,terminal_output +3361,15039460,"TERMINAL",0,0,"2162216",,terminal_output +3362,15040527,"TERMINAL",0,0,"3273327",,terminal_output +3363,15041652,"TERMINAL",0,0,"4384438",,terminal_output +3364,15042678,"TERMINAL",0,0,"5495549",,terminal_output +3365,15043702,"TERMINAL",0,0,"652066530",,terminal_output +3366,15044728,"TERMINAL",0,0,"7617761",,terminal_output +3367,15045851,"TERMINAL",0,0,"8728872",,terminal_output +3368,15046802,"TERMINAL",0,0,"9839983",,terminal_output +3369,15047835,"TERMINAL",0,0,"202051111205",,terminal_output +3370,15048868,"TERMINAL",0,0,"2162216",,terminal_output +3371,15049950,"TERMINAL",0,0,"3273327",,terminal_output +3372,15050971,"TERMINAL",0,0,"4384438",,terminal_output +3373,15051987,"TERMINAL",0,0,"5495549",,terminal_output +3374,15053026,"TERMINAL",0,0,"653066540",,terminal_output +3375,15054087,"TERMINAL",0,0,"7617761",,terminal_output +3376,15055122,"TERMINAL",0,0,"8728872",,terminal_output +3377,15056192,"TERMINAL",0,0,"9839983",,terminal_output +3378,15057194,"TERMINAL",0,0,"3094202094",,terminal_output +3379,15058236,"TERMINAL",0,0,"130511305",,terminal_output +3380,15059370,"TERMINAL",0,0,"2162216",,terminal_output +3381,15060393,"TERMINAL",0,0,"3273327",,terminal_output +3382,15061382,"TERMINAL",0,0,"4384438",,terminal_output +3383,15062425,"TERMINAL",0,0,"5495549",,terminal_output +3384,15063473,"TERMINAL",0,0,"654066550",,terminal_output +3385,15064517,"TERMINAL",0,0,"7617761",,terminal_output +3386,15065613,"TERMINAL",0,0,"8728872",,terminal_output +3387,15066608,"TERMINAL",0,0,"9839983",,terminal_output +3388,15067646,"TERMINAL",0,0,"4094303094",,terminal_output +3389,15068791,"TERMINAL",0,0,"140511405",,terminal_output +3390,15069812,"TERMINAL",0,0,"2162216",,terminal_output +3391,15070791,"TERMINAL",0,0,"3273327",,terminal_output +3392,15071832,"TERMINAL",0,0,"4495549",,terminal_output +3393,15072877,"TERMINAL",0,0,"65506658:00",,terminal_output +3394,15073928,"TERMINAL",0,0,"7617761",,terminal_output +3395,15074977,"TERMINAL",0,0,"8728872",,terminal_output +3396,15076162,"TERMINAL",0,0,"9839983",,terminal_output +3397,15077186,"TERMINAL",0,0,"5094404094",,terminal_output +3398,15078413,"TERMINAL",0,0,"150511505",,terminal_output +3399,15079157,"TERMINAL",0,0,"2162216",,terminal_output +3400,15080205,"TERMINAL",0,0,"3273327",,terminal_output +3401,15081282,"TERMINAL",0,0,"4384438",,terminal_output +3402,15082306,"TERMINAL",0,0,"5495549",,terminal_output +3403,15083344,"TERMINAL",0,0,"651:00:0066510",,terminal_output +3404,15084468,"TERMINAL",0,0,"7617761",,terminal_output +3405,15085435,"TERMINAL",0,0,"8728872",,terminal_output +3406,15086479,"TERMINAL",0,0,"9839983",,terminal_output +3407,15087584,"TERMINAL",0,0,"9:0094505094",,terminal_output +3408,15088578,"TERMINAL",0,0,"19:005118:005",,terminal_output +3409,15089619,"TERMINAL",0,0,"2162216",,terminal_output +3410,15090698,"TERMINAL",0,0,"3273327",,terminal_output +3411,15091725,"TERMINAL",0,0,"4384438",,terminal_output +3412,15092754,"TERMINAL",0,0,"5495549",,terminal_output +3413,15093880,"TERMINAL",0,0,"651066520",,terminal_output +3414,15095003,"TERMINAL",0,0,"7728872",,terminal_output +3415,15095911,"TERMINAL",0,0,"9839983",,terminal_output +3416,15096955,"TERMINAL",0,0,"10946:008:0094",,terminal_output +3417,15098075,"TERMINAL",0,0,"110511105",,terminal_output +3418,15099030,"TERMINAL",0,0,"2162216",,terminal_output +3419,15100122,"TERMINAL",0,0,"3273327",,terminal_output +3420,15101148,"TERMINAL",0,0,"4384438",,terminal_output +3421,15102172,"TERMINAL",0,0,"5495549",,terminal_output +3422,15103400,"TERMINAL",0,0,"652066530",,terminal_output +3423,15104318,"TERMINAL",0,0,"7617761",,terminal_output +3424,15105346,"TERMINAL",0,0,"8728872",,terminal_output +3425,15106369,"TERMINAL",0,0,"9839983",,terminal_output +3426,15107497,"TERMINAL",0,0,"2094101094",,terminal_output +3427,15108451,"TERMINAL",0,0,"120511205",,terminal_output +3428,15109544,"TERMINAL",0,0,"2162216",,terminal_output +3429,15110546,"TERMINAL",0,0,"3273327",,terminal_output +3430,15111638,"TERMINAL",0,0,"4384438",,terminal_output +3431,15112637,"TERMINAL",0,0,"5495549",,terminal_output +3432,15113669,"TERMINAL",0,0,"653066540",,terminal_output +3433,15114766,"TERMINAL",0,0,"7617761",,terminal_output +3434,15115750,"TERMINAL",0,0,"8728872",,terminal_output +3435,15116795,"TERMINAL",0,0,"9839983",,terminal_output +3436,15117830,"TERMINAL",0,0,"303052121305",,terminal_output +3437,15118911,"TERMINAL",0,0,"2162216",,terminal_output +3438,15119989,"TERMINAL",0,0,"3273327",,terminal_output +3439,15120975,"TERMINAL",0,0,"4384438",,terminal_output +3440,15122000,"TERMINAL",0,0,"5495549",,terminal_output +3441,15123064,"TERMINAL",0,0,"654066550",,terminal_output +3442,15124129,"TERMINAL",0,0,"7617761",,terminal_output +3443,15125314,"TERMINAL",0,0,"8728872",,terminal_output +3444,15126462,"TERMINAL",0,0,"9839983",,terminal_output +3445,15127505,"TERMINAL",0,0,"4094303094",,terminal_output +3446,15128538,"TERMINAL",0,0,"140511405",,terminal_output +3447,15129583,"TERMINAL",0,0,"2162216",,terminal_output +3448,15130636,"TERMINAL",0,0,"3273327",,terminal_output +3449,15131765,"TERMINAL",0,0,"4384438",,terminal_output +3450,15132718,"TERMINAL",0,0,"5495549",,terminal_output +3451,15133915,"TERMINAL",0,0,"65506659:00",,terminal_output +3452,15134803,"TERMINAL",0,0,"7617761",,terminal_output +3453,15135859,"TERMINAL",0,0,"8839983",,terminal_output +3454,15136926,"TERMINAL",0,0,"5094404094",,terminal_output +3455,15137932,"TERMINAL",0,0,"150511505",,terminal_output +3456,15139044,"TERMINAL",0,0,"2162216",,terminal_output +3457,15140018,"TERMINAL",0,0,"3273327",,terminal_output +3458,15141081,"TERMINAL",0,0,"4384438",,terminal_output +3459,15142104,"TERMINAL",0,0,"5495549",,terminal_output +3460,15143224,"TERMINAL",0,0,"651:0066510",,terminal_output +3461,15144209,"TERMINAL",0,0,"7617761",,terminal_output +3462,15145222,"TERMINAL",0,0,"8728872",,terminal_output +3463,15146437,"TERMINAL",0,0,"9839983",,terminal_output +3464,15147328,"TERMINAL",0,0,"50:0094505094",,terminal_output +3465,15148356,"TERMINAL",0,0,"150:005119:005",,terminal_output +3466,15149391,"TERMINAL",0,0,"2162216",,terminal_output +3467,15150433,"TERMINAL",0,0,"3273327",,terminal_output +3468,15151466,"TERMINAL",0,0,"4384438",,terminal_output +3469,15152509,"TERMINAL",0,0,"5495549",,terminal_output +3470,15153551,"TERMINAL",0,0,"651066520",,terminal_output +3471,15154705,"TERMINAL",0,0,"7617761",,terminal_output +3472,15155730,"TERMINAL",0,0,"8728872",,terminal_output +3473,15156769,"TERMINAL",0,0,"9839983",,terminal_output +3474,15157782,"TERMINAL",0,0,"10947:009:0094",,terminal_output +3475,15158801,"TERMINAL",0,0,"110511105",,terminal_output +3476,15159999,"TERMINAL",0,0,"2162216",,terminal_output +3477,15160838,"TERMINAL",0,0,"3384438",,terminal_output +3478,15161881,"TERMINAL",0,0,"5495549",,terminal_output +3479,15162920,"TERMINAL",0,0,"652066530",,terminal_output +3480,15164024,"TERMINAL",0,0,"7617761",,terminal_output +3481,15165048,"TERMINAL",0,0,"8728872",,terminal_output +3482,15166070,"TERMINAL",0,0,"9839983",,terminal_output +3483,15167082,"TERMINAL",0,0,"2094101094",,terminal_output +3484,15168130,"TERMINAL",0,0,"120511205",,terminal_output +3485,15169170,"TERMINAL",0,0,"2162216",,terminal_output +3486,15170372,"TERMINAL",0,0,"3273327",,terminal_output +3487,15171403,"TERMINAL",0,0,"4384438",,terminal_output +3488,15172351,"TERMINAL",0,0,"5495549",,terminal_output +3489,15173353,"TERMINAL",0,0,"653066540",,terminal_output +3490,15174390,"TERMINAL",0,0,"7617761",,terminal_output +3491,15175448,"TERMINAL",0,0,"8728872",,terminal_output +3492,15176482,"TERMINAL",0,0,"9839983",,terminal_output +3493,15177521,"TERMINAL",0,0,"3094202094",,terminal_output +3494,15178665,"TERMINAL",0,0,"130511305",,terminal_output +3495,15179641,"TERMINAL",0,0,"2162216",,terminal_output +3496,15180710,"TERMINAL",0,0,"3273327",,terminal_output +3497,15181677,"TERMINAL",0,0,"4384438",,terminal_output +3498,15182759,"TERMINAL",0,0,"5495549",,terminal_output +3499,15183783,"TERMINAL",0,0,"654066550",,terminal_output +3500,15184792,"TERMINAL",0,0,"7617761",,terminal_output +3501,15185939,"TERMINAL",0,0,"8839983",,terminal_output +3502,15186882,"TERMINAL",0,0,"4094303094",,terminal_output +3503,15187930,"TERMINAL",0,0,"140511405",,terminal_output +3504,15189014,"TERMINAL",0,0,"2162216",,terminal_output +3505,15190029,"TERMINAL",0,0,"3273327",,terminal_output +3506,15191168,"TERMINAL",0,0,"4384438",,terminal_output +3507,15192181,"TERMINAL",0,0,"5495549",,terminal_output +3508,15193207,"TERMINAL",0,0,"655066550:00",,terminal_output +3509,15194169,"TERMINAL",0,0,"7617761",,terminal_output +3510,15195357,"TERMINAL",0,0,"8728872",,terminal_output +3511,15196270,"TERMINAL",0,0,"9839983",,terminal_output +3512,15197402,"TERMINAL",0,0,"5094404094",,terminal_output +3513,15198632,"TERMINAL",0,0,"150511505",,terminal_output +3514,15199398,"TERMINAL",0,0,"2162216",,terminal_output +3515,15200438,"TERMINAL",0,0,"3273327",,terminal_output +3516,15201482,"TERMINAL",0,0,"4384438",,terminal_output +3517,15202530,"TERMINAL",0,0,"5495549",,terminal_output +3518,15203582,"TERMINAL",0,0,"652:0066510",,terminal_output +3519,15204675,"TERMINAL",0,0,"7617761",,terminal_output +3520,15205672,"TERMINAL",0,0,"8728872",,terminal_output +3521,15206715,"TERMINAL",0,0,"9839983",,terminal_output +3522,15207762,"TERMINAL",0,0,"1:0094505094",,terminal_output +3523,15208872,"TERMINAL",0,0,"11:0051150:005",,terminal_output +3524,15209851,"TERMINAL",0,0,"2273327",,terminal_output +3525,15210919,"TERMINAL",0,0,"4384438",,terminal_output +3526,15211935,"TERMINAL",0,0,"5495549",,terminal_output +3527,15213076,"TERMINAL",0,0,"651066520",,terminal_output +3528,15214038,"TERMINAL",0,0,"7617761",,terminal_output +3529,15215123,"TERMINAL",0,0,"8728872",,terminal_output +3530,15216253,"TERMINAL",0,0,"9839983",,terminal_output +3531,15217153,"TERMINAL",0,0,"10948:0050:0094",,terminal_output +3532,15218294,"TERMINAL",0,0,"110511105",,terminal_output +3533,15219266,"TERMINAL",0,0,"2162216",,terminal_output +3534,15220385,"TERMINAL",0,0,"3273327",,terminal_output +3535,15221366,"TERMINAL",0,0,"4384438",,terminal_output +3536,15222400,"TERMINAL",0,0,"5495549",,terminal_output +3537,15223444,"TERMINAL",0,0,"652066530",,terminal_output +3538,15224480,"TERMINAL",0,0,"7617761",,terminal_output +3539,15225566,"TERMINAL",0,0,"8728872",,terminal_output +3540,15226561,"TERMINAL",0,0,"9839983",,terminal_output +3541,15227610,"TERMINAL",0,0,"2094101094",,terminal_output +3542,15228736,"TERMINAL",0,0,"120511205",,terminal_output +3543,15229690,"TERMINAL",0,0,"2162216",,terminal_output +3544,15230785,"TERMINAL",0,0,"3273327",,terminal_output +3545,15231812,"TERMINAL",0,0,"4384438",,terminal_output +3546,15232837,"TERMINAL",0,0,"553066540",,terminal_output +3547,15234063,"TERMINAL",0,0,"7617761",,terminal_output +3548,15235086,"TERMINAL",0,0,"8728872",,terminal_output +3549,15235942,"TERMINAL",0,0,"9839983",,terminal_output +3550,15236978,"TERMINAL",0,0,"3094202094",,terminal_output +3551,15238019,"TERMINAL",0,0,"130511305",,terminal_output +3552,15239142,"TERMINAL",0,0,"2162216",,terminal_output +3553,15240215,"TERMINAL",0,0,"3273327",,terminal_output +3554,15241234,"TERMINAL",0,0,"4384438",,terminal_output +3555,15242268,"TERMINAL",0,0,"5495549",,terminal_output +3556,15243282,"TERMINAL",0,0,"654066550",,terminal_output +3557,15244343,"TERMINAL",0,0,"7617761",,terminal_output +3558,15245358,"TERMINAL",0,0,"8728872",,terminal_output +3559,15246362,"TERMINAL",0,0,"9839983",,terminal_output +3560,15247396,"TERMINAL",0,0,"4094303094",,terminal_output +3561,15248611,"TERMINAL",0,0,"140511405",,terminal_output +3562,15249490,"TERMINAL",0,0,"2162216",,terminal_output +3563,15250537,"TERMINAL",0,0,"3273327",,terminal_output +3564,15251676,"TERMINAL",0,0,"4384438",,terminal_output +3565,15252701,"TERMINAL",0,0,"5495549",,terminal_output +3566,15253725,"TERMINAL",0,0,"65506651:00",,terminal_output +3567,15254709,"TERMINAL",0,0,"7617761",,terminal_output +3568,15255771,"TERMINAL",0,0,"8728872",,terminal_output +3569,15256805,"TERMINAL",0,0,"9839983",,terminal_output +3570,15257842,"TERMINAL",0,0,"505054141505",,terminal_output +3571,15258948,"TERMINAL",0,0,"2162216",,terminal_output +3572,15259928,"TERMINAL",0,0,"3273327",,terminal_output +3573,15260971,"TERMINAL",0,0,"4384438",,terminal_output +3574,15262010,"TERMINAL",0,0,"5495549",,terminal_output +3575,15263145,"TERMINAL",0,0,"653:0066510",,terminal_output +3576,15264092,"TERMINAL",0,0,"7617761",,terminal_output +3577,15265143,"TERMINAL",0,0,"8728872",,terminal_output +3578,15266224,"TERMINAL",0,0,"9839983",,terminal_output +3579,15267244,"TERMINAL",0,0,"2:0094505094",,terminal_output +3580,15268273,"TERMINAL",0,0,"12:005111:005",,terminal_output +3581,15269389,"TERMINAL",0,0,"2162216",,terminal_output +3582,15270392,"TERMINAL",0,0,"3273327",,terminal_output +3583,15271399,"TERMINAL",0,0,"4384438",,terminal_output +3584,15272569,"TERMINAL",0,0,"5495549",,terminal_output +3585,15273481,"TERMINAL",0,0,"651066520",,terminal_output +3586,15274535,"TERMINAL",0,0,"7617761",,terminal_output +3587,15275583,"TERMINAL",0,0,"8728872",,terminal_output +3588,15276767,"TERMINAL",0,0,"9839983",,terminal_output +3589,15277690,"TERMINAL",0,0,"10949:001:0094",,terminal_output +3590,15278726,"TERMINAL",0,0,"110511105",,terminal_output +3591,15279764,"TERMINAL",0,0,"2162216",,terminal_output +3592,15280814,"TERMINAL",0,0,"3273327",,terminal_output +3593,15281989,"TERMINAL",0,0,"5495549",,terminal_output +3594,15282904,"TERMINAL",0,0,"652066530",,terminal_output +3595,15284036,"TERMINAL",0,0,"7617761",,terminal_output +3596,15285057,"TERMINAL",0,0,"8728872",,terminal_output +3597,15286085,"TERMINAL",0,0,"9839983",,terminal_output +3598,15287109,"TERMINAL",0,0,"2094101094",,terminal_output +3599,15288235,"TERMINAL",0,0,"120511205",,terminal_output +3600,15289163,"TERMINAL",0,0,"2162216",,terminal_output +3601,15290285,"TERMINAL",0,0,"3273327",,terminal_output +3602,15291304,"TERMINAL",0,0,"4384438",,terminal_output +3603,15292363,"TERMINAL",0,0,"5495549",,terminal_output +3604,15293344,"TERMINAL",0,0,"653066540",,terminal_output +3605,15294387,"TERMINAL",0,0,"7617761",,terminal_output +3606,15295434,"TERMINAL",0,0,"8728872",,terminal_output +3607,15296486,"TERMINAL",0,0,"9839983",,terminal_output +3608,15297535,"TERMINAL",0,0,"3094202094",,terminal_output +3609,15298679,"TERMINAL",0,0,"130511305",,terminal_output +3610,15299703,"TERMINAL",0,0,"2162216",,terminal_output +3611,15300675,"TERMINAL",0,0,"3273327",,terminal_output +3612,15301751,"TERMINAL",0,0,"4384438",,terminal_output +3613,15302828,"TERMINAL",0,0,"5495549",,terminal_output +3614,15303804,"TERMINAL",0,0,"654066550",,terminal_output +3615,15304926,"TERMINAL",0,0,"7728872",,terminal_output +3616,15306051,"TERMINAL",0,0,"9839983",,terminal_output +3617,15306934,"TERMINAL",0,0,"4094303094",,terminal_output +3618,15307978,"TERMINAL",0,0,"140511405",,terminal_output +3619,15309022,"TERMINAL",0,0,"2162216",,terminal_output +3620,15310251,"TERMINAL",0,0,"3273327",,terminal_output +3621,15311103,"TERMINAL",0,0,"4384438",,terminal_output +3622,15312148,"TERMINAL",0,0,"5495549",,terminal_output +3623,15313366,"TERMINAL",0,0,"65506652:00",,terminal_output +3624,15314249,"TERMINAL",0,0,"7617761",,terminal_output +3625,15315382,"TERMINAL",0,0,"8728872",,terminal_output +3626,15316396,"TERMINAL",0,0,"9839983",,terminal_output +3627,15317420,"TERMINAL",0,0,"5094404094",,terminal_output +3628,15318415,"TERMINAL",0,0,"150511505",,terminal_output +3629,15319463,"TERMINAL",0,0,"2162216",,terminal_output +3630,15320502,"TERMINAL",0,0,"3273327",,terminal_output +3631,15321546,"TERMINAL",0,0,"4384438",,terminal_output +3632,15322647,"TERMINAL",0,0,"5495549",,terminal_output +3633,15323682,"TERMINAL",0,0,"654:0066510",,terminal_output +3634,15324693,"TERMINAL",0,0,"7617761",,terminal_output +3635,15325815,"TERMINAL",0,0,"8728872",,terminal_output +3636,15326769,"TERMINAL",0,0,"9839983",,terminal_output +3637,15327839,"TERMINAL",0,0,"3:0094505094",,terminal_output +3638,15328852,"TERMINAL",0,0,"13:016222:016",,terminal_output +3639,15329913,"TERMINAL",0,0,"3273327",,terminal_output +3640,15331140,"TERMINAL",0,0,"4384438",,terminal_output +3641,15331987,"TERMINAL",0,0,"5495549",,terminal_output +3642,15333091,"TERMINAL",0,0,"651066520",,terminal_output +3643,15334074,"TERMINAL",0,0,"7617761",,terminal_output +3644,15335141,"TERMINAL",0,0,"8728872",,terminal_output +3645,15336265,"TERMINAL",0,0,"9839983",,terminal_output +3646,15337285,"TERMINAL",0,0,"109440:002:0094",,terminal_output +3647,15338308,"TERMINAL",0,0,"110511105",,terminal_output +3648,15339371,"TERMINAL",0,0,"2162216",,terminal_output +3649,15340357,"TERMINAL",0,0,"3273327",,terminal_output +3650,15341382,"TERMINAL",0,0,"4384438",,terminal_output +3651,15342407,"TERMINAL",0,0,"5495549",,terminal_output +3652,15343432,"TERMINAL",0,0,"652066530",,terminal_output +3653,15344481,"TERMINAL",0,0,"7617761",,terminal_output +3654,15345522,"TERMINAL",0,0,"8728872",,terminal_output +3655,15346566,"TERMINAL",0,0,"9839983",,terminal_output +3656,15347598,"TERMINAL",0,0,"2094101094",,terminal_output +3657,15348641,"TERMINAL",0,0,"120511205",,terminal_output +3658,15349786,"TERMINAL",0,0,"2162216",,terminal_output +3659,15350801,"TERMINAL",0,0,"3273327",,terminal_output +3660,15351780,"TERMINAL",0,0,"4384438",,terminal_output +3661,15352817,"TERMINAL",0,0,"553066540",,terminal_output +3662,15353873,"TERMINAL",0,0,"7617761",,terminal_output +3663,15355102,"TERMINAL",0,0,"8728872",,terminal_output +3664,15355943,"TERMINAL",0,0,"9839983",,terminal_output +3665,15356993,"TERMINAL",0,0,"3094202094",,terminal_output +3666,15358037,"TERMINAL",0,0,"130511305",,terminal_output +3667,15359197,"TERMINAL",0,0,"2162216",,terminal_output +3668,15360120,"TERMINAL",0,0,"3273327",,terminal_output +3669,15361157,"TERMINAL",0,0,"4384438",,terminal_output +3670,15362195,"TERMINAL",0,0,"5495549",,terminal_output +3671,15363293,"TERMINAL",0,0,"654066550",,terminal_output +3672,15364300,"TERMINAL",0,0,"7617761",,terminal_output +3673,15365550,"TERMINAL",0,0,"8728872",,terminal_output +3674,15366573,"TERMINAL",0,0,"9839983",,terminal_output +3675,15367695,"TERMINAL",0,0,"4094303094",,terminal_output +3676,15368721,"TERMINAL",0,0,"140511405",,terminal_output +3677,15369745,"TERMINAL",0,0,"2162216",,terminal_output +3678,15370770,"TERMINAL",0,0,"3273327",,terminal_output +3679,15371895,"TERMINAL",0,0,"4384438",,terminal_output +3680,15372843,"TERMINAL",0,0,"55506653:00",,terminal_output +3681,15373883,"TERMINAL",0,0,"7617761",,terminal_output +3682,15374967,"TERMINAL",0,0,"8728872",,terminal_output +3683,15375996,"TERMINAL",0,0,"9839983",,terminal_output +3684,15377023,"TERMINAL",0,0,"5094404094",,terminal_output +3685,15378060,"TERMINAL",0,0,"150511505",,terminal_output +3686,15379117,"TERMINAL",0,0,"2162216",,terminal_output +3687,15380170,"TERMINAL",0,0,"3273327",,terminal_output +3688,15381318,"TERMINAL",0,0,"4384438",,terminal_output +3689,15382437,"TERMINAL",0,0,"5495549",,terminal_output +3690,15383401,"TERMINAL",0,0,"655:0066510",,terminal_output +3691,15384387,"TERMINAL",0,0,"7617761",,terminal_output +3692,15385381,"TERMINAL",0,0,"8728872",,terminal_output +3693,15386422,"TERMINAL",0,0,"9839983",,terminal_output +3694,15387463,"TERMINAL",0,0,"4:0094505094",,terminal_output +3695,15388515,"TERMINAL",0,0,"14:005113:005",,terminal_output +3696,15389562,"TERMINAL",0,0,"2162216",,terminal_output +3697,15390636,"TERMINAL",0,0,"3273327",,terminal_output +3698,15391661,"TERMINAL",0,0,"4384438",,terminal_output +3699,15392833,"TERMINAL",0,0,"5495549",,terminal_output +3700,15393807,"TERMINAL",0,0,"651066520",,terminal_output +3701,15394832,"TERMINAL",0,0,"7617761",,terminal_output +3702,15395857,"TERMINAL",0,0,"8839983",,terminal_output +3703,15396867,"TERMINAL",0,0,"10941:003:0094",,terminal_output +3704,15397900,"TERMINAL",0,0,"110511105",,terminal_output +3705,15398966,"TERMINAL",0,0,"2162216",,terminal_output +3706,15400032,"TERMINAL",0,0,"3273327",,terminal_output +3707,15401066,"TERMINAL",0,0,"4384438",,terminal_output +3708,15402104,"TERMINAL",0,0,"5495549",,terminal_output +3709,15403112,"TERMINAL",0,0,"652066530",,terminal_output +3710,15404157,"TERMINAL",0,0,"7617761",,terminal_output +3711,15405278,"TERMINAL",0,0,"8728872",,terminal_output +3712,15406303,"TERMINAL",0,0,"9839983",,terminal_output +3713,15407328,"TERMINAL",0,0,"2094101094",,terminal_output +3714,15408348,"TERMINAL",0,0,"120511205",,terminal_output +3715,15409327,"TERMINAL",0,0,"2162216",,terminal_output +3716,15410436,"TERMINAL",0,0,"3273327",,terminal_output +3717,15411424,"TERMINAL",0,0,"4384438",,terminal_output +3718,15412461,"TERMINAL",0,0,"5495549",,terminal_output +3719,15413506,"TERMINAL",0,0,"653066540",,terminal_output +3720,15414543,"TERMINAL",0,0,"7617761",,terminal_output +3721,15415588,"TERMINAL",0,0,"8728872",,terminal_output +3722,15416627,"TERMINAL",0,0,"9839983",,terminal_output +3723,15417659,"TERMINAL",0,0,"3094202094",,terminal_output +3724,15418999,"TERMINAL",0,0,"130511305",,terminal_output +3725,15419830,"TERMINAL",0,0,"2162216",,terminal_output +3726,15420850,"TERMINAL",0,0,"3273327",,terminal_output +3727,15421865,"TERMINAL",0,0,"4495549",,terminal_output +3728,15422871,"TERMINAL",0,0,"654066550",,terminal_output +3729,15423916,"TERMINAL",0,0,"7617761",,terminal_output +3730,15425145,"TERMINAL",0,0,"8728872",,terminal_output +3731,15426064,"TERMINAL",0,0,"9839983",,terminal_output +3732,15427044,"TERMINAL",0,0,"4094303094",,terminal_output +3733,15428117,"TERMINAL",0,0,"140511405",,terminal_output +3734,15429143,"TERMINAL",0,0,"2162216",,terminal_output +3735,15430265,"TERMINAL",0,0,"3273327",,terminal_output +3736,15431228,"TERMINAL",0,0,"4384438",,terminal_output +3737,15432292,"TERMINAL",0,0,"5495549",,terminal_output +3738,15433443,"TERMINAL",0,0,"65506654:00",,terminal_output +3739,15434359,"TERMINAL",0,0,"7617761",,terminal_output +3740,15435404,"TERMINAL",0,0,"8728872",,terminal_output +3741,15436434,"TERMINAL",0,0,"9839983",,terminal_output +3742,15437477,"TERMINAL",0,0,"5094404094",,terminal_output +3743,15438516,"TERMINAL",0,0,"150511505",,terminal_output +3744,15439563,"TERMINAL",0,0,"2162216",,terminal_output +3745,15440607,"TERMINAL",0,0,"3273327",,terminal_output +3746,15441648,"TERMINAL",0,0,"4384438",,terminal_output +3747,15442755,"TERMINAL",0,0,"5495549",,terminal_output +3748,15443759,"TERMINAL",0,0,"656:0066510",,terminal_output +3749,15444812,"TERMINAL",0,0,"7617761",,terminal_output +3750,15445838,"TERMINAL",0,0,"8839983",,terminal_output +3751,15446883,"TERMINAL",0,0,"5:0094505094",,terminal_output +3752,15447932,"TERMINAL",0,0,"15:005114:005",,terminal_output +3753,15449107,"TERMINAL",0,0,"2162216",,terminal_output +3754,15450034,"TERMINAL",0,0,"3273327",,terminal_output +3755,15451050,"TERMINAL",0,0,"4384438",,terminal_output +3756,15452090,"TERMINAL",0,0,"5495549",,terminal_output +3757,15453135,"TERMINAL",0,0,"651066520",,terminal_output +3758,15454225,"TERMINAL",0,0,"7617761",,terminal_output +3759,15455249,"TERMINAL",0,0,"8728872",,terminal_output +3760,15456288,"TERMINAL",0,0,"9839983",,terminal_output +3761,15457504,"TERMINAL",0,0,"10942:004:0094",,terminal_output +3762,15458528,"TERMINAL",0,0,"110511105",,terminal_output +3763,15459452,"TERMINAL",0,0,"2162216",,terminal_output +3764,15460473,"TERMINAL",0,0,"3273327",,terminal_output +3765,15461507,"TERMINAL",0,0,"4384438",,terminal_output +3766,15462562,"TERMINAL",0,0,"5495549",,terminal_output +3767,15463559,"TERMINAL",0,0,"652066530",,terminal_output +3768,15464780,"TERMINAL",0,0,"7617761",,terminal_output +3769,15465647,"TERMINAL",0,0,"8728872",,terminal_output +3770,15466718,"TERMINAL",0,0,"9839983",,terminal_output +3771,15467745,"TERMINAL",0,0,"2094101094",,terminal_output +3772,15468773,"TERMINAL",0,0,"120511205",,terminal_output +3773,15469893,"TERMINAL",0,0,"2273327",,terminal_output +3774,15471025,"TERMINAL",0,0,"4384438",,terminal_output +3775,15472043,"TERMINAL",0,0,"5495549",,terminal_output +3776,15472947,"TERMINAL",0,0,"653066540",,terminal_output +3777,15474091,"TERMINAL",0,0,"7617761",,terminal_output +3778,15475115,"TERMINAL",0,0,"8728872",,terminal_output +3779,15476141,"TERMINAL",0,0,"9839983",,terminal_output +3780,15477163,"TERMINAL",0,0,"3094202094",,terminal_output +3781,15478187,"TERMINAL",0,0,"130511305",,terminal_output +3782,15479257,"TERMINAL",0,0,"2162216",,terminal_output +3783,15480274,"TERMINAL",0,0,"3273327",,terminal_output +3784,15481325,"TERMINAL",0,0,"4384438",,terminal_output +3785,15482417,"TERMINAL",0,0,"5495549",,terminal_output +3786,15483513,"TERMINAL",0,0,"654066550",,terminal_output +3787,15484434,"TERMINAL",0,0,"7617761",,terminal_output +3788,15485480,"TERMINAL",0,0,"8728872",,terminal_output +3789,15486588,"TERMINAL",0,0,"9839983",,terminal_output +3790,15487559,"TERMINAL",0,0,"4094303094",,terminal_output +3791,15488632,"TERMINAL",0,0,"140511405",,terminal_output +3792,15489654,"TERMINAL",0,0,"2162216",,terminal_output +3793,15490782,"TERMINAL",0,0,"3273327",,terminal_output +3794,15491816,"TERMINAL",0,0,"4384438",,terminal_output +3795,15492831,"TERMINAL",0,0,"5495549",,terminal_output +3796,15493960,"TERMINAL",0,0,"66517765:01",,terminal_output +3797,15494879,"TERMINAL",0,0,"8728872",,terminal_output +3798,15495905,"TERMINAL",0,0,"9839983",,terminal_output +3799,15496952,"TERMINAL",0,0,"5094404094",,terminal_output +3800,15498156,"TERMINAL",0,0,"150511505",,terminal_output +3801,15499078,"TERMINAL",0,0,"2162216",,terminal_output +3802,15500210,"TERMINAL",0,0,"3273327",,terminal_output +3803,15501129,"TERMINAL",0,0,"4384438",,terminal_output +3804,15502152,"TERMINAL",0,0,"5495549",,terminal_output +3805,15503196,"TERMINAL",0,0,"657:0066510",,terminal_output +3806,15504247,"TERMINAL",0,0,"7617761",,terminal_output +3807,15505326,"TERMINAL",0,0,"8728872",,terminal_output +3808,15506450,"TERMINAL",0,0,"9839983",,terminal_output +3809,15507372,"TERMINAL",0,0,"6:0094505094",,terminal_output +3810,15508448,"TERMINAL",0,0,"16:005115:005",,terminal_output +3811,15509522,"TERMINAL",0,0,"2162216",,terminal_output +3812,15510550,"TERMINAL",0,0,"3273327",,terminal_output +3813,15511527,"TERMINAL",0,0,"4384438",,terminal_output +3814,15512575,"TERMINAL",0,0,"5495549",,terminal_output +3815,15513620,"TERMINAL",0,0,"651066520",,terminal_output +3816,15514663,"TERMINAL",0,0,"7617761",,terminal_output +3817,15515776,"TERMINAL",0,0,"8728872",,terminal_output +3818,15516791,"TERMINAL",0,0,"9839983",,terminal_output +3819,15517825,"TERMINAL",0,0,"10943:005:0094",,terminal_output +3820,15518844,"TERMINAL",0,0,"111622116",,terminal_output +3821,15519969,"TERMINAL",0,0,"3273327",,terminal_output +3822,15520992,"TERMINAL",0,0,"4384438",,terminal_output +3823,15522035,"TERMINAL",0,0,"5495549",,terminal_output +3824,15523145,"TERMINAL",0,0,"652066530",,terminal_output +3825,15524123,"TERMINAL",0,0,"7617761",,terminal_output +3826,15525295,"TERMINAL",0,0,"8728872",,terminal_output +3827,15526216,"TERMINAL",0,0,"9839983",,terminal_output +3828,15527351,"TERMINAL",0,0,"2094101094",,terminal_output +3829,15528364,"TERMINAL",0,0,"120511205",,terminal_output +3830,15529416,"TERMINAL",0,0,"2162216",,terminal_output +3831,15530442,"TERMINAL",0,0,"3273327",,terminal_output +3832,15531436,"TERMINAL",0,0,"4384438",,terminal_output +3833,15532480,"TERMINAL",0,0,"5495549",,terminal_output +3834,15533520,"TERMINAL",0,0,"653066540",,terminal_output +3835,15534570,"TERMINAL",0,0,"7617761",,terminal_output +3836,15535611,"TERMINAL",0,0,"8728872",,terminal_output +3837,15536655,"TERMINAL",0,0,"9839983",,terminal_output +3838,15537726,"TERMINAL",0,0,"3094202094",,terminal_output +3839,15538815,"TERMINAL",0,0,"130511305",,terminal_output +3840,15539835,"TERMINAL",0,0,"2162216",,terminal_output +3841,15540865,"TERMINAL",0,0,"3384438",,terminal_output +3842,15541983,"TERMINAL",0,0,"5495549",,terminal_output +3843,15542940,"TERMINAL",0,0,"654066550",,terminal_output +3844,15544134,"TERMINAL",0,0,"7617761",,terminal_output +3845,15545158,"TERMINAL",0,0,"8728872",,terminal_output +3846,15546181,"TERMINAL",0,0,"9839983",,terminal_output +3847,15547315,"TERMINAL",0,0,"4094303094",,terminal_output +3848,15548176,"TERMINAL",0,0,"140511405",,terminal_output +3849,15549211,"TERMINAL",0,0,"2162216",,terminal_output +3850,15550294,"TERMINAL",0,0,"3273327",,terminal_output +3851,15551508,"TERMINAL",0,0,"4384438",,terminal_output +3852,15552435,"TERMINAL",0,0,"5495549",,terminal_output +3853,15553558,"TERMINAL",0,0,"65506656:00",,terminal_output +3854,15554477,"TERMINAL",0,0,"7617761",,terminal_output +3855,15555504,"TERMINAL",0,0,"8728872",,terminal_output +3856,15556532,"TERMINAL",0,0,"9839983",,terminal_output +3857,15557572,"TERMINAL",0,0,"5094404094",,terminal_output +3858,15558673,"TERMINAL",0,0,"150511505",,terminal_output +3859,15559657,"TERMINAL",0,0,"2162216",,terminal_output +3860,15560722,"TERMINAL",0,0,"3273327",,terminal_output +3861,15562054,"TERMINAL",0,0,"4384438",,terminal_output +3862,15562849,"TERMINAL",0,0,"5495549",,terminal_output +3863,15563898,"TERMINAL",0,0,"668:0177611",,terminal_output +3864,15564878,"TERMINAL",0,0,"8728872",,terminal_output +3865,15565946,"TERMINAL",0,0,"9839983",,terminal_output +3866,15566972,"TERMINAL",0,0,"7:0094505094",,terminal_output +3867,15568098,"TERMINAL",0,0,"17:005116:005",,terminal_output +3868,15569119,"TERMINAL",0,0,"2162216",,terminal_output +3869,15570099,"TERMINAL",0,0,"3273327",,terminal_output +3870,15571204,"TERMINAL",0,0,"4384438",,terminal_output +3871,15572198,"TERMINAL",0,0,"5495549",,terminal_output +3872,15573320,"TERMINAL",0,0,"651066520",,terminal_output +3873,15574344,"TERMINAL",0,0,"7617761",,terminal_output +3874,15575368,"TERMINAL",0,0,"8728872",,terminal_output +3875,15576382,"TERMINAL",0,0,"9839983",,terminal_output +3876,15577835,"TERMINAL",0,0,"10944:006:0094",,terminal_output +3877,15578643,"TERMINAL",0,0,"110511105",,terminal_output +3878,15579504,"TERMINAL",0,0,"2162216",,terminal_output +3879,15580542,"TERMINAL",0,0,"3273327",,terminal_output +3880,15581581,"TERMINAL",0,0,"4384438",,terminal_output +3881,15582627,"TERMINAL",0,0,"5495549",,terminal_output +3882,15583765,"TERMINAL",0,0,"652066530",,terminal_output +3883,15584701,"TERMINAL",0,0,"7617761",,terminal_output +3884,15585751,"TERMINAL",0,0,"8728872",,terminal_output +3885,15586801,"TERMINAL",0,0,"9839983",,terminal_output +3886,15587835,"TERMINAL",0,0,"202051111205",,terminal_output +3887,15588886,"TERMINAL",0,0,"2162216",,terminal_output +3888,15590115,"TERMINAL",0,0,"3273327",,terminal_output +3889,15590986,"TERMINAL",0,0,"4384438",,terminal_output +3890,15592006,"TERMINAL",0,0,"5495549",,terminal_output +3891,15593085,"TERMINAL",0,0,"653066540",,terminal_output +3892,15594115,"TERMINAL",0,0,"7617761",,terminal_output +3893,15595235,"TERMINAL",0,0,"8728872",,terminal_output +3894,15596257,"TERMINAL",0,0,"9839983",,terminal_output +3895,15597222,"TERMINAL",0,0,"3094202094",,terminal_output +3896,15598305,"TERMINAL",0,0,"130511305",,terminal_output +3897,15599328,"TERMINAL",0,0,"2162216",,terminal_output +3898,15600351,"TERMINAL",0,0,"3273327",,terminal_output +3899,15601585,"TERMINAL",0,0,"4384438",,terminal_output +3900,15602428,"TERMINAL",0,0,"5495549",,terminal_output +3901,15603631,"TERMINAL",0,0,"654066550",,terminal_output +3902,15604508,"TERMINAL",0,0,"7617761",,terminal_output +3903,15605559,"TERMINAL",0,0,"8728872",,terminal_output +3904,15606587,"TERMINAL",0,0,"9839983",,terminal_output +3905,15607627,"TERMINAL",0,0,"4094303094",,terminal_output +3906,15608669,"TERMINAL",0,0,"140511405",,terminal_output +3907,15609777,"TERMINAL",0,0,"2162216",,terminal_output +3908,15610761,"TERMINAL",0,0,"3273327",,terminal_output +3909,15611822,"TERMINAL",0,0,"4384438",,terminal_output +3910,15612841,"TERMINAL",0,0,"55506657:00",,terminal_output +3911,15613991,"TERMINAL",0,0,"7617761",,terminal_output +3912,15615106,"TERMINAL",0,0,"8728872",,terminal_output +3913,15616018,"TERMINAL",0,0,"9839983",,terminal_output +3914,15617012,"TERMINAL",0,0,"5094404094",,terminal_output +3915,15618075,"TERMINAL",0,0,"150511505",,terminal_output +3916,15619109,"TERMINAL",0,0,"2162216",,terminal_output +3917,15620157,"TERMINAL",0,0,"3273327",,terminal_output +3918,15621450,"TERMINAL",0,0,"4384438",,terminal_output +3919,15622284,"TERMINAL",0,0,"5495549",,terminal_output +3920,15623498,"TERMINAL",0,0,"659:0066510",,terminal_output +3921,15624472,"TERMINAL",0,0,"7617761",,terminal_output +3922,15625385,"TERMINAL",0,0,"8728872",,terminal_output +3923,15626498,"TERMINAL",0,0,"9839983",,terminal_output +3924,15627497,"TERMINAL",0,0,"8:0094505094",,terminal_output +3925,15628562,"TERMINAL",0,0,"18:005117:005",,terminal_output +3926,15629579,"TERMINAL",0,0,"2162216",,terminal_output +3927,15630631,"TERMINAL",0,0,"3273327",,terminal_output +3928,15631687,"TERMINAL",0,0,"4384438",,terminal_output +3929,15632806,"TERMINAL",0,0,"5495549",,terminal_output +3930,15633837,"TERMINAL",0,0,"651066520",,terminal_output +3931,15634806,"TERMINAL",0,0,"7617761",,terminal_output +3932,15635861,"TERMINAL",0,0,"8839983",,terminal_output +3933,15636899,"TERMINAL",0,0,"10945:007:0094",,terminal_output +3934,15637944,"TERMINAL",0,0,"110511105",,terminal_output +3935,15639062,"TERMINAL",0,0,"2162216",,terminal_output +3936,15640034,"TERMINAL",0,0,"3273327",,terminal_output +3937,15641220,"TERMINAL",0,0,"4384438",,terminal_output +3938,15642134,"TERMINAL",0,0,"5495549",,terminal_output +3939,15643268,"TERMINAL",0,0,"652066530",,terminal_output +3940,15644226,"TERMINAL",0,0,"7617761",,terminal_output +3941,15645450,"TERMINAL",0,0,"8728872",,terminal_output +3942,15646330,"TERMINAL",0,0,"9839983",,terminal_output +3943,15647458,"TERMINAL",0,0,"2094101094",,terminal_output +3944,15648690,"TERMINAL",0,0,"120511205",,terminal_output +3945,15649508,"TERMINAL",0,0,"2162216",,terminal_output +3946,15650485,"TERMINAL",0,0,"3273327",,terminal_output +3947,15651557,"TERMINAL",0,0,"4384438",,terminal_output +3948,15652588,"TERMINAL",0,0,"5495549",,terminal_output +3949,15653641,"TERMINAL",0,0,"653066540",,terminal_output +3950,15654681,"TERMINAL",0,0,"7617761",,terminal_output +3951,15655736,"TERMINAL",0,0,"8728872",,terminal_output +3952,15656779,"TERMINAL",0,0,"9839983",,terminal_output +3953,15657824,"TERMINAL",0,0,"303052121305",,terminal_output +3954,15659027,"TERMINAL",0,0,"2162216",,terminal_output +3955,15659950,"TERMINAL",0,0,"3273327",,terminal_output +3956,15661182,"TERMINAL",0,0,"4384438",,terminal_output +3957,15662005,"TERMINAL",0,0,"5495549",,terminal_output +3958,15663048,"TERMINAL",0,0,"654066550",,terminal_output +3959,15664188,"TERMINAL",0,0,"7617761",,terminal_output +3960,15665283,"TERMINAL",0,0,"8728872",,terminal_output +3961,15666305,"TERMINAL",0,0,"9839983",,terminal_output +3962,15667526,"TERMINAL",0,0,"4094303094",,terminal_output +3963,15668348,"TERMINAL",0,0,"140511405",,terminal_output +3964,15669372,"TERMINAL",0,0,"2162216",,terminal_output +3965,15670605,"TERMINAL",0,0,"3273327",,terminal_output +3966,15671473,"TERMINAL",0,0,"4384438",,terminal_output +3967,15672546,"TERMINAL",0,0,"5495549",,terminal_output +3968,15673504,"TERMINAL",0,0,"65506658:00",,terminal_output +3969,15674546,"TERMINAL",0,0,"7617761",,terminal_output +3970,15675597,"TERMINAL",0,0,"8728872",,terminal_output +3971,15676637,"TERMINAL",0,0,"9839983",,terminal_output +3972,15677680,"TERMINAL",0,0,"5094404094",,terminal_output +3973,15678722,"TERMINAL",0,0,"150511505",,terminal_output +3974,15679814,"TERMINAL",0,0,"2162216",,terminal_output +3975,15680809,"TERMINAL",0,0,"3273327",,terminal_output +3976,15681865,"TERMINAL",0,0,"4495549",,terminal_output +3977,15682920,"TERMINAL",0,0,"6510:0066510",,terminal_output +3978,15683958,"TERMINAL",0,0,"7617761",,terminal_output +3979,15685038,"TERMINAL",0,0,"8728872",,terminal_output +3980,15686061,"TERMINAL",0,0,"9839983",,terminal_output +3981,15687200,"TERMINAL",0,0,"9:0094505094",,terminal_output +3982,15688211,"TERMINAL",0,0,"19:005118:005",,terminal_output +3983,15689190,"TERMINAL",0,0,"2162216",,terminal_output +3984,15690276,"TERMINAL",0,0,"3273327",,terminal_output +3985,15691280,"TERMINAL",0,0,"4384438",,terminal_output +3986,15692332,"TERMINAL",0,0,"5495549",,terminal_output +3987,15693535,"TERMINAL",0,0,"651066520",,terminal_output +3988,15694492,"TERMINAL",0,0,"7617761",,terminal_output +3989,15695484,"TERMINAL",0,0,"8728872",,terminal_output +3990,15696610,"TERMINAL",0,0,"9839983",,terminal_output +3991,15697639,"TERMINAL",0,0,"10946:008:0094",,terminal_output +3992,15698658,"TERMINAL",0,0,"110511105",,terminal_output +3993,15699680,"TERMINAL",0,0,"2162216",,terminal_output +3994,15700810,"TERMINAL",0,0,"3273327",,terminal_output +3995,15701761,"TERMINAL",0,0,"4384438",,terminal_output +3996,15702837,"TERMINAL",0,0,"5495549",,terminal_output +3997,15703986,"TERMINAL",0,0,"662177631",,terminal_output +3998,15704903,"TERMINAL",0,0,"8728872",,terminal_output +3999,15706028,"TERMINAL",0,0,"9839983",,terminal_output +4000,15706983,"TERMINAL",0,0,"2094101094",,terminal_output +4001,15708077,"TERMINAL",0,0,"120511205",,terminal_output +4002,15709082,"TERMINAL",0,0,"2162216",,terminal_output +4003,15710330,"TERMINAL",0,0,"3273327",,terminal_output +4004,15711175,"TERMINAL",0,0,"4384438",,terminal_output +4005,15712276,"TERMINAL",0,0,"5495549",,terminal_output +4006,15713298,"TERMINAL",0,0,"653066540",,terminal_output +4007,15714327,"TERMINAL",0,0,"7617761",,terminal_output +4008,15715490,"TERMINAL",0,0,"8728872",,terminal_output +4009,15716467,"TERMINAL",0,0,"9839983",,terminal_output +4010,15717478,"TERMINAL",0,0,"3094202094",,terminal_output +4011,15718532,"TERMINAL",0,0,"130511305",,terminal_output +4012,15719561,"TERMINAL",0,0,"2162216",,terminal_output +4013,15720615,"TERMINAL",0,0,"3273327",,terminal_output +4014,15721692,"TERMINAL",0,0,"4384438",,terminal_output +4015,15722702,"TERMINAL",0,0,"5495549",,terminal_output +4016,15723748,"TERMINAL",0,0,"654066550",,terminal_output +4017,15724872,"TERMINAL",0,0,"7617761",,terminal_output +4018,15725895,"TERMINAL",0,0,"8839983",,terminal_output +4019,15726885,"TERMINAL",0,0,"4094303094",,terminal_output +4020,15728207,"TERMINAL",0,0,"140511405",,terminal_output +4021,15729026,"TERMINAL",0,0,"2162216",,terminal_output +4022,15730196,"TERMINAL",0,0,"3273327",,terminal_output +4023,15731219,"TERMINAL",0,0,"4384438",,terminal_output +4024,15732255,"TERMINAL",0,0,"5495549",,terminal_output +4025,15733269,"TERMINAL",0,0,"65506659:00",,terminal_output +4026,15734296,"TERMINAL",0,0,"7617761",,terminal_output +4027,15735352,"TERMINAL",0,0,"8728872",,terminal_output +4028,15736543,"TERMINAL",0,0,"9839983",,terminal_output +4029,15737571,"TERMINAL",0,0,"5094404094",,terminal_output +4030,15738491,"TERMINAL",0,0,"150511505",,terminal_output +4031,15739538,"TERMINAL",0,0,"2162216",,terminal_output +4032,15740592,"TERMINAL",0,0,"3273327",,terminal_output +4033,15741665,"TERMINAL",0,0,"4384438",,terminal_output +4034,15742841,"TERMINAL",0,0,"5495549",,terminal_output +4035,15743727,"TERMINAL",0,0,"651:0066510",,terminal_output +4036,15744897,"TERMINAL",0,0,"7617761",,terminal_output +4037,15745864,"TERMINAL",0,0,"8839983",,terminal_output +4038,15746862,"TERMINAL",0,0,"5:00:0094505094",,terminal_output +4039,15747912,"TERMINAL",0,0,"11:00:005119:005",,terminal_output +4040,15749014,"TERMINAL",0,0,"2162216",,terminal_output +4041,15750045,"TERMINAL",0,0,"3273327",,terminal_output +4042,15751086,"TERMINAL",0,0,"4384438",,terminal_output +4043,15752110,"TERMINAL",0,0,"5495549",,terminal_output +4044,15753238,"TERMINAL",0,0,"651066520",,terminal_output +4045,15754190,"TERMINAL",0,0,"7617761",,terminal_output +4046,15755284,"TERMINAL",0,0,"8728872",,terminal_output +4047,15756411,"TERMINAL",0,0,"9839983",,terminal_output +4048,15757436,"TERMINAL",0,0,"10947:009:0094",,terminal_output +4049,15758552,"TERMINAL",0,0,"110511105",,terminal_output +4050,15759544,"TERMINAL",0,0,"2162216",,terminal_output +4051,15760528,"TERMINAL",0,0,"3273327",,terminal_output +4052,15761551,"TERMINAL",0,0,"4384438",,terminal_output +4053,15762584,"TERMINAL",0,0,"5495549",,terminal_output +4054,15763633,"TERMINAL",0,0,"652066530",,terminal_output +4055,15764694,"TERMINAL",0,0,"7617761",,terminal_output +4056,15765832,"TERMINAL",0,0,"8728872",,terminal_output +4057,15766805,"TERMINAL",0,0,"9839983",,terminal_output +4058,15767844,"TERMINAL",0,0,"202051111205",,terminal_output +4059,15768903,"TERMINAL",0,0,"2162216",,terminal_output +4060,15770030,"TERMINAL",0,0,"3273327",,terminal_output +4061,15771052,"TERMINAL",0,0,"4384438",,terminal_output +4062,15772032,"TERMINAL",0,0,"5495549",,terminal_output +4063,15773070,"TERMINAL",0,0,"653066540",,terminal_output +4064,15774127,"TERMINAL",0,0,"7617761",,terminal_output +4065,15775177,"TERMINAL",0,0,"8728872",,terminal_output +4066,15776214,"TERMINAL",0,0,"9839983",,terminal_output +4067,15777261,"TERMINAL",0,0,"3094202094",,terminal_output +4068,15778310,"TERMINAL",0,0,"130511305",,terminal_output +4069,15779490,"TERMINAL",0,0,"2162216",,terminal_output +4070,15780580,"TERMINAL",0,0,"3273327",,terminal_output +4071,15781462,"TERMINAL",0,0,"4384438",,terminal_output +4072,15782522,"TERMINAL",0,0,"5495549",,terminal_output +4073,15783558,"TERMINAL",0,0,"654066550",,terminal_output +4074,15784604,"TERMINAL",0,0,"7617761",,terminal_output +4075,15785662,"TERMINAL",0,0,"8728872",,terminal_output +4076,15786711,"TERMINAL",0,0,"9839983",,terminal_output +4077,15787772,"TERMINAL",0,0,"4094303094",,terminal_output +4078,15788872,"TERMINAL",0,0,"140511405",,terminal_output +4079,15789854,"TERMINAL",0,0,"2273327",,terminal_output +4080,15790918,"TERMINAL",0,0,"4384438",,terminal_output +4081,15791962,"TERMINAL",0,0,"5495549",,terminal_output +4082,15792994,"TERMINAL",0,0,"65506653:00:00",,terminal_output +4083,15794098,"TERMINAL",0,0,"7617761",,terminal_output +4084,15795089,"TERMINAL",0,0,"8728872",,terminal_output +4085,15796149,"TERMINAL",0,0,"9839983",,terminal_output +4086,15797184,"TERMINAL",0,0,"5094404094",,terminal_output +4087,15798255,"TERMINAL",0,0,"150511505",,terminal_output +4088,15799284,"TERMINAL",0,0,"2162216",,terminal_output +4089,15800330,"TERMINAL",0,0,"3273327",,terminal_output +4090,15801388,"TERMINAL",0,0,"4384438",,terminal_output +4091,15802598,"TERMINAL",0,0,"5495549",,terminal_output +4092,15803513,"TERMINAL",0,0,"652:0066510",,terminal_output +4093,15804570,"TERMINAL",0,0,"7617761",,terminal_output +4094,15805671,"TERMINAL",0,0,"8728872",,terminal_output +4095,15806645,"TERMINAL",0,0,"9839983",,terminal_output +4096,15807668,"TERMINAL",0,0,"1:0094505094",,terminal_output +4097,15808736,"TERMINAL",0,0,"11:005113:00:005",,terminal_output +4098,15809765,"TERMINAL",0,0,"2162216",,terminal_output +4099,15810886,"TERMINAL",0,0,"3273327",,terminal_output +4100,15811910,"TERMINAL",0,0,"4495549",,terminal_output +4101,15812903,"TERMINAL",0,0,"651066520",,terminal_output +4102,15813950,"TERMINAL",0,0,"7617761",,terminal_output +4103,15815004,"TERMINAL",0,0,"8728872",,terminal_output +4104,15816082,"TERMINAL",0,0,"9839983",,terminal_output +4105,15817094,"TERMINAL",0,0,"10948:003:00:0094",,terminal_output +4106,15818160,"TERMINAL",0,0,"110511105",,terminal_output +4107,15819190,"TERMINAL",0,0,"2162216",,terminal_output +4108,15820232,"TERMINAL",0,0,"3273327",,terminal_output +4109,15821292,"TERMINAL",0,0,"4384438",,terminal_output +4110,15822357,"TERMINAL",0,0,"5495549",,terminal_output +4111,15823376,"TERMINAL",0,0,"652066530",,terminal_output +4112,15824419,"TERMINAL",0,0,"7617761",,terminal_output +4113,15825456,"TERMINAL",0,0,"8728872",,terminal_output +4114,15826546,"TERMINAL",0,0,"9839983",,terminal_output +4115,15827550,"TERMINAL",0,0,"2094101094",,terminal_output +4116,15828608,"TERMINAL",0,0,"120511205",,terminal_output +4117,15829644,"TERMINAL",0,0,"2162216",,terminal_output +4118,15830690,"TERMINAL",0,0,"3273327",,terminal_output +4119,15831747,"TERMINAL",0,0,"4384438",,terminal_output +4120,15832798,"TERMINAL",0,0,"5495549",,terminal_output +4121,15833848,"TERMINAL",0,0,"663177641",,terminal_output +4122,15834895,"TERMINAL",0,0,"8728872",,terminal_output +4123,15835943,"TERMINAL",0,0,"9839983",,terminal_output +4124,15836985,"TERMINAL",0,0,"3094202094",,terminal_output +4125,15838048,"TERMINAL",0,0,"130511305",,terminal_output +4126,15839098,"TERMINAL",0,0,"2162216",,terminal_output +4127,15840141,"TERMINAL",0,0,"3273327",,terminal_output +4128,15841186,"TERMINAL",0,0,"4384438",,terminal_output +4129,15842230,"TERMINAL",0,0,"5495549",,terminal_output +4130,15843285,"TERMINAL",0,0,"654066550",,terminal_output +4131,15844358,"TERMINAL",0,0,"7617761",,terminal_output +4132,15845405,"TERMINAL",0,0,"8728872",,terminal_output +4133,15846444,"TERMINAL",0,0,"9839983",,terminal_output +4134,15847536,"TERMINAL",0,0,"4094303094",,terminal_output +4135,15848553,"TERMINAL",0,0,"140511405",,terminal_output +4136,15849584,"TERMINAL",0,0,"2162216",,terminal_output +4137,15850622,"TERMINAL",0,0,"3273327",,terminal_output +4138,15851664,"TERMINAL",0,0,"4384438",,terminal_output +4139,15852713,"TERMINAL",0,0,"5495549",,terminal_output +4140,15853757,"TERMINAL",0,0,"65506651:00",,terminal_output +4141,15854807,"TERMINAL",0,0,"7617761",,terminal_output +4142,15855861,"TERMINAL",0,0,"8839983",,terminal_output +4143,15856905,"TERMINAL",0,0,"5094404094",,terminal_output +4144,15857953,"TERMINAL",0,0,"150511505",,terminal_output +4145,15858983,"TERMINAL",0,0,"2162216",,terminal_output +4146,15860036,"TERMINAL",0,0,"3273327",,terminal_output +4147,15861089,"TERMINAL",0,0,"4384438",,terminal_output +4148,15862134,"TERMINAL",0,0,"5495549",,terminal_output +4149,15863184,"TERMINAL",0,0,"653:0066510",,terminal_output +4150,15864235,"TERMINAL",0,0,"7617761",,terminal_output +4151,15865262,"TERMINAL",0,0,"8728872",,terminal_output +4152,15866313,"TERMINAL",0,0,"9839983",,terminal_output +4153,15867370,"TERMINAL",0,0,"2:0094505094",,terminal_output +4154,15868412,"TERMINAL",0,0,"12:005111:005",,terminal_output +4155,15869456,"TERMINAL",0,0,"2162216",,terminal_output +4156,15870526,"TERMINAL",0,0,"3273327",,terminal_output +4157,15871545,"TERMINAL",0,0,"4384438",,terminal_output +4158,15872588,"TERMINAL",0,0,"5495549",,terminal_output +4159,15873628,"TERMINAL",0,0,"651066520",,terminal_output +4160,15874677,"TERMINAL",0,0,"7617761",,terminal_output +4161,15875729,"TERMINAL",0,0,"8728872",,terminal_output +4162,15876972,"TERMINAL",0,0,"9839983",,terminal_output +4163,15877829,"TERMINAL",0,0,"101059:011:01105",,terminal_output +4164,15878880,"TERMINAL",0,0,"2162216",,terminal_output +4165,15879930,"TERMINAL",0,0,"3273327",,terminal_output +4166,15880972,"TERMINAL",0,0,"4384438",,terminal_output +4167,15882019,"TERMINAL",0,0,"5495549",,terminal_output +4168,15883069,"TERMINAL",0,0,"652066530",,terminal_output +4169,15884119,"TERMINAL",0,0,"7617761",,terminal_output +4170,15885175,"TERMINAL",0,0,"8728872",,terminal_output +4171,15886216,"TERMINAL",0,0,"9839983",,terminal_output +4172,15887256,"TERMINAL",0,0,"2094101094",,terminal_output +4173,15888307,"TERMINAL",0,0,"120511205",,terminal_output +4174,15889355,"TERMINAL",0,0,"2162216",,terminal_output +4175,15890409,"TERMINAL",0,0,"3273327",,terminal_output +4176,15891453,"TERMINAL",0,0,"4384438",,terminal_output +4177,15892532,"TERMINAL",0,0,"5495549",,terminal_output +4178,15893550,"TERMINAL",0,0,"653066540",,terminal_output +4179,15894592,"TERMINAL",0,0,"7617761",,terminal_output +4180,15895642,"TERMINAL",0,0,"8728872",,terminal_output +4181,15896734,"TERMINAL",0,0,"9839983",,terminal_output +4182,15897751,"TERMINAL",0,0,"3094202094",,terminal_output +4183,15898793,"TERMINAL",0,0,"130511305",,terminal_output +4184,15899839,"TERMINAL",0,0,"2273327",,terminal_output +4185,15900895,"TERMINAL",0,0,"4384438",,terminal_output +4186,15901972,"TERMINAL",0,0,"5495549",,terminal_output +4187,15903001,"TERMINAL",0,0,"654066550",,terminal_output +4188,15904053,"TERMINAL",0,0,"7617761",,terminal_output +4189,15905104,"TERMINAL",0,0,"8728872",,terminal_output +4190,15906152,"TERMINAL",0,0,"9839983",,terminal_output +4191,15907200,"TERMINAL",0,0,"4094303094",,terminal_output +4192,15908249,"TERMINAL",0,0,"140511405",,terminal_output +4193,15909289,"TERMINAL",0,0,"2162216",,terminal_output +4194,15910339,"TERMINAL",0,0,"3273327",,terminal_output +4195,15911386,"TERMINAL",0,0,"4384438",,terminal_output +4196,15912428,"TERMINAL",0,0,"5495549",,terminal_output +4197,15913479,"TERMINAL",0,0,"65506652:00",,terminal_output +4198,15914533,"TERMINAL",0,0,"7617761",,terminal_output +4199,15915588,"TERMINAL",0,0,"8728872",,terminal_output +4200,15916630,"TERMINAL",0,0,"9839983",,terminal_output +4201,15917680,"TERMINAL",0,0,"5094404094",,terminal_output +4202,15918722,"TERMINAL",0,0,"150511505",,terminal_output +4203,15919770,"TERMINAL",0,0,"2162216",,terminal_output +4204,15920827,"TERMINAL",0,0,"3384438",,terminal_output +4205,15921860,"TERMINAL",0,0,"5495549",,terminal_output +4206,15922913,"TERMINAL",0,0,"654:0066510",,terminal_output +4207,15923964,"TERMINAL",0,0,"7617761",,terminal_output +4208,15925005,"TERMINAL",0,0,"8728872",,terminal_output +4209,15926059,"TERMINAL",0,0,"9839983",,terminal_output +4210,15927104,"TERMINAL",0,0,"3:0094505094",,terminal_output +4211,15928159,"TERMINAL",0,0,"13:005112:005",,terminal_output +4212,15929203,"TERMINAL",0,0,"2162216",,terminal_output +4213,15930255,"TERMINAL",0,0,"3273327",,terminal_output +4214,15931325,"TERMINAL",0,0,"4384438",,terminal_output +4215,15932329,"TERMINAL",0,0,"5495549",,terminal_output +4216,15933377,"TERMINAL",0,0,"651066520",,terminal_output +4217,15934416,"TERMINAL",0,0,"7617761",,terminal_output +4218,15935462,"TERMINAL",0,0,"8728872",,terminal_output +4219,15936554,"TERMINAL",0,0,"9839983",,terminal_output +4220,15937573,"TERMINAL",0,0,"109450:002:0094",,terminal_output +4221,15938609,"TERMINAL",0,0,"110511105",,terminal_output +4222,15939666,"TERMINAL",0,0,"2162216",,terminal_output +4223,15940710,"TERMINAL",0,0,"3273327",,terminal_output +4224,15941758,"TERMINAL",0,0,"4384438",,terminal_output +4225,15942805,"TERMINAL",0,0,"5495549",,terminal_output +4226,15943860,"TERMINAL",0,0,"662177631",,terminal_output +4227,15944902,"TERMINAL",0,0,"8728872",,terminal_output +4228,15945953,"TERMINAL",0,0,"9839983",,terminal_output +4229,15947009,"TERMINAL",0,0,"2094101094",,terminal_output +4230,15948065,"TERMINAL",0,0,"120511205",,terminal_output +4231,15949106,"TERMINAL",0,0,"2162216",,terminal_output +4232,15950153,"TERMINAL",0,0,"3273327",,terminal_output +4233,15951207,"TERMINAL",0,0,"4384438",,terminal_output +4234,15952259,"TERMINAL",0,0,"5495549",,terminal_output +4235,15953309,"TERMINAL",0,0,"653066540",,terminal_output +4236,15954367,"TERMINAL",0,0,"7617761",,terminal_output +4237,15955404,"TERMINAL",0,0,"8728872",,terminal_output +4238,15956481,"TERMINAL",0,0,"9839983",,terminal_output +4239,15957564,"TERMINAL",0,0,"3094202094",,terminal_output +4240,15958586,"TERMINAL",0,0,"130511305",,terminal_output +4241,15959629,"TERMINAL",0,0,"2162216",,terminal_output +4242,15960681,"TERMINAL",0,0,"3273327",,terminal_output +4243,15961726,"TERMINAL",0,0,"4384438",,terminal_output +4244,15962768,"TERMINAL",0,0,"5495549",,terminal_output +4245,15963813,"TERMINAL",0,0,"664177651",,terminal_output +4246,15964860,"TERMINAL",0,0,"8728872",,terminal_output +4247,15965904,"TERMINAL",0,0,"9839983",,terminal_output +4248,15966976,"TERMINAL",0,0,"4094303094",,terminal_output +4249,15967994,"TERMINAL",0,0,"140511405",,terminal_output +4250,15969035,"TERMINAL",0,0,"2162216",,terminal_output +4251,15970083,"TERMINAL",0,0,"3273327",,terminal_output +4252,15971132,"TERMINAL",0,0,"4384438",,terminal_output +4253,15972185,"TERMINAL",0,0,"5495549",,terminal_output +4254,15973228,"TERMINAL",0,0,"65506653:00",,terminal_output +4255,15974276,"TERMINAL",0,0,"7617761",,terminal_output +4256,15975324,"TERMINAL",0,0,"8728872",,terminal_output +4257,15976378,"TERMINAL",0,0,"9839983",,terminal_output +4258,15977437,"TERMINAL",0,0,"5094404094",,terminal_output +4259,15978488,"TERMINAL",0,0,"150511505",,terminal_output +4260,15979559,"TERMINAL",0,0,"2162216",,terminal_output +4261,15980576,"TERMINAL",0,0,"3273327",,terminal_output +4262,15981621,"TERMINAL",0,0,"4384438",,terminal_output +4263,15982669,"TERMINAL",0,0,"5495549",,terminal_output +4264,15983720,"TERMINAL",0,0,"655:0066510",,terminal_output +4265,15984774,"TERMINAL",0,0,"7617761",,terminal_output +4266,15985820,"TERMINAL",0,0,"8839983",,terminal_output +4267,15986874,"TERMINAL",0,0,"4:0094505094",,terminal_output +4268,15987923,"TERMINAL",0,0,"14:005113:005",,terminal_output +4269,15988981,"TERMINAL",0,0,"2162216",,terminal_output +4270,15990027,"TERMINAL",0,0,"3273327",,terminal_output +4271,15991094,"TERMINAL",0,0,"4384438",,terminal_output +4272,15992126,"TERMINAL",0,0,"5495549",,terminal_output +4273,15993161,"TERMINAL",0,0,"651066520",,terminal_output +4274,15994207,"TERMINAL",0,0,"7617761",,terminal_output +4275,15995255,"TERMINAL",0,0,"8728872",,terminal_output +4276,15996300,"TERMINAL",0,0,"9839983",,terminal_output +4277,15997351,"TERMINAL",0,0,"10941:003:0094",,terminal_output +4278,15998408,"TERMINAL",0,0,"110511105",,terminal_output +4279,15999448,"TERMINAL",0,0,"2162216",,terminal_output +4280,16000504,"TERMINAL",0,0,"3273327",,terminal_output +4281,16001585,"TERMINAL",0,0,"4384438",,terminal_output +4282,16002583,"TERMINAL",0,0,"5495549",,terminal_output +4283,16003631,"TERMINAL",0,0,"652066530",,terminal_output +4284,16004680,"TERMINAL",0,0,"7617761",,terminal_output +4285,16005727,"TERMINAL",0,0,"8728872",,terminal_output +4286,16006769,"TERMINAL",0,0,"9839983",,terminal_output +4287,16008165,"TERMINAL",0,0,"202051111205",,terminal_output +4288,16009196,"TERMINAL",0,0,"2162216",,terminal_output +4289,16010247,"TERMINAL",0,0,"3273327",,terminal_output +4290,16011286,"TERMINAL",0,0,"4384438",,terminal_output +4291,16012340,"TERMINAL",0,0,"5495549",,terminal_output +4292,16013388,"TERMINAL",0,0,"653066540",,terminal_output +4293,16014438,"TERMINAL",0,0,"7617761",,terminal_output +4294,16015489,"TERMINAL",0,0,"8728872",,terminal_output +4295,16016553,"TERMINAL",0,0,"9839983",,terminal_output +4296,16017577,"TERMINAL",0,0,"3094202094",,terminal_output +4297,16018626,"TERMINAL",0,0,"130511305",,terminal_output +4298,16019676,"TERMINAL",0,0,"2162216",,terminal_output +4299,16020752,"TERMINAL",0,0,"3273327",,terminal_output +4300,16021801,"TERMINAL",0,0,"4384438",,terminal_output +4301,16022842,"TERMINAL",0,0,"554066550",,terminal_output +4302,16023890,"TERMINAL",0,0,"7617761",,terminal_output +4303,16024930,"TERMINAL",0,0,"8728872",,terminal_output +4304,16025984,"TERMINAL",0,0,"9839983",,terminal_output +4305,16027024,"TERMINAL",0,0,"4094303094",,terminal_output +4306,16028067,"TERMINAL",0,0,"140511405",,terminal_output +4307,16029118,"TERMINAL",0,0,"2162216",,terminal_output +4308,16030157,"TERMINAL",0,0,"3273327",,terminal_output +4309,16031202,"TERMINAL",0,0,"4384438",,terminal_output +4310,16032254,"TERMINAL",0,0,"5495549",,terminal_output +4311,16033293,"TERMINAL",0,0,"65506654:00",,terminal_output +4312,16034342,"TERMINAL",0,0,"7617761",,terminal_output +4313,16035390,"TERMINAL",0,0,"8728872",,terminal_output +4314,16036441,"TERMINAL",0,0,"9839983",,terminal_output +4315,16037492,"TERMINAL",0,0,"5094404094",,terminal_output +4316,16038582,"TERMINAL",0,0,"150511505",,terminal_output +4317,16039592,"TERMINAL",0,0,"2162216",,terminal_output +4318,16040645,"TERMINAL",0,0,"3273327",,terminal_output +4319,16041694,"TERMINAL",0,0,"4384438",,terminal_output +4320,16042738,"TERMINAL",0,0,"5495549",,terminal_output +4321,16043789,"TERMINAL",0,0,"656:0066510",,terminal_output +4322,16044840,"TERMINAL",0,0,"7728872",,terminal_output +4323,16045887,"TERMINAL",0,0,"9839983",,terminal_output +4324,16046941,"TERMINAL",0,0,"5:0094505094",,terminal_output +4325,16047997,"TERMINAL",0,0,"15:005114:005",,terminal_output +4326,16049043,"TERMINAL",0,0,"2162216",,terminal_output +4327,16050092,"TERMINAL",0,0,"3273327",,terminal_output +4328,16051129,"TERMINAL",0,0,"4384438",,terminal_output +4329,16052189,"TERMINAL",0,0,"5495549",,terminal_output +4330,16053235,"TERMINAL",0,0,"651066520",,terminal_output +4331,16054279,"TERMINAL",0,0,"7617761",,terminal_output +4332,16055338,"TERMINAL",0,0,"8728872",,terminal_output +4333,16056380,"TERMINAL",0,0,"9839983",,terminal_output +4334,16057431,"TERMINAL",0,0,"10942:004:0094",,terminal_output +4335,16058468,"TERMINAL",0,0,"110511105",,terminal_output +4336,16059522,"TERMINAL",0,0,"2162216",,terminal_output +4337,16060617,"TERMINAL",0,0,"3273327",,terminal_output +4338,16061631,"TERMINAL",0,0,"4384438",,terminal_output +4339,16062680,"TERMINAL",0,0,"5495549",,terminal_output +4340,16063726,"TERMINAL",0,0,"652066530",,terminal_output +4341,16064778,"TERMINAL",0,0,"7617761",,terminal_output +4342,16065823,"TERMINAL",0,0,"8839983",,terminal_output +4343,16066874,"TERMINAL",0,0,"2094101094",,terminal_output +4344,16067917,"TERMINAL",0,0,"120511205",,terminal_output +4345,16068955,"TERMINAL",0,0,"2162216",,terminal_output +4346,16070079,"TERMINAL",0,0,"3273327",,terminal_output +4347,16071052,"TERMINAL",0,0,"4384438",,terminal_output +4348,16072101,"TERMINAL",0,0,"5495549",,terminal_output +4349,16073150,"TERMINAL",0,0,"653066540",,terminal_output +4350,16074191,"TERMINAL",0,0,"7617761",,terminal_output +4351,16075244,"TERMINAL",0,0,"8728872",,terminal_output +4352,16076295,"TERMINAL",0,0,"9839983",,terminal_output +4353,16077342,"TERMINAL",0,0,"3094202094",,terminal_output +4354,16078394,"TERMINAL",0,0,"130511305",,terminal_output +4355,16079436,"TERMINAL",0,0,"2162216",,terminal_output +4356,16080486,"TERMINAL",0,0,"3273327",,terminal_output +4357,16081571,"TERMINAL",0,0,"4384438",,terminal_output +4358,16082578,"TERMINAL",0,0,"5495549",,terminal_output +4359,16083626,"TERMINAL",0,0,"654066550",,terminal_output +4360,16084676,"TERMINAL",0,0,"7617761",,terminal_output +4361,16085729,"TERMINAL",0,0,"8728872",,terminal_output +4362,16086771,"TERMINAL",0,0,"9839983",,terminal_output +4363,16087816,"TERMINAL",0,0,"4094303094",,terminal_output +4364,16088862,"TERMINAL",0,0,"241622416",,terminal_output +4365,16089996,"TERMINAL",0,0,"3273327",,terminal_output +4366,16091046,"TERMINAL",0,0,"4384438",,terminal_output +4367,16092086,"TERMINAL",0,0,"5495549",,terminal_output +4368,16093133,"TERMINAL",0,0,"65506655:00",,terminal_output +4369,16094176,"TERMINAL",0,0,"7617761",,terminal_output +4370,16095230,"TERMINAL",0,0,"8728872",,terminal_output +4371,16096274,"TERMINAL",0,0,"9839983",,terminal_output +4372,16097321,"TERMINAL",0,0,"5094404094",,terminal_output +4373,16098370,"TERMINAL",0,0,"150511505",,terminal_output +4374,16099421,"TERMINAL",0,0,"2162216",,terminal_output +4375,16100470,"TERMINAL",0,0,"3273327",,terminal_output +4376,16101517,"TERMINAL",0,0,"4384438",,terminal_output +4377,16102598,"TERMINAL",0,0,"5495549",,terminal_output +4378,16103608,"TERMINAL",0,0,"657:0066510",,terminal_output +4379,16104648,"TERMINAL",0,0,"7617761",,terminal_output +4380,16105697,"TERMINAL",0,0,"8728872",,terminal_output +4381,16106741,"TERMINAL",0,0,"9839983",,terminal_output +4382,16107832,"TERMINAL",0,0,"6:0094505094",,terminal_output +4383,16108850,"TERMINAL",0,0,"16:016225:016",,terminal_output +4384,16109901,"TERMINAL",0,0,"3273327",,terminal_output +4385,16110933,"TERMINAL",0,0,"4384438",,terminal_output +4386,16111990,"TERMINAL",0,0,"5495549",,terminal_output +4387,16113058,"TERMINAL",0,0,"651066520",,terminal_output +4388,16114117,"TERMINAL",0,0,"7617761",,terminal_output +4389,16115124,"TERMINAL",0,0,"8728872",,terminal_output +4390,16116182,"TERMINAL",0,0,"9839983",,terminal_output +4391,16117221,"TERMINAL",0,0,"10943:005:0094",,terminal_output +4392,16118274,"TERMINAL",0,0,"110511105",,terminal_output +4393,16119318,"TERMINAL",0,0,"2162216",,terminal_output +4394,16120364,"TERMINAL",0,0,"3273327",,terminal_output +4395,16121411,"TERMINAL",0,0,"4384438",,terminal_output +4396,16122452,"TERMINAL",0,0,"5495549",,terminal_output +4397,16123501,"TERMINAL",0,0,"652066530",,terminal_output +4398,16124548,"TERMINAL",0,0,"7617761",,terminal_output +4399,16125631,"TERMINAL",0,0,"8728872",,terminal_output +4400,16126647,"TERMINAL",0,0,"9839983",,terminal_output +4401,16127692,"TERMINAL",0,0,"2094101094",,terminal_output +4402,16128737,"TERMINAL",0,0,"120511205",,terminal_output +4403,16129791,"TERMINAL",0,0,"2162216",,terminal_output +4404,16130837,"TERMINAL",0,0,"3384438",,terminal_output +4405,16131896,"TERMINAL",0,0,"5495549",,terminal_output +4406,16132943,"TERMINAL",0,0,"653066540",,terminal_output +4407,16133994,"TERMINAL",0,0,"7617761",,terminal_output +4408,16135048,"TERMINAL",0,0,"8728872",,terminal_output +4409,16136099,"TERMINAL",0,0,"9839983",,terminal_output +4410,16137146,"TERMINAL",0,0,"3094202094",,terminal_output +4411,16138187,"TERMINAL",0,0,"130511305",,terminal_output +4412,16139246,"TERMINAL",0,0,"2162216",,terminal_output +4413,16140284,"TERMINAL",0,0,"3273327",,terminal_output +4414,16141318,"TERMINAL",0,0,"4384438",,terminal_output +4415,16142374,"TERMINAL",0,0,"5495549",,terminal_output +4416,16143415,"TERMINAL",0,0,"654066550",,terminal_output +4417,16144460,"TERMINAL",0,0,"7617761",,terminal_output +4418,16145513,"TERMINAL",0,0,"8728872",,terminal_output +4419,16146571,"TERMINAL",0,0,"9839983",,terminal_output +4420,16147624,"TERMINAL",0,0,"4094303094",,terminal_output +4421,16148669,"TERMINAL",0,0,"140511405",,terminal_output +4422,16149717,"TERMINAL",0,0,"2162216",,terminal_output +4423,16150768,"TERMINAL",0,0,"3273327",,terminal_output +4424,16151812,"TERMINAL",0,0,"4384438",,terminal_output +4425,16152857,"TERMINAL",0,0,"55506656:00",,terminal_output +4426,16153904,"TERMINAL",0,0,"7617761",,terminal_output +4427,16154950,"TERMINAL",0,0,"8728872",,terminal_output +4428,16156006,"TERMINAL",0,0,"9839983",,terminal_output +4429,16157062,"TERMINAL",0,0,"5094404094",,terminal_output +4430,16158122,"TERMINAL",0,0,"150511505",,terminal_output +4431,16159174,"TERMINAL",0,0,"2162216",,terminal_output +4432,16160229,"TERMINAL",0,0,"3273327",,terminal_output +4433,16161272,"TERMINAL",0,0,"4384438",,terminal_output +4434,16162332,"TERMINAL",0,0,"5495549",,terminal_output +4435,16163377,"TERMINAL",0,0,"658:0066510",,terminal_output +4436,16164426,"TERMINAL",0,0,"7617761",,terminal_output +4437,16165472,"TERMINAL",0,0,"8728872",,terminal_output +4438,16166515,"TERMINAL",0,0,"9839983",,terminal_output +4439,16167555,"TERMINAL",0,0,"7:0094505094",,terminal_output +4440,16168646,"TERMINAL",0,0,"17:005116:005",,terminal_output +4441,16169633,"TERMINAL",0,0,"2162216",,terminal_output +4442,16170691,"TERMINAL",0,0,"3273327",,terminal_output +4443,16171742,"TERMINAL",0,0,"4384438",,terminal_output +4444,16172791,"TERMINAL",0,0,"5495549",,terminal_output +4445,16173829,"TERMINAL",0,0,"661177621",,terminal_output +4446,16174871,"TERMINAL",0,0,"8728872",,terminal_output +4447,16175914,"TERMINAL",0,0,"9839983",,terminal_output +4448,16176978,"TERMINAL",0,0,"10944:006:0094",,terminal_output +4449,16177996,"TERMINAL",0,0,"110511105",,terminal_output +4450,16179032,"TERMINAL",0,0,"2162216",,terminal_output +4451,16180104,"TERMINAL",0,0,"3273327",,terminal_output +4452,16181152,"TERMINAL",0,0,"4384438",,terminal_output +4453,16182196,"TERMINAL",0,0,"5495549",,terminal_output +4454,16183256,"TERMINAL",0,0,"652066530",,terminal_output +4455,16184309,"TERMINAL",0,0,"7617761",,terminal_output +4456,16185348,"TERMINAL",0,0,"8728872",,terminal_output +4457,16186404,"TERMINAL",0,0,"9839983",,terminal_output +4458,16187448,"TERMINAL",0,0,"2094101094",,terminal_output +4459,16188508,"TERMINAL",0,0,"120511205",,terminal_output +4460,16189544,"TERMINAL",0,0,"2162216",,terminal_output +4461,16190632,"TERMINAL",0,0,"3273327",,terminal_output +4462,16191632,"TERMINAL",0,0,"4384438",,terminal_output +4463,16192688,"TERMINAL",0,0,"5495549",,terminal_output +4464,16193738,"TERMINAL",0,0,"653066540",,terminal_output +4465,16194790,"TERMINAL",0,0,"7617761",,terminal_output +4466,16195838,"TERMINAL",0,0,"8839983",,terminal_output +4467,16196880,"TERMINAL",0,0,"3094202094",,terminal_output +4468,16197947,"TERMINAL",0,0,"130511305",,terminal_output +4469,16198974,"TERMINAL",0,0,"2162216",,terminal_output +4470,16200111,"TERMINAL",0,0,"3273327",,terminal_output +4471,16201069,"TERMINAL",0,0,"4384438",,terminal_output +4472,16202122,"TERMINAL",0,0,"5495549",,terminal_output +4473,16203166,"TERMINAL",0,0,"654066550",,terminal_output +4474,16204215,"TERMINAL",0,0,"7617761",,terminal_output +4475,16205269,"TERMINAL",0,0,"8728872",,terminal_output +4476,16206321,"TERMINAL",0,0,"9839983",,terminal_output +4477,16207373,"TERMINAL",0,0,"4094303094",,terminal_output +4478,16208421,"TERMINAL",0,0,"140511405",,terminal_output +4479,16209481,"TERMINAL",0,0,"2162216",,terminal_output +4480,16210530,"TERMINAL",0,0,"3273327",,terminal_output +4481,16211579,"TERMINAL",0,0,"4384438",,terminal_output +4482,16212637,"TERMINAL",0,0,"5495549",,terminal_output +4483,16213688,"TERMINAL",0,0,"65506657:00",,terminal_output +4484,16214744,"TERMINAL",0,0,"7617761",,terminal_output +4485,16215784,"TERMINAL",0,0,"8728872",,terminal_output +4486,16216841,"TERMINAL",0,0,"994404094",,terminal_output +4487,16217882,"TERMINAL",0,0,"5150511505",,terminal_output +4488,16218932,"TERMINAL",0,0,"2162216",,terminal_output +4489,16220001,"TERMINAL",0,0,"3273327",,terminal_output +4490,16221035,"TERMINAL",0,0,"4384438",,terminal_output +4491,16222088,"TERMINAL",0,0,"5495549",,terminal_output +4492,16223150,"TERMINAL",0,0,"659:0066510",,terminal_output +4493,16224189,"TERMINAL",0,0,"7617761",,terminal_output +4494,16225231,"TERMINAL",0,0,"8728872",,terminal_output +4495,16226278,"TERMINAL",0,0,"9839983",,terminal_output +4496,16227328,"TERMINAL",0,0,"8:0094505094",,terminal_output +4497,16228356,"TERMINAL",0,0,"18:005117:005",,terminal_output +4498,16229407,"TERMINAL",0,0,"2162216",,terminal_output +4499,16230462,"TERMINAL",0,0,"3273327",,terminal_output +4500,16231510,"TERMINAL",0,0,"4384438",,terminal_output +4501,16232571,"TERMINAL",0,0,"5495549",,terminal_output +4502,16233649,"TERMINAL",0,0,"651066520",,terminal_output +4503,16234652,"TERMINAL",0,0,"7617761",,terminal_output +4504,16235699,"TERMINAL",0,0,"8728872",,terminal_output +4505,16236754,"TERMINAL",0,0,"9839983",,terminal_output +4506,16237797,"TERMINAL",0,0,"10945:007:0094",,terminal_output +4507,16238844,"TERMINAL",0,0,"111622116",,terminal_output +4508,16239894,"TERMINAL",0,0,"3273327",,terminal_output +4509,16240951,"TERMINAL",0,0,"4384438",,terminal_output +4510,16242049,"TERMINAL",0,0,"5495549",,terminal_output +4511,16243061,"TERMINAL",0,0,"652066530",,terminal_output +4512,16244120,"TERMINAL",0,0,"7617761",,terminal_output +4513,16245170,"TERMINAL",0,0,"8728872",,terminal_output +4514,16246214,"TERMINAL",0,0,"9839983",,terminal_output +4515,16247267,"TERMINAL",0,0,"2094101094",,terminal_output +4516,16248325,"TERMINAL",0,0,"120511205",,terminal_output +4517,16249366,"TERMINAL",0,0,"2162216",,terminal_output +4518,16250415,"TERMINAL",0,0,"3273327",,terminal_output +4519,16251460,"TERMINAL",0,0,"4384438",,terminal_output +4520,16252522,"TERMINAL",0,0,"5495549",,terminal_output +4521,16253571,"TERMINAL",0,0,"653066540",,terminal_output +4522,16254655,"TERMINAL",0,0,"7617761",,terminal_output +4523,16255688,"TERMINAL",0,0,"8728872",,terminal_output +4524,16256724,"TERMINAL",0,0,"9839983",,terminal_output +4525,16257762,"TERMINAL",0,0,"3094202094",,terminal_output +4526,16258814,"TERMINAL",0,0,"130511305",,terminal_output +4527,16259876,"TERMINAL",0,0,"2273327",,terminal_output +4528,16260921,"TERMINAL",0,0,"4384438",,terminal_output +4529,16261973,"TERMINAL",0,0,"5495549",,terminal_output +4530,16263021,"TERMINAL",0,0,"654066550",,terminal_output +4531,16264078,"TERMINAL",0,0,"7617761",,terminal_output +4532,16265120,"TERMINAL",0,0,"8728872",,terminal_output +4533,16266173,"TERMINAL",0,0,"9839983",,terminal_output +4534,16267221,"TERMINAL",0,0,"4094303094",,terminal_output +4535,16268266,"TERMINAL",0,0,"140511405",,terminal_output +4536,16269306,"TERMINAL",0,0,"2162216",,terminal_output +4537,16270363,"TERMINAL",0,0,"3273327",,terminal_output +4538,16271405,"TERMINAL",0,0,"4384438",,terminal_output +4539,16272456,"TERMINAL",0,0,"5495549",,terminal_output +4540,16273508,"TERMINAL",0,0,"65506658:00",,terminal_output +4541,16274557,"TERMINAL",0,0,"7617761",,terminal_output +4542,16275620,"TERMINAL",0,0,"8728872",,terminal_output +4543,16276682,"TERMINAL",0,0,"9839983",,terminal_output +4544,16277709,"TERMINAL",0,0,"5094404094",,terminal_output +4545,16278764,"TERMINAL",0,0,"150511505",,terminal_output +4546,16279802,"TERMINAL",0,0,"2162216",,terminal_output +4547,16280858,"TERMINAL",0,0,"3384438",,terminal_output +4548,16281904,"TERMINAL",0,0,"5495549",,terminal_output +4549,16282952,"TERMINAL",0,0,"6520:0066510",,terminal_output +4550,16283991,"TERMINAL",0,0,"7617761",,terminal_output +4551,16285036,"TERMINAL",0,0,"8728872",,terminal_output +4552,16286073,"TERMINAL",0,0,"9839983",,terminal_output +4553,16287129,"TERMINAL",0,0,"9:0094505094",,terminal_output +4554,16288174,"TERMINAL",0,0,"19:005118:005",,terminal_output +4555,16289225,"TERMINAL",0,0,"2162216",,terminal_output +4556,16290272,"TERMINAL",0,0,"3273327",,terminal_output +4557,16291322,"TERMINAL",0,0,"4384438",,terminal_output +4558,16292377,"TERMINAL",0,0,"5495549",,terminal_output +4559,16293435,"TERMINAL",0,0,"651066520",,terminal_output +4560,16294483,"TERMINAL",0,0,"7617761",,terminal_output +4561,16295544,"TERMINAL",0,0,"8728872",,terminal_output +4562,16296593,"TERMINAL",0,0,"9839983",,terminal_output +4563,16297652,"TERMINAL",0,0,"10946:008:0094",,terminal_output +4564,16298700,"TERMINAL",0,0,"110511105",,terminal_output +4565,16299742,"TERMINAL",0,0,"2162216",,terminal_output +4566,16300817,"TERMINAL",0,0,"3273327",,terminal_output +4567,16301868,"TERMINAL",0,0,"4495549",,terminal_output +4568,16302920,"TERMINAL",0,0,"652066530",,terminal_output +4569,16303974,"TERMINAL",0,0,"7617761",,terminal_output +4570,16305025,"TERMINAL",0,0,"8728872",,terminal_output +4571,16306072,"TERMINAL",0,0,"9839983",,terminal_output +4572,16307117,"TERMINAL",0,0,"2094101094",,terminal_output +4573,16308178,"TERMINAL",0,0,"120511205",,terminal_output +4574,16309249,"TERMINAL",0,0,"2162216",,terminal_output +4575,16310297,"TERMINAL",0,0,"3273327",,terminal_output +4576,16311337,"TERMINAL",0,0,"4384438",,terminal_output +4577,16312389,"TERMINAL",0,0,"5495549",,terminal_output +4578,16313439,"TERMINAL",0,0,"653066540",,terminal_output +4579,16314487,"TERMINAL",0,0,"7617761",,terminal_output +4580,16315530,"TERMINAL",0,0,"8728872",,terminal_output +4581,16316600,"TERMINAL",0,0,"9839983",,terminal_output +4582,16317670,"TERMINAL",0,0,"3094202094",,terminal_output +4583,16318697,"TERMINAL",0,0,"130511305",,terminal_output +4584,16319734,"TERMINAL",0,0,"2162216",,terminal_output +4585,16320781,"TERMINAL",0,0,"3273327",,terminal_output +4586,16321832,"TERMINAL",0,0,"4495549",,terminal_output +4587,16322875,"TERMINAL",0,0,"654066550",,terminal_output +4588,16323923,"TERMINAL",0,0,"7617761",,terminal_output +4589,16325004,"TERMINAL",0,0,"8728872",,terminal_output +4590,16326023,"TERMINAL",0,0,"9839983",,terminal_output +4591,16327074,"TERMINAL",0,0,"4094303094",,terminal_output +4592,16328128,"TERMINAL",0,0,"140511405",,terminal_output +4593,16329183,"TERMINAL",0,0,"2162216",,terminal_output +4594,16330237,"TERMINAL",0,0,"3273327",,terminal_output +4595,16331278,"TERMINAL",0,0,"4384438",,terminal_output +4596,16332325,"TERMINAL",0,0,"5495549",,terminal_output +4597,16333368,"TERMINAL",0,0,"65506659:00",,terminal_output +4598,16334422,"TERMINAL",0,0,"7617761",,terminal_output +4599,16335489,"TERMINAL",0,0,"8728872",,terminal_output +4600,16336532,"TERMINAL",0,0,"9839983",,terminal_output +4601,16337573,"TERMINAL",0,0,"5094404094",,terminal_output +4602,16338625,"TERMINAL",0,0,"150511505",,terminal_output +4603,16339716,"TERMINAL",0,0,"2162216",,terminal_output +4604,16340714,"TERMINAL",0,0,"3273327",,terminal_output +4605,16341756,"TERMINAL",0,0,"4384438",,terminal_output +4606,16342807,"TERMINAL",0,0,"5495549",,terminal_output +4607,16343857,"TERMINAL",0,0,"661:0177611",,terminal_output +4608,16344899,"TERMINAL",0,0,"8728872",,terminal_output +4609,16345947,"TERMINAL",0,0,"9839983",,terminal_output +4610,16347023,"TERMINAL",0,0,"10:0094505094",,terminal_output +4611,16348043,"TERMINAL",0,0,"110:005119:005",,terminal_output +4612,16349079,"TERMINAL",0,0,"2162216",,terminal_output +4613,16350128,"TERMINAL",0,0,"3273327",,terminal_output +4614,16351181,"TERMINAL",0,0,"4384438",,terminal_output +4615,16352230,"TERMINAL",0,0,"5495549",,terminal_output +4616,16353280,"TERMINAL",0,0,"651066520",,terminal_output +4617,16354321,"TERMINAL",0,0,"7617761",,terminal_output +4618,16355367,"TERMINAL",0,0,"8728872",,terminal_output +4619,16356420,"TERMINAL",0,0,"9839983",,terminal_output +4620,16357461,"TERMINAL",0,0,"10947:009:0094",,terminal_output +4621,16358508,"TERMINAL",0,0,"110511105",,terminal_output +4622,16359550,"TERMINAL",0,0,"2162216",,terminal_output +4623,16360596,"TERMINAL",0,0,"3273327",,terminal_output +4624,16361649,"TERMINAL",0,0,"4384438",,terminal_output +4625,16362690,"TERMINAL",0,0,"5495549",,terminal_output +4626,16363743,"TERMINAL",0,0,"652066530",,terminal_output +4627,16364792,"TERMINAL",0,0,"7617761",,terminal_output +4628,16365844,"TERMINAL",0,0,"8839983",,terminal_output +4629,16366908,"TERMINAL",0,0,"2094101094",,terminal_output +4630,16367947,"TERMINAL",0,0,"120511205",,terminal_output +4631,16368997,"TERMINAL",0,0,"2162216",,terminal_output +4632,16370046,"TERMINAL",0,0,"3273327",,terminal_output +4633,16371096,"TERMINAL",0,0,"4384438",,terminal_output +4634,16372147,"TERMINAL",0,0,"5495549",,terminal_output +4635,16373211,"TERMINAL",0,0,"653066540",,terminal_output +4636,16374265,"TERMINAL",0,0,"7617761",,terminal_output +4637,16375322,"TERMINAL",0,0,"8728872",,terminal_output +4638,16376372,"TERMINAL",0,0,"9839983",,terminal_output +4639,16377432,"TERMINAL",0,0,"3094202094",,terminal_output +4640,16378487,"TERMINAL",0,0,"130511305",,terminal_output +4641,16379536,"TERMINAL",0,0,"2162216",,terminal_output +4642,16380574,"TERMINAL",0,0,"3273327",,terminal_output +4643,16381621,"TERMINAL",0,0,"4384438",,terminal_output +4644,16382712,"TERMINAL",0,0,"5495549",,terminal_output +4645,16383710,"TERMINAL",0,0,"654066550",,terminal_output +4646,16384754,"TERMINAL",0,0,"7617761",,terminal_output +4647,16385801,"TERMINAL",0,0,"8728872",,terminal_output +4648,16386852,"TERMINAL",0,0,"994303094",,terminal_output +4649,16387904,"TERMINAL",0,0,"4140511405",,terminal_output +4650,16388961,"TERMINAL",0,0,"2162216",,terminal_output +4651,16390012,"TERMINAL",0,0,"3273327",,terminal_output +4652,16391065,"TERMINAL",0,0,"4384438",,terminal_output +4653,16392115,"TERMINAL",0,0,"5495549",,terminal_output +4654,16393165,"TERMINAL",0,0,"655066510:00",,terminal_output +4655,16394208,"TERMINAL",0,0,"7617761",,terminal_output +4656,16395251,"TERMINAL",0,0,"8728872",,terminal_output +4657,16396298,"TERMINAL",0,0,"9839983",,terminal_output +4658,16397350,"TERMINAL",0,0,"5094404094",,terminal_output +4659,16398417,"TERMINAL",0,0,"150511505",,terminal_output +4660,16399448,"TERMINAL",0,0,"2162216",,terminal_output +4661,16400499,"TERMINAL",0,0,"3273327",,terminal_output +4662,16401540,"TERMINAL",0,0,"4384438",,terminal_output +4663,16402590,"TERMINAL",0,0,"5495549",,terminal_output +4664,16403638,"TERMINAL",0,0,"652:0066510",,terminal_output +4665,16404712,"TERMINAL",0,0,"7617761",,terminal_output +4666,16405736,"TERMINAL",0,0,"8728872",,terminal_output +4667,16406782,"TERMINAL",0,0,"9839983",,terminal_output +4668,16407833,"TERMINAL",0,0,"1:001:005515110:005",,terminal_output +4669,16408880,"TERMINAL",0,0,"2162216",,terminal_output +4670,16409933,"TERMINAL",0,0,"3273327",,terminal_output +4671,16410978,"TERMINAL",0,0,"4384438",,terminal_output +4672,16412033,"TERMINAL",0,0,"5495549",,terminal_output +4673,16413087,"TERMINAL",0,0,"651066520",,terminal_output +4674,16414137,"TERMINAL",0,0,"7617761",,terminal_output +4675,16415183,"TERMINAL",0,0,"8728872",,terminal_output +4676,16416225,"TERMINAL",0,0,"9839983",,terminal_output +4677,16417288,"TERMINAL",0,0,"10948:0010:0094",,terminal_output +4678,16418347,"TERMINAL",0,0,"110511105",,terminal_output +4679,16419396,"TERMINAL",0,0,"2162216",,terminal_output +4680,16420443,"TERMINAL",0,0,"3273327",,terminal_output +4681,16421508,"TERMINAL",0,0,"4384438",,terminal_output +4682,16422561,"TERMINAL",0,0,"5495549",,terminal_output +4683,16423617,"TERMINAL",0,0,"652066530",,terminal_output +4684,16424703,"TERMINAL",0,0,"7617761",,terminal_output +4685,16425712,"TERMINAL",0,0,"8728872",,terminal_output +4686,16426762,"TERMINAL",0,0,"9839983",,terminal_output +4687,16427807,"TERMINAL",0,0,"2094101094",,terminal_output +4688,16428856,"TERMINAL",0,0,"121622216",,terminal_output +4689,16429906,"TERMINAL",0,0,"3273327",,terminal_output +4690,16430958,"TERMINAL",0,0,"4384438",,terminal_output +4691,16432019,"TERMINAL",0,0,"5495549",,terminal_output +4692,16433066,"TERMINAL",0,0,"653066540",,terminal_output +4693,16434136,"TERMINAL",0,0,"7617761",,terminal_output +4694,16435169,"TERMINAL",0,0,"8728872",,terminal_output +4695,16436221,"TERMINAL",0,0,"9839983",,terminal_output +4696,16437272,"TERMINAL",0,0,"3094202094",,terminal_output +4697,16438317,"TERMINAL",0,0,"130511305",,terminal_output +4698,16439371,"TERMINAL",0,0,"2162216",,terminal_output +4699,16440418,"TERMINAL",0,0,"3273327",,terminal_output +4700,16441468,"TERMINAL",0,0,"4384438",,terminal_output +4701,16442517,"TERMINAL",0,0,"5495549",,terminal_output +4702,16443573,"TERMINAL",0,0,"654066550",,terminal_output +4703,16444623,"TERMINAL",0,0,"7617761",,terminal_output +4704,16445710,"TERMINAL",0,0,"8728872",,terminal_output +4705,16446736,"TERMINAL",0,0,"9839983",,terminal_output +4706,16447827,"TERMINAL",0,0,"4094303094",,terminal_output +4707,16448844,"TERMINAL",0,0,"141622416",,terminal_output +4708,16449895,"TERMINAL",0,0,"3273327",,terminal_output +4709,16451594,"TERMINAL",0,0,"4384438",,terminal_output +4710,16452647,"TERMINAL",0,0,"5495549",,terminal_output +4711,16453737,"TERMINAL",0,0,"65506651:00",,terminal_output +4712,16454759,"TERMINAL",0,0,"7617761",,terminal_output +4713,16455810,"TERMINAL",0,0,"8728872",,terminal_output +4714,16456868,"TERMINAL",0,0,"994404094",,terminal_output +4715,16457918,"TERMINAL",0,0,"5150511505",,terminal_output +4716,16458972,"TERMINAL",0,0,"2162216",,terminal_output +4717,16460018,"TERMINAL",0,0,"3273327",,terminal_output +4718,16461061,"TERMINAL",0,0,"4384438",,terminal_output +4719,16462109,"TERMINAL",0,0,"5495549",,terminal_output +4720,16463171,"TERMINAL",0,0,"653:0066510",,terminal_output +4721,16464209,"TERMINAL",0,0,"7617761",,terminal_output +4722,16465263,"TERMINAL",0,0,"8728872",,terminal_output +4723,16466313,"TERMINAL",0,0,"9839983",,terminal_output +4724,16467366,"TERMINAL",0,0,"2:0094505094",,terminal_output +4725,16468419,"TERMINAL",0,0,"12:005111:005",,terminal_output +4726,16469470,"TERMINAL",0,0,"2162216",,terminal_output +4727,16470520,"TERMINAL",0,0,"3273327",,terminal_output +4728,16471574,"TERMINAL",0,0,"4384438",,terminal_output +4729,16472619,"TERMINAL",0,0,"5495549",,terminal_output +4730,16473674,"TERMINAL",0,0,"651066520",,terminal_output +4731,16474756,"TERMINAL",0,0,"7617761",,terminal_output +4732,16475777,"TERMINAL",0,0,"8728872",,terminal_output +4733,16476831,"TERMINAL",0,0,"9949:001:0094",,terminal_output +4734,16477876,"TERMINAL",0,0,"1110511105",,terminal_output +4735,16478924,"TERMINAL",0,0,"2162216",,terminal_output +4736,16479986,"TERMINAL",0,0,"3273327",,terminal_output +4737,16481036,"TERMINAL",0,0,"4384438",,terminal_output +4738,16482097,"TERMINAL",0,0,"5495549",,terminal_output +4739,16483143,"TERMINAL",0,0,"652066530",,terminal_output +4740,16484193,"TERMINAL",0,0,"7617761",,terminal_output +4741,16485264,"TERMINAL",0,0,"8728872",,terminal_output +4742,16486311,"TERMINAL",0,0,"9839983",,terminal_output +4743,16487363,"TERMINAL",0,0,"2094101094",,terminal_output +4744,16488415,"TERMINAL",0,0,"120511205",,terminal_output +4745,16489471,"TERMINAL",0,0,"2162216",,terminal_output +4746,16490519,"TERMINAL",0,0,"3273327",,terminal_output +4747,16491593,"TERMINAL",0,0,"4384438",,terminal_output +4748,16492613,"TERMINAL",0,0,"5495549",,terminal_output +4749,16493657,"TERMINAL",0,0,"653066540",,terminal_output +4750,16494697,"TERMINAL",0,0,"7617761",,terminal_output +4751,16495747,"TERMINAL",0,0,"8728872",,terminal_output +4752,16496799,"TERMINAL",0,0,"9839983",,terminal_output +4753,16497852,"TERMINAL",0,0,"303052121305",,terminal_output +4754,16498904,"TERMINAL",0,0,"2162216",,terminal_output +4755,16499962,"TERMINAL",0,0,"3273327",,terminal_output +4756,16501020,"TERMINAL",0,0,"4384438",,terminal_output +4757,16502063,"TERMINAL",0,0,"5495549",,terminal_output +4758,16503113,"TERMINAL",0,0,"654066550",,terminal_output +4759,16504169,"TERMINAL",0,0,"7617761",,terminal_output +4760,16505215,"TERMINAL",0,0,"8728872",,terminal_output +4761,16506274,"TERMINAL",0,0,"9839983",,terminal_output +4762,16507331,"TERMINAL",0,0,"4094303094",,terminal_output +4763,16508378,"TERMINAL",0,0,"140511405",,terminal_output +4764,16509431,"TERMINAL",0,0,"2162216",,terminal_output +4765,16510478,"TERMINAL",0,0,"3273327",,terminal_output +4766,16511536,"TERMINAL",0,0,"4384438",,terminal_output +4767,16512593,"TERMINAL",0,0,"5495549",,terminal_output +4768,16513646,"TERMINAL",0,0,"65506652:00",,terminal_output +4769,16514696,"TERMINAL",0,0,"7617761",,terminal_output +4770,16515782,"TERMINAL",0,0,"8728872",,terminal_output +4771,16516809,"TERMINAL",0,0,"9839983",,terminal_output +4772,16517866,"TERMINAL",0,0,"505054141505",,terminal_output +4773,16518917,"TERMINAL",0,0,"2162216",,terminal_output +4774,16519974,"TERMINAL",0,0,"3273327",,terminal_output +4775,16521021,"TERMINAL",0,0,"4384438",,terminal_output +4776,16522077,"TERMINAL",0,0,"5495549",,terminal_output +4777,16523127,"TERMINAL",0,0,"654:0066510",,terminal_output +4778,16524183,"TERMINAL",0,0,"7617761",,terminal_output +4779,16525218,"TERMINAL",0,0,"8728872",,terminal_output +4780,16526278,"TERMINAL",0,0,"9839983",,terminal_output +4781,16527327,"TERMINAL",0,0,"3:0094505094",,terminal_output +4782,16528380,"TERMINAL",0,0,"13:005112:005",,terminal_output +4783,16529437,"TERMINAL",0,0,"2162216",,terminal_output +4784,16530487,"TERMINAL",0,0,"3273327",,terminal_output +4785,16531532,"TERMINAL",0,0,"4384438",,terminal_output +4786,16532580,"TERMINAL",0,0,"5495549",,terminal_output +4787,16533630,"TERMINAL",0,0,"651066520",,terminal_output +4788,16534676,"TERMINAL",0,0,"7617761",,terminal_output +4789,16535767,"TERMINAL",0,0,"8728872",,terminal_output +4790,16536780,"TERMINAL",0,0,"9839983",,terminal_output +4791,16537830,"TERMINAL",0,0,"101053:00:012:01105",,terminal_output +4792,16538870,"TERMINAL",0,0,"2162216",,terminal_output +4793,16539920,"TERMINAL",0,0,"3273327",,terminal_output +4794,16540973,"TERMINAL",0,0,"4384438",,terminal_output +4795,16542039,"TERMINAL",0,0,"5495549",,terminal_output +4796,16543079,"TERMINAL",0,0,"652066530",,terminal_output +4797,16544142,"TERMINAL",0,0,"7617761",,terminal_output +4798,16545177,"TERMINAL",0,0,"8728872",,terminal_output +4799,16546228,"TERMINAL",0,0,"9839983",,terminal_output +4800,16547284,"TERMINAL",0,0,"2094101094",,terminal_output +4801,16548334,"TERMINAL",0,0,"120511205",,terminal_output +4802,16549398,"TERMINAL",0,0,"2162216",,terminal_output +4803,16550456,"TERMINAL",0,0,"3273327",,terminal_output +4804,16551510,"TERMINAL",0,0,"4384438",,terminal_output +4805,16552555,"TERMINAL",0,0,"5495549",,terminal_output +4806,16553602,"TERMINAL",0,0,"653066540",,terminal_output +4807,16554652,"TERMINAL",0,0,"7617761",,terminal_output +4808,16555774,"TERMINAL",0,0,"8728872",,terminal_output +4809,16556770,"TERMINAL",0,0,"9839983",,terminal_output +4810,16557833,"TERMINAL",0,0,"3094202094",,terminal_output +4811,16558846,"TERMINAL",0,0,"131622316",,terminal_output +4812,16559894,"TERMINAL",0,0,"3273327",,terminal_output +4813,16560957,"TERMINAL",0,0,"4384438",,terminal_output +4814,16562022,"TERMINAL",0,0,"5495549",,terminal_output +4815,16563057,"TERMINAL",0,0,"654066550",,terminal_output +4816,16564141,"TERMINAL",0,0,"7617761",,terminal_output +4817,16565166,"TERMINAL",0,0,"8728872",,terminal_output +4818,16566226,"TERMINAL",0,0,"9839983",,terminal_output +4819,16567315,"TERMINAL",0,0,"4094303094",,terminal_output +4820,16568375,"TERMINAL",0,0,"140511405",,terminal_output +4821,16569421,"TERMINAL",0,0,"2162216",,terminal_output +4822,16570468,"TERMINAL",0,0,"3273327",,terminal_output +4823,16571517,"TERMINAL",0,0,"4384438",,terminal_output +4824,16572743,"TERMINAL",0,0,"5495549",,terminal_output +4825,16573613,"TERMINAL",0,0,"65506653:00",,terminal_output +4826,16574669,"TERMINAL",0,0,"7617761",,terminal_output +4827,16575712,"TERMINAL",0,0,"8728872",,terminal_output +4828,16576807,"TERMINAL",0,0,"9839983",,terminal_output +4829,16577821,"TERMINAL",0,0,"505054141505",,terminal_output +4830,16578864,"TERMINAL",0,0,"2162216",,terminal_output +4831,16579916,"TERMINAL",0,0,"3273327",,terminal_output +4832,16580959,"TERMINAL",0,0,"4384438",,terminal_output +4833,16582023,"TERMINAL",0,0,"5495549",,terminal_output +4834,16583068,"TERMINAL",0,0,"655:0066510",,terminal_output +4835,16584143,"TERMINAL",0,0,"7617761",,terminal_output +4836,16585158,"TERMINAL",0,0,"8728872",,terminal_output +4837,16586208,"TERMINAL",0,0,"9839983",,terminal_output +4838,16587257,"TERMINAL",0,0,"4:0094505094",,terminal_output +4839,16588305,"TERMINAL",0,0,"14:005113:005",,terminal_output +4840,16589360,"TERMINAL",0,0,"2162216",,terminal_output +4841,16590420,"TERMINAL",0,0,"3273327",,terminal_output +4842,16591460,"TERMINAL",0,0,"4384438",,terminal_output +4843,16592510,"TERMINAL",0,0,"5495549",,terminal_output +4844,16593557,"TERMINAL",0,0,"651066520",,terminal_output +4845,16594614,"TERMINAL",0,0,"7617761",,terminal_output +4846,16595664,"TERMINAL",0,0,"8728872",,terminal_output +4847,16596713,"TERMINAL",0,0,"9839983",,terminal_output +4848,16597790,"TERMINAL",0,0,"10941:003:0094",,terminal_output +4849,16598852,"TERMINAL",0,0,"111622116",,terminal_output +4850,16599872,"TERMINAL",0,0,"3273327",,terminal_output +4851,16600916,"TERMINAL",0,0,"4384438",,terminal_output +4852,16601968,"TERMINAL",0,0,"5495549",,terminal_output +4853,16603026,"TERMINAL",0,0,"652066530",,terminal_output +4854,16604070,"TERMINAL",0,0,"7617761",,terminal_output +4855,16605114,"TERMINAL",0,0,"8728872",,terminal_output +4856,16606169,"TERMINAL",0,0,"9839983",,terminal_output +4857,16607225,"TERMINAL",0,0,"2094101094",,terminal_output +4858,16608276,"TERMINAL",0,0,"120511205",,terminal_output +4859,16609329,"TERMINAL",0,0,"2162216",,terminal_output +4860,16610371,"TERMINAL",0,0,"3273327",,terminal_output +4861,16611430,"TERMINAL",0,0,"4384438",,terminal_output +4862,16612492,"TERMINAL",0,0,"5495549",,terminal_output +4863,16613523,"TERMINAL",0,0,"653066540",,terminal_output +4864,16614582,"TERMINAL",0,0,"7617761",,terminal_output +4865,16615627,"TERMINAL",0,0,"8728872",,terminal_output +4866,16616716,"TERMINAL",0,0,"9839983",,terminal_output +4867,16617856,"TERMINAL",0,0,"303052121305",,terminal_output +4868,16618910,"TERMINAL",0,0,"2162216",,terminal_output +4869,16619984,"TERMINAL",0,0,"3273327",,terminal_output +4870,16621020,"TERMINAL",0,0,"4384438",,terminal_output +4871,16622082,"TERMINAL",0,0,"5495549",,terminal_output +4872,16623125,"TERMINAL",0,0,"654066550",,terminal_output +4873,16624172,"TERMINAL",0,0,"7617761",,terminal_output +4874,16625220,"TERMINAL",0,0,"8728872",,terminal_output +4875,16626276,"TERMINAL",0,0,"9839983",,terminal_output +4876,16627321,"TERMINAL",0,0,"4094303094",,terminal_output +4877,16628375,"TERMINAL",0,0,"140511405",,terminal_output +4878,16629428,"TERMINAL",0,0,"2162216",,terminal_output +4879,16630482,"TERMINAL",0,0,"3273327",,terminal_output +4880,16631531,"TERMINAL",0,0,"4384438",,terminal_output +4881,16632577,"TERMINAL",0,0,"5495549",,terminal_output +4882,16633630,"TERMINAL",0,0,"65506654:00",,terminal_output +4883,16634688,"TERMINAL",0,0,"7617761",,terminal_output +4884,16635744,"TERMINAL",0,0,"8728872",,terminal_output +4885,16636825,"TERMINAL",0,0,"9839983",,terminal_output +4886,16637840,"TERMINAL",0,0,"505054141505",,terminal_output +4887,16638884,"TERMINAL",0,0,"2162216",,terminal_output +4888,16639937,"TERMINAL",0,0,"3273327",,terminal_output +4889,16640985,"TERMINAL",0,0,"4384438",,terminal_output +4890,16642038,"TERMINAL",0,0,"5495549",,terminal_output +4891,16643090,"TERMINAL",0,0,"656:0066510",,terminal_output +4892,16644152,"TERMINAL",0,0,"7617761",,terminal_output +4893,16645189,"TERMINAL",0,0,"8728872",,terminal_output +4894,16646235,"TERMINAL",0,0,"9839983",,terminal_output +4895,16647287,"TERMINAL",0,0,"5:0094505094",,terminal_output +4896,16648341,"TERMINAL",0,0,"15:005114:005",,terminal_output +4897,16649399,"TERMINAL",0,0,"2162216",,terminal_output +4898,16650449,"TERMINAL",0,0,"3273327",,terminal_output +4899,16651490,"TERMINAL",0,0,"4384438",,terminal_output +4900,16652542,"TERMINAL",0,0,"5495549",,terminal_output +4901,16653586,"TERMINAL",0,0,"651066520",,terminal_output +4902,16654641,"TERMINAL",0,0,"7617761",,terminal_output +4903,16655686,"TERMINAL",0,0,"8728872",,terminal_output +4904,16656743,"TERMINAL",0,0,"9839983",,terminal_output +4905,16657792,"TERMINAL",0,0,"10942:004:0094",,terminal_output +4906,16658838,"TERMINAL",0,0,"111622116",,terminal_output +4907,16659895,"TERMINAL",0,0,"3273327",,terminal_output +4908,16660940,"TERMINAL",0,0,"4384438",,terminal_output +4909,16661988,"TERMINAL",0,0,"5495549",,terminal_output +4910,16663055,"TERMINAL",0,0,"652066530",,terminal_output +4911,16664086,"TERMINAL",0,0,"7617761",,terminal_output +4912,16665199,"TERMINAL",0,0,"8728872",,terminal_output +4913,16666249,"TERMINAL",0,0,"9839983",,terminal_output +4914,16667300,"TERMINAL",0,0,"2094101094",,terminal_output +4915,16668427,"TERMINAL",0,0,"120511205",,terminal_output +4916,16669473,"TERMINAL",0,0,"2162216",,terminal_output +4917,16670520,"TERMINAL",0,0,"3273327",,terminal_output +4918,16671572,"TERMINAL",0,0,"4384438",,terminal_output +4919,16672610,"TERMINAL",0,0,"5495549",,terminal_output +4920,16673652,"TERMINAL",0,0,"653066540",,terminal_output +4921,16674705,"TERMINAL",0,0,"7617761",,terminal_output +4922,16675796,"TERMINAL",0,0,"8728872",,terminal_output +4923,16676803,"TERMINAL",0,0,"9839983",,terminal_output +4924,16677874,"TERMINAL",0,0,"303052121305",,terminal_output +4925,16678919,"TERMINAL",0,0,"2162216",,terminal_output +4926,16679961,"TERMINAL",0,0,"3273327",,terminal_output +4927,16681013,"TERMINAL",0,0,"4384438",,terminal_output +4928,16682064,"TERMINAL",0,0,"5495549",,terminal_output +4929,16683113,"TERMINAL",0,0,"654066550",,terminal_output +4930,16684168,"TERMINAL",0,0,"7617761",,terminal_output +4931,16685216,"TERMINAL",0,0,"8728872",,terminal_output +4932,16686272,"TERMINAL",0,0,"9839983",,terminal_output +4933,16687324,"TERMINAL",0,0,"4094303094",,terminal_output +4934,16688374,"TERMINAL",0,0,"140511405",,terminal_output +4935,16689428,"TERMINAL",0,0,"2162216",,terminal_output +4936,16690479,"TERMINAL",0,0,"3273327",,terminal_output +4937,16691528,"TERMINAL",0,0,"4384438",,terminal_output +4938,16692636,"TERMINAL",0,0,"5495549",,terminal_output +4939,16693680,"TERMINAL",0,0,"65506655:00",,terminal_output +4940,16694727,"TERMINAL",0,0,"7617761",,terminal_output +4941,16695796,"TERMINAL",0,0,"8728872",,terminal_output +4942,16696828,"TERMINAL",0,0,"994404094",,terminal_output +4943,16697863,"TERMINAL",0,0,"5150511505",,terminal_output +4944,16698908,"TERMINAL",0,0,"2162216",,terminal_output +4945,16700059,"TERMINAL",0,0,"3273327",,terminal_output +4946,16701005,"TERMINAL",0,0,"4384438",,terminal_output +4947,16702062,"TERMINAL",0,0,"5495549",,terminal_output +4948,16703103,"TERMINAL",0,0,"657:0066510",,terminal_output +4949,16704152,"TERMINAL",0,0,"7617761",,terminal_output +4950,16705189,"TERMINAL",0,0,"8728872",,terminal_output +4951,16706241,"TERMINAL",0,0,"9839983",,terminal_output +4952,16707274,"TERMINAL",0,0,"6:0094505094",,terminal_output +4953,16708328,"TERMINAL",0,0,"16:005115:005",,terminal_output +4954,16709370,"TERMINAL",0,0,"2162216",,terminal_output +4955,16710426,"TERMINAL",0,0,"3273327",,terminal_output +4956,16711472,"TERMINAL",0,0,"4384438",,terminal_output +4957,16712514,"TERMINAL",0,0,"5495549",,terminal_output +4958,16713558,"TERMINAL",0,0,"651066520",,terminal_output +4959,16714608,"TERMINAL",0,0,"7617761",,terminal_output +4960,16715656,"TERMINAL",0,0,"8728872",,terminal_output +4961,16716706,"TERMINAL",0,0,"9839983",,terminal_output +4962,16717752,"TERMINAL",0,0,"10943:005:0094",,terminal_output +4963,16718837,"TERMINAL",0,0,"110511105",,terminal_output +4964,16719843,"TERMINAL",0,0,"2273327",,terminal_output +4965,16720898,"TERMINAL",0,0,"4384438",,terminal_output +4966,16721938,"TERMINAL",0,0,"5495549",,terminal_output +4967,16722983,"TERMINAL",0,0,"652066530",,terminal_output +4968,16724032,"TERMINAL",0,0,"7617761",,terminal_output +4969,16725079,"TERMINAL",0,0,"8728872",,terminal_output +4970,16726144,"TERMINAL",0,0,"9839983",,terminal_output +4971,16727194,"TERMINAL",0,0,"2094101094",,terminal_output +4972,16728254,"TERMINAL",0,0,"120511205",,terminal_output +4973,16729307,"TERMINAL",0,0,"2162216",,terminal_output +4974,16730347,"TERMINAL",0,0,"3273327",,terminal_output +4975,16731405,"TERMINAL",0,0,"4384438",,terminal_output +4976,16732455,"TERMINAL",0,0,"5495549",,terminal_output +4977,16733499,"TERMINAL",0,0,"653066540",,terminal_output +4978,16734544,"TERMINAL",0,0,"7617761",,terminal_output +4979,16735595,"TERMINAL",0,0,"8728872",,terminal_output +4980,16736648,"TERMINAL",0,0,"9839983",,terminal_output +4981,16737698,"TERMINAL",0,0,"3094202094",,terminal_output +4982,16738751,"TERMINAL",0,0,"130511305",,terminal_output +4983,16739830,"TERMINAL",0,0,"2162216",,terminal_output +4984,16740853,"TERMINAL",0,0,"3384438",,terminal_output +4985,16741915,"TERMINAL",0,0,"5495549",,terminal_output +4986,16742962,"TERMINAL",0,0,"654066550",,terminal_output +4987,16744020,"TERMINAL",0,0,"7617761",,terminal_output +4988,16745066,"TERMINAL",0,0,"8728872",,terminal_output +4989,16746110,"TERMINAL",0,0,"9839983",,terminal_output +4990,16747158,"TERMINAL",0,0,"4094303094",,terminal_output +4991,16748201,"TERMINAL",0,0,"140511405",,terminal_output +4992,16749249,"TERMINAL",0,0,"2162216",,terminal_output +4993,16750300,"TERMINAL",0,0,"3273327",,terminal_output +4994,16751353,"TERMINAL",0,0,"4384438",,terminal_output +4995,16752397,"TERMINAL",0,0,"5495549",,terminal_output +4996,16753451,"TERMINAL",0,0,"65506656:00",,terminal_output +4997,16754510,"TERMINAL",0,0,"7617761",,terminal_output +4998,16755557,"TERMINAL",0,0,"8728872",,terminal_output +4999,16756615,"TERMINAL",0,0,"9839983",,terminal_output +5000,16757665,"TERMINAL",0,0,"5094404094",,terminal_output +5001,16758721,"TERMINAL",0,0,"150511505",,terminal_output +5002,16759767,"TERMINAL",0,0,"2162216",,terminal_output +5003,16760856,"TERMINAL",0,0,"3384438",,terminal_output +5004,16761869,"TERMINAL",0,0,"5495549",,terminal_output +5005,16762918,"TERMINAL",0,0,"658:0066510",,terminal_output +5006,16763966,"TERMINAL",0,0,"7617761",,terminal_output +5007,16765015,"TERMINAL",0,0,"8728872",,terminal_output +5008,16766060,"TERMINAL",0,0,"9839983",,terminal_output +5009,16767118,"TERMINAL",0,0,"7:0094505094",,terminal_output +5010,16768168,"TERMINAL",0,0,"17:005116:005",,terminal_output +5011,16769221,"TERMINAL",0,0,"2162216",,terminal_output +5012,16770263,"TERMINAL",0,0,"3273327",,terminal_output +5013,16771309,"TERMINAL",0,0,"4384438",,terminal_output +5014,16772360,"TERMINAL",0,0,"5495549",,terminal_output +5015,16773406,"TERMINAL",0,0,"651066520",,terminal_output +5016,16774461,"TERMINAL",0,0,"7617761",,terminal_output +5017,16775500,"TERMINAL",0,0,"8728872",,terminal_output +5018,16776553,"TERMINAL",0,0,"9839983",,terminal_output +5019,16777606,"TERMINAL",0,0,"10944:006:0094",,terminal_output +5020,16778647,"TERMINAL",0,0,"110511105",,terminal_output +5021,16779695,"TERMINAL",0,0,"2162216",,terminal_output +5022,16780753,"TERMINAL",0,0,"3273327",,terminal_output +5023,16781826,"TERMINAL",0,0,"4384438",,terminal_output +5024,16782844,"TERMINAL",0,0,"552066530",,terminal_output +5025,16783890,"TERMINAL",0,0,"7617761",,terminal_output +5026,16784940,"TERMINAL",0,0,"8728872",,terminal_output +5027,16785983,"TERMINAL",0,0,"9839983",,terminal_output +5028,16787052,"TERMINAL",0,0,"2094101094",,terminal_output +5029,16788129,"TERMINAL",0,0,"120511205",,terminal_output +5030,16789160,"TERMINAL",0,0,"2162216",,terminal_output +5031,16790190,"TERMINAL",0,0,"3273327",,terminal_output +5032,16791242,"TERMINAL",0,0,"4384438",,terminal_output +5033,16792282,"TERMINAL",0,0,"5495549",,terminal_output +5034,16793321,"TERMINAL",0,0,"653066540",,terminal_output +5035,16794361,"TERMINAL",0,0,"7617761",,terminal_output +5036,16795404,"TERMINAL",0,0,"8728872",,terminal_output +5037,16796456,"TERMINAL",0,0,"9839983",,terminal_output +5038,16797501,"TERMINAL",0,0,"3094202094",,terminal_output +5039,16798541,"TERMINAL",0,0,"130511305",,terminal_output +5040,16799584,"TERMINAL",0,0,"2162216",,terminal_output +5041,16800634,"TERMINAL",0,0,"3273327",,terminal_output +5042,16801681,"TERMINAL",0,0,"4384438",,terminal_output +5043,16802732,"TERMINAL",0,0,"5495549",,terminal_output +5044,16803778,"TERMINAL",0,0,"654066550",,terminal_output +5045,16804867,"TERMINAL",0,0,"7728872",,terminal_output +5046,16805880,"TERMINAL",0,0,"9839983",,terminal_output +5047,16806936,"TERMINAL",0,0,"4094303094",,terminal_output +5048,16807984,"TERMINAL",0,0,"140511405",,terminal_output +5049,16809032,"TERMINAL",0,0,"2162216",,terminal_output +5050,16810082,"TERMINAL",0,0,"3273327",,terminal_output +5051,16811117,"TERMINAL",0,0,"4384438",,terminal_output +5052,16812165,"TERMINAL",0,0,"5495549",,terminal_output +5053,16813213,"TERMINAL",0,0,"65506657:00",,terminal_output +5054,16814261,"TERMINAL",0,0,"7617761",,terminal_output +5055,16815315,"TERMINAL",0,0,"8728872",,terminal_output +5056,16816364,"TERMINAL",0,0,"9839983",,terminal_output +5057,16817414,"TERMINAL",0,0,"5094404094",,terminal_output +5058,16818465,"TERMINAL",0,0,"150511505",,terminal_output +5059,16819569,"TERMINAL",0,0,"2162216",,terminal_output +5060,16820564,"TERMINAL",0,0,"3273327",,terminal_output +5061,16821614,"TERMINAL",0,0,"4384438",,terminal_output +5062,16822685,"TERMINAL",0,0,"5495549",,terminal_output +5063,16823725,"TERMINAL",0,0,"659:0066510",,terminal_output +5064,16824767,"TERMINAL",0,0,"7617761",,terminal_output +5065,16825873,"TERMINAL",0,0,"8839983",,terminal_output +5066,16826874,"TERMINAL",0,0,"8:0094505094",,terminal_output +5067,16827918,"TERMINAL",0,0,"18:005117:005",,terminal_output +5068,16828968,"TERMINAL",0,0,"2162216",,terminal_output +5069,16830017,"TERMINAL",0,0,"3273327",,terminal_output +5070,16831067,"TERMINAL",0,0,"4384438",,terminal_output +5071,16832118,"TERMINAL",0,0,"5495549",,terminal_output +5072,16833158,"TERMINAL",0,0,"651066520",,terminal_output +5073,16834209,"TERMINAL",0,0,"7617761",,terminal_output +5074,16835261,"TERMINAL",0,0,"8728872",,terminal_output +5075,16836309,"TERMINAL",0,0,"9839983",,terminal_output +5076,16837347,"TERMINAL",0,0,"10945:007:0094",,terminal_output +5077,16838399,"TERMINAL",0,0,"110511105",,terminal_output +5078,16839444,"TERMINAL",0,0,"2162216",,terminal_output +5079,16840487,"TERMINAL",0,0,"3273327",,terminal_output +5080,16841533,"TERMINAL",0,0,"4384438",,terminal_output +5081,16842583,"TERMINAL",0,0,"5495549",,terminal_output +5082,16843623,"TERMINAL",0,0,"652066530",,terminal_output +5083,16844661,"TERMINAL",0,0,"7617761",,terminal_output +5084,16845706,"TERMINAL",0,0,"8728872",,terminal_output +5085,16846752,"TERMINAL",0,0,"9839983",,terminal_output +5086,16847804,"TERMINAL",0,0,"2094101094",,terminal_output +5087,16848886,"TERMINAL",0,0,"121622216",,terminal_output +5088,16849902,"TERMINAL",0,0,"3273327",,terminal_output +5089,16850954,"TERMINAL",0,0,"4384438",,terminal_output +5090,16852039,"TERMINAL",0,0,"5495549",,terminal_output +5091,16853051,"TERMINAL",0,0,"653066540",,terminal_output +5092,16854101,"TERMINAL",0,0,"7617761",,terminal_output +5093,16855143,"TERMINAL",0,0,"8728872",,terminal_output +5094,16856180,"TERMINAL",0,0,"9839983",,terminal_output +5095,16857231,"TERMINAL",0,0,"3094202094",,terminal_output +5096,16858273,"TERMINAL",0,0,"130511305",,terminal_output +5097,16859312,"TERMINAL",0,0,"2162216",,terminal_output +5098,16860364,"TERMINAL",0,0,"3273327",,terminal_output +5099,16861414,"TERMINAL",0,0,"4384438",,terminal_output +5100,16862462,"TERMINAL",0,0,"5495549",,terminal_output +5101,16863503,"TERMINAL",0,0,"654066550",,terminal_output +5102,16864550,"TERMINAL",0,0,"7617761",,terminal_output +5103,16865599,"TERMINAL",0,0,"8728872",,terminal_output +5104,16866651,"TERMINAL",0,0,"9839983",,terminal_output +5105,16867700,"TERMINAL",0,0,"4094303094",,terminal_output +5106,16868750,"TERMINAL",0,0,"140511405",,terminal_output +5107,16869797,"TERMINAL",0,0,"2162216",,terminal_output +5108,16870883,"TERMINAL",0,0,"3384438",,terminal_output +5109,16871895,"TERMINAL",0,0,"5495549",,terminal_output +5110,16872980,"TERMINAL",0,0,"65506658:00",,terminal_output +5111,16873986,"TERMINAL",0,0,"7617761",,terminal_output +5112,16875048,"TERMINAL",0,0,"8728872",,terminal_output +5113,16876085,"TERMINAL",0,0,"9839983",,terminal_output +5114,16877129,"TERMINAL",0,0,"5094404094",,terminal_output +5115,16878173,"TERMINAL",0,0,"150511505",,terminal_output +5116,16879227,"TERMINAL",0,0,"2162216",,terminal_output +5117,16880275,"TERMINAL",0,0,"3273327",,terminal_output +5118,16881315,"TERMINAL",0,0,"4384438",,terminal_output +5119,16882360,"TERMINAL",0,0,"5495549",,terminal_output +5120,16883408,"TERMINAL",0,0,"6530:0066510",,terminal_output +5121,16884460,"TERMINAL",0,0,"7617761",,terminal_output +5122,16885501,"TERMINAL",0,0,"8728872",,terminal_output +5123,16886542,"TERMINAL",0,0,"9839983",,terminal_output +5124,16887601,"TERMINAL",0,0,"9:0094505094",,terminal_output +5125,16888653,"TERMINAL",0,0,"19:005118:005",,terminal_output +5126,16889710,"TERMINAL",0,0,"2162216",,terminal_output +5127,16890747,"TERMINAL",0,0,"3273327",,terminal_output +5128,16891794,"TERMINAL",0,0,"4384438",,terminal_output +5129,16892866,"TERMINAL",0,0,"551066520",,terminal_output +5130,16893905,"TERMINAL",0,0,"7617761",,terminal_output +5131,16894953,"TERMINAL",0,0,"8728872",,terminal_output +5132,16895991,"TERMINAL",0,0,"9839983",,terminal_output +5133,16897047,"TERMINAL",0,0,"10946:008:0094",,terminal_output +5134,16898093,"TERMINAL",0,0,"110511105",,terminal_output +5135,16899165,"TERMINAL",0,0,"2162216",,terminal_output +5136,16900186,"TERMINAL",0,0,"3273327",,terminal_output +5137,16901235,"TERMINAL",0,0,"4384438",,terminal_output +5138,16902284,"TERMINAL",0,0,"5495549",,terminal_output +5139,16903331,"TERMINAL",0,0,"652066530",,terminal_output +5140,16904374,"TERMINAL",0,0,"7617761",,terminal_output +5141,16905419,"TERMINAL",0,0,"8728872",,terminal_output +5142,16906481,"TERMINAL",0,0,"9839983",,terminal_output +5143,16907534,"TERMINAL",0,0,"2094101094",,terminal_output +5144,16908585,"TERMINAL",0,0,"120511205",,terminal_output +5145,16909626,"TERMINAL",0,0,"2162216",,terminal_output +5146,16910682,"TERMINAL",0,0,"3273327",,terminal_output +5147,16911730,"TERMINAL",0,0,"4384438",,terminal_output +5148,16912776,"TERMINAL",0,0,"5495549",,terminal_output +5149,16913823,"TERMINAL",0,0,"663177641",,terminal_output +5150,16914871,"TERMINAL",0,0,"8728872",,terminal_output +5151,16915922,"TERMINAL",0,0,"9839983",,terminal_output +5152,16916964,"TERMINAL",0,0,"3094202094",,terminal_output +5153,16918012,"TERMINAL",0,0,"130511305",,terminal_output +5154,16919057,"TERMINAL",0,0,"2162216",,terminal_output +5155,16920108,"TERMINAL",0,0,"3273327",,terminal_output +5156,16921159,"TERMINAL",0,0,"4384438",,terminal_output +5157,16922214,"TERMINAL",0,0,"5495549",,terminal_output +5158,16923266,"TERMINAL",0,0,"654066550",,terminal_output +5159,16924313,"TERMINAL",0,0,"7617761",,terminal_output +5160,16925364,"TERMINAL",0,0,"8728872",,terminal_output +5161,16926408,"TERMINAL",0,0,"9839983",,terminal_output +5162,16927456,"TERMINAL",0,0,"4094303094",,terminal_output +5163,16928508,"TERMINAL",0,0,"140511405",,terminal_output +5164,16929557,"TERMINAL",0,0,"2162216",,terminal_output +5165,16930604,"TERMINAL",0,0,"3273327",,terminal_output +5166,16931647,"TERMINAL",0,0,"4384438",,terminal_output +5167,16932692,"TERMINAL",0,0,"5495549",,terminal_output +5168,16933735,"TERMINAL",0,0,"65506659:00",,terminal_output +5169,16934787,"TERMINAL",0,0,"7617761",,terminal_output +5170,16935855,"TERMINAL",0,0,"8839983",,terminal_output +5171,16936932,"TERMINAL",0,0,"5094404094",,terminal_output +5172,16937955,"TERMINAL",0,0,"150511505",,terminal_output +5173,16938999,"TERMINAL",0,0,"2162216",,terminal_output +5174,16940049,"TERMINAL",0,0,"3273327",,terminal_output +5175,16941129,"TERMINAL",0,0,"4384438",,terminal_output +5176,16942150,"TERMINAL",0,0,"5495549",,terminal_output +5177,16943203,"TERMINAL",0,0,"651:0066510",,terminal_output +5178,16944471,"TERMINAL",0,0,"7617761",,terminal_output +5179,16945315,"TERMINAL",0,0,"8728872",,terminal_output +5180,16946361,"TERMINAL",0,0,"9839983",,terminal_output +5181,16947408,"TERMINAL",0,0,"20:0094505094",,terminal_output +5182,16948462,"TERMINAL",0,0,"120:005119:005",,terminal_output +5183,16949505,"TERMINAL",0,0,"2162216",,terminal_output +5184,16950560,"TERMINAL",0,0,"3273327",,terminal_output +5185,16951607,"TERMINAL",0,0,"4384438",,terminal_output +5186,16952661,"TERMINAL",0,0,"5495549",,terminal_output +5187,16953708,"TERMINAL",0,0,"651066520",,terminal_output +5188,16954762,"TERMINAL",0,0,"7617761",,terminal_output +5189,16955827,"TERMINAL",0,0,"8728872",,terminal_output +5190,16956886,"TERMINAL",0,0,"9947:009:0094",,terminal_output +5191,16957918,"TERMINAL",0,0,"1110511105",,terminal_output +5192,16958962,"TERMINAL",0,0,"2162216",,terminal_output +5193,16960025,"TERMINAL",0,0,"3273327",,terminal_output +5194,16961071,"TERMINAL",0,0,"4384438",,terminal_output +5195,16962100,"TERMINAL",0,0,"5495549",,terminal_output +5196,16963140,"TERMINAL",0,0,"652066530",,terminal_output +5197,16964194,"TERMINAL",0,0,"7617761",,terminal_output +5198,16965242,"TERMINAL",0,0,"8728872",,terminal_output +5199,16966307,"TERMINAL",0,0,"9839983",,terminal_output +5200,16967337,"TERMINAL",0,0,"2094101094",,terminal_output +5201,16968392,"TERMINAL",0,0,"120511205",,terminal_output +5202,16969451,"TERMINAL",0,0,"2162216",,terminal_output +5203,16970505,"TERMINAL",0,0,"3273327",,terminal_output +5204,16971550,"TERMINAL",0,0,"4384438",,terminal_output +5205,16972609,"TERMINAL",0,0,"5495549",,terminal_output +5206,16973653,"TERMINAL",0,0,"653066540",,terminal_output +5207,16974706,"TERMINAL",0,0,"7617761",,terminal_output +5208,16975755,"TERMINAL",0,0,"8728872",,terminal_output +5209,16976804,"TERMINAL",0,0,"9839983",,terminal_output +5210,16977849,"TERMINAL",0,0,"303052121305",,terminal_output +5211,16978940,"TERMINAL",0,0,"2162216",,terminal_output +5212,16979948,"TERMINAL",0,0,"3273327",,terminal_output +5213,16981002,"TERMINAL",0,0,"4384438",,terminal_output +5214,16982054,"TERMINAL",0,0,"5495549",,terminal_output +5215,16983098,"TERMINAL",0,0,"654066550",,terminal_output +5216,16984169,"TERMINAL",0,0,"7617761",,terminal_output +5217,16985197,"TERMINAL",0,0,"8728872",,terminal_output +5218,16986249,"TERMINAL",0,0,"9839983",,terminal_output +5219,16987300,"TERMINAL",0,0,"4094303094",,terminal_output +5220,16988348,"TERMINAL",0,0,"140511405",,terminal_output +5221,16989387,"TERMINAL",0,0,"2162216",,terminal_output +5222,16990437,"TERMINAL",0,0,"3273327",,terminal_output +5223,16991486,"TERMINAL",0,0,"[1;133H4384438",,terminal_output +5224,16992549,"TERMINAL",0,0,"5495549",,terminal_output +5225,16993593,"TERMINAL",0,0,"655066520:00",,terminal_output +5226,16994643,"TERMINAL",0,0,"7617761",,terminal_output +5227,16995687,"TERMINAL",0,0,"8728872",,terminal_output +5228,16996731,"TERMINAL",0,0,"9839983",,terminal_output +5229,16997809,"TERMINAL",0,0,"5094404094",,terminal_output +5230,16998826,"TERMINAL",0,0,"151622516",,terminal_output +5231,16999889,"TERMINAL",0,0,"3273327",,terminal_output +5232,17000945,"TERMINAL",0,0,"4384438",,terminal_output +5233,17002033,"TERMINAL",0,0,"5495549",,terminal_output +5234,17003051,"TERMINAL",0,0,"652:0066510",,terminal_output +5235,17004104,"TERMINAL",0,0,"7617761",,terminal_output +5236,17005149,"TERMINAL",0,0,"8728872",,terminal_output +5237,17006208,"TERMINAL",0,0,"9839983",,terminal_output +5238,17007258,"TERMINAL",0,0,"1:0094505094",,terminal_output +5239,17008321,"TERMINAL",0,0,"11:0051120:005",,terminal_output +5240,17009366,"TERMINAL",0,0,"2162216",,terminal_output +5241,17010414,"TERMINAL",0,0,"3273327",,terminal_output +5242,17011465,"TERMINAL",0,0,"4384438",,terminal_output +5243,17012513,"TERMINAL",0,0,"5495549",,terminal_output +5244,17013562,"TERMINAL",0,0,"651066520",,terminal_output +5245,17014614,"TERMINAL",0,0,"7617761",,terminal_output +5246,17015659,"TERMINAL",0,0,"8728872",,terminal_output +5247,17016707,"TERMINAL",0,0,"9839983",,terminal_output +5248,17017807,"TERMINAL",0,0,"10948:0020:0094",,terminal_output +5249,17018807,"TERMINAL",0,0,"110511105",,terminal_output +5250,17019865,"TERMINAL",0,0,"2273327",,terminal_output +5251,17020944,"TERMINAL",0,0,"4384438",,terminal_output +5252,17021995,"TERMINAL",0,0,"5495549",,terminal_output +5253,17023004,"TERMINAL",0,0,"652066530",,terminal_output +5254,17024049,"TERMINAL",0,0,"7617761",,terminal_output +5255,17025100,"TERMINAL",0,0,"8728872",,terminal_output +5256,17026144,"TERMINAL",0,0,"9839983",,terminal_output +5257,17027195,"TERMINAL",0,0,"2094101094",,terminal_output +5258,17028250,"TERMINAL",0,0,"120511205",,terminal_output +5259,17029298,"TERMINAL",0,0,"2162216",,terminal_output +5260,17030346,"TERMINAL",0,0,"3273327",,terminal_output +5261,17031388,"TERMINAL",0,0,"4384438",,terminal_output +5262,17032427,"TERMINAL",0,0,"5495549",,terminal_output +5263,17033480,"TERMINAL",0,0,"653066540",,terminal_output +5264,17034524,"TERMINAL",0,0,"7617761",,terminal_output +5265,17035569,"TERMINAL",0,0,"8728872",,terminal_output +5266,17036628,"TERMINAL",0,0,"9839983",,terminal_output +5267,17037668,"TERMINAL",0,0,"3094202094",,terminal_output +5268,17038709,"TERMINAL",0,0,"130511305",,terminal_output +5269,17039752,"TERMINAL",0,0,"2162216",,terminal_output +5270,17040800,"TERMINAL",0,0,"3273327",,terminal_output +5271,17041845,"TERMINAL",0,0,"4495549",,terminal_output +5272,17042942,"TERMINAL",0,0,"654066550",,terminal_output +5273,17043942,"TERMINAL",0,0,"7617761",,terminal_output +5274,17044998,"TERMINAL",0,0,"8728872",,terminal_output +5275,17046049,"TERMINAL",0,0,"9839983",,terminal_output +5276,17047096,"TERMINAL",0,0,"4094303094",,terminal_output +5277,17048150,"TERMINAL",0,0,"140511405",,terminal_output +5278,17049202,"TERMINAL",0,0,"2162216",,terminal_output +5279,17050253,"TERMINAL",0,0,"3273327",,terminal_output +5280,17051302,"TERMINAL",0,0,"4384438",,terminal_output +5281,17052349,"TERMINAL",0,0,"5495549",,terminal_output +5282,17053542,"TERMINAL",0,0,"65506651:00",,terminal_output +5283,17054584,"TERMINAL",0,0,"7617761",,terminal_output +5284,17055638,"TERMINAL",0,0,"8728872",,terminal_output +5285,17056674,"TERMINAL",0,0,"9839983",,terminal_output +5286,17057722,"TERMINAL",0,0,"5094404094",,terminal_output +5287,17058778,"TERMINAL",0,0,"150511505",,terminal_output +5288,17060035,"TERMINAL",0,0,"2273327",,terminal_output +5289,17061082,"TERMINAL",0,0,"4384438",,terminal_output +5290,17062128,"TERMINAL",0,0,"5495549",,terminal_output +5291,17063177,"TERMINAL",0,0,"653:0066510",,terminal_output +5292,17064240,"TERMINAL",0,0,"7617761",,terminal_output +5293,17065292,"TERMINAL",0,0,"8728872",,terminal_output +5294,17066330,"TERMINAL",0,0,"9839983",,terminal_output +5295,17067371,"TERMINAL",0,0,"2:0094505094",,terminal_output +5296,17068418,"TERMINAL",0,0,"12:005111:005",,terminal_output +5297,17069455,"TERMINAL",0,0,"2162216",,terminal_output +5298,17070508,"TERMINAL",0,0,"3273327",,terminal_output +5299,17071547,"TERMINAL",0,0,"4384438",,terminal_output +5300,17072593,"TERMINAL",0,0,"5495549",,terminal_output +5301,17073651,"TERMINAL",0,0,"651066520",,terminal_output +5302,17074697,"TERMINAL",0,0,"7617761",,terminal_output +5303,17075739,"TERMINAL",0,0,"8728872",,terminal_output +5304,17076789,"TERMINAL",0,0,"9839983",,terminal_output +5305,17077836,"TERMINAL",0,0,"101059:011:01105",,terminal_output +5306,17078890,"TERMINAL",0,0,"2162216",,terminal_output +5307,17079941,"TERMINAL",0,0,"3273327",,terminal_output +5308,17080985,"TERMINAL",0,0,"4384438",,terminal_output +5309,17082035,"TERMINAL",0,0,"5495549",,terminal_output +5310,17083226,"TERMINAL",0,0,"652066530",,terminal_output +5311,17084122,"TERMINAL",0,0,"7617761",,terminal_output +5312,17085173,"TERMINAL",0,0,"8728872",,terminal_output +5313,17086222,"TERMINAL",0,0,"9839983",,terminal_output +5314,17087276,"TERMINAL",0,0,"2094101094",,terminal_output +5315,17088333,"TERMINAL",0,0,"120511205",,terminal_output +5316,17089374,"TERMINAL",0,0,"2162216",,terminal_output +5317,17090428,"TERMINAL",0,0,"3273327",,terminal_output +5318,17091472,"TERMINAL",0,0,"4384438",,terminal_output +5319,17092526,"TERMINAL",0,0,"5495549",,terminal_output +5320,17093573,"TERMINAL",0,0,"653066540",,terminal_output +5321,17094614,"TERMINAL",0,0,"7617761",,terminal_output +5322,17095686,"TERMINAL",0,0,"8728872",,terminal_output +5323,17096712,"TERMINAL",0,0,"9839983",,terminal_output +5324,17097801,"TERMINAL",0,0,"3094202094",,terminal_output +5325,17098801,"TERMINAL",0,0,"130511305",,terminal_output +5326,17099850,"TERMINAL",0,0,"2273327",,terminal_output +5327,17100898,"TERMINAL",0,0,"4384438",,terminal_output +5328,17101973,"TERMINAL",0,0,"5495549",,terminal_output +5329,17103024,"TERMINAL",0,0,"654066550",,terminal_output +5330,17104057,"TERMINAL",0,0,"7617761",,terminal_output +5331,17105105,"TERMINAL",0,0,"8728872",,terminal_output +5332,17106150,"TERMINAL",0,0,"9839983",,terminal_output +5333,17107194,"TERMINAL",0,0,"4094303094",,terminal_output +5334,17108236,"TERMINAL",0,0,"140511405",,terminal_output +5335,17109277,"TERMINAL",0,0,"2162216",,terminal_output +5336,17110322,"TERMINAL",0,0,"3273327",,terminal_output +5337,17111369,"TERMINAL",0,0,"4384438",,terminal_output +5338,17112410,"TERMINAL",0,0,"5495549",,terminal_output +5339,17113455,"TERMINAL",0,0,"65506652:00",,terminal_output +5340,17114502,"TERMINAL",0,0,"7617761",,terminal_output +5341,17115553,"TERMINAL",0,0,"8728872",,terminal_output +5342,17116596,"TERMINAL",0,0,"9839983",,terminal_output +5343,17117640,"TERMINAL",0,0,"5094404094",,terminal_output +5344,17118698,"TERMINAL",0,0,"150511505",,terminal_output +5345,17119747,"TERMINAL",0,0,"2162216",,terminal_output +5346,17120842,"TERMINAL",0,0,"3273327",,terminal_output +5347,17121907,"TERMINAL",0,0,"4495549",,terminal_output +5348,17122883,"TERMINAL",0,0,"654:0066510",,terminal_output +5349,17123965,"TERMINAL",0,0,"7617761",,terminal_output +5350,17124975,"TERMINAL",0,0,"8728872",,terminal_output +5351,17126032,"TERMINAL",0,0,"9839983",,terminal_output +5352,17127077,"TERMINAL",0,0,"3:0094505094",,terminal_output +5353,17128128,"TERMINAL",0,0,"13:005112:005",,terminal_output +5354,17129189,"TERMINAL",0,0,"2162216",,terminal_output +5355,17130223,"TERMINAL",0,0,"3273327",,terminal_output +5356,17131273,"TERMINAL",0,0,"4384438",,terminal_output +5357,17132323,"TERMINAL",0,0,"5495549",,terminal_output +5358,17133368,"TERMINAL",0,0,"651066520",,terminal_output +5359,17134413,"TERMINAL",0,0,"7617761",,terminal_output +5360,17135463,"TERMINAL",0,0,"8728872",,terminal_output +5361,17136503,"TERMINAL",0,0,"9839983",,terminal_output +5362,17137556,"TERMINAL",0,0,"109410:002:0094",,terminal_output +5363,17138609,"TERMINAL",0,0,"110511105",,terminal_output +5364,17139655,"TERMINAL",0,0,"2162216",,terminal_output +5365,17140693,"TERMINAL",0,0,"3273327",,terminal_output +5366,17141741,"TERMINAL",0,0,"4384438",,terminal_output +5367,17142791,"TERMINAL",0,0,"5495549",,terminal_output +5368,17143849,"TERMINAL",0,0,"662177631",,terminal_output +5369,17144892,"TERMINAL",0,0,"8728872",,terminal_output +5370,17145980,"TERMINAL",0,0,"9839983",,terminal_output +5371,17146987,"TERMINAL",0,0,"2094101094",,terminal_output +5372,17148032,"TERMINAL",0,0,"120511205",,terminal_output +5373,17149084,"TERMINAL",0,0,"2162216",,terminal_output +5374,17150136,"TERMINAL",0,0,"3273327",,terminal_output +5375,17151183,"TERMINAL",0,0,"4384438",,terminal_output +5376,17152233,"TERMINAL",0,0,"5495549",,terminal_output +5377,17153282,"TERMINAL",0,0,"653066540",,terminal_output +5378,17154325,"TERMINAL",0,0,"7617761",,terminal_output +5379,17155381,"TERMINAL",0,0,"8728872",,terminal_output +5380,17156419,"TERMINAL",0,0,"9839983",,terminal_output +5381,17157471,"TERMINAL",0,0,"3094202094",,terminal_output +5382,17158498,"TERMINAL",0,0,"130511305",,terminal_output +5383,17159547,"TERMINAL",0,0,"2162216",,terminal_output +5384,17160594,"TERMINAL",0,0,"3273327",,terminal_output +5385,17161638,"TERMINAL",0,0,"4384438",,terminal_output +5386,17162689,"TERMINAL",0,0,"5495549",,terminal_output +5387,17163736,"TERMINAL",0,0,"654066550",,terminal_output +5388,17164781,"TERMINAL",0,0,"7617761",,terminal_output +5389,17165825,"TERMINAL",0,0,"8839983",,terminal_output +5390,17166872,"TERMINAL",0,0,"4094303094",,terminal_output +5391,17167920,"TERMINAL",0,0,"140511405",,terminal_output +5392,17169010,"TERMINAL",0,0,"2162216",,terminal_output +5393,17170021,"TERMINAL",0,0,"3273327",,terminal_output +5394,17171075,"TERMINAL",0,0,"4384438",,terminal_output +5395,17172122,"TERMINAL",0,0,"5495549",,terminal_output +5396,17173169,"TERMINAL",0,0,"65506653:00",,terminal_output +5397,17175253,"TERMINAL",0,0,"7728872",,terminal_output +5398,17176330,"TERMINAL",0,0,"9839983",,terminal_output +5399,17177373,"TERMINAL",0,0,"5094404094",,terminal_output +5400,17178426,"TERMINAL",0,0,"150511505",,terminal_output +5401,17179474,"TERMINAL",0,0,"2162216",,terminal_output +5402,17180527,"TERMINAL",0,0,"3273327",,terminal_output +5403,17181574,"TERMINAL",0,0,"4384438",,terminal_output +5404,17182628,"TERMINAL",0,0,"5495549",,terminal_output +5405,17183676,"TERMINAL",0,0,"655:0066510",,terminal_output +5406,17184727,"TERMINAL",0,0,"7617761",,terminal_output +5407,17185778,"TERMINAL",0,0,"8728872",,terminal_output +5408,17186826,"TERMINAL",0,0,"994505094",,terminal_output +5409,17187886,"TERMINAL",0,0,"4:014:005113:005",,terminal_output +5410,17188933,"TERMINAL",0,0,"2162216",,terminal_output +5411,17189967,"TERMINAL",0,0,"3273327",,terminal_output +5412,17191017,"TERMINAL",0,0,"4384438",,terminal_output +5413,17192060,"TERMINAL",0,0,"5495549",,terminal_output +5414,17193115,"TERMINAL",0,0,"651066520",,terminal_output +5415,17194195,"TERMINAL",0,0,"7617761",,terminal_output +5416,17195213,"TERMINAL",0,0,"8728872",,terminal_output +5417,17196264,"TERMINAL",0,0,"9839983",,terminal_output +5418,17197320,"TERMINAL",0,0,"10941:003:0094",,terminal_output +5419,17198360,"TERMINAL",0,0,"110511105",,terminal_output +5420,17199415,"TERMINAL",0,0,"2162216",,terminal_output +5421,17200468,"TERMINAL",0,0,"3273327",,terminal_output +5422,17201508,"TERMINAL",0,0,"4384438",,terminal_output +5423,17202555,"TERMINAL",0,0,"5495549",,terminal_output +5424,17203602,"TERMINAL",0,0,"652066530",,terminal_output +5425,17204650,"TERMINAL",0,0,"7617761",,terminal_output +5426,17205697,"TERMINAL",0,0,"8728872",,terminal_output +5427,17206752,"TERMINAL",0,0,"9839983",,terminal_output +5428,17207796,"TERMINAL",0,0,"2094101094",,terminal_output +5429,17208839,"TERMINAL",0,0,"121622216",,terminal_output +5430,17209894,"TERMINAL",0,0,"3273327",,terminal_output +5431,17210944,"TERMINAL",0,0,"4384438",,terminal_output +5432,17212025,"TERMINAL",0,0,"5495549",,terminal_output +5433,17213054,"TERMINAL",0,0,"653066540",,terminal_output +5434,17214092,"TERMINAL",0,0,"7617761",,terminal_output +5435,17215138,"TERMINAL",0,0,"8728872",,terminal_output +5436,17216183,"TERMINAL",0,0,"9839983",,terminal_output +5437,17217232,"TERMINAL",0,0,"3094202094",,terminal_output +5438,17218280,"TERMINAL",0,0,"130511305",,terminal_output +5439,17219334,"TERMINAL",0,0,"2162216",,terminal_output +5440,17220388,"TERMINAL",0,0,"3273327",,terminal_output +5441,17221431,"TERMINAL",0,0,"4384438",,terminal_output +5442,17222485,"TERMINAL",0,0,"5495549",,terminal_output +5443,17223532,"TERMINAL",0,0,"654066550",,terminal_output +5444,17224582,"TERMINAL",0,0,"7617761",,terminal_output +5445,17225631,"TERMINAL",0,0,"8728872",,terminal_output +5446,17226670,"TERMINAL",0,0,"9839983",,terminal_output +5447,17227708,"TERMINAL",0,0,"4094303094",,terminal_output +5448,17228749,"TERMINAL",0,0,"140511405",,terminal_output +5449,17229794,"TERMINAL",0,0,"2162216",,terminal_output +5450,17230844,"TERMINAL",0,0,"3384438",,terminal_output +5451,17231891,"TERMINAL",0,0,"5495549",,terminal_output +5452,17232939,"TERMINAL",0,0,"65506654:00",,terminal_output +5453,17234013,"TERMINAL",0,0,"7617761",,terminal_output +5454,17235054,"TERMINAL",0,0,"8728872",,terminal_output +5455,17236076,"TERMINAL",0,0,"9839983",,terminal_output +5456,17237124,"TERMINAL",0,0,"5094404094",,terminal_output +5457,17238179,"TERMINAL",0,0,"150511505",,terminal_output +5458,17239207,"TERMINAL",0,0,"2162216",,terminal_output +5459,17240266,"TERMINAL",0,0,"3273327",,terminal_output +5460,17241308,"TERMINAL",0,0,"4384438",,terminal_output +5461,17242370,"TERMINAL",0,0,"5495549",,terminal_output +5462,17243384,"TERMINAL",0,0,"656:0066510",,terminal_output +5463,17244452,"TERMINAL",0,0,"7617761",,terminal_output +5464,17245483,"TERMINAL",0,0,"8728872",,terminal_output +5465,17246534,"TERMINAL",0,0,"9839983",,terminal_output +5466,17247580,"TERMINAL",0,0,"5:0094505094",,terminal_output +5467,17248626,"TERMINAL",0,0,"15:005114:005",,terminal_output +5468,17249671,"TERMINAL",0,0,"2162216",,terminal_output +5469,17250720,"TERMINAL",0,0,"3273327",,terminal_output +5470,17251769,"TERMINAL",0,0,"4384438",,terminal_output +5471,17252821,"TERMINAL",0,0,"551066520",,terminal_output +5472,17253871,"TERMINAL",0,0,"7617761",,terminal_output +5473,17254917,"TERMINAL",0,0,"8728872",,terminal_output +5474,17255969,"TERMINAL",0,0,"9839983",,terminal_output +5475,17257047,"TERMINAL",0,0,"10942:004:0094",,terminal_output +5476,17258074,"TERMINAL",0,0,"110511105",,terminal_output +5477,17259122,"TERMINAL",0,0,"2162216",,terminal_output +5478,17260176,"TERMINAL",0,0,"3273327",,terminal_output +5479,17261228,"TERMINAL",0,0,"4384438",,terminal_output +5480,17262277,"TERMINAL",0,0,"5495549",,terminal_output +5481,17263334,"TERMINAL",0,0,"652066530",,terminal_output +5482,17264380,"TERMINAL",0,0,"7617761",,terminal_output +5483,17265430,"TERMINAL",0,0,"8728872",,terminal_output +5484,17266473,"TERMINAL",0,0,"9839983",,terminal_output +5485,17267523,"TERMINAL",0,0,"2094101094",,terminal_output +5486,17268574,"TERMINAL",0,0,"120511205",,terminal_output +5487,17269629,"TERMINAL",0,0,"2162216",,terminal_output +5488,17270678,"TERMINAL",0,0,"3273327",,terminal_output +5489,17271727,"TERMINAL",0,0,"4384438",,terminal_output +5490,17272786,"TERMINAL",0,0,"5495549",,terminal_output +5491,17273830,"TERMINAL",0,0,"663177641",,terminal_output +5492,17274879,"TERMINAL",0,0,"8728872",,terminal_output +5493,17275927,"TERMINAL",0,0,"9839983",,terminal_output +5494,17276981,"TERMINAL",0,0,"3094202094",,terminal_output +5495,17278043,"TERMINAL",0,0,"130511305",,terminal_output +5496,17279060,"TERMINAL",0,0,"2162216",,terminal_output +5497,17280107,"TERMINAL",0,0,"3273327",,terminal_output +5498,17281155,"TERMINAL",0,0,"4384438",,terminal_output +5499,17282204,"TERMINAL",0,0,"5495549",,terminal_output +5500,17283258,"TERMINAL",0,0,"654066550",,terminal_output +5501,17284298,"TERMINAL",0,0,"7617761",,terminal_output +5502,17285341,"TERMINAL",0,0,"8728872",,terminal_output +5503,17286380,"TERMINAL",0,0,"9839983",,terminal_output +5504,17287428,"TERMINAL",0,0,"4094303094",,terminal_output +5505,17288485,"TERMINAL",0,0,"140511405",,terminal_output +5506,17289539,"TERMINAL",0,0,"2162216",,terminal_output +5507,17290592,"TERMINAL",0,0,"3273327",,terminal_output +5508,17291634,"TERMINAL",0,0,"4384438",,terminal_output +5509,17292686,"TERMINAL",0,0,"5495549",,terminal_output +5510,17293720,"TERMINAL",0,0,"65506655:00",,terminal_output +5511,17294769,"TERMINAL",0,0,"7617761",,terminal_output +5512,17295823,"TERMINAL",0,0,"8728872",,terminal_output +5513,17296855,"TERMINAL",0,0,"994404094",,terminal_output +5514,17297884,"TERMINAL",0,0,"5150511505",,terminal_output +5515,17298930,"TERMINAL",0,0,"2162216",,terminal_output +5516,17299969,"TERMINAL",0,0,"3273327",,terminal_output +5517,17301067,"TERMINAL",0,0,"4384438",,terminal_output +5518,17302080,"TERMINAL",0,0,"5495549",,terminal_output +5519,17303125,"TERMINAL",0,0,"657:0066510",,terminal_output +5520,17304205,"TERMINAL",0,0,"7617761",,terminal_output +5521,17305220,"TERMINAL",0,0,"8728872",,terminal_output +5522,17306256,"TERMINAL",0,0,"9839983",,terminal_output +5523,17307304,"TERMINAL",0,0,"6:0094505094",,terminal_output +5524,17308352,"TERMINAL",0,0,"16:005115:005",,terminal_output +5525,17309405,"TERMINAL",0,0,"2162216",,terminal_output +5526,17310446,"TERMINAL",0,0,"3273327",,terminal_output +5527,17311499,"TERMINAL",0,0,"4384438",,terminal_output +5528,17312547,"TERMINAL",0,0,"5495549",,terminal_output +5529,17313597,"TERMINAL",0,0,"651066520",,terminal_output +5530,17314649,"TERMINAL",0,0,"7617761",,terminal_output +5531,17315692,"TERMINAL",0,0,"8728872",,terminal_output +5532,17316727,"TERMINAL",0,0,"9839983",,terminal_output +5533,17317779,"TERMINAL",0,0,"10943:005:0094",,terminal_output +5534,17318820,"TERMINAL",0,0,"111622116",,terminal_output +5535,17319864,"TERMINAL",0,0,"3273327",,terminal_output +5536,17320918,"TERMINAL",0,0,"4384438",,terminal_output +5537,17321965,"TERMINAL",0,0,"5495549",,terminal_output +5538,17323038,"TERMINAL",0,0,"652066530",,terminal_output +5539,17324058,"TERMINAL",0,0,"7617761",,terminal_output +5540,17325108,"TERMINAL",0,0,"8728872",,terminal_output +5541,17326159,"TERMINAL",0,0,"9839983",,terminal_output +5542,17327199,"TERMINAL",0,0,"2094101094",,terminal_output +5543,17328245,"TERMINAL",0,0,"120511205",,terminal_output +5544,17329288,"TERMINAL",0,0,"2162216",,terminal_output +5545,17330346,"TERMINAL",0,0,"3273327",,terminal_output +5546,17331384,"TERMINAL",0,0,"4384438",,terminal_output +5547,17332436,"TERMINAL",0,0,"5495549",,terminal_output +5548,17333493,"TERMINAL",0,0,"653066540",,terminal_output +5549,17334542,"TERMINAL",0,0,"7617761",,terminal_output +5550,17335593,"TERMINAL",0,0,"8728872",,terminal_output +5551,17336647,"TERMINAL",0,0,"9839983",,terminal_output +5552,17337698,"TERMINAL",0,0,"3094202094",,terminal_output +5553,17338749,"TERMINAL",0,0,"130511305",,terminal_output +5554,17339806,"TERMINAL",0,0,"2162216",,terminal_output +5555,17340865,"TERMINAL",0,0,"3384438",,terminal_output +5556,17341913,"TERMINAL",0,0,"5495549",,terminal_output +5557,17342966,"TERMINAL",0,0,"654066550",,terminal_output +5558,17344003,"TERMINAL",0,0,"7617761",,terminal_output +5559,17345086,"TERMINAL",0,0,"8728872",,terminal_output +5560,17346092,"TERMINAL",0,0,"9839983",,terminal_output +5561,17347149,"TERMINAL",0,0,"4094303094",,terminal_output +5562,17348186,"TERMINAL",0,0,"140511405",,terminal_output +5563,17349220,"TERMINAL",0,0,"2162216",,terminal_output +5564,17350281,"TERMINAL",0,0,"3273327",,terminal_output +5565,17351316,"TERMINAL",0,0,"4384438",,terminal_output +5566,17352358,"TERMINAL",0,0,"5495549",,terminal_output +5567,17353411,"TERMINAL",0,0,"65506656:00",,terminal_output +5568,17354453,"TERMINAL",0,0,"7617761",,terminal_output +5569,17355505,"TERMINAL",0,0,"8728872",,terminal_output +5570,17356543,"TERMINAL",0,0,"9839983",,terminal_output +5571,17357591,"TERMINAL",0,0,"5094404094",,terminal_output +5572,17358646,"TERMINAL",0,0,"150511505",,terminal_output +5573,17359683,"TERMINAL",0,0,"2162216",,terminal_output +5574,17360734,"TERMINAL",0,0,"3273327",,terminal_output +5575,17361784,"TERMINAL",0,0,"4384438",,terminal_output +5576,17362830,"TERMINAL",0,0,"558:0066510",,terminal_output +5577,17363884,"TERMINAL",0,0,"7617761",,terminal_output +5578,17364934,"TERMINAL",0,0,"8728872",,terminal_output +5579,17365983,"TERMINAL",0,0,"9839983",,terminal_output +5580,17367029,"TERMINAL",0,0,"7:0094505094",,terminal_output +5581,17368072,"TERMINAL",0,0,"17:005116:005",,terminal_output +5582,17369118,"TERMINAL",0,0,"2162216",,terminal_output +5583,17370163,"TERMINAL",0,0,"3273327",,terminal_output +5584,17371219,"TERMINAL",0,0,"4384438",,terminal_output +5585,17372265,"TERMINAL",0,0,"5495549",,terminal_output +5586,17373315,"TERMINAL",0,0,"651066520",,terminal_output +5587,17374370,"TERMINAL",0,0,"7617761",,terminal_output +5588,17375421,"TERMINAL",0,0,"8728872",,terminal_output +5589,17376472,"TERMINAL",0,0,"9839983",,terminal_output +5590,17377517,"TERMINAL",0,0,"10944:006:0094",,terminal_output +5591,17378560,"TERMINAL",0,0,"110511105",,terminal_output +5592,17379612,"TERMINAL",0,0,"2162216",,terminal_output +5593,17380661,"TERMINAL",0,0,"3273327",,terminal_output +5594,17381713,"TERMINAL",0,0,"4384438",,terminal_output +5595,17382793,"TERMINAL",0,0,"5495549",,terminal_output +5596,17383800,"TERMINAL",0,0,"652066530",,terminal_output +5597,17384852,"TERMINAL",0,0,"7728872",,terminal_output +5598,17385894,"TERMINAL",0,0,"9839983",,terminal_output +5599,17386952,"TERMINAL",0,0,"2094101094",,terminal_output +5600,17388007,"TERMINAL",0,0,"120511205",,terminal_output +5601,17389073,"TERMINAL",0,0,"2162216",,terminal_output +5602,17390087,"TERMINAL",0,0,"3273327",,terminal_output +5603,17391134,"TERMINAL",0,0,"4384438",,terminal_output +5604,17392168,"TERMINAL",0,0,"5495549",,terminal_output +5605,17393227,"TERMINAL",0,0,"653066540",,terminal_output +5606,17394275,"TERMINAL",0,0,"7617761",,terminal_output +5607,17395325,"TERMINAL",0,0,"8728872",,terminal_output +5608,17396365,"TERMINAL",0,0,"9839983",,terminal_output +5609,17397397,"TERMINAL",0,0,"3094202094",,terminal_output +5610,17398439,"TERMINAL",0,0,"130511305",,terminal_output +5611,17399489,"TERMINAL",0,0,"2162216",,terminal_output +5612,17400543,"TERMINAL",0,0,"3273327",,terminal_output +5613,17401609,"TERMINAL",0,0,"4384438",,terminal_output +5614,17402657,"TERMINAL",0,0,"5495549",,terminal_output +5615,17403703,"TERMINAL",0,0,"654066550",,terminal_output +5616,17404758,"TERMINAL",0,0,"7617761",,terminal_output +5617,17405803,"TERMINAL",0,0,"8728872",,terminal_output +5618,17406974,"TERMINAL",0,0,"994303094",,terminal_output +5619,17407890,"TERMINAL",0,0,"4140511405",,terminal_output +5620,17408955,"TERMINAL",0,0,"2162216",,terminal_output +5621,17409986,"TERMINAL",0,0,"3273327",,terminal_output +5622,17411038,"TERMINAL",0,0,"4384438",,terminal_output +5623,17412083,"TERMINAL",0,0,"5495549",,terminal_output +5624,17413130,"TERMINAL",0,0,"65506657:00",,terminal_output +5625,17414176,"TERMINAL",0,0,"7617761",,terminal_output +5626,17415218,"TERMINAL",0,0,"8728872",,terminal_output +5627,17416262,"TERMINAL",0,0,"9839983",,terminal_output +5628,17417310,"TERMINAL",0,0,"5094404094",,terminal_output +5629,17418365,"TERMINAL",0,0,"150511505",,terminal_output +5630,17419422,"TERMINAL",0,0,"2162216",,terminal_output +5631,17420467,"TERMINAL",0,0,"3273327",,terminal_output +5632,17421509,"TERMINAL",0,0,"4384438",,terminal_output +5633,17422553,"TERMINAL",0,0,"5495549",,terminal_output +5634,17423614,"TERMINAL",0,0,"659:0066510",,terminal_output +5635,17424646,"TERMINAL",0,0,"7617761",,terminal_output +5636,17425696,"TERMINAL",0,0,"8728872",,terminal_output +5637,17426749,"TERMINAL",0,0,"9839983",,terminal_output +5638,17427796,"TERMINAL",0,0,"8:0094505094",,terminal_output +5639,17428837,"TERMINAL",0,0,"18:016227:016",,terminal_output +5640,17429884,"TERMINAL",0,0,"3273327",,terminal_output +5641,17430927,"TERMINAL",0,0,"4384438",,terminal_output +5642,17431983,"TERMINAL",0,0,"5495549",,terminal_output +5643,17433033,"TERMINAL",0,0,"651066520",,terminal_output +5644,17434114,"TERMINAL",0,0,"7617761",,terminal_output +5645,17435124,"TERMINAL",0,0,"8728872",,terminal_output +5646,17436174,"TERMINAL",0,0,"9839983",,terminal_output +5647,17437233,"TERMINAL",0,0,"10945:007:0094",,terminal_output +5648,17438284,"TERMINAL",0,0,"110511105",,terminal_output +5649,17439326,"TERMINAL",0,0,"2162216",,terminal_output +5650,17440368,"TERMINAL",0,0,"3273327",,terminal_output +5651,17441408,"TERMINAL",0,0,"4384438",,terminal_output +5652,17442454,"TERMINAL",0,0,"5495549",,terminal_output +5653,17443502,"TERMINAL",0,0,"652066530",,terminal_output +5654,17444541,"TERMINAL",0,0,"7617761",,terminal_output +5655,17445594,"TERMINAL",0,0,"8728872",,terminal_output +5656,17446644,"TERMINAL",0,0,"9839983",,terminal_output +5657,17447690,"TERMINAL",0,0,"2094101094",,terminal_output +5658,17448730,"TERMINAL",0,0,"120511205",,terminal_output +5659,17449800,"TERMINAL",0,0,"2162216",,terminal_output +5660,17450846,"TERMINAL",0,0,"3384438",,terminal_output +5661,17451885,"TERMINAL",0,0,"5495549",,terminal_output +5662,17452944,"TERMINAL",0,0,"653066540",,terminal_output +5663,17453974,"TERMINAL",0,0,"7617761",,terminal_output +5664,17455015,"TERMINAL",0,0,"8728872",,terminal_output +5665,17456099,"TERMINAL",0,0,"9839983",,terminal_output +5666,17457114,"TERMINAL",0,0,"3094202094",,terminal_output +5667,17458171,"TERMINAL",0,0,"130511305",,terminal_output +5668,17459220,"TERMINAL",0,0,"2162216",,terminal_output +5669,17460259,"TERMINAL",0,0,"3273327",,terminal_output +5670,17461306,"TERMINAL",0,0,"4384438",,terminal_output +5671,17462355,"TERMINAL",0,0,"5495549",,terminal_output +5672,17463396,"TERMINAL",0,0,"654066550",,terminal_output +5673,17464454,"TERMINAL",0,0,"7617761",,terminal_output +5674,17465499,"TERMINAL",0,0,"8728872",,terminal_output +5675,17466548,"TERMINAL",0,0,"9839983",,terminal_output +5676,17467590,"TERMINAL",0,0,"4094303094",,terminal_output +5677,17468638,"TERMINAL",0,0,"140511405",,terminal_output +5678,17469684,"TERMINAL",0,0,"2162216",,terminal_output +5679,17470725,"TERMINAL",0,0,"3273327",,terminal_output +5680,17471767,"TERMINAL",0,0,"4384438",,terminal_output +5681,17472812,"TERMINAL",0,0,"5495549",,terminal_output +5682,17473856,"TERMINAL",0,0,"66517768:01",,terminal_output +5683,17474903,"TERMINAL",0,0,"8728872",,terminal_output +5684,17475938,"TERMINAL",0,0,"9839983",,terminal_output +5685,17476999,"TERMINAL",0,0,"5094404094",,terminal_output +5686,17478034,"TERMINAL",0,0,"150511505",,terminal_output +5687,17479120,"TERMINAL",0,0,"2162216",,terminal_output +5688,17480130,"TERMINAL",0,0,"3273327",,terminal_output +5689,17481173,"TERMINAL",0,0,"4384438",,terminal_output +5690,17482215,"TERMINAL",0,0,"5495549",,terminal_output +5691,17483259,"TERMINAL",0,0,"6540:0066510",,terminal_output +5692,17484306,"TERMINAL",0,0,"7617761",,terminal_output +5693,17485349,"TERMINAL",0,0,"8728872",,terminal_output +5694,17486387,"TERMINAL",0,0,"9839983",,terminal_output +5695,17487436,"TERMINAL",0,0,"9:0094505094",,terminal_output +5696,17488490,"TERMINAL",0,0,"19:005118:005",,terminal_output +5697,17489535,"TERMINAL",0,0,"2162216",,terminal_output +5698,17490574,"TERMINAL",0,0,"3273327",,terminal_output +5699,17491616,"TERMINAL",0,0,"4384438",,terminal_output +5700,17492668,"TERMINAL",0,0,"5495549",,terminal_output +5701,17493711,"TERMINAL",0,0,"651066520",,terminal_output +5702,17494769,"TERMINAL",0,0,"7617761",,terminal_output +5703,17495808,"TERMINAL",0,0,"8728872",,terminal_output +5704,17496846,"TERMINAL",0,0,"9946:008:0094",,terminal_output +5705,17498311,"TERMINAL",0,0,"1110511105",,terminal_output +5706,17499372,"TERMINAL",0,0,"2162216",,terminal_output +5707,17500410,"TERMINAL",0,0,"3273327",,terminal_output +5708,17501460,"TERMINAL",0,0,"4384438",,terminal_output +5709,17502510,"TERMINAL",0,0,"5495549",,terminal_output +5710,17503550,"TERMINAL",0,0,"652066530",,terminal_output +5711,17504599,"TERMINAL",0,0,"7617761",,terminal_output +5712,17505646,"TERMINAL",0,0,"8728872",,terminal_output +5713,17506690,"TERMINAL",0,0,"9839983",,terminal_output +5714,17507735,"TERMINAL",0,0,"2094101094",,terminal_output +5715,17508779,"TERMINAL",0,0,"120511205",,terminal_output +5716,17509831,"TERMINAL",0,0,"2273327",,terminal_output +5717,17510894,"TERMINAL",0,0,"4384438",,terminal_output +5718,17511951,"TERMINAL",0,0,"5495549",,terminal_output +5719,17512990,"TERMINAL",0,0,"653066540",,terminal_output +5720,17514030,"TERMINAL",0,0,"7617761",,terminal_output +5721,17515122,"TERMINAL",0,0,"8728872",,terminal_output +5722,17516126,"TERMINAL",0,0,"9839983",,terminal_output +5723,17517175,"TERMINAL",0,0,"3094202094",,terminal_output +5724,17518211,"TERMINAL",0,0,"130511305",,terminal_output +5725,17519256,"TERMINAL",0,0,"2162216",,terminal_output +5726,17520310,"TERMINAL",0,0,"3273327",,terminal_output +5727,17521356,"TERMINAL",0,0,"4384438",,terminal_output +5728,17522395,"TERMINAL",0,0,"5495549",,terminal_output +5729,17523454,"TERMINAL",0,0,"654066550",,terminal_output +5730,17524504,"TERMINAL",0,0,"7617761",,terminal_output +5731,17525560,"TERMINAL",0,0,"8728872",,terminal_output +5732,17526615,"TERMINAL",0,0,"9839983",,terminal_output +5733,17527669,"TERMINAL",0,0,"4094303094",,terminal_output +5734,17528720,"TERMINAL",0,0,"140511405",,terminal_output +5735,17529773,"TERMINAL",0,0,"2162216",,terminal_output +5736,17530813,"TERMINAL",0,0,"3384438",,terminal_output +5737,17531861,"TERMINAL",0,0,"5495549",,terminal_output +5738,17532905,"TERMINAL",0,0,"65506659:00",,terminal_output +5739,17533958,"TERMINAL",0,0,"7617761",,terminal_output +5740,17535014,"TERMINAL",0,0,"8728872",,terminal_output +5741,17536063,"TERMINAL",0,0,"9839983",,terminal_output +5742,17537120,"TERMINAL",0,0,"5094404094",,terminal_output +5743,17538157,"TERMINAL",0,0,"150511505",,terminal_output +5744,17539218,"TERMINAL",0,0,"2162216",,terminal_output +5745,17540265,"TERMINAL",0,0,"3273327",,terminal_output +5746,17541312,"TERMINAL",0,0,"4384438",,terminal_output +5747,17542342,"TERMINAL",0,0,"5495549",,terminal_output +5748,17543379,"TERMINAL",0,0,"651:0066510",,terminal_output +5749,17544430,"TERMINAL",0,0,"7617761",,terminal_output +5750,17545490,"TERMINAL",0,0,"8728872",,terminal_output +5751,17546574,"TERMINAL",0,0,"9839983",,terminal_output +5752,17547584,"TERMINAL",0,0,"30:0094505094",,terminal_output +5753,17548633,"TERMINAL",0,0,"130:005119:005",,terminal_output +5754,17549688,"TERMINAL",0,0,"2162216",,terminal_output +5755,17550739,"TERMINAL",0,0,"3273327",,terminal_output +5756,17551781,"TERMINAL",0,0,"4384438",,terminal_output +5757,17552833,"TERMINAL",0,0,"551066520",,terminal_output +5758,17553881,"TERMINAL",0,0,"7617761",,terminal_output +5759,17554928,"TERMINAL",0,0,"8728872",,terminal_output +5760,17555979,"TERMINAL",0,0,"9839983",,terminal_output +5761,17557081,"TERMINAL",0,0,"10947:009:0094",,terminal_output +5762,17558071,"TERMINAL",0,0,"110511105",,terminal_output +5763,17559148,"TERMINAL",0,0,"2162216",,terminal_output +5764,17560171,"TERMINAL",0,0,"3273327",,terminal_output +5765,17561221,"TERMINAL",0,0,"4384438",,terminal_output +5766,17562262,"TERMINAL",0,0,"5495549",,terminal_output +5767,17563310,"TERMINAL",0,0,"652066530",,terminal_output +5768,17564361,"TERMINAL",0,0,"7617761",,terminal_output +5769,17565414,"TERMINAL",0,0,"8728872",,terminal_output +5770,17566458,"TERMINAL",0,0,"9839983",,terminal_output +5771,17567515,"TERMINAL",0,0,"2094101094",,terminal_output +5772,17568551,"TERMINAL",0,0,"120511205",,terminal_output +5773,17569608,"TERMINAL",0,0,"2162216",,terminal_output +5774,17570653,"TERMINAL",0,0,"3273327",,terminal_output +5775,17571709,"TERMINAL",0,0,"4384438",,terminal_output +5776,17572795,"TERMINAL",0,0,"5495549",,terminal_output +5777,17573806,"TERMINAL",0,0,"653066540",,terminal_output +5778,17574851,"TERMINAL",0,0,"7728872",,terminal_output +5779,17575892,"TERMINAL",0,0,"9839983",,terminal_output +5780,17576955,"TERMINAL",0,0,"3094202094",,terminal_output +5781,17577989,"TERMINAL",0,0,"130511305",,terminal_output +5782,17579038,"TERMINAL",0,0,"2162216",,terminal_output +5783,17580139,"TERMINAL",0,0,"3273327",,terminal_output +5784,17581145,"TERMINAL",0,0,"4384438",,terminal_output +5785,17582192,"TERMINAL",0,0,"5495549",,terminal_output +5786,17583248,"TERMINAL",0,0,"654066550",,terminal_output +5787,17584290,"TERMINAL",0,0,"7617761",,terminal_output +5788,17585332,"TERMINAL",0,0,"8728872",,terminal_output +5789,17586371,"TERMINAL",0,0,"9839983",,terminal_output +5790,17587417,"TERMINAL",0,0,"4094303094",,terminal_output +5791,17588460,"TERMINAL",0,0,"140511405",,terminal_output +5792,17589510,"TERMINAL",0,0,"2162216",,terminal_output +5793,17590549,"TERMINAL",0,0,"3273327",,terminal_output +5794,17591602,"TERMINAL",0,0,"4384438",,terminal_output +5795,17592649,"TERMINAL",0,0,"5495549",,terminal_output +5796,17593704,"TERMINAL",0,0,"655066530:00",,terminal_output +5797,17594751,"TERMINAL",0,0,"7617761",,terminal_output +5798,17595800,"TERMINAL",0,0,"8728872",,terminal_output +5799,17596845,"TERMINAL",0,0,"994404094",,terminal_output +5800,17597895,"TERMINAL",0,0,"5150511505",,terminal_output +5801,17598949,"TERMINAL",0,0,"2162216",,terminal_output +5802,17599985,"TERMINAL",0,0,"3273327",,terminal_output +5803,17601044,"TERMINAL",0,0,"4384438",,terminal_output +5804,17602130,"TERMINAL",0,0,"5495549",,terminal_output +5805,17603132,"TERMINAL",0,0,"652:0066510",,terminal_output +5806,17604224,"TERMINAL",0,0,"7617761",,terminal_output +5807,17605236,"TERMINAL",0,0,"8728872",,terminal_output +5808,17606296,"TERMINAL",0,0,"9839983",,terminal_output +5809,17607339,"TERMINAL",0,0,"1:0094505094",,terminal_output +5810,17608390,"TERMINAL",0,0,"11:0051130:005",,terminal_output +5811,17609448,"TERMINAL",0,0,"2162216",,terminal_output +5812,17610497,"TERMINAL",0,0,"3273327",,terminal_output +5813,17611550,"TERMINAL",0,0,"4384438",,terminal_output +5814,17612595,"TERMINAL",0,0,"5495549",,terminal_output +5815,17613652,"TERMINAL",0,0,"651066520",,terminal_output +5816,17614705,"TERMINAL",0,0,"7617761",,terminal_output +5817,17615759,"TERMINAL",0,0,"8728872",,terminal_output +5818,17616810,"TERMINAL",0,0,"9839983",,terminal_output +5819,17617861,"TERMINAL",0,0,"101058:0130:01105",,terminal_output +5820,17618904,"TERMINAL",0,0,"2162216",,terminal_output +5821,17619962,"TERMINAL",0,0,"3273327",,terminal_output +5822,17621010,"TERMINAL",0,0,"4384438",,terminal_output +5823,17622087,"TERMINAL",0,0,"5495549",,terminal_output +5824,17623147,"TERMINAL",0,0,"652066530",,terminal_output +5825,17624176,"TERMINAL",0,0,"7617761",,terminal_output +5826,17625206,"TERMINAL",0,0,"8728872",,terminal_output +5827,17626257,"TERMINAL",0,0,"9839983",,terminal_output +5828,17627300,"TERMINAL",0,0,"2094101094",,terminal_output +5829,17628347,"TERMINAL",0,0,"120511205",,terminal_output +5830,17629395,"TERMINAL",0,0,"2162216",,terminal_output +5831,17630440,"TERMINAL",0,0,"3273327",,terminal_output +5832,17631498,"TERMINAL",0,0,"4384438",,terminal_output +5833,17632546,"TERMINAL",0,0,"5495549",,terminal_output +5834,17633596,"TERMINAL",0,0,"653066540",,terminal_output +5835,17634644,"TERMINAL",0,0,"7617761",,terminal_output +5836,17635698,"TERMINAL",0,0,"8728872",,terminal_output +5837,17636742,"TERMINAL",0,0,"9839983",,terminal_output +5838,17637782,"TERMINAL",0,0,"3094202094",,terminal_output +5839,17638825,"TERMINAL",0,0,"131622316",,terminal_output +5840,17639874,"TERMINAL",0,0,"3273327",,terminal_output +5841,17640932,"TERMINAL",0,0,"4384438",,terminal_output +5842,17642041,"TERMINAL",0,0,"5495549",,terminal_output +5843,17643028,"TERMINAL",0,0,"654066550",,terminal_output +5844,17644076,"TERMINAL",0,0,"7617761",,terminal_output +5845,17645169,"TERMINAL",0,0,"8728872",,terminal_output +5846,17646166,"TERMINAL",0,0,"9839983",,terminal_output +5847,17647220,"TERMINAL",0,0,"4094303094",,terminal_output +5848,17648268,"TERMINAL",0,0,"140511405",,terminal_output +5849,17649323,"TERMINAL",0,0,"2162216",,terminal_output +5850,17650368,"TERMINAL",0,0,"3273327",,terminal_output +5851,17651419,"TERMINAL",0,0,"4384438",,terminal_output +5852,17652465,"TERMINAL",0,0,"5495549",,terminal_output +5853,17653513,"TERMINAL",0,0,"65506651:00",,terminal_output +5854,17654559,"TERMINAL",0,0,"7617761",,terminal_output +5855,17655605,"TERMINAL",0,0,"8728872",,terminal_output +5856,17656987,"TERMINAL",0,0,"994404094",,terminal_output +5857,17658026,"TERMINAL",0,0,"5150511505",,terminal_output +5858,17659106,"TERMINAL",0,0,"2162216",,terminal_output +5859,17660156,"TERMINAL",0,0,"3273327",,terminal_output +5860,17661169,"TERMINAL",0,0,"4384438",,terminal_output +5861,17662219,"TERMINAL",0,0,"5495549",,terminal_output +5862,17663258,"TERMINAL",0,0,"653:0066510",,terminal_output +5863,17664301,"TERMINAL",0,0,"7617761",,terminal_output +5864,17665353,"TERMINAL",0,0,"8728872",,terminal_output +5865,17666393,"TERMINAL",0,0,"9839983",,terminal_output +5866,17667456,"TERMINAL",0,0,"2:0094505094",,terminal_output +5867,17668502,"TERMINAL",0,0,"12:005111:005",,terminal_output +5868,17669548,"TERMINAL",0,0,"2162216",,terminal_output +5869,17670605,"TERMINAL",0,0,"3273327",,terminal_output +5870,17671658,"TERMINAL",0,0,"4384438",,terminal_output +5871,17672702,"TERMINAL",0,0,"5495549",,terminal_output +5872,17673755,"TERMINAL",0,0,"651066520",,terminal_output +5873,17674824,"TERMINAL",0,0,"7617761",,terminal_output +5874,17675847,"TERMINAL",0,0,"8839983",,terminal_output +5875,17676889,"TERMINAL",0,0,"10949:001:0094",,terminal_output +5876,17677945,"TERMINAL",0,0,"110511105",,terminal_output +5877,17678990,"TERMINAL",0,0,"2162216",,terminal_output +5878,17680037,"TERMINAL",0,0,"3273327",,terminal_output +5879,17681083,"TERMINAL",0,0,"4384438",,terminal_output +5880,17682164,"TERMINAL",0,0,"5495549",,terminal_output +5881,17683168,"TERMINAL",0,0,"652066530",,terminal_output +5882,17684232,"TERMINAL",0,0,"7617761",,terminal_output +5883,17685272,"TERMINAL",0,0,"8728872",,terminal_output +5884,17686323,"TERMINAL",0,0,"9839983",,terminal_output +5885,17687369,"TERMINAL",0,0,"2094101094",,terminal_output +5886,17688411,"TERMINAL",0,0,"120511205",,terminal_output +5887,17689455,"TERMINAL",0,0,"2162216",,terminal_output +5888,17690510,"TERMINAL",0,0,"3273327",,terminal_output +5889,17691563,"TERMINAL",0,0,"4384438",,terminal_output +5890,17692604,"TERMINAL",0,0,"5495549",,terminal_output +5891,17693652,"TERMINAL",0,0,"653066540",,terminal_output +5892,17694697,"TERMINAL",0,0,"7617761",,terminal_output +5893,17695741,"TERMINAL",0,0,"8728872",,terminal_output +5894,17696796,"TERMINAL",0,0,"9839983",,terminal_output +5895,17697846,"TERMINAL",0,0,"303052121305",,terminal_output +5896,17698894,"TERMINAL",0,0,"2162216",,terminal_output +5897,17699953,"TERMINAL",0,0,"3273327",,terminal_output +5898,17700997,"TERMINAL",0,0,"4384438",,terminal_output +5899,17702082,"TERMINAL",0,0,"5495549",,terminal_output +5900,17703104,"TERMINAL",0,0,"654066550",,terminal_output +5901,17704162,"TERMINAL",0,0,"7617761",,terminal_output +5902,17705250,"TERMINAL",0,0,"8728872",,terminal_output +5903,17706264,"TERMINAL",0,0,"9839983",,terminal_output +5904,17707313,"TERMINAL",0,0,"4094303094",,terminal_output +5905,17708365,"TERMINAL",0,0,"140511405",,terminal_output +5906,17709413,"TERMINAL",0,0,"2162216",,terminal_output +5907,17710460,"TERMINAL",0,0,"3273327",,terminal_output +5908,17711504,"TERMINAL",0,0,"4384438",,terminal_output +5909,17712559,"TERMINAL",0,0,"5495549",,terminal_output +5910,17713615,"TERMINAL",0,0,"65506652:00",,terminal_output +5911,17714659,"TERMINAL",0,0,"7617761",,terminal_output +5912,17715711,"TERMINAL",0,0,"8728872",,terminal_output +5913,17716757,"TERMINAL",0,0,"9839983",,terminal_output +5914,17717808,"TERMINAL",0,0,"5094404094",,terminal_output +5915,17718854,"TERMINAL",0,0,"151622516",,terminal_output +5916,17719901,"TERMINAL",0,0,"3273327",,terminal_output +5917,17720947,"TERMINAL",0,0,"4384438",,terminal_output +5918,17721983,"TERMINAL",0,0,"5495549",,terminal_output +5919,17723023,"TERMINAL",0,0,"654:0066510",,terminal_output +5920,17724068,"TERMINAL",0,0,"7617761",,terminal_output +5921,17725110,"TERMINAL",0,0,"8728872",,terminal_output +5922,17726152,"TERMINAL",0,0,"9839983",,terminal_output +5923,17727229,"TERMINAL",0,0,"3:0094505094",,terminal_output +5924,17728251,"TERMINAL",0,0,"13:005112:005",,terminal_output +5925,17729292,"TERMINAL",0,0,"2162216",,terminal_output +5926,17730339,"TERMINAL",0,0,"3273327",,terminal_output +5927,17731393,"TERMINAL",0,0,"4384438",,terminal_output +5928,17732443,"TERMINAL",0,0,"5495549",,terminal_output +5929,17733490,"TERMINAL",0,0,"651066520",,terminal_output +5930,17734534,"TERMINAL",0,0,"7617761",,terminal_output +5931,17735584,"TERMINAL",0,0,"8728872",,terminal_output +5932,17736645,"TERMINAL",0,0,"9839983",,terminal_output +5933,17737679,"TERMINAL",0,0,"109420:002:0094",,terminal_output +5934,17738725,"TERMINAL",0,0,"110511105",,terminal_output +5935,17739780,"TERMINAL",0,0,"2162216",,terminal_output +5936,17740825,"TERMINAL",0,0,"3384438",,terminal_output +5937,17741868,"TERMINAL",0,0,"5495549",,terminal_output +5938,17742912,"TERMINAL",0,0,"652066530",,terminal_output +5939,17743947,"TERMINAL",0,0,"7617761",,terminal_output +5940,17745000,"TERMINAL",0,0,"8728872",,terminal_output +5941,17746039,"TERMINAL",0,0,"9839983",,terminal_output +5942,17747096,"TERMINAL",0,0,"2094101094",,terminal_output +5943,17748128,"TERMINAL",0,0,"120511205",,terminal_output +5944,17749175,"TERMINAL",0,0,"2162216",,terminal_output +5945,17750251,"TERMINAL",0,0,"3273327",,terminal_output +5946,17751273,"TERMINAL",0,0,"4384438",,terminal_output +5947,17752321,"TERMINAL",0,0,"5495549",,terminal_output +5948,17753368,"TERMINAL",0,0,"653066540",,terminal_output +5949,17754412,"TERMINAL",0,0,"7617761",,terminal_output +5950,17755458,"TERMINAL",0,0,"8728872",,terminal_output +5951,17756509,"TERMINAL",0,0,"9839983",,terminal_output +5952,17757557,"TERMINAL",0,0,"3094202094",,terminal_output +5953,17758612,"TERMINAL",0,0,"130511305",,terminal_output +5954,17759643,"TERMINAL",0,0,"2162216",,terminal_output +5955,17760698,"TERMINAL",0,0,"3273327",,terminal_output +5956,17761750,"TERMINAL",0,0,"4384438",,terminal_output +5957,17762800,"TERMINAL",0,0,"5495549",,terminal_output +5958,17763854,"TERMINAL",0,0,"664177651",,terminal_output +5959,17764892,"TERMINAL",0,0,"8728872",,terminal_output +5960,17765937,"TERMINAL",0,0,"9839983",,terminal_output +5961,17766981,"TERMINAL",0,0,"4094303094",,terminal_output +5962,17768025,"TERMINAL",0,0,"140511405",,terminal_output +5963,17769072,"TERMINAL",0,0,"2162216",,terminal_output +5964,17770122,"TERMINAL",0,0,"3273327",,terminal_output +5965,17771166,"TERMINAL",0,0,"4384438",,terminal_output +5966,17772250,"TERMINAL",0,0,"5495549",,terminal_output +5967,17773270,"TERMINAL",0,0,"65506653:00",,terminal_output +5968,17774312,"TERMINAL",0,0,"7617761",,terminal_output +5969,17775360,"TERMINAL",0,0,"8728872",,terminal_output +5970,17776405,"TERMINAL",0,0,"9839983",,terminal_output +5971,17777460,"TERMINAL",0,0,"5094404094",,terminal_output +5972,17778517,"TERMINAL",0,0,"150511505",,terminal_output +5973,17779569,"TERMINAL",0,0,"2162216",,terminal_output +5974,17780602,"TERMINAL",0,0,"3273327",,terminal_output +5975,17781661,"TERMINAL",0,0,"4384438",,terminal_output +5976,17782701,"TERMINAL",0,0,"5495549",,terminal_output +5977,17783738,"TERMINAL",0,0,"655:0066510",,terminal_output +5978,17784801,"TERMINAL",0,0,"7617761",,terminal_output +5979,17785853,"TERMINAL",0,0,"8839983",,terminal_output +5980,17786897,"TERMINAL",0,0,"4:0094505094",,terminal_output +5981,17787940,"TERMINAL",0,0,"14:005113:005",,terminal_output +5982,17788984,"TERMINAL",0,0,"2162216",,terminal_output +5983,17790033,"TERMINAL",0,0,"3273327",,terminal_output +5984,17791082,"TERMINAL",0,0,"4384438",,terminal_output +5985,17792128,"TERMINAL",0,0,"5495549",,terminal_output +5986,17793184,"TERMINAL",0,0,"651066520",,terminal_output +5987,17794240,"TERMINAL",0,0,"7617761",,terminal_output +5988,17795283,"TERMINAL",0,0,"8728872",,terminal_output +5989,17796330,"TERMINAL",0,0,"9839983",,terminal_output +5990,17797382,"TERMINAL",0,0,"10941:003:0094",,terminal_output +5991,17798427,"TERMINAL",0,0,"110511105",,terminal_output +5992,17799474,"TERMINAL",0,0,"2162216",,terminal_output +5993,17800515,"TERMINAL",0,0,"3273327",,terminal_output +5994,17801566,"TERMINAL",0,0,"4384438",,terminal_output +5995,17802618,"TERMINAL",0,0,"5495549",,terminal_output +5996,17803662,"TERMINAL",0,0,"652066530",,terminal_output +5997,17804713,"TERMINAL",0,0,"7617761",,terminal_output +5998,17805760,"TERMINAL",0,0,"8728872",,terminal_output +5999,17806812,"TERMINAL",0,0,"9839983",,terminal_output +6000,17807857,"TERMINAL",0,0,"202051111205",,terminal_output +6001,17808913,"TERMINAL",0,0,"2162216",,terminal_output +6002,17809962,"TERMINAL",0,0,"3273327",,terminal_output +6003,17811014,"TERMINAL",0,0,"4384438",,terminal_output +6004,17812096,"TERMINAL",0,0,"5495549",,terminal_output +6005,17813104,"TERMINAL",0,0,"653066540",,terminal_output +6006,17814159,"TERMINAL",0,0,"7617761",,terminal_output +6007,17815210,"TERMINAL",0,0,"8728872",,terminal_output +6008,17816293,"TERMINAL",0,0,"9839983",,terminal_output +6009,17817304,"TERMINAL",0,0,"3094202094",,terminal_output +6010,17818361,"TERMINAL",0,0,"130511305",,terminal_output +6011,17819397,"TERMINAL",0,0,"2162216",,terminal_output +6012,17820445,"TERMINAL",0,0,"3273327",,terminal_output +6013,17821489,"TERMINAL",0,0,"4384438",,terminal_output +6014,17822539,"TERMINAL",0,0,"5495549",,terminal_output +6015,17823584,"TERMINAL",0,0,"654066550",,terminal_output +6016,17824645,"TERMINAL",0,0,"7617761",,terminal_output +6017,17825700,"TERMINAL",0,0,"8728872",,terminal_output +6018,17826727,"TERMINAL",0,0,"9839983",,terminal_output +6019,17827780,"TERMINAL",0,0,"4094303094",,terminal_output +6020,17828821,"TERMINAL",0,0,"141622416",,terminal_output +6021,17829856,"TERMINAL",0,0,"3273327",,terminal_output +6022,17830909,"TERMINAL",0,0,"4384438",,terminal_output +6023,17831952,"TERMINAL",0,0,"5495549",,terminal_output +6024,17833004,"TERMINAL",0,0,"65506654:00",,terminal_output +6025,17834073,"TERMINAL",0,0,"7617761",,terminal_output +6026,17835115,"TERMINAL",0,0,"8728872",,terminal_output +6027,17836160,"TERMINAL",0,0,"9839983",,terminal_output +6028,17837197,"TERMINAL",0,0,"5094404094",,terminal_output +6029,17838277,"TERMINAL",0,0,"150511505",,terminal_output +6030,17839288,"TERMINAL",0,0,"2162216",,terminal_output +6031,17840329,"TERMINAL",0,0,"3273327",,terminal_output +6032,17841370,"TERMINAL",0,0,"4384438",,terminal_output +6033,17842407,"TERMINAL",0,0,"5495549",,terminal_output +6034,17843456,"TERMINAL",0,0,"656:0066510",,terminal_output +6035,17844506,"TERMINAL",0,0,"7617761",,terminal_output +6036,17845556,"TERMINAL",0,0,"8728872",,terminal_output +6037,17846599,"TERMINAL",0,0,"9839983",,terminal_output +6038,17847805,"TERMINAL",0,0,"5:0094505094",,terminal_output +6039,17848685,"TERMINAL",0,0,"15:005114:005",,terminal_output +6040,17849738,"TERMINAL",0,0,"2162216",,terminal_output +6041,17850770,"TERMINAL",0,0,"3273327",,terminal_output +6042,17851815,"TERMINAL",0,0,"4384438",,terminal_output +6043,17852854,"TERMINAL",0,0,"551066520",,terminal_output +6044,17853892,"TERMINAL",0,0,"7617761",,terminal_output +6045,17854937,"TERMINAL",0,0,"8728872",,terminal_output +6046,17855978,"TERMINAL",0,0,"9839983",,terminal_output +6047,17857030,"TERMINAL",0,0,"10942:004:0094",,terminal_output +6048,17858070,"TERMINAL",0,0,"110511105",,terminal_output +6049,17859118,"TERMINAL",0,0,"2162216",,terminal_output +6050,17860154,"TERMINAL",0,0,"3273327",,terminal_output +6051,17861193,"TERMINAL",0,0,"4384438",,terminal_output +6052,17862232,"TERMINAL",0,0,"5495549",,terminal_output +6053,17863313,"TERMINAL",0,0,"652066530",,terminal_output +6054,17864331,"TERMINAL",0,0,"7617761",,terminal_output +6055,17865376,"TERMINAL",0,0,"8728872",,terminal_output +6056,17866420,"TERMINAL",0,0,"9839983",,terminal_output +6057,17867459,"TERMINAL",0,0,"2094101094",,terminal_output +6058,17868531,"TERMINAL",0,0,"120511205",,terminal_output +6059,17869547,"TERMINAL",0,0,"2162216",,terminal_output +6060,17870584,"TERMINAL",0,0,"3273327",,terminal_output +6061,17871639,"TERMINAL",0,0,"4384438",,terminal_output +6062,17872692,"TERMINAL",0,0,"5495549",,terminal_output +6063,17873742,"TERMINAL",0,0,"653066540",,terminal_output +6064,17874780,"TERMINAL",0,0,"7617761",,terminal_output +6065,17875819,"TERMINAL",0,0,"8839983",,terminal_output +6066,17876861,"TERMINAL",0,0,"3094202094",,terminal_output +6067,17877904,"TERMINAL",0,0,"130511305",,terminal_output +6068,17879043,"TERMINAL",0,0,"2162216",,terminal_output +6069,17879997,"TERMINAL",0,0,"3273327",,terminal_output +6070,17881036,"TERMINAL",0,0,"4384438",,terminal_output +6071,17882091,"TERMINAL",0,0,"5495549",,terminal_output +6072,17883127,"TERMINAL",0,0,"654066550",,terminal_output +6073,17884162,"TERMINAL",0,0,"7617761",,terminal_output +6074,17885205,"TERMINAL",0,0,"8728872",,terminal_output +6075,17886250,"TERMINAL",0,0,"9839983",,terminal_output +6076,17887326,"TERMINAL",0,0,"4094303094",,terminal_output +6077,17888342,"TERMINAL",0,0,"140511405",,terminal_output +6078,17889381,"TERMINAL",0,0,"2162216",,terminal_output +6079,17890422,"TERMINAL",0,0,"3273327",,terminal_output +6080,17891463,"TERMINAL",0,0,"4384438",,terminal_output +6081,17892504,"TERMINAL",0,0,"5495549",,terminal_output +6082,17893548,"TERMINAL",0,0,"65506655:00",,terminal_output +6083,17894587,"TERMINAL",0,0,"7617761",,terminal_output +6084,17895636,"TERMINAL",0,0,"8728872",,terminal_output +6085,17896677,"TERMINAL",0,0,"9839983",,terminal_output +6086,17897724,"TERMINAL",0,0,"5094404094",,terminal_output +6087,17898784,"TERMINAL",0,0,"150511505",,terminal_output +6088,17899816,"TERMINAL",0,0,"2162216",,terminal_output +6089,17900857,"TERMINAL",0,0,"3384438",,terminal_output +6090,17901900,"TERMINAL",0,0,"5495549",,terminal_output +6091,17902944,"TERMINAL",0,0,"657:0066510",,terminal_output +6092,17903989,"TERMINAL",0,0,"7617761",,terminal_output +6093,17905032,"TERMINAL",0,0,"8728872",,terminal_output +6094,17906076,"TERMINAL",0,0,"9839983",,terminal_output +6095,17907117,"TERMINAL",0,0,"6:0094505094",,terminal_output +6096,17908159,"TERMINAL",0,0,"16:005115:005",,terminal_output +6097,17909242,"TERMINAL",0,0,"2162216",,terminal_output +6098,17910294,"TERMINAL",0,0,"3273327",,terminal_output +6099,17911396,"TERMINAL",0,0,"4384438",,terminal_output +6100,17912337,"TERMINAL",0,0,"5495549",,terminal_output +6101,17913388,"TERMINAL",0,0,"651066520",,terminal_output +6102,17914425,"TERMINAL",0,0,"7617761",,terminal_output +6103,17915468,"TERMINAL",0,0,"8728872",,terminal_output +6104,17916513,"TERMINAL",0,0,"9839983",,terminal_output +6105,17917561,"TERMINAL",0,0,"10943:005:0094",,terminal_output +6106,17918608,"TERMINAL",0,0,"110511105",,terminal_output +6107,17919646,"TERMINAL",0,0,"2162216",,terminal_output +6108,17920696,"TERMINAL",0,0,"3273327",,terminal_output +6109,17921732,"TERMINAL",0,0,"4384438",,terminal_output +6110,17922776,"TERMINAL",0,0,"5495549",,terminal_output +6111,17923824,"TERMINAL",0,0,"662177631",,terminal_output +6112,17924880,"TERMINAL",0,0,"8728872",,terminal_output +6113,17925915,"TERMINAL",0,0,"9839983",,terminal_output +6114,17926958,"TERMINAL",0,0,"2094101094",,terminal_output +6115,17927996,"TERMINAL",0,0,"120511205",,terminal_output +6116,17929035,"TERMINAL",0,0,"2162216",,terminal_output +6117,17930079,"TERMINAL",0,0,"3273327",,terminal_output +6118,17931128,"TERMINAL",0,0,"4384438",,terminal_output +6119,17932172,"TERMINAL",0,0,"5495549",,terminal_output +6120,17933222,"TERMINAL",0,0,"653066540",,terminal_output +6121,17934267,"TERMINAL",0,0,"7617761",,terminal_output +6122,17935333,"TERMINAL",0,0,"8728872",,terminal_output +6123,17936343,"TERMINAL",0,0,"9839983",,terminal_output +6124,17937381,"TERMINAL",0,0,"3094202094",,terminal_output +6125,17938424,"TERMINAL",0,0,"130511305",,terminal_output +6126,17939465,"TERMINAL",0,0,"2162216",,terminal_output +6127,17940499,"TERMINAL",0,0,"3273327",,terminal_output +6128,17941540,"TERMINAL",0,0,"4384438",,terminal_output +6129,17942655,"TERMINAL",0,0,"5495549",,terminal_output +6130,17943623,"TERMINAL",0,0,"654066550",,terminal_output +6131,17944664,"TERMINAL",0,0,"7617761",,terminal_output +6132,17945706,"TERMINAL",0,0,"8728872",,terminal_output +6133,17946773,"TERMINAL",0,0,"9839983",,terminal_output +6134,17947810,"TERMINAL",0,0,"4094303094",,terminal_output +6135,17948861,"TERMINAL",0,0,"141622416",,terminal_output +6136,17949902,"TERMINAL",0,0,"3273327",,terminal_output +6137,17950938,"TERMINAL",0,0,"4384438",,terminal_output +6138,17951980,"TERMINAL",0,0,"5495549",,terminal_output +6139,17953031,"TERMINAL",0,0,"65506656:00",,terminal_output +6140,17954068,"TERMINAL",0,0,"7617761",,terminal_output +6141,17955107,"TERMINAL",0,0,"8728872",,terminal_output +6142,17956145,"TERMINAL",0,0,"9839983",,terminal_output +6143,17957190,"TERMINAL",0,0,"5094404094",,terminal_output +6144,17958227,"TERMINAL",0,0,"150511505",,terminal_output +6145,17959267,"TERMINAL",0,0,"2162216",,terminal_output +6146,17960337,"TERMINAL",0,0,"3273327",,terminal_output +6147,17961356,"TERMINAL",0,0,"4384438",,terminal_output +6148,17962408,"TERMINAL",0,0,"5495549",,terminal_output +6149,17963440,"TERMINAL",0,0,"658:0066510",,terminal_output +6150,17964481,"TERMINAL",0,0,"7617761",,terminal_output +6151,17965521,"TERMINAL",0,0,"8728872",,terminal_output +6152,17966564,"TERMINAL",0,0,"9839983",,terminal_output +6153,17967612,"TERMINAL",0,0,"7:0094505094",,terminal_output +6154,17968657,"TERMINAL",0,0,"17:005116:005",,terminal_output +6155,17969696,"TERMINAL",0,0,"2162216",,terminal_output +6156,17970745,"TERMINAL",0,0,"3273327",,terminal_output +6157,17971790,"TERMINAL",0,0,"4384438",,terminal_output +6158,17972836,"TERMINAL",0,0,"551066520",,terminal_output +6159,17973876,"TERMINAL",0,0,"7617761",,terminal_output +6160,17974927,"TERMINAL",0,0,"8728872",,terminal_output +6161,17975975,"TERMINAL",0,0,"9839983",,terminal_output +6162,17977019,"TERMINAL",0,0,"10944:006:0094",,terminal_output +6163,17978064,"TERMINAL",0,0,"110511105",,terminal_output +6164,17979096,"TERMINAL",0,0,"2162216",,terminal_output +6165,17980142,"TERMINAL",0,0,"3273327",,terminal_output +6166,17981188,"TERMINAL",0,0,"4384438",,terminal_output +6167,17982239,"TERMINAL",0,0,"5495549",,terminal_output +6168,17983282,"TERMINAL",0,0,"652066530",,terminal_output +6169,17984328,"TERMINAL",0,0,"7617761",,terminal_output +6170,17985370,"TERMINAL",0,0,"8728872",,terminal_output +6171,17986409,"TERMINAL",0,0,"9839983",,terminal_output +6172,17987451,"TERMINAL",0,0,"2094101094",,terminal_output +6173,17988487,"TERMINAL",0,0,"120511205",,terminal_output +6174,17989532,"TERMINAL",0,0,"2162216",,terminal_output +6175,17990582,"TERMINAL",0,0,"3273327",,terminal_output +6176,17991623,"TERMINAL",0,0,"4384438",,terminal_output +6177,17992662,"TERMINAL",0,0,"5495549",,terminal_output +6178,17993711,"TERMINAL",0,0,"653066540",,terminal_output +6179,17994753,"TERMINAL",0,0,"7617761",,terminal_output +6180,17995802,"TERMINAL",0,0,"8728872",,terminal_output +6181,17996840,"TERMINAL",0,0,"994202094",,terminal_output +6182,17997889,"TERMINAL",0,0,"3130511305",,terminal_output +6183,17998925,"TERMINAL",0,0,"2162216",,terminal_output +6184,17999972,"TERMINAL",0,0,"3273327",,terminal_output +6185,18001011,"TERMINAL",0,0,"4384438",,terminal_output +6186,18002053,"TERMINAL",0,0,"5495549",,terminal_output +6187,18003092,"TERMINAL",0,0,"654066550",,terminal_output +6188,18004132,"TERMINAL",0,0,"7617761",,terminal_output +6189,18005177,"TERMINAL",0,0,"8728872",,terminal_output +6190,18006225,"TERMINAL",0,0,"9839983",,terminal_output +6191,18007277,"TERMINAL",0,0,"4094303094",,terminal_output +6192,18008352,"TERMINAL",0,0,"140511405",,terminal_output +6193,18009357,"TERMINAL",0,0,"2162216",,terminal_output +6194,18010432,"TERMINAL",0,0,"3273327",,terminal_output +6195,18011472,"TERMINAL",0,0,"4384438",,terminal_output +6196,18012527,"TERMINAL",0,0,"5495549",,terminal_output +6197,18013571,"TERMINAL",0,0,"65506657:00",,terminal_output +6198,18014612,"TERMINAL",0,0,"7617761",,terminal_output +6199,18015666,"TERMINAL",0,0,"8728872",,terminal_output +6200,18016719,"TERMINAL",0,0,"9839983",,terminal_output +6201,18017966,"TERMINAL",0,0,"505054141505",,terminal_output +6202,18019053,"TERMINAL",0,0,"2162216",,terminal_output +6203,18020092,"TERMINAL",0,0,"3273327",,terminal_output +6204,18021129,"TERMINAL",0,0,"4384438",,terminal_output +6205,18022167,"TERMINAL",0,0,"5495549",,terminal_output +6206,18023207,"TERMINAL",0,0,"659:0066510",,terminal_output +6207,18024256,"TERMINAL",0,0,"7617761",,terminal_output +6208,18025288,"TERMINAL",0,0,"8728872",,terminal_output +6209,18026375,"TERMINAL",0,0,"9839983",,terminal_output +6210,18027377,"TERMINAL",0,0,"8:0094505094",,terminal_output +6211,18028425,"TERMINAL",0,0,"18:005117:005",,terminal_output +6212,18029475,"TERMINAL",0,0,"2162216",,terminal_output +6213,18030514,"TERMINAL",0,0,"3273327",,terminal_output +6214,18031562,"TERMINAL",0,0,"4384438",,terminal_output +6215,18032606,"TERMINAL",0,0,"5495549",,terminal_output +6216,18033646,"TERMINAL",0,0,"651066520",,terminal_output +6217,18034690,"TERMINAL",0,0,"7617761",,terminal_output +6218,18035741,"TERMINAL",0,0,"8728872",,terminal_output +6219,18036831,"TERMINAL",0,0,"9945:007:0094",,terminal_output +6220,18037880,"TERMINAL",0,0,"1110511105",,terminal_output +6221,18038928,"TERMINAL",0,0,"2162216",,terminal_output +6222,18039969,"TERMINAL",0,0,"3273327",,terminal_output +6223,18041019,"TERMINAL",0,0,"4384438",,terminal_output +6224,18042108,"TERMINAL",0,0,"5495549",,terminal_output +6225,18043111,"TERMINAL",0,0,"652066530",,terminal_output +6226,18044150,"TERMINAL",0,0,"7617761",,terminal_output +6227,18045200,"TERMINAL",0,0,"8728872",,terminal_output +6228,18046239,"TERMINAL",0,0,"9839983",,terminal_output +6229,18047280,"TERMINAL",0,0,"2094101094",,terminal_output +6230,18048366,"TERMINAL",0,0,"120511205",,terminal_output +6231,18049370,"TERMINAL",0,0,"2162216",,terminal_output +6232,18050426,"TERMINAL",0,0,"3273327",,terminal_output +6233,18051465,"TERMINAL",0,0,"4384438",,terminal_output +6234,18052514,"TERMINAL",0,0,"5495549",,terminal_output +6235,18053557,"TERMINAL",0,0,"653066540",,terminal_output +6236,18054598,"TERMINAL",0,0,"7617761",,terminal_output +6237,18055647,"TERMINAL",0,0,"8728872",,terminal_output +6238,18056716,"TERMINAL",0,0,"9839983",,terminal_output +6239,18057730,"TERMINAL",0,0,"3094202094",,terminal_output +6240,18058775,"TERMINAL",0,0,"130511305",,terminal_output +6241,18059810,"TERMINAL",0,0,"2162216",,terminal_output +6242,18060870,"TERMINAL",0,0,"4384438",,terminal_output +6243,18061907,"TERMINAL",0,0,"5495549",,terminal_output +6244,18062951,"TERMINAL",0,0,"654066550",,terminal_output +6245,18063988,"TERMINAL",0,0,"7617761",,terminal_output +6246,18065037,"TERMINAL",0,0,"8728872",,terminal_output +6247,18066089,"TERMINAL",0,0,"9839983",,terminal_output +6248,18067136,"TERMINAL",0,0,"4094303094",,terminal_output +6249,18068175,"TERMINAL",0,0,"140511405",,terminal_output +6250,18069213,"TERMINAL",0,0,"2162216",,terminal_output +6251,18070254,"TERMINAL",0,0,"3273327",,terminal_output +6252,18071296,"TERMINAL",0,0,"4384438",,terminal_output +6253,18072377,"TERMINAL",0,0,"5495549",,terminal_output +6254,18073398,"TERMINAL",0,0,"65506658:00",,terminal_output +6255,18074440,"TERMINAL",0,0,"7617761",,terminal_output +6256,18075488,"TERMINAL",0,0,"8728872",,terminal_output +6257,18076531,"TERMINAL",0,0,"9839983",,terminal_output +6258,18077575,"TERMINAL",0,0,"5094404094",,terminal_output +6259,18078609,"TERMINAL",0,0,"150511505",,terminal_output +6260,18079651,"TERMINAL",0,0,"2162216",,terminal_output +6261,18080690,"TERMINAL",0,0,"3273327",,terminal_output +6262,18081739,"TERMINAL",0,0,"4384438",,terminal_output +6263,18082772,"TERMINAL",0,0,"5495549",,terminal_output +6264,18083820,"TERMINAL",0,0,"6650:0177611",,terminal_output +6265,18084864,"TERMINAL",0,0,"8728872",,terminal_output +6266,18085905,"TERMINAL",0,0,"9839983",,terminal_output +6267,18086942,"TERMINAL",0,0,"9:0094505094",,terminal_output +6268,18087989,"TERMINAL",0,0,"19:005118:005",,terminal_output +6269,18089031,"TERMINAL",0,0,"2162216",,terminal_output +6270,18090073,"TERMINAL",0,0,"3273327",,terminal_output +6271,18091123,"TERMINAL",0,0,"4384438",,terminal_output +6272,18092159,"TERMINAL",0,0,"5495549",,terminal_output +6273,18093204,"TERMINAL",0,0,"651066520",,terminal_output +6274,18094265,"TERMINAL",0,0,"7617761",,terminal_output +6275,18095280,"TERMINAL",0,0,"8728872",,terminal_output +6276,18096333,"TERMINAL",0,0,"9839983",,terminal_output +6277,18097419,"TERMINAL",0,0,"10946:008:0094",,terminal_output +6278,18098415,"TERMINAL",0,0,"110511105",,terminal_output +6279,18099465,"TERMINAL",0,0,"2162216",,terminal_output +6280,18100502,"TERMINAL",0,0,"3273327",,terminal_output +6281,18101547,"TERMINAL",0,0,"4384438",,terminal_output +6282,18102594,"TERMINAL",0,0,"5495549",,terminal_output +6283,18103641,"TERMINAL",0,0,"652066530",,terminal_output +6284,18104676,"TERMINAL",0,0,"7617761",,terminal_output +6285,18105720,"TERMINAL",0,0,"8728872",,terminal_output +6286,18106770,"TERMINAL",0,0,"9839983",,terminal_output +6287,18107814,"TERMINAL",0,0,"2094101094",,terminal_output +6288,18108846,"TERMINAL",0,0,"121622216",,terminal_output +6289,18109889,"TERMINAL",0,0,"3273327",,terminal_output +6290,18110925,"TERMINAL",0,0,"4384438",,terminal_output +6291,18111970,"TERMINAL",0,0,"5495549",,terminal_output +6292,18113006,"TERMINAL",0,0,"653066540",,terminal_output +6293,18114049,"TERMINAL",0,0,"7617761",,terminal_output +6294,18115089,"TERMINAL",0,0,"8728872",,terminal_output +6295,18116129,"TERMINAL",0,0,"9839983",,terminal_output +6296,18117173,"TERMINAL",0,0,"3094202094",,terminal_output +6297,18118206,"TERMINAL",0,0,"130511305",,terminal_output +6298,18119274,"TERMINAL",0,0,"2162216",,terminal_output +6299,18120294,"TERMINAL",0,0,"3273327",,terminal_output +6300,18121351,"TERMINAL",0,0,"4384438",,terminal_output +6301,18122425,"TERMINAL",0,0,"5495549",,terminal_output +6302,18123461,"TERMINAL",0,0,"654066550",,terminal_output +6303,18124499,"TERMINAL",0,0,"7617761",,terminal_output +6304,18125540,"TERMINAL",0,0,"8728872",,terminal_output +6305,18126588,"TERMINAL",0,0,"9839983",,terminal_output +6306,18127631,"TERMINAL",0,0,"4094303094",,terminal_output +6307,18128674,"TERMINAL",0,0,"140511405",,terminal_output +6308,18129714,"TERMINAL",0,0,"2162216",,terminal_output +6309,18130751,"TERMINAL",0,0,"3273327",,terminal_output +6310,18131787,"TERMINAL",0,0,"4384438",,terminal_output +6311,18132840,"TERMINAL",0,0,"55506659:00",,terminal_output +6312,18133883,"TERMINAL",0,0,"7617761",,terminal_output +6313,18134927,"TERMINAL",0,0,"8728872",,terminal_output +6314,18135967,"TERMINAL",0,0,"9839983",,terminal_output +6315,18137007,"TERMINAL",0,0,"5094404094",,terminal_output +6316,18138351,"TERMINAL",0,0,"150511505",,terminal_output +6317,18139419,"TERMINAL",0,0,"2162216",,terminal_output +6318,18140474,"TERMINAL",0,0,"3273327",,terminal_output +6319,18141522,"TERMINAL",0,0,"4384438",,terminal_output +6320,18142580,"TERMINAL",0,0,"5495549",,terminal_output +6321,18143622,"TERMINAL",0,0,"651:0066510",,terminal_output +6322,18144664,"TERMINAL",0,0,"7617761",,terminal_output +6323,18145710,"TERMINAL",0,0,"8728872",,terminal_output +6324,18146761,"TERMINAL",0,0,"9839983",,terminal_output +6325,18147805,"TERMINAL",0,0,"40:0094505094",,terminal_output +6326,18148851,"TERMINAL",0,0,"140:016229:016",,terminal_output +6327,18149890,"TERMINAL",0,0,"3273327",,terminal_output +6328,18150946,"TERMINAL",0,0,"4384438",,terminal_output +6329,18151992,"TERMINAL",0,0,"5495549",,terminal_output +6330,18153035,"TERMINAL",0,0,"651066520",,terminal_output +6331,18154084,"TERMINAL",0,0,"7617761",,terminal_output +6332,18155128,"TERMINAL",0,0,"8728872",,terminal_output +6333,18156168,"TERMINAL",0,0,"9839983",,terminal_output +6334,18157212,"TERMINAL",0,0,"10947:009:0094",,terminal_output +6335,18158259,"TERMINAL",0,0,"110511105",,terminal_output +6336,18159308,"TERMINAL",0,0,"2162216",,terminal_output +6337,18160354,"TERMINAL",0,0,"3273327",,terminal_output +6338,18161440,"TERMINAL",0,0,"4384438",,terminal_output +6339,18162449,"TERMINAL",0,0,"5495549",,terminal_output +6340,18163500,"TERMINAL",0,0,"652066530",,terminal_output +6341,18164548,"TERMINAL",0,0,"7617761",,terminal_output +6342,18165594,"TERMINAL",0,0,"8728872",,terminal_output +6343,18166647,"TERMINAL",0,0,"9839983",,terminal_output +6344,18167693,"TERMINAL",0,0,"2094101094",,terminal_output +6345,18168745,"TERMINAL",0,0,"120511205",,terminal_output +6346,18169788,"TERMINAL",0,0,"2162216",,terminal_output +6347,18170839,"TERMINAL",0,0,"3384438",,terminal_output +6348,18171888,"TERMINAL",0,0,"5495549",,terminal_output +6349,18172934,"TERMINAL",0,0,"653066540",,terminal_output +6350,18173978,"TERMINAL",0,0,"7617761",,terminal_output +6351,18175023,"TERMINAL",0,0,"8728872",,terminal_output +6352,18176184,"TERMINAL",0,0,"9839983",,terminal_output +6353,18177111,"TERMINAL",0,0,"3094202094",,terminal_output +6354,18178334,"TERMINAL",0,0,"130511305",,terminal_output +6355,18179380,"TERMINAL",0,0,"2162216",,terminal_output +6356,18180453,"TERMINAL",0,0,"3273327",,terminal_output +6357,18181466,"TERMINAL",0,0,"4384438",,terminal_output +6358,18182508,"TERMINAL",0,0,"5495549",,terminal_output +6359,18183562,"TERMINAL",0,0,"654066550",,terminal_output +6360,18184607,"TERMINAL",0,0,"7617761",,terminal_output +6361,18185649,"TERMINAL",0,0,"8728872",,terminal_output +6362,18186692,"TERMINAL",0,0,"9839983",,terminal_output +6363,18187735,"TERMINAL",0,0,"4094303094",,terminal_output +6364,18188769,"TERMINAL",0,0,"140511405",,terminal_output +6365,18189812,"TERMINAL",0,0,"2273327",,terminal_output +6366,18190863,"TERMINAL",0,0,"4384438",,terminal_output +6367,18191904,"TERMINAL",0,0,"5495549",,terminal_output +6368,18192943,"TERMINAL",0,0,"655066540:00",,terminal_output +6369,18193987,"TERMINAL",0,0,"7617761",,terminal_output +6370,18195032,"TERMINAL",0,0,"8728872",,terminal_output +6371,18196078,"TERMINAL",0,0,"9839983",,terminal_output +6372,18197125,"TERMINAL",0,0,"5094404094",,terminal_output +6373,18198168,"TERMINAL",0,0,"150511505",,terminal_output +6374,18199213,"TERMINAL",0,0,"2162216",,terminal_output +6375,18200265,"TERMINAL",0,0,"3273327",,terminal_output +6376,18201309,"TERMINAL",0,0,"4384438",,terminal_output +6377,18202348,"TERMINAL",0,0,"5495549",,terminal_output +6378,18203396,"TERMINAL",0,0,"652:0066510",,terminal_output +6379,18204467,"TERMINAL",0,0,"7617761",,terminal_output +6380,18205475,"TERMINAL",0,0,"8728872",,terminal_output +6381,18206527,"TERMINAL",0,0,"9839983",,terminal_output +6382,18207563,"TERMINAL",0,0,"1:0094505094",,terminal_output +6383,18208601,"TERMINAL",0,0,"11:0051140:005",,terminal_output +6384,18209644,"TERMINAL",0,0,"2162216",,terminal_output +6385,18210695,"TERMINAL",0,0,"3273327",,terminal_output +6386,18211736,"TERMINAL",0,0,"4384438",,terminal_output +6387,18212781,"TERMINAL",0,0,"5495549",,terminal_output +6388,18213868,"TERMINAL",0,0,"661177621",,terminal_output +6389,18214911,"TERMINAL",0,0,"8728872",,terminal_output +6390,18215953,"TERMINAL",0,0,"9839983",,terminal_output +6391,18217002,"TERMINAL",0,0,"10948:0040:0094",,terminal_output +6392,18218040,"TERMINAL",0,0,"110511105",,terminal_output +6393,18219089,"TERMINAL",0,0,"2162216",,terminal_output +6394,18220118,"TERMINAL",0,0,"3273327",,terminal_output +6395,18221159,"TERMINAL",0,0,"4384438",,terminal_output +6396,18222202,"TERMINAL",0,0,"5495549",,terminal_output +6397,18223248,"TERMINAL",0,0,"652066530",,terminal_output +6398,18224282,"TERMINAL",0,0,"7617761",,terminal_output +6399,18225332,"TERMINAL",0,0,"8728872",,terminal_output +6400,18226374,"TERMINAL",0,0,"9839983",,terminal_output +6401,18227454,"TERMINAL",0,0,"2094101094",,terminal_output +6402,18228466,"TERMINAL",0,0,"120511205",,terminal_output +6403,18229512,"TERMINAL",0,0,"2162216",,terminal_output +6404,18230556,"TERMINAL",0,0,"3273327",,terminal_output +6405,18231606,"TERMINAL",0,0,"4384438",,terminal_output +6406,18232648,"TERMINAL",0,0,"5495549",,terminal_output +6407,18233691,"TERMINAL",0,0,"653066540",,terminal_output +6408,18234743,"TERMINAL",0,0,"7617761",,terminal_output +6409,18235797,"TERMINAL",0,0,"8728872",,terminal_output +6410,18236839,"TERMINAL",0,0,"994202094",,terminal_output +6411,18237879,"TERMINAL",0,0,"3130511305",,terminal_output +6412,18238921,"TERMINAL",0,0,"2162216",,terminal_output +6413,18239957,"TERMINAL",0,0,"3273327",,terminal_output +6414,18241011,"TERMINAL",0,0,"4384438",,terminal_output +6415,18242065,"TERMINAL",0,0,"5495549",,terminal_output +6416,18243105,"TERMINAL",0,0,"654066550",,terminal_output +6417,18244147,"TERMINAL",0,0,"7617761",,terminal_output +6418,18245187,"TERMINAL",0,0,"8728872",,terminal_output +6419,18246225,"TERMINAL",0,0,"9839983",,terminal_output +6420,18247270,"TERMINAL",0,0,"4094303094",,terminal_output +6421,18248312,"TERMINAL",0,0,"140511405",,terminal_output +6422,18249348,"TERMINAL",0,0,"2162216",,terminal_output +6423,18250390,"TERMINAL",0,0,"3273327",,terminal_output +6424,18251444,"TERMINAL",0,0,"4384438",,terminal_output +6425,18252476,"TERMINAL",0,0,"5495549",,terminal_output +6426,18253515,"TERMINAL",0,0,"65506651:00",,terminal_output +6427,18254553,"TERMINAL",0,0,"7617761",,terminal_output +6428,18255592,"TERMINAL",0,0,"8728872",,terminal_output +6429,18256637,"TERMINAL",0,0,"9839983",,terminal_output +6430,18257687,"TERMINAL",0,0,"5094404094",,terminal_output +6431,18258732,"TERMINAL",0,0,"150511505",,terminal_output +6432,18259775,"TERMINAL",0,0,"2162216",,terminal_output +6433,18260825,"TERMINAL",0,0,"3384438",,terminal_output +6434,18261860,"TERMINAL",0,0,"5495549",,terminal_output +6435,18262910,"TERMINAL",0,0,"653:0066510",,terminal_output +6436,18263956,"TERMINAL",0,0,"7617761",,terminal_output +6437,18265005,"TERMINAL",0,0,"8728872",,terminal_output +6438,18266057,"TERMINAL",0,0,"9839983",,terminal_output +6439,18267102,"TERMINAL",0,0,"2:0094505094",,terminal_output +6440,18268149,"TERMINAL",0,0,"12:005111:005",,terminal_output +6441,18269189,"TERMINAL",0,0,"2162216",,terminal_output +6442,18270253,"TERMINAL",0,0,"3273327",,terminal_output +6443,18271280,"TERMINAL",0,0,"4384438",,terminal_output +6444,18272331,"TERMINAL",0,0,"5495549",,terminal_output +6445,18273375,"TERMINAL",0,0,"651066520",,terminal_output +6446,18274417,"TERMINAL",0,0,"7617761",,terminal_output +6447,18275497,"TERMINAL",0,0,"8728872",,terminal_output +6448,18276516,"TERMINAL",0,0,"9839983",,terminal_output +6449,18277562,"TERMINAL",0,0,"10949:001:0094",,terminal_output +6450,18278618,"TERMINAL",0,0,"110511105",,terminal_output +6451,18279647,"TERMINAL",0,0,"2162216",,terminal_output +6452,18280692,"TERMINAL",0,0,"3273327",,terminal_output +6453,18281734,"TERMINAL",0,0,"4384438",,terminal_output +6454,18282771,"TERMINAL",0,0,"5495549",,terminal_output +6455,18283813,"TERMINAL",0,0,"652066530",,terminal_output +6456,18284852,"TERMINAL",0,0,"8728872",,terminal_output +6457,18285894,"TERMINAL",0,0,"9839983",,terminal_output +6458,18286936,"TERMINAL",0,0,"2094101094",,terminal_output +6459,18287978,"TERMINAL",0,0,"120511205",,terminal_output +6460,18289025,"TERMINAL",0,0,"2162216",,terminal_output +6461,18290075,"TERMINAL",0,0,"3273327",,terminal_output +6462,18291112,"TERMINAL",0,0,"4384438",,terminal_output +6463,18292152,"TERMINAL",0,0,"5495549",,terminal_output +6464,18293193,"TERMINAL",0,0,"653066540",,terminal_output +6465,18294246,"TERMINAL",0,0,"7617761",,terminal_output +6466,18295286,"TERMINAL",0,0,"8728872",,terminal_output +6467,18296331,"TERMINAL",0,0,"9839983",,terminal_output +6468,18297375,"TERMINAL",0,0,"3094202094",,terminal_output +6469,18298419,"TERMINAL",0,0,"130511305",,terminal_output +6470,18299488,"TERMINAL",0,0,"2162216",,terminal_output +6471,18300518,"TERMINAL",0,0,"3273327",,terminal_output +6472,18301565,"TERMINAL",0,0,"4384438",,terminal_output +6473,18302609,"TERMINAL",0,0,"5495549",,terminal_output +6474,18303643,"TERMINAL",0,0,"654066550",,terminal_output +6475,18304686,"TERMINAL",0,0,"7617761",,terminal_output +6476,18305721,"TERMINAL",0,0,"8728872",,terminal_output +6477,18306768,"TERMINAL",0,0,"9839983",,terminal_output +6478,18307812,"TERMINAL",0,0,"4094303094",,terminal_output +6479,18308856,"TERMINAL",0,0,"241622416",,terminal_output +6480,18309903,"TERMINAL",0,0,"3273327",,terminal_output +6481,18310947,"TERMINAL",0,0,"4384438",,terminal_output +6482,18311986,"TERMINAL",0,0,"5495549",,terminal_output +6483,18313021,"TERMINAL",0,0,"65506652:00",,terminal_output +6484,18314068,"TERMINAL",0,0,"7617761",,terminal_output +6485,18315110,"TERMINAL",0,0,"8728872",,terminal_output +6486,18316160,"TERMINAL",0,0,"9839983",,terminal_output +6487,18317202,"TERMINAL",0,0,"5094404094",,terminal_output +6488,18318246,"TERMINAL",0,0,"150511505",,terminal_output +6489,18319298,"TERMINAL",0,0,"2162216",,terminal_output +6490,18320347,"TERMINAL",0,0,"3273327",,terminal_output +6491,18321381,"TERMINAL",0,0,"4384438",,terminal_output +6492,18322431,"TERMINAL",0,0,"5495549",,terminal_output +6493,18323505,"TERMINAL",0,0,"654:0066510",,terminal_output +6494,18324524,"TERMINAL",0,0,"7617761",,terminal_output +6495,18325571,"TERMINAL",0,0,"8728872",,terminal_output +6496,18326620,"TERMINAL",0,0,"9839983",,terminal_output +6497,18327665,"TERMINAL",0,0,"3:0094505094",,terminal_output +6498,18328697,"TERMINAL",0,0,"13:005112:005",,terminal_output +6499,18329740,"TERMINAL",0,0,"2162216",,terminal_output +6500,18330790,"TERMINAL",0,0,"3273327",,terminal_output +6501,18331833,"TERMINAL",0,0,"4495549",,terminal_output +6502,18332873,"TERMINAL",0,0,"651066520",,terminal_output +6503,18333925,"TERMINAL",0,0,"7617761",,terminal_output +6504,18334966,"TERMINAL",0,0,"8728872",,terminal_output +6505,18336012,"TERMINAL",0,0,"9839983",,terminal_output +6506,18337054,"TERMINAL",0,0,"109430:002:0094",,terminal_output +6507,18338105,"TERMINAL",0,0,"110511105",,terminal_output +6508,18339148,"TERMINAL",0,0,"2162216",,terminal_output +6509,18340189,"TERMINAL",0,0,"3273327",,terminal_output +6510,18341233,"TERMINAL",0,0,"4384438",,terminal_output +6511,18342270,"TERMINAL",0,0,"5495549",,terminal_output +6512,18343314,"TERMINAL",0,0,"652066530",,terminal_output +6513,18344351,"TERMINAL",0,0,"7617761",,terminal_output +6514,18345390,"TERMINAL",0,0,"8728872",,terminal_output +6515,18346433,"TERMINAL",0,0,"9839983",,terminal_output +6516,18347512,"TERMINAL",0,0,"2094101094",,terminal_output +6517,18348553,"TERMINAL",0,0,"120511205",,terminal_output +6518,18349561,"TERMINAL",0,0,"2162216",,terminal_output +6519,18350600,"TERMINAL",0,0,"3273327",,terminal_output +6520,18351642,"TERMINAL",0,0,"4384438",,terminal_output +6521,18352685,"TERMINAL",0,0,"5495549",,terminal_output +6522,18353728,"TERMINAL",0,0,"653066540",,terminal_output +6523,18354771,"TERMINAL",0,0,"7617761",,terminal_output +6524,18355817,"TERMINAL",0,0,"8839983",,terminal_output +6525,18356858,"TERMINAL",0,0,"3094202094",,terminal_output +6526,18357941,"TERMINAL",0,0,"130511305",,terminal_output +6527,18358955,"TERMINAL",0,0,"2162216",,terminal_output +6528,18359996,"TERMINAL",0,0,"3273327",,terminal_output +6529,18361037,"TERMINAL",0,0,"4384438",,terminal_output +6530,18362083,"TERMINAL",0,0,"5495549",,terminal_output +6531,18363119,"TERMINAL",0,0,"654066550",,terminal_output +6532,18364164,"TERMINAL",0,0,"7617761",,terminal_output +6533,18365205,"TERMINAL",0,0,"8728872",,terminal_output +6534,18366262,"TERMINAL",0,0,"9839983",,terminal_output +6535,18367289,"TERMINAL",0,0,"4094303094",,terminal_output +6536,18368522,"TERMINAL",0,0,"140511405",,terminal_output +6537,18369527,"TERMINAL",0,0,"2162216",,terminal_output +6538,18370575,"TERMINAL",0,0,"3273327",,terminal_output +6539,18371622,"TERMINAL",0,0,"4384438",,terminal_output +6540,18372669,"TERMINAL",0,0,"5495549",,terminal_output +6541,18373719,"TERMINAL",0,0,"65506653:00",,terminal_output +6542,18374763,"TERMINAL",0,0,"7617761",,terminal_output +6543,18375799,"TERMINAL",0,0,"8728872",,terminal_output +6544,18376853,"TERMINAL",0,0,"994404094",,terminal_output +6545,18377896,"TERMINAL",0,0,"5150511505",,terminal_output +6546,18379045,"TERMINAL",0,0,"2162216",,terminal_output +6547,18380085,"TERMINAL",0,0,"3273327",,terminal_output +6548,18381131,"TERMINAL",0,0,"4384438",,terminal_output +6549,18382186,"TERMINAL",0,0,"5495549",,terminal_output +6550,18383214,"TERMINAL",0,0,"655:0066510",,terminal_output +6551,18384294,"TERMINAL",0,0,"7617761",,terminal_output +6552,18385291,"TERMINAL",0,0,"8728872",,terminal_output +6553,18386335,"TERMINAL",0,0,"9839983",,terminal_output +6554,18387382,"TERMINAL",0,0,"4:0094505094",,terminal_output +6555,18388432,"TERMINAL",0,0,"14:005113:005",,terminal_output +6556,18389473,"TERMINAL",0,0,"2162216",,terminal_output +6557,18390543,"TERMINAL",0,0,"3273327",,terminal_output +6558,18391572,"TERMINAL",0,0,"4384438",,terminal_output +6559,18392615,"TERMINAL",0,0,"5495549",,terminal_output +6560,18393654,"TERMINAL",0,0,"651066520",,terminal_output +6561,18394683,"TERMINAL",0,0,"7617761",,terminal_output +6562,18395740,"TERMINAL",0,0,"8728872",,terminal_output +6563,18396784,"TERMINAL",0,0,"9839983",,terminal_output +6564,18397829,"TERMINAL",0,0,"101051:013:01105",,terminal_output +6565,18398875,"TERMINAL",0,0,"2162216",,terminal_output +6566,18399914,"TERMINAL",0,0,"3273327",,terminal_output +6567,18400954,"TERMINAL",0,0,"4384438",,terminal_output +6568,18402015,"TERMINAL",0,0,"5495549",,terminal_output +6569,18403199,"TERMINAL",0,0,"652066530",,terminal_output +6570,18404086,"TERMINAL",0,0,"7617761",,terminal_output +6571,18405128,"TERMINAL",0,0,"8728872",,terminal_output +6572,18406168,"TERMINAL",0,0,"9839983",,terminal_output +6573,18407216,"TERMINAL",0,0,"2094101094",,terminal_output +6574,18408262,"TERMINAL",0,0,"120511205",,terminal_output +6575,18409300,"TERMINAL",0,0,"2162216",,terminal_output +6576,18410338,"TERMINAL",0,0,"3273327",,terminal_output +6577,18411381,"TERMINAL",0,0,"4384438",,terminal_output +6578,18412422,"TERMINAL",0,0,"5495549",,terminal_output +6579,18413467,"TERMINAL",0,0,"653066540",,terminal_output +6580,18414540,"TERMINAL",0,0,"7617761",,terminal_output +6581,18415549,"TERMINAL",0,0,"8728872",,terminal_output +6582,18416631,"TERMINAL",0,0,"9839983",,terminal_output +6583,18417642,"TERMINAL",0,0,"3094202094",,terminal_output +6584,18418676,"TERMINAL",0,0,"130511305",,terminal_output +6585,18419718,"TERMINAL",0,0,"2162216",,terminal_output +6586,18420755,"TERMINAL",0,0,"3273327",,terminal_output +6587,18421796,"TERMINAL",0,0,"4384438",,terminal_output +6588,18422842,"TERMINAL",0,0,"554066550",,terminal_output +6589,18423885,"TERMINAL",0,0,"7617761",,terminal_output +6590,18424936,"TERMINAL",0,0,"8728872",,terminal_output +6591,18425977,"TERMINAL",0,0,"9839983",,terminal_output +6592,18427027,"TERMINAL",0,0,"4094303094",,terminal_output +6593,18428068,"TERMINAL",0,0,"140511405",,terminal_output +6594,18429122,"TERMINAL",0,0,"2162216",,terminal_output +6595,18430166,"TERMINAL",0,0,"3273327",,terminal_output +6596,18431210,"TERMINAL",0,0,"4384438",,terminal_output +6597,18432259,"TERMINAL",0,0,"5495549",,terminal_output +6598,18433311,"TERMINAL",0,0,"65506654:00",,terminal_output +6599,18434354,"TERMINAL",0,0,"7617761",,terminal_output +6600,18435400,"TERMINAL",0,0,"8728872",,terminal_output +6601,18436444,"TERMINAL",0,0,"9839983",,terminal_output +6602,18437492,"TERMINAL",0,0,"5094404094",,terminal_output +6603,18438566,"TERMINAL",0,0,"150511505",,terminal_output +6604,18439584,"TERMINAL",0,0,"2162216",,terminal_output +6605,18440630,"TERMINAL",0,0,"3273327",,terminal_output +6606,18441677,"TERMINAL",0,0,"4384438",,terminal_output +6607,18442721,"TERMINAL",0,0,"5495549",,terminal_output +6608,18443768,"TERMINAL",0,0,"656:0066510",,terminal_output +6609,18444813,"TERMINAL",0,0,"7617761",,terminal_output +6610,18445861,"TERMINAL",0,0,"9839983",,terminal_output +6611,18446909,"TERMINAL",0,0,"5:0094505094",,terminal_output +6612,18447942,"TERMINAL",0,0,"15:005114:005",,terminal_output +6613,18448988,"TERMINAL",0,0,"2162216",,terminal_output +6614,18450032,"TERMINAL",0,0,"3273327",,terminal_output +6615,18451078,"TERMINAL",0,0,"4384438",,terminal_output +6616,18452147,"TERMINAL",0,0,"5495549",,terminal_output +6617,18453163,"TERMINAL",0,0,"651066520",,terminal_output +6618,18454204,"TERMINAL",0,0,"7617761",,terminal_output +6619,18455253,"TERMINAL",0,0,"8728872",,terminal_output +6620,18456293,"TERMINAL",0,0,"9839983",,terminal_output +6621,18457327,"TERMINAL",0,0,"10942:004:0094",,terminal_output +6622,18458374,"TERMINAL",0,0,"110511105",,terminal_output +6623,18459413,"TERMINAL",0,0,"2162216",,terminal_output +6624,18460454,"TERMINAL",0,0,"3273327",,terminal_output +6625,18461495,"TERMINAL",0,0,"4384438",,terminal_output +6626,18462581,"TERMINAL",0,0,"5495549",,terminal_output +6627,18463574,"TERMINAL",0,0,"652066530",,terminal_output +6628,18464626,"TERMINAL",0,0,"7617761",,terminal_output +6629,18465660,"TERMINAL",0,0,"8728872",,terminal_output +6630,18466697,"TERMINAL",0,0,"9839983",,terminal_output +6631,18467744,"TERMINAL",0,0,"2094101094",,terminal_output +6632,18468781,"TERMINAL",0,0,"120511205",,terminal_output +6633,18469824,"TERMINAL",0,0,"2273327",,terminal_output +6634,18470861,"TERMINAL",0,0,"4384438",,terminal_output +6635,18471899,"TERMINAL",0,0,"5495549",,terminal_output +6636,18472945,"TERMINAL",0,0,"653066540",,terminal_output +6637,18473987,"TERMINAL",0,0,"7617761",,terminal_output +6638,18475024,"TERMINAL",0,0,"8728872",,terminal_output +6639,18476068,"TERMINAL",0,0,"9839983",,terminal_output +6640,18477116,"TERMINAL",0,0,"3094202094",,terminal_output +6641,18478155,"TERMINAL",0,0,"130511305",,terminal_output +6642,18479190,"TERMINAL",0,0,"2162216",,terminal_output +6643,18480242,"TERMINAL",0,0,"3273327",,terminal_output +6644,18481289,"TERMINAL",0,0,"4384438",,terminal_output +6645,18482335,"TERMINAL",0,0,"5495549",,terminal_output +6646,18483369,"TERMINAL",0,0,"654066550",,terminal_output +6647,18484426,"TERMINAL",0,0,"7617761",,terminal_output +6648,18485467,"TERMINAL",0,0,"8728872",,terminal_output +6649,18486515,"TERMINAL",0,0,"9839983",,terminal_output +6650,18487583,"TERMINAL",0,0,"4094303094",,terminal_output +6651,18488589,"TERMINAL",0,0,"140511405",,terminal_output +6652,18489637,"TERMINAL",0,0,"2162216",,terminal_output +6653,18490677,"TERMINAL",0,0,"3273327",,terminal_output +6654,18491721,"TERMINAL",0,0,"4384438",,terminal_output +6655,18492801,"TERMINAL",0,0,"5495549",,terminal_output +6656,18493784,"TERMINAL",0,0,"65506655:00",,terminal_output +6657,18494821,"TERMINAL",0,0,"7728872",,terminal_output +6658,18495880,"TERMINAL",0,0,"9839983",,terminal_output +6659,18496920,"TERMINAL",0,0,"5094404094",,terminal_output +6660,18497958,"TERMINAL",0,0,"150511505",,terminal_output +6661,18499022,"TERMINAL",0,0,"2162216",,terminal_output +6662,18500047,"TERMINAL",0,0,"3273327",,terminal_output +6663,18501092,"TERMINAL",0,0,"4384438",,terminal_output +6664,18502145,"TERMINAL",0,0,"5495549",,terminal_output +6665,18503162,"TERMINAL",0,0,"657:0066510",,terminal_output +6666,18504200,"TERMINAL",0,0,"7617761",,terminal_output +6667,18505253,"TERMINAL",0,0,"8728872",,terminal_output +6668,18506293,"TERMINAL",0,0,"9839983",,terminal_output +6669,18507346,"TERMINAL",0,0,"6:0094505094",,terminal_output +6670,18508381,"TERMINAL",0,0,"16:005115:005",,terminal_output +6671,18509413,"TERMINAL",0,0,"2162216",,terminal_output +6672,18510456,"TERMINAL",0,0,"3273327",,terminal_output +6673,18511496,"TERMINAL",0,0,"4384438",,terminal_output +6674,18512533,"TERMINAL",0,0,"5495549",,terminal_output +6675,18513602,"TERMINAL",0,0,"651066520",,terminal_output +6676,18514624,"TERMINAL",0,0,"7617761",,terminal_output +6677,18515672,"TERMINAL",0,0,"8728872",,terminal_output +6678,18516714,"TERMINAL",0,0,"9839983",,terminal_output +6679,18517794,"TERMINAL",0,0,"10943:005:0094",,terminal_output +6680,18518801,"TERMINAL",0,0,"110511105",,terminal_output +6681,18519844,"TERMINAL",0,0,"2273327",,terminal_output +6682,18520894,"TERMINAL",0,0,"4384438",,terminal_output +6683,18521932,"TERMINAL",0,0,"5495549",,terminal_output +6684,18523011,"TERMINAL",0,0,"652066530",,terminal_output +6685,18524041,"TERMINAL",0,0,"7617761",,terminal_output +6686,18525090,"TERMINAL",0,0,"8728872",,terminal_output +6687,18526117,"TERMINAL",0,0,"9839983",,terminal_output +6688,18527161,"TERMINAL",0,0,"2094101094",,terminal_output +6689,18528199,"TERMINAL",0,0,"120511205",,terminal_output +6690,18529241,"TERMINAL",0,0,"2162216",,terminal_output +6691,18530291,"TERMINAL",0,0,"3273327",,terminal_output +6692,18531417,"TERMINAL",0,0,"4384438",,terminal_output +6693,18532459,"TERMINAL",0,0,"5495549",,terminal_output +6694,18533510,"TERMINAL",0,0,"653066540",,terminal_output +6695,18534552,"TERMINAL",0,0,"7617761",,terminal_output +6696,18535648,"TERMINAL",0,0,"8728872",,terminal_output +6697,18536641,"TERMINAL",0,0,"9839983",,terminal_output +6698,18537762,"TERMINAL",0,0,"3094202094",,terminal_output +6699,18538725,"TERMINAL",0,0,"130511305",,terminal_output +6700,18539778,"TERMINAL",0,0,"2162216",,terminal_output +6701,18540824,"TERMINAL",0,0,"3384438",,terminal_output +6702,18541886,"TERMINAL",0,0,"5495549",,terminal_output +6703,18542920,"TERMINAL",0,0,"654066550",,terminal_output +6704,18543969,"TERMINAL",0,0,"7617761",,terminal_output +6705,18545007,"TERMINAL",0,0,"8728872",,terminal_output +6706,18546049,"TERMINAL",0,0,"9839983",,terminal_output +6707,18547139,"TERMINAL",0,0,"4094303094",,terminal_output +6708,18548162,"TERMINAL",0,0,"140511405",,terminal_output +6709,18549185,"TERMINAL",0,0,"2162216",,terminal_output +6710,18550234,"TERMINAL",0,0,"3273327",,terminal_output +6711,18551280,"TERMINAL",0,0,"4384438",,terminal_output +6712,18552328,"TERMINAL",0,0,"5495549",,terminal_output +6713,18553373,"TERMINAL",0,0,"65506656:00",,terminal_output +6714,18554417,"TERMINAL",0,0,"7617761",,terminal_output +6715,18555468,"TERMINAL",0,0,"8728872",,terminal_output +6716,18556511,"TERMINAL",0,0,"9839983",,terminal_output +6717,18557578,"TERMINAL",0,0,"5094404094",,terminal_output +6718,18558637,"TERMINAL",0,0,"150511505",,terminal_output +6719,18559663,"TERMINAL",0,0,"2162216",,terminal_output +6720,18560705,"TERMINAL",0,0,"3273327",,terminal_output +6721,18561756,"TERMINAL",0,0,"4384438",,terminal_output +6722,18562802,"TERMINAL",0,0,"5495549",,terminal_output +6723,18563852,"TERMINAL",0,0,"668:0177611",,terminal_output +6724,18564901,"TERMINAL",0,0,"8728872",,terminal_output +6725,18565950,"TERMINAL",0,0,"9839983",,terminal_output +6726,18567028,"TERMINAL",0,0,"7:0094505094",,terminal_output +6727,18568044,"TERMINAL",0,0,"17:005116:005",,terminal_output +6728,18569092,"TERMINAL",0,0,"2162216",,terminal_output +6729,18570142,"TERMINAL",0,0,"3273327",,terminal_output +6730,18571187,"TERMINAL",0,0,"4384438",,terminal_output +6731,18572239,"TERMINAL",0,0,"5495549",,terminal_output +6732,18573278,"TERMINAL",0,0,"651066520",,terminal_output +6733,18574322,"TERMINAL",0,0,"7617761",,terminal_output +6734,18575370,"TERMINAL",0,0,"8728872",,terminal_output +6735,18576406,"TERMINAL",0,0,"9839983",,terminal_output +6736,18577450,"TERMINAL",0,0,"10944:006:0094",,terminal_output +6737,18578496,"TERMINAL",0,0,"110511105",,terminal_output +6738,18579548,"TERMINAL",0,0,"2162216",,terminal_output +6739,18580626,"TERMINAL",0,0,"3273327",,terminal_output +6740,18581644,"TERMINAL",0,0,"4384438",,terminal_output +6741,18582686,"TERMINAL",0,0,"5495549",,terminal_output +6742,18583728,"TERMINAL",0,0,"652066530",,terminal_output +6743,18584777,"TERMINAL",0,0,"7617761",,terminal_output +6744,18585828,"TERMINAL",0,0,"8839983",,terminal_output +6745,18586869,"TERMINAL",0,0,"2094101094",,terminal_output +6746,18587920,"TERMINAL",0,0,"120511205",,terminal_output +6747,18588986,"TERMINAL",0,0,"2162216",,terminal_output +6748,18590012,"TERMINAL",0,0,"3273327",,terminal_output +6749,18591069,"TERMINAL",0,0,"4384438",,terminal_output +6750,18592102,"TERMINAL",0,0,"5495549",,terminal_output +6751,18593143,"TERMINAL",0,0,"653066540",,terminal_output +6752,18594192,"TERMINAL",0,0,"7617761",,terminal_output +6753,18595240,"TERMINAL",0,0,"8728872",,terminal_output +6754,18596286,"TERMINAL",0,0,"9839983",,terminal_output +6755,18597328,"TERMINAL",0,0,"3094202094",,terminal_output +6756,18598376,"TERMINAL",0,0,"130511305",,terminal_output +6757,18599416,"TERMINAL",0,0,"2162216",,terminal_output +6758,18600473,"TERMINAL",0,0,"3273327",,terminal_output +6759,18601514,"TERMINAL",0,0,"4384438",,terminal_output +6760,18602559,"TERMINAL",0,0,"5495549",,terminal_output +6761,18603629,"TERMINAL",0,0,"654066550",,terminal_output +6762,18604656,"TERMINAL",0,0,"7617761",,terminal_output +6763,18605702,"TERMINAL",0,0,"8728872",,terminal_output +6764,18606746,"TERMINAL",0,0,"9839983",,terminal_output +6765,18607792,"TERMINAL",0,0,"4094303094",,terminal_output +6766,18608842,"TERMINAL",0,0,"141622416",,terminal_output +6767,18609892,"TERMINAL",0,0,"3273327",,terminal_output +6768,18610936,"TERMINAL",0,0,"4384438",,terminal_output +6769,18611979,"TERMINAL",0,0,"5495549",,terminal_output +6770,18613024,"TERMINAL",0,0,"65506657:00",,terminal_output +6771,18614081,"TERMINAL",0,0,"7617761",,terminal_output +6772,18615118,"TERMINAL",0,0,"8728872",,terminal_output +6773,18616158,"TERMINAL",0,0,"9839983",,terminal_output +6774,18617209,"TERMINAL",0,0,"5094404094",,terminal_output +6775,18618255,"TERMINAL",0,0,"150511505",,terminal_output +6776,18619321,"TERMINAL",0,0,"2162216",,terminal_output +6777,18620355,"TERMINAL",0,0,"3273327",,terminal_output +6778,18621398,"TERMINAL",0,0,"4384438",,terminal_output +6779,18622442,"TERMINAL",0,0,"5495549",,terminal_output +6780,18623481,"TERMINAL",0,0,"659:0066510",,terminal_output +6781,18624519,"TERMINAL",0,0,"7617761",,terminal_output +6782,18625567,"TERMINAL",0,0,"8728872",,terminal_output +6783,18626611,"TERMINAL",0,0,"9839983",,terminal_output +6784,18627659,"TERMINAL",0,0,"8:0094505094",,terminal_output +6785,18628704,"TERMINAL",0,0,"18:005117:005",,terminal_output +6786,18629814,"TERMINAL",0,0,"2162216",,terminal_output +6787,18630790,"TERMINAL",0,0,"3273327",,terminal_output +6788,18631399,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +6789,18632074,"TERMINAL",0,0,"idle",,terminal_command +6790,18632142,"TERMINAL",0,0,"]633;E;2025-08-14 15:48:05 idle;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;CPartition dev_cpuonly : 11 nodes idle\r\nPartition cpuonly : 241 nodes idle\r\nPartition dev_accelerated : 0 nodes idle\r\nPartition accelerated : 24 nodes idle\r\nPartition dev_accelerated-h100 : 1 nodes idle\r\nPartition accelerated-h100 : 1 nodes idle\r\nPartition large : 8 nodes idle\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +6791,19545511,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",0,0,"",shellscript,tab +6792,19550608,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1442,0,"",shellscript,selection_mouse +6793,19552223,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1701,0,"",shellscript,selection_mouse +6794,19552400,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes.sbatch",1695,12,"SLURM_JOB_ID",shellscript,selection_mouse +6795,19564074,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes copy.sbatch",0,0,"",shellscript,tab +6796,19570138,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=4\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=192 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-8-nodes-$slurm_job_id \\n --tags tokenizer big-run 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid",shellscript,tab +6797,19571205,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",0,0,"",shellscript,tab +6798,19572316,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes copy.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-8-nodes-$slurm_job_id \\n --tags tokenizer big-run 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid",shellscript,tab +6799,19577970,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-8-nodes-$slurm_job_id \\n --tags tokenizer big-run 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid",shellscript,tab +6800,19579807,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",460,0,"",shellscript,selection_mouse +6801,19580706,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",443,29,"",shellscript,content +6802,19581605,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",455,0,"servation=llmtum\n#SBATCH --re",shellscript,content +6803,19581627,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",460,0,"",shellscript,selection_command +6804,19605107,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1449,0,"",shellscript,selection_mouse +6805,19605975,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1448,0,"",shellscript,selection_command +6806,19608136,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1558,0,"\n",shellscript,content +6807,19608487,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1559,0,"l",shellscript,content +6808,19608488,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1560,0,"",shellscript,selection_keyboard +6809,19609126,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1559,1,"",shellscript,content +6810,19610188,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1433,0,"",shellscript,selection_command +6811,19610707,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1558,0,"\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id",shellscript,content +6812,19610742,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1559,0,"",shellscript,selection_command +6813,19611088,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1560,0,"",shellscript,selection_command +6814,19611595,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1561,0,"",shellscript,selection_command +6815,19611602,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1562,0,"",shellscript,selection_command +6816,19611629,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1563,0,"",shellscript,selection_command +6817,19611708,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1564,0,"",shellscript,selection_command +6818,19611709,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1565,0,"",shellscript,selection_command +6819,19611722,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1566,0,"",shellscript,selection_command +6820,19611752,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1567,0,"",shellscript,selection_command +6821,19611789,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1568,0,"",shellscript,selection_command +6822,19611834,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1569,0,"",shellscript,selection_command +6823,19611870,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1570,0,"",shellscript,selection_command +6824,19611871,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1571,0,"",shellscript,selection_command +6825,19611944,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1572,0,"",shellscript,selection_command +6826,19611944,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1573,0,"",shellscript,selection_command +6827,19612063,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1574,0,"",shellscript,selection_command +6828,19612865,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1574,110,"",shellscript,content +6829,19612891,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1573,0,"",shellscript,selection_command +6830,19613292,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1574,0,"",shellscript,selection_command +6831,19613815,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1574,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973",shellscript,content +6832,19614602,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1688,0,"",shellscript,selection_command +6833,19618936,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,0,"",shellscript,selection_mouse +6834,19619169,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,2,"$r",shellscript,selection_mouse +6835,19619169,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,5,"$rest",shellscript,selection_mouse +6836,19619169,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,7,"$restor",shellscript,selection_mouse +6837,19619169,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,8,"$restore",shellscript,selection_mouse +6838,19619169,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,10,"$restore_c",shellscript,selection_mouse +6839,19619203,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,12,"$restore_ckp",shellscript,selection_mouse +6840,19619203,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,13,"$restore_ckpt",shellscript,selection_mouse +6841,19619228,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,14,"$restore_ckpt_",shellscript,selection_mouse +6842,19619242,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,15,"$restore_ckpt_f",shellscript,selection_mouse +6843,19619275,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,16,"$restore_ckpt_fl",shellscript,selection_mouse +6844,19619369,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,17,"$restore_ckpt_fla",shellscript,selection_mouse +6845,19619496,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,18,"$restore_ckpt_flag",shellscript,selection_mouse +6846,19620191,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,18,"",shellscript,content +6847,19622162,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1790,0,"-",shellscript,content +6848,19622163,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1791,0,"",shellscript,selection_keyboard +6849,19622268,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1791,0,"-",shellscript,content +6850,19622269,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1792,0,"",shellscript,selection_keyboard +6851,19625661,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1792,0,"r",shellscript,content +6852,19625662,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1793,0,"",shellscript,selection_keyboard +6853,19625850,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1793,0,"e",shellscript,content +6854,19625851,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1794,0,"",shellscript,selection_keyboard +6855,19626060,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1794,0,"s",shellscript,content +6856,19626061,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1795,0,"",shellscript,selection_keyboard +6857,19626160,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1795,0,"t",shellscript,content +6858,19626161,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1796,0,"",shellscript,selection_keyboard +6859,19626356,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1796,0,"o",shellscript,content +6860,19626356,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1797,0,"",shellscript,selection_keyboard +6861,19626419,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1797,0,"r",shellscript,content +6862,19626420,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1798,0,"",shellscript,selection_keyboard +6863,19626576,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1798,0,"e",shellscript,content +6864,19626577,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1799,0,"",shellscript,selection_keyboard +6865,19626759,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1799,0,"_",shellscript,content +6866,19626760,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1800,0,"",shellscript,selection_keyboard +6867,19627539,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1800,0,"c",shellscript,content +6868,19627539,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1801,0,"",shellscript,selection_keyboard +6869,19628623,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1801,0,"k",shellscript,content +6870,19628624,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1802,0,"",shellscript,selection_keyboard +6871,19628781,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1802,0,"p",shellscript,content +6872,19628782,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1803,0,"",shellscript,selection_keyboard +6873,19628869,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1803,0,"t",shellscript,content +6874,19628870,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1804,0,"",shellscript,selection_keyboard +6875,19640825,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1682,0,"",shellscript,selection_mouse +6876,19640981,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1682,1,"3",shellscript,selection_mouse +6877,19641001,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1682,2,"34",shellscript,selection_mouse +6878,19641014,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1682,3,"342",shellscript,selection_mouse +6879,19641037,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1682,4,"3421",shellscript,selection_mouse +6880,19641073,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1682,5,"34219",shellscript,selection_mouse +6881,19641151,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1682,6,"342197",shellscript,selection_mouse +6882,19641298,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1682,7,"3421973",shellscript,selection_mouse +6883,19644579,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1464,0,"",shellscript,selection_mouse +6884,19644787,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1463,0,"",shellscript,selection_command +6885,19645335,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1433,0,"",shellscript,selection_command +6886,19645702,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1433,0,"#",shellscript,content +6887,19645703,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1434,0,"",shellscript,selection_keyboard +6888,19645805,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1434,0," ",shellscript,content +6889,19645806,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1435,0,"",shellscript,selection_keyboard +6890,19646171,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1434,0,"",shellscript,selection_command +6891,19649371,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,0,"",shellscript,selection_mouse +6892,19649524,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,1,"$",shellscript,selection_mouse +6893,19649572,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,2,"$S",shellscript,selection_mouse +6894,19649573,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,3,"$SL",shellscript,selection_mouse +6895,19649654,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,4,"$SLU",shellscript,selection_mouse +6896,19649655,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,5,"$SLUR",shellscript,selection_mouse +6897,19649655,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,6,"$SLURM",shellscript,selection_mouse +6898,19649665,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,7,"$SLURM_",shellscript,selection_mouse +6899,19649676,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,8,"$SLURM_J",shellscript,selection_mouse +6900,19649710,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,9,"$SLURM_JO",shellscript,selection_mouse +6901,19649798,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,10,"$SLURM_JOB",shellscript,selection_mouse +6902,19649830,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,11,"$SLURM_JOB_",shellscript,selection_mouse +6903,19649874,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,12,"$SLURM_JOB_I",shellscript,selection_mouse +6904,19650161,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,13,"$SLURM_JOB_ID",shellscript,selection_mouse +6905,19650749,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,13,"",shellscript,content +6906,19651292,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1824,0,"3421973",shellscript,content +6907,19655750,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2030,0,"",shellscript,selection_mouse +6908,19656582,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2029,1,"",shellscript,content +6909,19656659,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2029,0,"4",shellscript,content +6910,19656659,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2030,0,"",shellscript,selection_keyboard +6911,19661753,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2029,1,"",shellscript,content +6912,19661811,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2029,0,"2",shellscript,content +6913,19661811,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2030,0,"",shellscript,selection_keyboard +6914,19683052,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1425,0,"",shellscript,selection_mouse +6915,19683191,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1419,12,"SLURM_JOB_ID",shellscript,selection_mouse +6916,19684259,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1411,0,"",shellscript,selection_mouse +6917,19684396,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1405,12,"slurm_job_id",shellscript,selection_mouse +6918,19687138,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2050,0,"",shellscript,selection_mouse +6919,19688278,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2049,1,"d",shellscript,selection_mouse +6920,19688279,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2047,3,"_id",shellscript,selection_mouse +6921,19688319,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2046,4,"b_id",shellscript,selection_mouse +6922,19688320,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2045,5,"ob_id",shellscript,selection_mouse +6923,19688320,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2044,6,"job_id",shellscript,selection_mouse +6924,19688357,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2043,7,"_job_id",shellscript,selection_mouse +6925,19688394,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2042,8,"m_job_id",shellscript,selection_mouse +6926,19688473,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2041,9,"rm_job_id",shellscript,selection_mouse +6927,19688480,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2040,10,"urm_job_id",shellscript,selection_mouse +6928,19688522,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2039,11,"lurm_job_id",shellscript,selection_mouse +6929,19688611,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2038,12,"slurm_job_id",shellscript,selection_mouse +6930,19688744,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",2037,13,"$slurm_job_id",shellscript,selection_mouse +6931,19691349,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,0,"",shellscript,selection_mouse +6932,19691449,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,1,"$",shellscript,selection_mouse +6933,19691480,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,3,"$SL",shellscript,selection_mouse +6934,19691480,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,5,"$SLUR",shellscript,selection_mouse +6935,19691502,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,6,"$SLURM",shellscript,selection_mouse +6936,19691503,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,7,"$SLURM_",shellscript,selection_mouse +6937,19691518,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,8,"$SLURM_J",shellscript,selection_mouse +6938,19691554,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,9,"$SLURM_JO",shellscript,selection_mouse +6939,19691591,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,10,"$SLURM_JOB",shellscript,selection_mouse +6940,19691625,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,11,"$SLURM_JOB_",shellscript,selection_mouse +6941,19691661,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,12,"$SLURM_JOB_I",shellscript,selection_mouse +6942,19691697,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,13,"$SLURM_JOB_ID",shellscript,selection_mouse +6943,19692110,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,13,"",shellscript,content +6944,19692611,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1418,0,"3421973",shellscript,content +6945,19694298,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1822,0,"",shellscript,selection_mouse +6946,19695294,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1822,0,"$",shellscript,content +6947,19695295,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1823,0,"",shellscript,selection_keyboard +6948,19695891,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1822,1,"",shellscript,content +6949,19696060,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1821,1,"",shellscript,content +6950,19696186,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1820,1,"",shellscript,content +6951,19696353,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1819,1,"",shellscript,content +6952,19696427,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1818,1,"",shellscript,content +6953,19696664,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1818,1,"",shellscript,content +6954,19696832,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1818,1,"",shellscript,content +6955,19696953,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1818,1,"",shellscript,content +6956,19697642,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1818,0,"$",shellscript,content +6957,19697643,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1819,0,"",shellscript,selection_keyboard +6958,19697880,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1819,0,"s",shellscript,content +6959,19697881,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1820,0,"",shellscript,selection_keyboard +6960,19697970,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1820,0,"l",shellscript,content +6961,19697971,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1821,0,"",shellscript,selection_keyboard +6962,19698811,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1819,2,"slurm_job_id",shellscript,content +6963,19707902,"TERMINAL",0,0,"bash",,terminal_focus +6964,19721047,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",0,0,"",shellscript,tab +6965,19723290,"TERMINAL",0,0,"bash",,terminal_focus +6966,19725194,"TERMINAL",0,0,"queue",,terminal_command +6967,19725291,"TERMINAL",0,0,"]633;E;2025-08-14 16:06:18 queue;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 16:06:18 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-22:06:17\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-02:17:22\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 3:53:08\t 1 hkn07133421982 accelerat train_to tum_cte0 R 4:05:08\t 4 hkn[0724,0726,0728,0731]3421973 accelerat train_to tum_cte0 R 4:05:17\t 2 hkn[0722-0723]3421964 accelerat train_to tum_cte0 R 4:05:32\t 1 hkn0720",,terminal_output +6968,19726506,"TERMINAL",0,0,"9839983",,terminal_output +6969,19727370,"TERMINAL",0,0,"2094101094",,terminal_output +6970,19728467,"TERMINAL",0,0,"120511205",,terminal_output +6971,19729452,"TERMINAL",0,0,"2162216",,terminal_output +6972,19730500,"TERMINAL",0,0,"3273327",,terminal_output +6973,19731618,"TERMINAL",0,0,"4384438",,terminal_output +6974,19732592,"TERMINAL",0,0,"5495549",,terminal_output +6975,19733641,"TERMINAL",0,0,"653066540",,terminal_output +6976,19734582,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +6977,19737008,"TERMINAL",0,0,"scancel 3421973",,terminal_command +6978,19737049,"TERMINAL",0,0,"]633;E;2025-08-14 16:06:30 scancel 3421973;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +6979,19745085,"TERMINAL",0,0,"logs",,terminal_command +6980,19745115,"TERMINAL",0,0,"]633;E;2025-08-14 16:06:38 logs;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +6981,19746058,"TERMINAL",0,0,"bash",,terminal_focus +6982,19748274,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",,terminal_command +6983,19748298,"TERMINAL",0,0,"]633;E;2025-08-14 16:06:41 sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Csbatch: error: Unable to open file slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_franz/maskgit/dynamics-cotraining]633;D;1",,terminal_output +6984,19754667,"TERMINAL",0,0,"dev",,terminal_command +6985,19756325,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",,terminal_command +6986,19757855,"TERMINAL",0,0,"bash",,terminal_focus +6987,19759125,"TERMINAL",0,0,"ls",,terminal_command +6988,19759176,"TERMINAL",0,0,"]633;E;2025-08-14 16:06:52 ls;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C",,terminal_output +6989,19759418,"TERMINAL",0,0,"atari train_lam_action_space_scaling_50_3329804.log\r\nbig_run train_lam_action_space_scaling_50_3331286.log\r\nbig-runs train_lam_action_space_scaling_6_3318549.log\r\ncausal train_lam_action_space_scaling_6_3320178.log\r\ncoinrun train_lam_action_space_scaling_6_3321528.log\r\nmaskgit train_lam_action_space_scaling_6_3329790.log\r\nmaskgit-maskprob-fix train_lam_action_space_scaling_6_3329805.log\r\npreprocess train_lam_action_space_scaling_6_3331287.log\r\ntrain_dyn_causal_180M_3372931.log train_lam_action_space_scaling_8_3318550.log\r\ntrain_dyn_causal_180M_3372963.log train_lam_action_space_scaling_8_3329791.log\r\ntrain_dyn_causal_180M_3372969.log train_lam_action_space_scaling_8_3329806.log\r\ntrain_dyn_causal_180M_3373107.log train_lam_action_space_scaling_8_3331288.log\r\ntrain_dyn_causal_255M_3372932.log train_lam_minecraft_overfit_sample_3309655.log\r\ntrain_dyn_causal_255M_3372970.log train_lam_model_size_scaling_38M_3317098.log\r\ntrain_dyn_causal_255M_3373108.log train_lam_model_size_scaling_38M_3317115.log\r\ntrain_dyn_causal_356M_3372934.log train_lam_model_size_scaling_38M_3317231.log\r\ntrain_dyn_causal_356M_3372971.log train_tokenizer_batch_size_scaling_16_node_3321526.log\r\ntrain_dyn_causal_356M_3373109.log train_tokenizer_batch_size_scaling_1_node_3318551.log\r\ntrain_dyn_causal_500M_3372936.log train_tokenizer_batch_size_scaling_2_node_3318552.log\r\ntrain_dyn_causal_500M_3372972.log train_tokenizer_batch_size_scaling_2_node_3330806.log\r\ntrain_dyn_causal_500M_3373110.log train_tokenizer_batch_size_scaling_2_node_3330848.log\r\ntrain_dyn_new_arch-bugfixed-spatial-shift_3359343.log train_tokenizer_batch_size_scaling_2_node_3331282.log\r\ntrain_dyn_new_arch-bugfixed-temporal-shift_3359349.log train_tokenizer_batch_size_scaling_4_node_3318553.log\r\ntrain_dyn_yolorun_3333026.log train_tokenizer_batch_size_scaling_4_node_3320175.log\r\ntrain_dyn_yolorun_3333448.log train_tokenizer_batch_size_scaling_4_node_3321524.log\r\ntrain_dyn_yolorun_3335345.log train_tokenizer_batch_size_scaling_8_node_3320176.log\r\ntrain_dyn_yolorun_3335362.log train_tokenizer_batch_size_scaling_8_node_3321525.log\r\ntrain_dyn_yolorun_3348592.log train_tokenizer_minecraft_overfit_sample_3309656.log\r\ntrain_dyn_yolorun_new_arch_3351743.log train_tokenizer_model_size_scaling_127M_3317233.log\r\ntrain_dyn_yolorun_new_arch_3352103.log train_tokenizer_model_size_scaling_127M_3318554.log\r\ntrain_dyn_yolorun_new_arch_3352115.log train_tokenizer_model_size_scaling_140M_3313562.log\r\ntrain_dyn_yolorun_new_arch_3358457.log train_tokenizer_model_size_scaling_140M_3316019.log\r\ntrain_lam_action_space_scaling_10_3320179.log train_tokenizer_model_size_scaling_200M_3313563.log\r\ntrain_lam_action_space_scaling_10_3321529.log train_tokenizer_model_size_scaling_200M_3316020.log\r\ntrain_lam_action_space_scaling_10_3329786.log train_tokenizer_model_size_scaling_227M_3317234.log\r\ntrain_lam_action_space_scaling_10_3329801.log train_tokenizer_model_size_scaling_227M_3318555.log\r\ntrain_lam_action_space_scaling_10_3331283.log train_tokenizer_model_size_scaling_227M_3320173.log\r\ntrain_lam_action_space_scaling_12_3318546.log train_tokenizer_model_size_scaling_227M_3321523.log\r\ntrain_lam_action_space_scaling_12_3320177.log train_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_lam_action_space_scaling_12_3321527.log train_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_lam_action_space_scaling_12_3329787.log train_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_lam_action_space_scaling_12_3329802.log train_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_lam_action_space_scaling_12_3331284.log train_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_lam_action_space_scaling_20_3318547.log train_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_lam_action_space_scaling_20_3329788.log train_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_lam_action_space_scaling_20_3329803.log train_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_lam_action_space_scaling_20_3331285.log train_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_lam_action_space_scaling_50_3320180.log train_tokenizer_model_size_scaling_80M_3316026.log\r\ntrain_lam_action_space_scaling_50_3329789.log yoloruns\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +6990,19767630,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",0,0,"",shellscript,tab +6991,19771984,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",0,0,"",shellscript,tab +6992,19772463,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",225,0,"",shellscript,selection_mouse +6993,19775653,"TERMINAL",0,0,"queue",,terminal_command +6994,19775697,"TERMINAL",0,0,"]633;E;2025-08-14 16:07:08 queue;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C",,terminal_output +6995,19775773,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 16:07:08 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3421973 accelerat train_to tum_cte0 CG 4:05:29\t 2 hkn[0722-0723]3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-22:07:07\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-02:18:12\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3423052 accelerat train_to tum_cte0 R\t0:18\t 2 hkn[0714-0715]3422008 accelerat train_dy tum_cte0 R 3:53:58\t 1 hkn07133421982 accelerat train_to tum_cte0 R 4:05:58\t 4 hkn[0724,0726,0728,0731]3421964 accelerat train_to tum_cte0 R 4:06:22\t 1 hkn0720",,terminal_output +6996,19776778,"TERMINAL",0,0,"9839993",,terminal_output +6997,19777817,"TERMINAL",0,0,"10105214:016:015",,terminal_output +6998,19778927,"TERMINAL",0,0,"2162226",,terminal_output +6999,19779935,"TERMINAL",0,0,"3273337",,terminal_output +7000,19780974,"TERMINAL",0,0,"4384448",,terminal_output +7001,19782204,"TERMINAL",0,0,"5495559",,terminal_output +7002,19783139,"TERMINAL",0,0,"\r652066630",,terminal_output +7003,19783214,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +7004,19785277,"TERMINAL",0,0,"scancel 3421982",,terminal_command +7005,19785305,"TERMINAL",0,0,"]633;E;2025-08-14 16:07:18 scancel 3421982;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +7006,19794913,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",0,0,"",shellscript,tab +7007,19802261,"TERMINAL",0,0,"cd big-runs/",,terminal_command +7008,19803309,"TERMINAL",0,0,"ls",,terminal_command +7009,19806236,"TERMINAL",0,0,"cd ..",,terminal_command +7010,19806272,"TERMINAL",0,0,"]633;E;2025-08-14 16:07:39 cd ..;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +7011,19807852,"TERMINAL",0,0,"cd big_run/",,terminal_command +7012,19808209,"TERMINAL",0,0,"ls",,terminal_command +7013,19808275,"TERMINAL",0,0,"]633;E;2025-08-14 16:07:41 ls;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;Ctokenizer tokenizer-lr-scaling\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run]633;D;0",,terminal_output +7014,19809914,"TERMINAL",0,0,"cd tokenizer",,terminal_command +7015,19809954,"TERMINAL",0,0,"]633;E;2025-08-14 16:07:43 cd tokenizer;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer]633;D;0",,terminal_output +7016,19810230,"TERMINAL",0,0,"ls",,terminal_command +7017,19810317,"TERMINAL",0,0,"]633;E;2025-08-14 16:07:43 ls;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;Ctrain_tokenizer_1e-4_3388151.log train_tokenizer_1e-4_3404607.log train_tokenizer_1e-4_3421973.log train_tokenizer_1e-4_3423052.log\r\ntrain_tokenizer_1e-4_3388153.log train_tokenizer_1e-4_3412401.log train_tokenizer_1e-4_3421982.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer]633;D;0",,terminal_output +7018,19812786,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",0,0,"",shellscript,tab +7019,19816234,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",0,0,"",shellscript,tab +7020,19818645,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes.sbatch",0,0,"",shellscript,tab +7021,19822428,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",0,0,"",shellscript,tab +7022,19823035,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",0,0,"",shellscript,tab +7023,19827086,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",0,0,"",shellscript,tab +7024,19838838,"TERMINAL",0,0,"ls",,terminal_command +7025,19838941,"TERMINAL",0,0,"]633;E;2025-08-14 16:08:12 ls;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;Ctrain_tokenizer_1e-4_3388151.log train_tokenizer_1e-4_3404607.log train_tokenizer_1e-4_3421973.log train_tokenizer_1e-4_3423052.log\r\ntrain_tokenizer_1e-4_3388153.log train_tokenizer_1e-4_3412401.log train_tokenizer_1e-4_3421982.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer]633;D;0",,terminal_output +7026,19841148,"TERMINAL",0,0,"queue",,terminal_command +7027,19841215,"TERMINAL",0,0,"]633;E;2025-08-14 16:08:14 queue;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 16:08:14 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-22:08:13\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-02:19:18\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3423052 accelerat train_to tum_cte0 R\t1:24\t 2 hkn[0714-0715]3422008 accelerat train_dy tum_cte0 R 3:55:04\t 1 hkn07133421964 accelerat train_to tum_cte0 R 4:07:28\t 1 hkn0720",,terminal_output +7028,19842255,"TERMINAL",0,0,"549559",,terminal_output +7029,19843316,"TERMINAL",0,0,"65206630",,terminal_output +7030,19844100,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer]633;D;0",,terminal_output +7031,19849725,"TERMINAL",0,0,"tail -f train_tokenizer_1e-4_3423052.log",,terminal_command +7032,19849759,"TERMINAL",0,0,"]633;E;2025-08-14 16:08:22 tail -f train_tokenizer_1e-4_3423052.log ;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;CWARNING:absl:Missing metrics for step 14000\r\nWARNING:absl:Missing metrics for step 14000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 14000\r\nWARNING:absl:Missing metrics for step 14000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 14000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\r\n",,terminal_output +7033,19855071,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/train_tokenizer_1e-4_3423052.log",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=3421973\n\n# CHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973\n\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n --restore_ckpt \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-2-nodes-$slurm_job_id \\n --tags tokenizer big-run 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pidSLURM_JOB_USER=tum_cte0515\nSLURM_TASKS_PER_NODE=4(x2)\nSLURM_JOB_UID=999226\nSLURM_TASK_PID=3589358\nSLURM_JOB_GPUS=0,1,2,3\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\nSLURMD_NODENAME=hkn0714\nSLURM_JOB_START_TIME=1755180410\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1755353210\nSLURM_CPUS_ON_NODE=24\nSLURM_JOB_CPUS_PER_NODE=24(x2)\nSLURM_GPUS_ON_NODE=4\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=accelerated\nSLURM_TRES_PER_TASK=cpu=5\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=2\nSLURM_JOBID=3423052\nSLURM_JOB_QOS=normal\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=5\nSLURM_JOB_RESERVATION=llmtum\nSLURM_NTASKS=8\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e8.hkn0714\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_NODELIST=hkn[0714-0715]\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=8\nSLURM_NNODES=2\nSLURM_SUBMIT_HOST=hkn1993.localdomain\nSLURM_JOB_ID=3423052\nSLURM_NODEID=0\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_tokenizer_1e-4\nSLURM_NTASKS_PER_NODE=4\nSLURM_JOB_GID=502226\nSLURM_JOB_NODELIST=hkn[0714-0715]\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nwandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\nwandb: Tracking run with wandb version 0.19.11\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250814_160751-3421973\nwandb: Run `wandb offline` to turn off syncing.\nwandb: Resuming run tokenizer-2-nodes-3421973\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/3421973\nWARNING:absl:Missing metrics for step 13000\nWARNING:absl:Missing metrics for step 13000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/013000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 13000\nWARNING:absl:Missing metrics for step 13000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/013000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 13000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/013000/metrics/metrics not found.\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/013000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 13000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/013000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 13000\nWARNING:absl:Missing metrics for step 13000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/013000/metrics/metrics not found.\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/013000/metrics/metrics not found.\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/013000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 12000\nWARNING:absl:Missing metrics for step 12000\nWARNING:absl:Missing metrics for step 12000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/012000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 12000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/012000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 12000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/012000/metrics/metrics not found.\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/012000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 12000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/012000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 12000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/012000/metrics/metrics not found.\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/012000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 14000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 12000\nWARNING:absl:Missing metrics for step 14000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/012000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 14000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 14000\nWARNING:absl:Missing metrics for step 14000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 14000\nWARNING:absl:Missing metrics for step 14000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\nWARNING:absl:Missing metrics for step 14000\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973/014000/metrics/metrics not found.\n",log,tab +7034,19856638,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/train_tokenizer_1e-4_3423052.log",834,0,"",log,selection_mouse +7035,19856644,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer/train_tokenizer_1e-4_3423052.log",833,0,"",log,selection_command +7036,19897819,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +7037,19898727,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +7038,19899744,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +7039,19900788,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +7040,19904762,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +7041,19906778,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['decoder', 'encoder', 'vq']\r\nParameter counts:\r\n{'decoder': 16858736, 'encoder': 16858752, 'vq': 32768, 'total': 33750256}\r\nRestored dataloader and model state from step 14000\r\nStarting training from step 14000...\r\nRunning on 8 devices.\r\nCounting all components: ['decoder', 'encoder', 'vq']\r\nParameter counts:\r\n{'decoder': 16858736, 'encoder': 16858752, 'vq': 32768, 'total': 33750256}\r\nRestored dataloader and model state from step 14000\r\nStarting training from step 14000...\r\nRunning on 8 devices.\r\nCounting all components: ['decoder', 'encoder', 'vq']\r\nParameter counts:\r\n{'decoder': 16858736, 'encoder': 16858752, 'vq': 32768, 'total': 33750256}\r\nRestored dataloader and model state from step 14000\r\nStarting training from step 14000...\r\nRunning on 8 devices.\r\nCounting all components: ['decoder', 'encoder', 'vq']\r\nParameter counts:\r\n{'decoder': 16858736, 'encoder': 16858752, 'vq': 32768, 'total': 33750256}\r\nRestored dataloader and model state from step 14000\r\nStarting training from step 14000...\r\nRunning on 8 devices.\r\nCounting all components: ['decoder', 'encoder', 'vq']\r\nParameter counts:\r\n{'decoder': 16858736, 'encoder': 16858752, 'vq': 32768, 'total': 33750256}\r\nRestored dataloader and model state from step 14000\r\nStarting training from step 14000...\r\nRunning on 8 devices.\r\nCounting all components: ['decoder', 'encoder', 'vq']\r\nParameter counts:\r\n{'decoder': 16858736, 'encoder': 16858752, 'vq': 32768, 'total': 33750256}\r\nRestored dataloader and model state from step 14000\r\nStarting training from step 14000...\r\nRunning on 8 devices.\r\nCounting all components: ['decoder', 'encoder', 'vq']\r\nParameter counts:\r\n{'decoder': 16858736, 'encoder': 16858752, 'vq': 32768, 'total': 33750256}\r\nRestored dataloader and model state from step 14000\r\nStarting training from step 14000...\r\nRunning on 8 devices.\r\nCounting all components: ['decoder', 'encoder', 'vq']\r\nParameter counts:\r\n{'decoder': 16858736, 'encoder': 16858752, 'vq': 32768, 'total': 33750256}\r\nRestored dataloader and model state from step 14000\r\nStarting training from step 14000...\r\n",,terminal_output +7042,19930891,"TERMINAL",0,0,"2025-08-14 16:09:43.161302: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-08-14 16:09:43.161454: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +7043,19932805,"TERMINAL",0,0,"2025-08-14 16:09:45.352547: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +7044,19933860,"TERMINAL",0,0,"2025-08-14 16:09:46.905137: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +7045,20195951,"TERMINAL",0,0,"Step 14000, loss: 0.005812517367303371\r\nStep 14001, loss: 0.005395861808210611\r\nStep 14002, loss: 0.006233721040189266\r\nStep 14003, loss: 0.006271838676184416\r\nStep 14004, loss: 0.0058039287105202675\r\nStep 14005, loss: 0.005527175962924957\r\nStep 14006, loss: 0.0051373583264648914\r\nStep 14007, loss: 0.005339019000530243\r\nStep 14008, loss: 0.005341792479157448\r\nStep 14009, loss: 0.006364726927131414\r\nStep 14010, loss: 0.005711421370506287\r\nStep 14011, loss: 0.005796034820377827\r\nStep 14012, loss: 0.005808668676763773\r\nStep 14013, loss: 0.005712124519050121\r\nStep 14014, loss: 0.005829066503793001\r\nStep 14015, loss: 0.005777937360107899\r\nStep 14016, loss: 0.005655390676110983\r\nStep 14017, loss: 0.005510229617357254\r\nStep 14018, loss: 0.005274210125207901\r\nStep 14019, loss: 0.00648468779399991\r\nStep 14020, loss: 0.005588251631706953\r\nStep 14021, loss: 0.00559698510915041\r\nStep 14022, loss: 0.005708655342459679\r\nStep 14023, loss: 0.0058304802514612675\r\nStep 14024, loss: 0.005690033081918955\r\nStep 14025, loss: 0.005009351763874292\r\nStep 14000, loss: 0.005812517367303371\r\nStep 14001, loss: 0.005395861808210611\r\nStep 14002, loss: 0.006233721040189266\r\nStep 14003, loss: 0.006271838676184416\r\nStep 14004, loss: 0.0058039287105202675\r\nStep 14005, loss: 0.005527175962924957\r\nStep 14006, loss: 0.0051373583264648914\r\nStep 14007, loss: 0.005339019000530243\r\nStep 14008, loss: 0.005341792479157448\r\nStep 14009, loss: 0.006364726927131414\r\nStep 14010, loss: 0.005711421370506287\r\nStep 14011, loss: 0.005796034820377827\r\nStep 14012, loss: 0.005808668676763773\r\nStep 14013, loss: 0.005712124519050121\r\nStep 14014, loss: 0.005829066503793001\r\nStep 14015, loss: 0.005777937360107899\r\nStep 14016, loss: 0.005655390676110983\r\nStep 14017, loss: 0.005510229617357254\r\nStep 14018, loss: 0.005274210125207901\r\nStep 14019, loss: 0.00648468779399991\r\nStep 14020, loss: 0.005588251631706953\r\nStep 14021, loss: 0.00559698510915041\r\nStep 14022, loss: 0.005708655342459679\r\nStep 14023, loss: 0.0058304802514612675\r\nStep 14024, loss: 0.005690033081918955\r\nStep 14025, loss: 0.005009351763874292\r\nStep 14026, loss: 0.0051519679836928844\r\nStep 14027, loss: 0.005798495374619961\r\nStep 14028, loss: 0.005797643680125475\r\nStep 14029, loss: 0.005946979857981205\r\nStep 14030, loss: 0.005064364057034254\r\nStep 14031, loss: 0.005285544320940971\r\nStep 14032, loss: 0.006247749552130699\r\nStep 14033, loss: 0.005247312597930431\r\nStep 14034, loss: 0.0056716445833444595\r\nStep 14035, loss: 0.005997042171657085\r\nStep 14036, loss: 0.005055141169577837\r\nStep 14037, loss: 0.005161302629858255\r\nStep 14038, loss: 0.006230906583368778\r\nStep 14039, loss: 0.005674421321600676\r\nStep 14040, loss: 0.005266168154776096\r\nStep 14041, loss: 0.005765286274254322\r\nStep 14042, loss: 0.005393369123339653\r\nStep 14043, loss: 0.005357636138796806\r\nStep 14044, loss: 0.005934067536145449\r\nStep 14045, loss: 0.0053665414452552795\r\nStep 14046, loss: 0.0053523811511695385\r\nStep 14047, loss: 0.005093598272651434\r\nStep 14048, loss: 0.004940506536513567\r\nStep 14049, loss: 0.005741895642131567\r\nStep 14050, loss: 0.006130511872470379\r\nStep 14051, loss: 0.005829012952744961\r\nStep 14026, loss: 0.0051519679836928844\r\nStep 14027, loss: 0.005798495374619961\r\nStep 14028, loss: 0.005797643680125475\r\nStep 14029, loss: 0.005946979857981205\r\nStep 14030, loss: 0.005064364057034254\r\nStep 14031, loss: 0.005285544320940971\r\nStep 14032, loss: 0.006247749552130699\r\nStep 14033, loss: 0.005247312597930431\r\nStep 14034, loss: 0.0056716445833444595\r\nStep 14035, loss: 0.005997042171657085\r\nStep 14036, loss: 0.005055141169577837\r\nStep 14037, loss: 0.005161302629858255\r\nStep 14038, loss: 0.006230906583368778\r\nStep 14039, loss: 0.005674421321600676\r\nStep 14040, loss: 0.005266168154776096\r\nStep 14041, loss: 0.005765286274254322\r\nStep 14042, loss: 0.005393369123339653\r\nStep 14043, loss: 0.005357636138796806\r\nStep 14044, loss: 0.005934067536145449\r\nStep 14045, loss: 0.0053665414452552795\r\nStep 14046, loss: 0.0053523811511695385\r\nStep 14047, loss: 0.005093598272651434\r\nStep 14048, loss: 0.004940506536513567\r\nStep 14049, loss: 0.005741895642131567\r\nStep 14050, loss: 0.006130511872470379\r\nStep 14051, loss: 0.005829012952744961\r\nStep 14052, loss: 0.0056651271879673\r\nStep 14053, loss: 0.0057511054910719395\r\nStep 14054, loss: 0.005653360392898321\r\nStep 14055, loss: 0.006465739570558071\r\nStep 14056, loss: 0.005385654512792826\r\nStep 14057, loss: 0.005513523705303669\r\nStep 14058, loss: 0.005006593186408281\r\nStep 14059, loss: 0.006068247836083174\r\nStep 14060, loss: 0.005818826146423817\r\nStep 14061, loss: 0.005489379167556763\r\nStep 14062, loss: 0.0056139761582016945\r\nStep 14063, loss: 0.00583872152492404\r\nStep 14064, loss: 0.00543252844363451\r\nStep 14065, loss: 0.005481886677443981\r\nStep 14066, loss: 0.005980655550956726\r\nStep 14067, loss: 0.005767150782048702\r\nStep 14068, loss: 0.005873392801731825\r\nStep 14069, loss: 0.005222206469625235\r\nStep 14070, loss: 0.005088838282972574\r\nStep 14071, loss: 0.005561526864767075\r\nStep 14072, loss: 0.005540770944207907\r\nStep 14073, loss: 0.005777228157967329\r\nStep 14074, loss: 0.005579386837780476\r\nStep 14075, loss: 0.006171700544655323\r\nStep 14076, loss: 0.005738664418458939\r\nStep 14077, loss: 0.005083790980279446\r\nStep 14052, loss: 0.0056651271879673\r\nStep 14053, loss: 0.0057511054910719395\r\nStep 14054, loss: 0.005653360392898321\r\nStep 14055, loss: 0.006465739570558071\r\nStep 14056, loss: 0.005385654512792826\r\nStep 14057, loss: 0.005513523705303669\r\nStep 14058, loss: 0.005006593186408281\r\nStep 14059, loss: 0.006068247836083174\r\nStep 14060, loss: 0.005818826146423817\r\nStep 14061, loss: 0.005489379167556763\r\nStep 14062, loss: 0.0056139761582016945\r\nStep 14063, loss: 0.00583872152492404\r\nStep 14064, loss: 0.00543252844363451\r\nStep 14065, loss: 0.005481886677443981\r\nStep 14066, loss: 0.005980655550956726\r\nStep 14067, loss: 0.005767150782048702\r\nStep 14068, loss: 0.005873392801731825\r\nStep 14069, loss: 0.005222206469625235\r\nStep 14070, loss: 0.005088838282972574\r\nStep 14071, loss: 0.005561526864767075\r\nStep 14072, loss: 0.005540770944207907\r\nStep 14073, loss: 0.005777228157967329\r\nStep 14074, loss: 0.005579386837780476\r\nStep 14075, loss: 0.006171700544655323\r\nStep 14076, loss: 0.005738664418458939\r\nStep 14077, loss: 0.005083790980279446\r\nStep 14078, loss: 0.005408609751611948\r\nStep 14079, loss: 0.00533793680369854\r\nStep 14080, loss: 0.005665861535817385\r\nStep 14081, loss: 0.005766092333942652\r\nStep 14082, loss: 0.005746547598391771\r\nStep 14083, loss: 0.005311197135597467\r\nStep 14084, loss: 0.005504858214408159\r\nStep 14085, loss: 0.0054124463349580765\r\nStep 14086, loss: 0.005440876819193363\r\nStep 14087, loss: 0.005958800204098225\r\nStep 14088, loss: 0.00549651961773634\r\nStep 14089, loss: 0.005657386500388384\r\nStep 14090, loss: 0.006051063537597656\r\nStep 14091, loss: 0.005685704294592142\r\nStep 14092, loss: 0.0051639024168252945\r\nStep 14093, loss: 0.005636001005768776\r\nStep 14094, loss: 0.005396587774157524\r\nStep 14095, loss: 0.005838281940668821\r\nStep 14096, loss: 0.005267083644866943\r\nStep 14097, loss: 0.005533750634640455\r\nStep 14098, loss: 0.005617189686745405\r\nStep 14099, loss: 0.005358037538826466\r\nStep 14100, loss: 0.0055032591335475445\r\nStep 14101, loss: 0.005565322004258633\r\nStep 14102, loss: 0.005121121183037758\r\nStep 14103, loss: 0.005191965028643608\r\nStep 14078, loss: 0.005408609751611948\r\nStep 14079, loss: 0.00533793680369854\r\nStep 14080, loss: 0.005665861535817385\r\nStep 14081, loss: 0.005766092333942652\r\nStep 14082, loss: 0.005746547598391771\r\nStep 14083, loss: 0.005311197135597467\r\nStep 14084, loss: 0.005504858214408159\r\nStep 14085, loss: 0.0054124463349580765\r\nStep 14086, loss: 0.005440876819193363\r\nStep 14087, loss: 0.005958800204098225\r\nStep 14088, loss: 0.00549651961773634\r\nStep 14089, loss: 0.005657386500388384\r\nStep 14090, loss: 0.006051063537597656\r\nStep 14091, loss: 0.005685704294592142\r\nStep 14092, loss: 0.0051639024168252945\r\nStep 14093, loss: 0.005636001005768776\r\nStep 14094, loss: 0.005396587774157524\r\nStep 14095, loss: 0.005838281940668821\r\nStep 14096, loss: 0.005267083644866943\r\nStep 14097, loss: 0.005533750634640455\r\nStep 14098, loss: 0.005617189686745405\r\nStep 14099, loss: 0.005358037538826466\r\nStep 14100, loss: 0.0055032591335475445\r\nStep 14101, loss: 0.005565322004258633\r\nStep 14102, loss: 0.005121121183037758\r\nStep 14103, loss: 0.005191965028643608\r\nStep 14000, loss: 0.005812517367303371\r\nStep 14001, loss: 0.005395861808210611\r\nStep 14002, loss: 0.006233721040189266\r\nStep 14003, loss: 0.006271838676184416\r\nStep 14004, loss: 0.0058039287105202675\r\nStep 14005, loss: 0.005527175962924957\r\nStep 14006, loss: 0.0051373583264648914\r\nStep 14007, loss: 0.005339019000530243\r\nStep 14008, loss: 0.005341792479157448\r\nStep 14009, loss: 0.006364726927131414\r\nStep 14010, loss: 0.005711421370506287\r\nStep 14011, loss: 0.005796034820377827\r\nStep 14012, loss: 0.005808668676763773\r\nStep 14013, loss: 0.005712124519050121\r\nStep 14014, loss: 0.005829066503793001\r\nStep 14015, loss: 0.005777937360107899\r\nStep 14016, loss: 0.005655390676110983\r\nStep 14017, loss: 0.005510229617357254\r\nStep 14018, loss: 0.005274210125207901\r\nStep 14019, loss: 0.00648468779399991\r\nStep 14020, loss: 0.005588251631706953\r\nStep 14021, loss: 0.00559698510915041\r\nStep 14022, loss: 0.005708655342459679\r\nStep 14023, loss: 0.0058304802514612675\r\nStep 14024, loss: 0.005690033081918955\r\nStep 14025, loss: 0.005009351763874292\r\nStep 14000, loss: 0.005812517367303371\r\nStep 14001, loss: 0.005395861808210611\r\nStep 14002, loss: 0.006233721040189266\r\nStep 14003, loss: 0.006271838676184416\r\nStep 14004, loss: 0.0058039287105202675\r\nStep 14005, loss: 0.005527175962924957\r\nStep 14006, loss: 0.0051373583264648914\r\nStep 14007, loss: 0.005339019000530243\r\nStep 14008, loss: 0.005341792479157448\r\nStep 14009, loss: 0.006364726927131414\r\nStep 14010, loss: 0.005711421370506287\r\nStep 14011, loss: 0.005796034820377827\r\nStep 14012, loss: 0.005808668676763773\r\nStep 14013, loss: 0.005712124519050121\r\nStep 14014, loss: 0.005829066503793001\r\nStep 14015, loss: 0.005777937360107899\r\nStep 14016, loss: 0.005655390676110983\r\nStep 14017, loss: 0.005510229617357254\r\nStep 14018, loss: 0.005274210125207901\r\nStep 14019, loss: 0.00648468779399991\r\nStep 14020, loss: 0.005588251631706953\r\nStep 14021, loss: 0.00559698510915041\r\nStep 14022, loss: 0.005708655342459679\r\nStep 14023, loss: 0.0058304802514612675\r\nStep 14024, loss: 0.005690033081918955\r\nStep 14025, loss: 0.005009351763874292\r\nStep 14026, loss: 0.0051519679836928844\r\nStep 14027, loss: 0.005798495374619961\r\nStep 14028, loss: 0.005797643680125475\r\nStep 14029, loss: 0.005946979857981205\r\nStep 14030, loss: 0.005064364057034254\r\nStep 14031, loss: 0.005285544320940971\r\nStep 14032, loss: 0.006247749552130699\r\nStep 14033, loss: 0.005247312597930431\r\nStep 14034, loss: 0.0056716445833444595\r\nStep 14035, loss: 0.005997042171657085\r\nStep 14036, loss: 0.005055141169577837\r\nStep 14037, loss: 0.005161302629858255\r\nStep 14038, loss: 0.006230906583368778\r\nStep 14039, loss: 0.005674421321600676\r\nStep 14040, loss: 0.005266168154776096\r\nStep 14041, loss: 0.005765286274254322\r\nStep 14042, loss: 0.005393369123339653\r\nStep 14043, loss: 0.005357636138796806\r\nStep 14044, loss: 0.005934067536145449\r\nStep 14045, loss: 0.0053665414452552795\r\nStep 14046, loss: 0.0053523811511695385\r\nStep 14047, loss: 0.005093598272651434\r\nStep 14048, loss: 0.004940506536513567\r\nStep 14049, loss: 0.005741895642131567\r\nStep 14050, loss: 0.006130511872470379\r\nStep 14051, loss: 0.005829012952744961\r\nStep 14026, loss: 0.0051519679836928844\r\nStep 14027, loss: 0.005798495374619961\r\nStep 14028, loss: 0.005797643680125475\r\nStep 14029, loss: 0.005946979857981205\r\nStep 14030, loss: 0.005064364057034254\r\nStep 14031, loss: 0.005285544320940971\r\nStep 14032, loss: 0.006247749552130699\r\nStep 14033, loss: 0.005247312597930431\r\nStep 14034, loss: 0.0056716445833444595\r\nStep 14035, loss: 0.005997042171657085\r\nStep 14036, loss: 0.005055141169577837\r\nStep 14037, loss: 0.005161302629858255\r\nStep 14038, loss: 0.006230906583368778\r\nStep 14039, loss: 0.005674421321600676\r\nStep 14040, loss: 0.005266168154776096\r\nStep 14041, loss: 0.005765286274254322\r\nStep 14042, loss: 0.005393369123339653\r\nStep 14043, loss: 0.005357636138796806\r\nStep 14044, loss: 0.005934067536145449\r\nStep 14045, loss: 0.0053665414452552795\r\nStep 14046, loss: 0.0053523811511695385\r\nStep 14047, loss: 0.005093598272651434\r\nStep 14048, loss: 0.004940506536513567\r\nStep 14049, loss: 0.005741895642131567\r\nStep 14050, loss: 0.006130511872470379\r\nStep 14051, loss: 0.005829012952744961\r\nStep 14052, loss: 0.0056651271879673\r\nStep 14053, loss: 0.0057511054910719395\r\nStep 14054, loss: 0.005653360392898321\r\nStep 14055, loss: 0.006465739570558071\r\nStep 14056, loss: 0.005385654512792826\r\nStep 14057, loss: 0.005513523705303669\r\nStep 14058, loss: 0.005006593186408281\r\nStep 14059, loss: 0.006068247836083174\r\nStep 14060, loss: 0.005818826146423817\r\nStep 14061, loss: 0.005489379167556763\r\nStep 14062, loss: 0.0056139761582016945\r\nStep 14063, loss: 0.00583872152492404\r\nStep 14064, loss: 0.00543252844363451\r\nStep 14065, loss: 0.005481886677443981\r\nStep 14066, loss: 0.005980655550956726\r\nStep 14067, loss: 0.005767150782048702\r\nStep 14068, loss: 0.005873392801731825\r\nStep 14069, loss: 0.005222206469625235\r\nStep 14070, loss: 0.005088838282972574\r\nStep 14071, loss: 0.005561526864767075\r\nStep 14072, loss: 0.005540770944207907\r\nStep 14073, loss: 0.005777228157967329\r\nStep 14074, loss: 0.005579386837780476\r\nStep 14075, loss: 0.006171700544655323\r\nStep 14076, loss: 0.005738664418458939\r\nStep 14077, loss: 0.005083790980279446\r\nStep 14052, loss: 0.0056651271879673\r\nStep 14053, loss: 0.0057511054910719395\r\nStep 14054, loss: 0.005653360392898321\r\nStep 14055, loss: 0.006465739570558071\r\nStep 14056, loss: 0.005385654512792826\r\nStep 14057, loss: 0.005513523705303669\r\nStep 14058, loss: 0.005006593186408281\r\nStep 14059, loss: 0.006068247836083174\r\nStep 14060, loss: 0.005818826146423817\r\nStep 14061, loss: 0.005489379167556763\r\nStep 14062, loss: 0.0056139761582016945\r\nStep 14063, loss: 0.00583872152492404\r\nStep 14064, loss: 0.00543252844363451\r\nStep 14065, loss: 0.005481886677443981\r\nStep 14066, loss: 0.005980655550956726\r\nStep 14067, loss: 0.005767150782048702\r\nStep 14068, loss: 0.005873392801731825\r\nStep 14069, loss: 0.005222206469625235\r\nStep 14070, loss: 0.005088838282972574\r\nStep 14071, loss: 0.005561526864767075\r\nStep 14072, loss: 0.005540770944207907\r\nStep 14073, loss: 0.005777228157967329\r\nStep 14074, loss: 0.005579386837780476\r\nStep 14075, loss: 0.006171700544655323\r\nStep 14076, loss: 0.005738664418458939\r\nStep 14077, loss: 0.005083790980279446\r\nStep 14078, loss: 0.005408609751611948\r\nStep 14079, loss: 0.00533793680369854\r\nStep 14080, loss: 0.005665861535817385\r\nStep 14081, loss: 0.005766092333942652\r\nStep 14082, loss: 0.005746547598391771\r\nStep 14083, loss: 0.005311197135597467\r\nStep 14084, loss: 0.005504858214408159\r\nStep 14085, loss: 0.0054124463349580765\r\nStep 14086, loss: 0.005440876819193363\r\nStep 14087, loss: 0.005958800204098225\r\nStep 14088, loss: 0.00549651961773634\r\nStep 14089, loss: 0.005657386500388384\r\nStep 14090, loss: 0.006051063537597656\r\nStep 14091, loss: 0.005685704294592142\r\nStep 14092, loss: 0.0051639024168252945\r\nStep 14093, loss: 0.005636001005768776\r\nStep 14094, loss: 0.005396587774157524\r\nStep 14095, loss: 0.005838281940668821\r\nStep 14096, loss: 0.005267083644866943\r\nStep 14097, loss: 0.005533750634640455\r\nStep 14098, loss: 0.005617189686745405\r\nStep 14099, loss: 0.005358037538826466\r\nStep 14100, loss: 0.0055032591335475445\r\nStep 14101, loss: 0.005565322004258633\r\nStep 14102, loss: 0.005121121183037758\r\nStep 14103, loss: 0.005191965028643608\r\nStep 14078, loss: 0.005408609751611948\r\nStep 14079, loss: 0.00533793680369854\r\nStep 14080, loss: 0.005665861535817385\r\nStep 14081, loss: 0.005766092333942652\r\nStep 14082, loss: 0.005746547598391771\r\nStep 14083, loss: 0.005311197135597467\r\nStep 14084, loss: 0.005504858214408159\r\nStep 14085, loss: 0.0054124463349580765\r\nStep 14086, loss: 0.005440876819193363\r\nStep 14087, loss: 0.005958800204098225\r\nStep 14088, loss: 0.00549651961773634\r\nStep 14089, loss: 0.005657386500388384\r\nStep 14090, loss: 0.006051063537597656\r\nStep 14091, loss: 0.005685704294592142\r\nStep 14092, loss: 0.0051639024168252945\r\nStep 14093, loss: 0.005636001005768776\r\nStep 14094, loss: 0.005396587774157524\r\nStep 14095, loss: 0.005838281940668821\r\nStep 14096, loss: 0.005267083644866943\r\nStep 14097, loss: 0.005533750634640455\r\nStep 14098, loss: 0.005617189686745405\r\nStep 14099, loss: 0.005358037538826466\r\nStep 14100, loss: 0.0055032591335475445\r\nStep 14101, loss: 0.005565322004258633\r\nStep 14102, loss: 0.005121121183037758\r\nStep 14103, loss: 0.005191965028643608\r\nStep 14104, loss: 0.0059740617871284485\r\nStep 14105, loss: 0.005372120067477226\r\nStep 14106, loss: 0.0058049592189490795\r\nStep 14107, loss: 0.0053164539858698845\r\nStep 14108, loss: 0.005402512848377228\r\nStep 14109, loss: 0.006003755610436201\r\nStep 14110, loss: 0.006096096709370613\r\nStep 14111, loss: 0.005075401160866022\r\nStep 14112, loss: 0.0057980772107839584\r\nStep 14113, loss: 0.005073919426649809\r\nStep 14114, loss: 0.006226831581443548\r\nStep 14115, loss: 0.005437750834971666\r\nStep 14116, loss: 0.005633506923913956\r\nStep 14117, loss: 0.005700716748833656\r\nStep 14118, loss: 0.005500595550984144\r\nStep 14119, loss: 0.00533545995131135\r\nStep 14120, loss: 0.005449911113828421\r\nStep 14121, loss: 0.005522486753761768\r\nStep 14122, loss: 0.005895093083381653\r\nStep 14123, loss: 0.005890729371458292\r\nStep 14124, loss: 0.00601503299549222\r\nStep 14125, loss: 0.005567929241806269\r\nStep 14126, loss: 0.005374021362513304\r\nStep 14127, loss: 0.005096671171486378\r\nStep 14128, loss: 0.006076371762901545\r\nStep 14129, loss: 0.005206948611885309\r\nStep 14000, loss: 0.005812517367303371\r\nStep 14001, loss: 0.005395861808210611\r\nStep 14002, loss: 0.006233721040189266\r\nStep 14003, loss: 0.006271838676184416\r\nStep 14004, loss: 0.0058039287105202675\r\nStep 14005, loss: 0.005527175962924957\r\nStep 14006, loss: 0.0051373583264648914\r\nStep 14007, loss: 0.005339019000530243\r\nStep 14008, loss: 0.005341792479157448\r\nStep 14009, loss: 0.006364726927131414\r\nStep 14010, loss: 0.005711421370506287\r\nStep 14011, loss: 0.005796034820377827\r\nStep 14012, loss: 0.005808668676763773\r\nStep 14013, loss: 0.005712124519050121\r\nStep 14014, loss: 0.005829066503793001\r\nStep 14015, loss: 0.005777937360107899\r\nStep 14016, loss: 0.005655390676110983\r\nStep 14017, loss: 0.005510229617357254\r\nStep 14018, loss: 0.005274210125207901\r\nStep 14019, loss: 0.00648468779399991\r\nStep 14020, loss: 0.005588251631706953\r\nStep 14021, loss: 0.00559698510915041\r\nStep 14022, loss: 0.005708655342459679\r\nStep 14023, loss: 0.0058304802514612675\r\nStep 14024, loss: 0.005690033081918955\r\nStep 14025, loss: 0.005009351763874292\r\nStep 14130, loss: 0.0055325026623904705\r\nStep 14131, loss: 0.005312476307153702\r\nStep 14132, loss: 0.0060998909175395966\r\nStep 14133, loss: 0.005691619589924812\r\nStep 14134, loss: 0.005320987198501825\r\nStep 14135, loss: 0.0060443882830441\r\nStep 14136, loss: 0.005440847482532263\r\nStep 14137, loss: 0.005523428786545992\r\nStep 14138, loss: 0.005344294011592865\r\nStep 14139, loss: 0.00552268885076046\r\nStep 14140, loss: 0.005881190299987793\r\nStep 14141, loss: 0.005549816880375147\r\nStep 14142, loss: 0.006060194689780474\r\nStep 14143, loss: 0.005829296540468931\r\nStep 14144, loss: 0.005755137652158737\r\nStep 14145, loss: 0.005770351737737656\r\nStep 14146, loss: 0.0050999876111745834\r\nStep 14147, loss: 0.005218647886067629\r\nStep 14148, loss: 0.005474107339978218\r\nStep 14149, loss: 0.006084137596189976\r\nStep 14150, loss: 0.005109443794935942\r\nStep 14151, loss: 0.005723423324525356\r\nStep 14152, loss: 0.005723710171878338\r\nStep 14153, loss: 0.00552727747708559\r\nStep 14154, loss: 0.004998225253075361\r\nStep 14155, loss: 0.005156924482434988\r\nStep 14026, loss: 0.0051519679836928844\r\nStep 14027, loss: 0.005798495374619961\r\nStep 14028, loss: 0.005797643680125475\r\nStep 14029, loss: 0.005946979857981205\r\nStep 14030, loss: 0.005064364057034254\r\nStep 14031, loss: 0.005285544320940971\r\nStep 14032, loss: 0.006247749552130699\r\nStep 14033, loss: 0.005247312597930431\r\nStep 14034, loss: 0.0056716445833444595\r\nStep 14035, loss: 0.005997042171657085\r\nStep 14036, loss: 0.005055141169577837\r\nStep 14037, loss: 0.005161302629858255\r\nStep 14038, loss: 0.006230906583368778\r\nStep 14039, loss: 0.005674421321600676\r\nStep 14040, loss: 0.005266168154776096\r\nStep 14041, loss: 0.005765286274254322\r\nStep 14042, loss: 0.005393369123339653\r\nStep 14043, loss: 0.005357636138796806\r\nStep 14044, loss: 0.005934067536145449\r\nStep 14045, loss: 0.0053665414452552795\r\nStep 14046, loss: 0.0053523811511695385\r\nStep 14047, loss: 0.005093598272651434\r\nStep 14048, loss: 0.004940506536513567\r\nStep 14049, loss: 0.005741895642131567\r\nStep 14050, loss: 0.006130511872470379\r\nStep 14051, loss: 0.005829012952744961\r\nStep 14156, loss: 0.005859788041561842\r\nStep 14157, loss: 0.005650458391755819\r\nStep 14158, loss: 0.005723326001316309\r\nStep 14159, loss: 0.00611317390576005\r\nStep 14160, loss: 0.005279841832816601\r\nStep 14161, loss: 0.005833033937960863\r\nStep 14162, loss: 0.005306457169353962\r\nStep 14163, loss: 0.0051093194633722305\r\nStep 14164, loss: 0.005190893076360226\r\nStep 14165, loss: 0.006194702349603176\r\nStep 14166, loss: 0.0054787928238511086\r\nStep 14167, loss: 0.00566356536000967\r\nStep 14168, loss: 0.005384392570704222\r\nStep 14169, loss: 0.0052908617071807384\r\nStep 14170, loss: 0.005024982150644064\r\nStep 14171, loss: 0.0057339356280863285\r\nStep 14172, loss: 0.005511219147592783\r\nStep 14173, loss: 0.005600320175290108\r\nStep 14174, loss: 0.00517815537750721\r\nStep 14175, loss: 0.0052972761914134026\r\nStep 14176, loss: 0.005538758356124163\r\nStep 14177, loss: 0.004929340444505215\r\nStep 14178, loss: 0.005274365656077862\r\nStep 14179, loss: 0.005560077261179686\r\nStep 14180, loss: 0.005800329148769379\r\nStep 14181, loss: 0.005059470888227224\r\nStep 14052, loss: 0.0056651271879673\r\nStep 14053, loss: 0.0057511054910719395\r\nStep 14054, loss: 0.005653360392898321\r\nStep 14055, loss: 0.006465739570558071\r\nStep 14056, loss: 0.005385654512792826\r\nStep 14057, loss: 0.005513523705303669\r\nStep 14058, loss: 0.005006593186408281\r\nStep 14059, loss: 0.006068247836083174\r\nStep 14060, loss: 0.005818826146423817\r\nStep 14061, loss: 0.005489379167556763\r\nStep 14062, loss: 0.0056139761582016945\r\nStep 14063, loss: 0.00583872152492404\r\nStep 14064, loss: 0.00543252844363451\r\nStep 14065, loss: 0.005481886677443981\r\nStep 14066, loss: 0.005980655550956726\r\nStep 14067, loss: 0.005767150782048702\r\nStep 14068, loss: 0.005873392801731825\r\nStep 14069, loss: 0.005222206469625235\r\nStep 14070, loss: 0.005088838282972574\r\nStep 14071, loss: 0.005561526864767075\r\nStep 14072, loss: 0.005540770944207907\r\nStep 14073, loss: 0.005777228157967329\r\nStep 14074, loss: 0.005579386837780476\r\nStep 14075, loss: 0.006171700544655323\r\nStep 14076, loss: 0.005738664418458939\r\nStep 14077, loss: 0.005083790980279446\r\nStep 14182, loss: 0.004967685788869858\r\nStep 14183, loss: 0.004787784069776535\r\nStep 14184, loss: 0.005978157743811607\r\nStep 14185, loss: 0.005130748730152845\r\nStep 14186, loss: 0.005504007451236248\r\nStep 14187, loss: 0.005300256423652172\r\nStep 14188, loss: 0.005019632168114185\r\nStep 14189, loss: 0.005699401721358299\r\nStep 14190, loss: 0.005882008001208305\r\nStep 14191, loss: 0.005467270966619253\r\nStep 14192, loss: 0.004971909336745739\r\nStep 14193, loss: 0.005435934290289879\r\nStep 14194, loss: 0.005090567283332348\r\nStep 14195, loss: 0.0049857087433338165\r\nStep 14196, loss: 0.0066054509952664375\r\nStep 14197, loss: 0.005820296239107847\r\nStep 14198, loss: 0.00521948654204607\r\nStep 14199, loss: 0.006478350609540939\r\nStep 14200, loss: 0.006314639933407307\r\nStep 14201, loss: 0.005547133274376392\r\nStep 14202, loss: 0.005370508413761854\r\nStep 14203, loss: 0.00628014886751771\r\nStep 14204, loss: 0.005909750238060951\r\nStep 14205, loss: 0.005536030046641827\r\nStep 14206, loss: 0.006537686102092266\r\nStep 14207, loss: 0.005798934958875179\r\nStep 14078, loss: 0.005408609751611948\r\nStep 14079, loss: 0.00533793680369854\r\nStep 14080, loss: 0.005665861535817385\r\nStep 14081, loss: 0.005766092333942652\r\nStep 14082, loss: 0.005746547598391771\r\nStep 14083, loss: 0.005311197135597467\r\nStep 14084, loss: 0.005504858214408159\r\nStep 14085, loss: 0.0054124463349580765\r\nStep 14086, loss: 0.005440876819193363\r\nStep 14087, loss: 0.005958800204098225\r\nStep 14088, loss: 0.00549651961773634\r\nStep 14089, loss: 0.005657386500388384\r\nStep 14090, loss: 0.006051063537597656\r\nStep 14091, loss: 0.005685704294592142\r\nStep 14092, loss: 0.0051639024168252945\r\nStep 14093, loss: 0.005636001005768776\r\nStep 14094, loss: 0.005396587774157524\r\nStep 14095, loss: 0.005838281940668821\r\nStep 14096, loss: 0.005267083644866943\r\nStep 14097, loss: 0.005533750634640455\r\nStep 14098, loss: 0.005617189686745405\r\nStep 14099, loss: 0.005358037538826466\r\nStep 14100, loss: 0.0055032591335475445\r\nStep 14101, loss: 0.005565322004258633\r\nStep 14102, loss: 0.005121121183037758\r\nStep 14103, loss: 0.005191965028643608\r\nStep 14000, loss: 0.005812517367303371\r\nStep 14001, loss: 0.005395861808210611\r\nStep 14002, loss: 0.006233721040189266\r\nStep 14003, loss: 0.006271838676184416\r\nStep 14004, loss: 0.0058039287105202675\r\nStep 14005, loss: 0.005527175962924957\r\nStep 14006, loss: 0.0051373583264648914\r\nStep 14007, loss: 0.005339019000530243\r\nStep 14008, loss: 0.005341792479157448\r\nStep 14009, loss: 0.006364726927131414\r\nStep 14010, loss: 0.005711421370506287\r\nStep 14011, loss: 0.005796034820377827\r\nStep 14012, loss: 0.005808668676763773\r\nStep 14013, loss: 0.005712124519050121\r\nStep 14014, loss: 0.005829066503793001\r\nStep 14015, loss: 0.005777937360107899\r\nStep 14016, loss: 0.005655390676110983\r\nStep 14017, loss: 0.005510229617357254\r\nStep 14018, loss: 0.005274210125207901\r\nStep 14019, loss: 0.00648468779399991\r\nStep 14020, loss: 0.005588251631706953\r\nStep 14021, loss: 0.00559698510915041\r\nStep 14022, loss: 0.005708655342459679\r\nStep 14023, loss: 0.0058304802514612675\r\nStep 14024, loss: 0.005690033081918955\r\nStep 14025, loss: 0.005009351763874292\r\nStep 14104, loss: 0.0059740617871284485\r\nStep 14105, loss: 0.005372120067477226\r\nStep 14106, loss: 0.0058049592189490795\r\nStep 14107, loss: 0.0053164539858698845\r\nStep 14108, loss: 0.005402512848377228\r\nStep 14109, loss: 0.006003755610436201\r\nStep 14110, loss: 0.006096096709370613\r\nStep 14111, loss: 0.005075401160866022\r\nStep 14112, loss: 0.0057980772107839584\r\nStep 14113, loss: 0.005073919426649809\r\nStep 14114, loss: 0.006226831581443548\r\nStep 14115, loss: 0.005437750834971666\r\nStep 14116, loss: 0.005633506923913956\r\nStep 14117, loss: 0.005700716748833656\r\nStep 14118, loss: 0.005500595550984144\r\nStep 14119, loss: 0.00533545995131135\r\nStep 14120, loss: 0.005449911113828421\r\nStep 14121, loss: 0.005522486753761768\r\nStep 14122, loss: 0.005895093083381653\r\nStep 14123, loss: 0.005890729371458292\r\nStep 14124, loss: 0.00601503299549222\r\nStep 14125, loss: 0.005567929241806269\r\nStep 14126, loss: 0.005374021362513304\r\nStep 14127, loss: 0.005096671171486378\r\nStep 14128, loss: 0.006076371762901545\r\nStep 14129, loss: 0.005206948611885309\r\nStep 14026, loss: 0.0051519679836928844\r\nStep 14027, loss: 0.005798495374619961\r\nStep 14028, loss: 0.005797643680125475\r\nStep 14029, loss: 0.005946979857981205\r\nStep 14030, loss: 0.005064364057034254\r\nStep 14031, loss: 0.005285544320940971\r\nStep 14032, loss: 0.006247749552130699\r\nStep 14033, loss: 0.005247312597930431\r\nStep 14034, loss: 0.0056716445833444595\r\nStep 14035, loss: 0.005997042171657085\r\nStep 14036, loss: 0.005055141169577837\r\nStep 14037, loss: 0.005161302629858255\r\nStep 14038, loss: 0.006230906583368778\r\nStep 14039, loss: 0.005674421321600676\r\nStep 14040, loss: 0.005266168154776096\r\nStep 14041, loss: 0.005765286274254322\r\nStep 14042, loss: 0.005393369123339653\r\nStep 14043, loss: 0.005357636138796806\r\nStep 14044, loss: 0.005934067536145449\r\nStep 14045, loss: 0.0053665414452552795\r\nStep 14046, loss: 0.0053523811511695385\r\nStep 14047, loss: 0.005093598272651434\r\nStep 14048, loss: 0.004940506536513567\r\nStep 14049, loss: 0.005741895642131567\r\nStep 14050, loss: 0.006130511872470379\r\nStep 14051, loss: 0.005829012952744961\r\nStep 14130, loss: 0.0055325026623904705\r\nStep 14131, loss: 0.005312476307153702\r\nStep 14132, loss: 0.0060998909175395966\r\nStep 14133, loss: 0.005691619589924812\r\nStep 14134, loss: 0.005320987198501825\r\nStep 14135, loss: 0.0060443882830441\r\nStep 14136, loss: 0.005440847482532263\r\nStep 14137, loss: 0.005523428786545992\r\nStep 14138, loss: 0.005344294011592865\r\nStep 14139, loss: 0.00552268885076046\r\nStep 14140, loss: 0.005881190299987793\r\nStep 14141, loss: 0.005549816880375147\r\nStep 14142, loss: 0.006060194689780474\r\nStep 14143, loss: 0.005829296540468931\r\nStep 14144, loss: 0.005755137652158737\r\nStep 14145, loss: 0.005770351737737656\r\nStep 14146, loss: 0.0050999876111745834\r\nStep 14147, loss: 0.005218647886067629\r\nStep 14148, loss: 0.005474107339978218\r\nStep 14149, loss: 0.006084137596189976\r\nStep 14150, loss: 0.005109443794935942\r\nStep 14151, loss: 0.005723423324525356\r\nStep 14152, loss: 0.005723710171878338\r\nStep 14153, loss: 0.00552727747708559\r\nStep 14154, loss: 0.004998225253075361\r\nStep 14155, loss: 0.005156924482434988\r\nStep 14052, loss: 0.0056651271879673\r\nStep 14053, loss: 0.0057511054910719395\r\nStep 14054, loss: 0.005653360392898321\r\nStep 14055, loss: 0.006465739570558071\r\nStep 14056, loss: 0.005385654512792826\r\nStep 14057, loss: 0.005513523705303669\r\nStep 14058, loss: 0.005006593186408281\r\nStep 14059, loss: 0.006068247836083174\r\nStep 14060, loss: 0.005818826146423817\r\nStep 14061, loss: 0.005489379167556763\r\nStep 14062, loss: 0.0056139761582016945\r\nStep 14063, loss: 0.00583872152492404\r\nStep 14064, loss: 0.00543252844363451\r\nStep 14065, loss: 0.005481886677443981\r\nStep 14066, loss: 0.005980655550956726\r\nStep 14067, loss: 0.005767150782048702\r\nStep 14068, loss: 0.005873392801731825\r\nStep 14069, loss: 0.005222206469625235\r\nStep 14070, loss: 0.005088838282972574\r\nStep 14071, loss: 0.005561526864767075\r\nStep 14072, loss: 0.005540770944207907\r\nStep 14073, loss: 0.005777228157967329\r\nStep 14074, loss: 0.005579386837780476\r\nStep 14075, loss: 0.006171700544655323\r\nStep 14076, loss: 0.005738664418458939\r\nStep 14077, loss: 0.005083790980279446\r\nStep 14156, loss: 0.005859788041561842\r\nStep 14157, loss: 0.005650458391755819\r\nStep 14158, loss: 0.005723326001316309\r\nStep 14159, loss: 0.00611317390576005\r\nStep 14160, loss: 0.005279841832816601\r\nStep 14161, loss: 0.005833033937960863\r\nStep 14162, loss: 0.005306457169353962\r\nStep 14163, loss: 0.0051093194633722305\r\nStep 14164, loss: 0.005190893076360226\r\nStep 14165, loss: 0.006194702349603176\r\nStep 14166, loss: 0.0054787928238511086\r\nStep 14167, loss: 0.00566356536000967\r\nStep 14168, loss: 0.005384392570704222\r\nStep 14169, loss: 0.0052908617071807384\r\nStep 14170, loss: 0.005024982150644064\r\nStep 14171, loss: 0.0057339356280863285\r\nStep 14172, loss: 0.005511219147592783\r\nStep 14173, loss: 0.005600320175290108\r\nStep 14174, loss: 0.00517815537750721\r\nStep 14175, loss: 0.0052972761914134026\r\nStep 14176, loss: 0.005538758356124163\r\nStep 14177, loss: 0.004929340444505215\r\nStep 14178, loss: 0.005274365656077862\r\nStep 14179, loss: 0.005560077261179686\r\nStep 14180, loss: 0.005800329148769379\r\nStep 14181, loss: 0.005059470888227224\r\nStep 14078, loss: 0.005408609751611948\r\nStep 14079, loss: 0.00533793680369854\r\nStep 14080, loss: 0.005665861535817385\r\nStep 14081, loss: 0.005766092333942652\r\nStep 14082, loss: 0.005746547598391771\r\nStep 14083, loss: 0.005311197135597467\r\nStep 14084, loss: 0.005504858214408159\r\nStep 14085, loss: 0.0054124463349580765\r\nStep 14086, loss: 0.005440876819193363\r\nStep 14087, loss: 0.005958800204098225\r\nStep 14088, loss: 0.00549651961773634\r\nStep 14089, loss: 0.005657386500388384\r\nStep 14090, loss: 0.006051063537597656\r\nStep 14091, loss: 0.005685704294592142\r\nStep 14092, loss: 0.0051639024168252945\r\nStep 14093, loss: 0.005636001005768776\r\nStep 14094, loss: 0.005396587774157524\r\nStep 14095, loss: 0.005838281940668821\r\nStep 14096, loss: 0.005267083644866943\r\nStep 14097, loss: 0.005533750634640455\r\nStep 14098, loss: 0.005617189686745405\r\nStep 14099, loss: 0.005358037538826466\r\nStep 14100, loss: 0.0055032591335475445\r\nStep 14101, loss: 0.005565322004258633\r\nStep 14102, loss: 0.005121121183037758\r\nStep 14103, loss: 0.005191965028643608\r\nStep 14182, loss: 0.004967685788869858\r\nStep 14183, loss: 0.004787784069776535\r\nStep 14184, loss: 0.005978157743811607\r\nStep 14185, loss: 0.005130748730152845\r\nStep 14186, loss: 0.005504007451236248\r\nStep 14187, loss: 0.005300256423652172\r\nStep 14188, loss: 0.005019632168114185\r\nStep 14189, loss: 0.005699401721358299\r\nStep 14190, loss: 0.005882008001208305\r\nStep 14191, loss: 0.005467270966619253\r\nStep 14192, loss: 0.004971909336745739\r\nStep 14193, loss: 0.005435934290289879\r\nStep 14194, loss: 0.005090567283332348\r\nStep 14195, loss: 0.0049857087433338165\r\nStep 14196, loss: 0.0066054509952664375\r\nStep 14197, loss: 0.005820296239107847\r\nStep 14198, loss: 0.00521948654204607\r\nStep 14199, loss: 0.006478350609540939\r\nStep 14200, loss: 0.006314639933407307\r\nStep 14201, loss: 0.005547133274376392\r\nStep 14202, loss: 0.005370508413761854\r\nStep 14203, loss: 0.00628014886751771\r\nStep 14204, loss: 0.005909750238060951\r\nStep 14205, loss: 0.005536030046641827\r\nStep 14206, loss: 0.006537686102092266\r\nStep 14207, loss: 0.005798934958875179\r\nStep 14000, loss: 0.005812517367303371\r\nStep 14001, loss: 0.005395861808210611\r\nStep 14002, loss: 0.006233721040189266\r\nStep 14003, loss: 0.006271838676184416\r\nStep 14004, loss: 0.0058039287105202675\r\nStep 14005, loss: 0.005527175962924957\r\nStep 14006, loss: 0.0051373583264648914\r\nStep 14007, loss: 0.005339019000530243\r\nStep 14008, loss: 0.005341792479157448\r\nStep 14009, loss: 0.006364726927131414\r\nStep 14010, loss: 0.005711421370506287\r\nStep 14011, loss: 0.005796034820377827\r\nStep 14012, loss: 0.005808668676763773\r\nStep 14013, loss: 0.005712124519050121\r\nStep 14014, loss: 0.005829066503793001\r\nStep 14015, loss: 0.005777937360107899\r\nStep 14016, loss: 0.005655390676110983\r\nStep 14017, loss: 0.005510229617357254\r\nStep 14018, loss: 0.005274210125207901\r\nStep 14019, loss: 0.00648468779399991\r\nStep 14020, loss: 0.005588251631706953\r\nStep 14021, loss: 0.00559698510915041\r\nStep 14022, loss: 0.005708655342459679\r\nStep 14023, loss: 0.0058304802514612675\r\nStep 14024, loss: 0.005690033081918955\r\nStep 14025, loss: 0.005009351763874292\r\nStep 14104, loss: 0.0059740617871284485\r\nStep 14105, loss: 0.005372120067477226\r\nStep 14106, loss: 0.0058049592189490795\r\nStep 14107, loss: 0.0053164539858698845\r\nStep 14108, loss: 0.005402512848377228\r\nStep 14109, loss: 0.006003755610436201\r\nStep 14110, loss: 0.006096096709370613\r\nStep 14111, loss: 0.005075401160866022\r\nStep 14112, loss: 0.0057980772107839584\r\nStep 14113, loss: 0.005073919426649809\r\nStep 14114, loss: 0.006226831581443548\r\nStep 14115, loss: 0.005437750834971666\r\nStep 14116, loss: 0.005633506923913956\r\nStep 14117, loss: 0.005700716748833656\r\nStep 14118, loss: 0.005500595550984144\r\nStep 14119, loss: 0.00533545995131135\r\nStep 14120, loss: 0.005449911113828421\r\nStep 14121, loss: 0.005522486753761768\r\nStep 14122, loss: 0.005895093083381653\r\nStep 14123, loss: 0.005890729371458292\r\nStep 14124, loss: 0.00601503299549222\r\nStep 14125, loss: 0.005567929241806269\r\nStep 14126, loss: 0.005374021362513304\r\nStep 14127, loss: 0.005096671171486378\r\nStep 14128, loss: 0.006076371762901545\r\nStep 14129, loss: 0.005206948611885309\r\nStep 14026, loss: 0.0051519679836928844\r\nStep 14027, loss: 0.005798495374619961\r\nStep 14028, loss: 0.005797643680125475\r\nStep 14029, loss: 0.005946979857981205\r\nStep 14030, loss: 0.005064364057034254\r\nStep 14031, loss: 0.005285544320940971\r\nStep 14032, loss: 0.006247749552130699\r\nStep 14033, loss: 0.005247312597930431\r\nStep 14034, loss: 0.0056716445833444595\r\nStep 14035, loss: 0.005997042171657085\r\nStep 14036, loss: 0.005055141169577837\r\nStep 14037, loss: 0.005161302629858255\r\nStep 14038, loss: 0.006230906583368778\r\nStep 14039, loss: 0.005674421321600676\r\nStep 14040, loss: 0.005266168154776096\r\nStep 14041, loss: 0.005765286274254322\r\nStep 14042, loss: 0.005393369123339653\r\nStep 14043, loss: 0.005357636138796806\r\nStep 14044, loss: 0.005934067536145449\r\nStep 14045, loss: 0.0053665414452552795\r\nStep 14046, loss: 0.0053523811511695385\r\nStep 14047, loss: 0.005093598272651434\r\nStep 14048, loss: 0.004940506536513567\r\nStep 14049, loss: 0.005741895642131567\r\nStep 14050, loss: 0.006130511872470379\r\nStep 14051, loss: 0.005829012952744961\r\nStep 14130, loss: 0.0055325026623904705\r\nStep 14131, loss: 0.005312476307153702\r\nStep 14132, loss: 0.0060998909175395966\r\nStep 14133, loss: 0.005691619589924812\r\nStep 14134, loss: 0.005320987198501825\r\nStep 14135, loss: 0.0060443882830441\r\nStep 14136, loss: 0.005440847482532263\r\nStep 14137, loss: 0.005523428786545992\r\nStep 14138, loss: 0.005344294011592865\r\nStep 14139, loss: 0.00552268885076046\r\nStep 14140, loss: 0.005881190299987793\r\nStep 14141, loss: 0.005549816880375147\r\nStep 14142, loss: 0.006060194689780474\r\nStep 14143, loss: 0.005829296540468931\r\nStep 14144, loss: 0.005755137652158737\r\nStep 14145, loss: 0.005770351737737656\r\nStep 14146, loss: 0.0050999876111745834\r\nStep 14147, loss: 0.005218647886067629\r\nStep 14148, loss: 0.005474107339978218\r\nStep 14149, loss: 0.006084137596189976\r\nStep 14150, loss: 0.005109443794935942\r\nStep 14151, loss: 0.005723423324525356\r\nStep 14152, loss: 0.005723710171878338\r\nStep 14153, loss: 0.00552727747708559\r\nStep 14154, loss: 0.004998225253075361\r\nStep 14155, loss: 0.005156924482434988\r\nStep 14052, loss: 0.0056651271879673\r\nStep 14053, loss: 0.0057511054910719395\r\nStep 14054, loss: 0.005653360392898321\r\nStep 14055, loss: 0.006465739570558071\r\nStep 14056, loss: 0.005385654512792826\r\nStep 14057, loss: 0.005513523705303669\r\nStep 14058, loss: 0.005006593186408281\r\nStep 14059, loss: 0.006068247836083174\r\nStep 14060, loss: 0.005818826146423817\r\nStep 14061, loss: 0.005489379167556763\r\nStep 14062, loss: 0.0056139761582016945\r\nStep 14063, loss: 0.00583872152492404\r\nStep 14064, loss: 0.00543252844363451\r\nStep 14065, loss: 0.005481886677443981\r\nStep 14066, loss: 0.005980655550956726\r\nStep 14067, loss: 0.005767150782048702\r\nStep 14068, loss: 0.005873392801731825\r\nStep 14069, loss: 0.005222206469625235\r\nStep 14070, loss: 0.005088838282972574\r\nStep 14071, loss: 0.005561526864767075\r\nStep 14072, loss: 0.005540770944207907\r\nStep 14073, loss: 0.005777228157967329\r\nStep 14074, loss: 0.005579386837780476\r\nStep 14075, loss: 0.006171700544655323\r\nStep 14076, loss: 0.005738664418458939\r\nStep 14077, loss: 0.005083790980279446\r\nStep 14156, loss: 0.005859788041561842\r\nStep 14157, loss: 0.005650458391755819\r\nStep 14158, loss: 0.005723326001316309\r\nStep 14159, loss: 0.00611317390576005\r\nStep 14160, loss: 0.005279841832816601\r\nStep 14161, loss: 0.005833033937960863\r\nStep 14162, loss: 0.005306457169353962\r\nStep 14163, loss: 0.0051093194633722305\r\nStep 14164, loss: 0.005190893076360226\r\nStep 14165, loss: 0.006194702349603176\r\nStep 14166, loss: 0.0054787928238511086\r\nStep 14167, loss: 0.00566356536000967\r\nStep 14168, loss: 0.005384392570704222\r\nStep 14169, loss: 0.0052908617071807384\r\nStep 14170, loss: 0.005024982150644064\r\nStep 14171, loss: 0.0057339356280863285\r\nStep 14172, loss: 0.005511219147592783\r\nStep 14173, loss: 0.005600320175290108\r\nStep 14174, loss: 0.00517815537750721\r\nStep 14175, loss: 0.0052972761914134026\r\nStep 14176, loss: 0.005538758356124163\r\nStep 14177, loss: 0.004929340444505215\r\nStep 14178, loss: 0.005274365656077862\r\nStep 14179, loss: 0.005560077261179686\r\nStep 14180, loss: 0.005800329148769379\r\nStep 14181, loss: 0.005059470888227224\r\nStep 14078, loss: 0.005408609751611948\r\nStep 14079, loss: 0.00533793680369854\r\nStep 14080, loss: 0.005665861535817385\r\nStep 14081, loss: 0.005766092333942652\r\nStep 14082, loss: 0.005746547598391771\r\nStep 14083, loss: 0.005311197135597467\r\nStep 14084, loss: 0.005504858214408159\r\nStep 14085, loss: 0.0054124463349580765\r\nStep 14086, loss: 0.005440876819193363\r\nStep 14087, loss: 0.005958800204098225\r\nStep 14088, loss: 0.00549651961773634\r\nStep 14089, loss: 0.005657386500388384\r\nStep 14090, loss: 0.006051063537597656\r\nStep 14091, loss: 0.005685704294592142\r\nStep 14092, loss: 0.0051639024168252945\r\nStep 14093, loss: 0.005636001005768776\r\nStep 14094, loss: 0.005396587774157524\r\nStep 14095, loss: 0.005838281940668821\r\nStep 14096, loss: 0.005267083644866943\r\nStep 14097, loss: 0.005533750634640455\r\nStep 14098, loss: 0.005617189686745405\r\nStep 14099, loss: 0.005358037538826466\r\nStep 14100, loss: 0.0055032591335475445\r\nStep 14101, loss: 0.005565322004258633\r\nStep 14102, loss: 0.005121121183037758\r\nStep 14103, loss: 0.005191965028643608\r\nStep 14182, loss: 0.004967685788869858\r\nStep 14183, loss: 0.004787784069776535\r\nStep 14184, loss: 0.005978157743811607\r\nStep 14185, loss: 0.005130748730152845\r\nStep 14186, loss: 0.005504007451236248\r\nStep 14187, loss: 0.005300256423652172\r\nStep 14188, loss: 0.005019632168114185\r\nStep 14189, loss: 0.005699401721358299\r\nStep 14190, loss: 0.005882008001208305\r\nStep 14191, loss: 0.005467270966619253\r\nStep 14192, loss: 0.004971909336745739\r\nStep 14193, loss: 0.005435934290289879\r\nStep 14194, loss: 0.005090567283332348\r\nStep 14195, loss: 0.0049857087433338165\r\nStep 14196, loss: 0.0066054509952664375\r\nStep 14197, loss: 0.005820296239107847\r\nStep 14198, loss: 0.00521948654204607\r\nStep 14199, loss: 0.006478350609540939\r\nStep 14200, loss: 0.006314639933407307\r\nStep 14201, loss: 0.005547133274376392\r\nStep 14202, loss: 0.005370508413761854\r\nStep 14203, loss: 0.00628014886751771\r\nStep 14204, loss: 0.005909750238060951\r\nStep 14205, loss: 0.005536030046641827\r\nStep 14206, loss: 0.006537686102092266\r\nStep 14207, loss: 0.005798934958875179\r\nStep 14104, loss: 0.0059740617871284485\r\nStep 14105, loss: 0.005372120067477226\r\nStep 14106, loss: 0.0058049592189490795\r\nStep 14107, loss: 0.0053164539858698845\r\nStep 14108, loss: 0.005402512848377228\r\nStep 14109, loss: 0.006003755610436201\r\nStep 14110, loss: 0.006096096709370613\r\nStep 14111, loss: 0.005075401160866022\r\nStep 14112, loss: 0.0057980772107839584\r\nStep 14113, loss: 0.005073919426649809\r\nStep 14114, loss: 0.006226831581443548\r\nStep 14115, loss: 0.005437750834971666\r\nStep 14116, loss: 0.005633506923913956\r\nStep 14117, loss: 0.005700716748833656\r\nStep 14118, loss: 0.005500595550984144\r\nStep 14119, loss: 0.00533545995131135\r\nStep 14120, loss: 0.005449911113828421\r\nStep 14121, loss: 0.005522486753761768\r\nStep 14122, loss: 0.005895093083381653\r\nStep 14123, loss: 0.005890729371458292\r\nStep 14124, loss: 0.00601503299549222\r\nStep 14125, loss: 0.005567929241806269\r\nStep 14126, loss: 0.005374021362513304\r\nStep 14127, loss: 0.005096671171486378\r\nStep 14128, loss: 0.006076371762901545\r\nStep 14129, loss: 0.005206948611885309\r\nStep 14104, loss: 0.0059740617871284485\r\nStep 14105, loss: 0.005372120067477226\r\nStep 14106, loss: 0.0058049592189490795\r\nStep 14107, loss: 0.0053164539858698845\r\nStep 14108, loss: 0.005402512848377228\r\nStep 14109, loss: 0.006003755610436201\r\nStep 14110, loss: 0.006096096709370613\r\nStep 14111, loss: 0.005075401160866022\r\nStep 14112, loss: 0.0057980772107839584\r\nStep 14113, loss: 0.005073919426649809\r\nStep 14114, loss: 0.006226831581443548\r\nStep 14115, loss: 0.005437750834971666\r\nStep 14116, loss: 0.005633506923913956\r\nStep 14117, loss: 0.005700716748833656\r\nStep 14118, loss: 0.005500595550984144\r\nStep 14119, loss: 0.00533545995131135\r\nStep 14120, loss: 0.005449911113828421\r\nStep 14121, loss: 0.005522486753761768\r\nStep 14122, loss: 0.005895093083381653\r\nStep 14123, loss: 0.005890729371458292\r\nStep 14124, loss: 0.00601503299549222\r\nStep 14125, loss: 0.005567929241806269\r\nStep 14126, loss: 0.005374021362513304\r\nStep 14127, loss: 0.005096671171486378\r\nStep 14128, loss: 0.006076371762901545\r\nStep 14129, loss: 0.005206948611885309\r\nStep 14130, loss: 0.0055325026623904705\r\nStep 14131, loss: 0.005312476307153702\r\nStep 14132, loss: 0.0060998909175395966\r\nStep 14133, loss: 0.005691619589924812\r\nStep 14134, loss: 0.005320987198501825\r\nStep 14135, loss: 0.0060443882830441\r\nStep 14136, loss: 0.005440847482532263\r\nStep 14137, loss: 0.005523428786545992\r\nStep 14138, loss: 0.005344294011592865\r\nStep 14139, loss: 0.00552268885076046\r\nStep 14140, loss: 0.005881190299987793\r\nStep 14141, loss: 0.005549816880375147\r\nStep 14142, loss: 0.006060194689780474\r\nStep 14143, loss: 0.005829296540468931\r\nStep 14144, loss: 0.005755137652158737\r\nStep 14145, loss: 0.005770351737737656\r\nStep 14146, loss: 0.0050999876111745834\r\nStep 14147, loss: 0.005218647886067629\r\nStep 14148, loss: 0.005474107339978218\r\nStep 14149, loss: 0.006084137596189976\r\nStep 14150, loss: 0.005109443794935942\r\nStep 14151, loss: 0.005723423324525356\r\nStep 14152, loss: 0.005723710171878338\r\nStep 14153, loss: 0.00552727747708559\r\nStep 14154, loss: 0.004998225253075361\r\nStep 14155, loss: 0.005156924482434988\r\nStep 14130, loss: 0.0055325026623904705\r\nStep 14131, loss: 0.005312476307153702\r\nStep 14132, loss: 0.0060998909175395966\r\nStep 14133, loss: 0.005691619589924812\r\nStep 14134, loss: 0.005320987198501825\r\nStep 14135, loss: 0.0060443882830441\r\nStep 14136, loss: 0.005440847482532263\r\nStep 14137, loss: 0.005523428786545992\r\nStep 14138, loss: 0.005344294011592865\r\nStep 14139, loss: 0.00552268885076046\r\nStep 14140, loss: 0.005881190299987793\r\nStep 14141, loss: 0.005549816880375147\r\nStep 14142, loss: 0.006060194689780474\r\nStep 14143, loss: 0.005829296540468931\r\nStep 14144, loss: 0.005755137652158737\r\nStep 14145, loss: 0.005770351737737656\r\nStep 14146, loss: 0.0050999876111745834\r\nStep 14147, loss: 0.005218647886067629\r\nStep 14148, loss: 0.005474107339978218\r\nStep 14149, loss: 0.006084137596189976\r\nStep 14150, loss: 0.005109443794935942\r\nStep 14151, loss: 0.005723423324525356\r\nStep 14152, loss: 0.005723710171878338\r\nStep 14153, loss: 0.00552727747708559\r\nStep 14154, loss: 0.004998225253075361\r\nStep 14155, loss: 0.005156924482434988\r\nStep 14156, loss: 0.005859788041561842\r\nStep 14157, loss: 0.005650458391755819\r\nStep 14158, loss: 0.005723326001316309\r\nStep 14159, loss: 0.00611317390576005\r\nStep 14160, loss: 0.005279841832816601\r\nStep 14161, loss: 0.005833033937960863\r\nStep 14162, loss: 0.005306457169353962\r\nStep 14163, loss: 0.0051093194633722305\r\nStep 14164, loss: 0.005190893076360226\r\nStep 14165, loss: 0.006194702349603176\r\nStep 14166, loss: 0.0054787928238511086\r\nStep 14167, loss: 0.00566356536000967\r\nStep 14168, loss: 0.005384392570704222\r\nStep 14169, loss: 0.0052908617071807384\r\nStep 14170, loss: 0.005024982150644064\r\nStep 14171, loss: 0.0057339356280863285\r\nStep 14172, loss: 0.005511219147592783\r\nStep 14173, loss: 0.005600320175290108\r\nStep 14174, loss: 0.00517815537750721\r\nStep 14175, loss: 0.0052972761914134026\r\nStep 14176, loss: 0.005538758356124163\r\nStep 14177, loss: 0.004929340444505215\r\nStep 14178, loss: 0.005274365656077862\r\nStep 14179, loss: 0.005560077261179686\r\nStep 14180, loss: 0.005800329148769379\r\nStep 14181, loss: 0.005059470888227224\r\nStep 14156, loss: 0.005859788041561842\r\nStep 14157, loss: 0.005650458391755819\r\nStep 14158, loss: 0.005723326001316309\r\nStep 14159, loss: 0.00611317390576005\r\nStep 14160, loss: 0.005279841832816601\r\nStep 14161, loss: 0.005833033937960863\r\nStep 14162, loss: 0.005306457169353962\r\nStep 14163, loss: 0.0051093194633722305\r\nStep 14164, loss: 0.005190893076360226\r\nStep 14165, loss: 0.006194702349603176\r\nStep 14166, loss: 0.0054787928238511086\r\nStep 14167, loss: 0.00566356536000967\r\nStep 14168, loss: 0.005384392570704222\r\nStep 14169, loss: 0.0052908617071807384\r\nStep 14170, loss: 0.005024982150644064\r\nStep 14171, loss: 0.0057339356280863285\r\nStep 14172, loss: 0.005511219147592783\r\nStep 14173, loss: 0.005600320175290108\r\nStep 14174, loss: 0.00517815537750721\r\nStep 14175, loss: 0.0052972761914134026\r\nStep 14176, loss: 0.005538758356124163\r\nStep 14177, loss: 0.004929340444505215\r\nStep 14178, loss: 0.005274365656077862\r\nStep 14179, loss: 0.005560077261179686\r\nStep 14180, loss: 0.005800329148769379\r\nStep 14181, loss: 0.005059470888227224\r\nStep 14182, loss: 0.004967685788869858\r\nStep 14183, loss: 0.004787784069776535\r\nStep 14184, loss: 0.005978157743811607\r\nStep 14185, loss: 0.005130748730152845\r\nStep 14186, loss: 0.005504007451236248\r\nStep 14187, loss: 0.005300256423652172\r\nStep 14188, loss: 0.005019632168114185\r\nStep 14189, loss: 0.005699401721358299\r\nStep 14190, loss: 0.005882008001208305\r\nStep 14191, loss: 0.005467270966619253\r\nStep 14192, loss: 0.004971909336745739\r\nStep 14193, loss: 0.005435934290289879\r\nStep 14194, loss: 0.005090567283332348\r\nStep 14195, loss: 0.0049857087433338165\r\nStep 14196, loss: 0.0066054509952664375\r\nStep 14197, loss: 0.005820296239107847\r\nStep 14198, loss: 0.00521948654204607\r\nStep 14199, loss: 0.006478350609540939\r\nStep 14200, loss: 0.006314639933407307\r\nStep 14201, loss: 0.005547133274376392\r\nStep 14202, loss: 0.005370508413761854\r\nStep 14203, loss: 0.00628014886751771\r\nStep 14204, loss: 0.005909750238060951\r\nStep 14205, loss: 0.005536030046641827\r\nStep 14206, loss: 0.006537686102092266\r\nStep 14207, loss: 0.005798934958875179\r\nStep 14182, loss: 0.004967685788869858\r\nStep 14183, loss: 0.004787784069776535\r\nStep 14184, loss: 0.005978157743811607\r\nStep 14185, loss: 0.005130748730152845\r\nStep 14186, loss: 0.005504007451236248\r\nStep 14187, loss: 0.005300256423652172\r\nStep 14188, loss: 0.005019632168114185\r\nStep 14189, loss: 0.005699401721358299\r\nStep 14190, loss: 0.005882008001208305\r\nStep 14191, loss: 0.005467270966619253\r\nStep 14192, loss: 0.004971909336745739\r\nStep 14193, loss: 0.005435934290289879\r\nStep 14194, loss: 0.005090567283332348\r\nStep 14195, loss: 0.0049857087433338165\r\nStep 14196, loss: 0.0066054509952664375\r\nStep 14197, loss: 0.005820296239107847\r\nStep 14198, loss: 0.00521948654204607\r\nStep 14199, loss: 0.006478350609540939\r\nStep 14200, loss: 0.006314639933407307\r\nStep 14201, loss: 0.005547133274376392\r\nStep 14202, loss: 0.005370508413761854\r\nStep 14203, loss: 0.00628014886751771\r\nStep 14204, loss: 0.005909750238060951\r\nStep 14205, loss: 0.005536030046641827\r\nStep 14206, loss: 0.006537686102092266\r\nStep 14207, loss: 0.005798934958875179\r\nStep 14208, loss: 0.005759541876614094\r\nStep 14208, loss: 0.005759541876614094\r\nStep 14104, loss: 0.0059740617871284485\r\nStep 14105, loss: 0.005372120067477226\r\nStep 14106, loss: 0.0058049592189490795\r\nStep 14107, loss: 0.0053164539858698845\r\nStep 14108, loss: 0.005402512848377228\r\nStep 14109, loss: 0.006003755610436201\r\nStep 14110, loss: 0.006096096709370613\r\nStep 14111, loss: 0.005075401160866022\r\nStep 14112, loss: 0.0057980772107839584\r\nStep 14113, loss: 0.005073919426649809\r\nStep 14114, loss: 0.006226831581443548\r\nStep 14115, loss: 0.005437750834971666\r\nStep 14116, loss: 0.005633506923913956\r\nStep 14117, loss: 0.005700716748833656\r\nStep 14118, loss: 0.005500595550984144\r\nStep 14119, loss: 0.00533545995131135\r\nStep 14120, loss: 0.005449911113828421\r\nStep 14121, loss: 0.005522486753761768\r\nStep 14122, loss: 0.005895093083381653\r\nStep 14123, loss: 0.005890729371458292\r\nStep 14124, loss: 0.00601503299549222\r\nStep 14125, loss: 0.005567929241806269\r\nStep 14126, loss: 0.005374021362513304\r\nStep 14127, loss: 0.005096671171486378\r\nStep 14128, loss: 0.006076371762901545\r\nStep 14129, loss: 0.005206948611885309\r\nStep 14208, loss: 0.005759541876614094\r\nStep 14130, loss: 0.0055325026623904705\r\nStep 14131, loss: 0.005312476307153702\r\nStep 14132, loss: 0.0060998909175395966\r\nStep 14133, loss: 0.005691619589924812\r\nStep 14134, loss: 0.005320987198501825\r\nStep 14135, loss: 0.0060443882830441\r\nStep 14136, loss: 0.005440847482532263\r\nStep 14137, loss: 0.005523428786545992\r\nStep 14138, loss: 0.005344294011592865\r\nStep 14139, loss: 0.00552268885076046\r\nStep 14140, loss: 0.005881190299987793\r\nStep 14141, loss: 0.005549816880375147\r\nStep 14142, loss: 0.006060194689780474\r\nStep 14143, loss: 0.005829296540468931\r\nStep 14144, loss: 0.005755137652158737\r\nStep 14145, loss: 0.005770351737737656\r\nStep 14146, loss: 0.0050999876111745834\r\nStep 14147, loss: 0.005218647886067629\r\nStep 14148, loss: 0.005474107339978218\r\nStep 14149, loss: 0.006084137596189976\r\nStep 14150, loss: 0.005109443794935942\r\nStep 14151, loss: 0.005723423324525356\r\nStep 14152, loss: 0.005723710171878338\r\nStep 14153, loss: 0.00552727747708559\r\nStep 14154, loss: 0.004998225253075361\r\nStep 14155, loss: 0.005156924482434988\r\nStep 14208, loss: 0.005759541876614094\r\nStep 14156, loss: 0.005859788041561842\r\nStep 14157, loss: 0.005650458391755819\r\nStep 14158, loss: 0.005723326001316309\r\nStep 14159, loss: 0.00611317390576005\r\nStep 14160, loss: 0.005279841832816601\r\nStep 14161, loss: 0.005833033937960863\r\nStep 14162, loss: 0.005306457169353962\r\nStep 14163, loss: 0.0051093194633722305\r\nStep 14164, loss: 0.005190893076360226\r\nStep 14165, loss: 0.006194702349603176\r\nStep 14166, loss: 0.0054787928238511086\r\nStep 14167, loss: 0.00566356536000967\r\nStep 14168, loss: 0.005384392570704222\r\nStep 14169, loss: 0.0052908617071807384\r\nStep 14170, loss: 0.005024982150644064\r\nStep 14171, loss: 0.0057339356280863285\r\nStep 14172, loss: 0.005511219147592783\r\nStep 14173, loss: 0.005600320175290108\r\nStep 14174, loss: 0.00517815537750721\r\nStep 14175, loss: 0.0052972761914134026\r\nStep 14176, loss: 0.005538758356124163\r\nStep 14177, loss: 0.004929340444505215\r\nStep 14178, loss: 0.005274365656077862\r\nStep 14179, loss: 0.005560077261179686\r\nStep 14180, loss: 0.005800329148769379\r\nStep 14181, loss: 0.005059470888227224\r\nStep 14000, loss: 0.005812517367303371\r\nStep 14001, loss: 0.005395861808210611\r\nStep 14002, loss: 0.006233721040189266\r\nStep 14003, loss: 0.006271838676184416\r\nStep 14004, loss: 0.0058039287105202675\r\nStep 14005, loss: 0.005527175962924957\r\nStep 14006, loss: 0.0051373583264648914\r\nStep 14007, loss: 0.005339019000530243\r\nStep 14008, loss: 0.005341792479157448\r\nStep 14009, loss: 0.006364726927131414\r\nStep 14010, loss: 0.005711421370506287\r\nStep 14011, loss: 0.005796034820377827\r\nStep 14012, loss: 0.005808668676763773\r\nStep 14013, loss: 0.005712124519050121\r\nStep 14014, loss: 0.005829066503793001\r\nStep 14015, loss: 0.005777937360107899\r\nStep 14016, loss: 0.005655390676110983\r\nStep 14017, loss: 0.005510229617357254\r\nStep 14018, loss: 0.005274210125207901\r\nStep 14019, loss: 0.00648468779399991\r\nStep 14020, loss: 0.005588251631706953\r\nStep 14021, loss: 0.00559698510915041\r\nStep 14022, loss: 0.005708655342459679\r\nStep 14023, loss: 0.0058304802514612675\r\nStep 14024, loss: 0.005690033081918955\r\nStep 14025, loss: 0.005009351763874292\r\nStep 14182, loss: 0.004967685788869858\r\nStep 14183, loss: 0.004787784069776535\r\nStep 14184, loss: 0.005978157743811607\r\nStep 14185, loss: 0.005130748730152845\r\nStep 14186, loss: 0.005504007451236248\r\nStep 14187, loss: 0.005300256423652172\r\nStep 14188, loss: 0.005019632168114185\r\nStep 14189, loss: 0.005699401721358299\r\nStep 14190, loss: 0.005882008001208305\r\nStep 14191, loss: 0.005467270966619253\r\nStep 14192, loss: 0.004971909336745739\r\nStep 14193, loss: 0.005435934290289879\r\nStep 14194, loss: 0.005090567283332348\r\nStep 14195, loss: 0.0049857087433338165\r\nStep 14196, loss: 0.0066054509952664375\r\nStep 14197, loss: 0.005820296239107847\r\nStep 14198, loss: 0.00521948654204607\r\nStep 14199, loss: 0.006478350609540939\r\nStep 14200, loss: 0.006314639933407307\r\nStep 14201, loss: 0.005547133274376392\r\nStep 14202, loss: 0.005370508413761854\r\nStep 14203, loss: 0.00628014886751771\r\nStep 14204, loss: 0.005909750238060951\r\nStep 14205, loss: 0.005536030046641827\r\nStep 14206, loss: 0.006537686102092266\r\nStep 14207, loss: 0.005798934958875179\r\nStep 14026, loss: 0.0051519679836928844\r\nStep 14027, loss: 0.005798495374619961\r\nStep 14028, loss: 0.005797643680125475\r\nStep 14029, loss: 0.005946979857981205\r\nStep 14030, loss: 0.005064364057034254\r\nStep 14031, loss: 0.005285544320940971\r\nStep 14032, loss: 0.006247749552130699\r\nStep 14033, loss: 0.005247312597930431\r\nStep 14034, loss: 0.0056716445833444595\r\nStep 14035, loss: 0.005997042171657085\r\nStep 14036, loss: 0.005055141169577837\r\nStep 14037, loss: 0.005161302629858255\r\nStep 14038, loss: 0.006230906583368778\r\nStep 14039, loss: 0.005674421321600676\r\nStep 14040, loss: 0.005266168154776096\r\nStep 14041, loss: 0.005765286274254322\r\nStep 14042, loss: 0.005393369123339653\r\nStep 14043, loss: 0.005357636138796806\r\nStep 14044, loss: 0.005934067536145449\r\nStep 14045, loss: 0.0053665414452552795\r\nStep 14046, loss: 0.0053523811511695385\r\nStep 14047, loss: 0.005093598272651434\r\nStep 14048, loss: 0.004940506536513567\r\nStep 14049, loss: 0.005741895642131567\r\nStep 14050, loss: 0.006130511872470379\r\nStep 14051, loss: 0.005829012952744961\r\nStep 14104, loss: 0.0059740617871284485\r\nStep 14105, loss: 0.005372120067477226\r\nStep 14106, loss: 0.0058049592189490795\r\nStep 14107, loss: 0.0053164539858698845\r\nStep 14108, loss: 0.005402512848377228\r\nStep 14109, loss: 0.006003755610436201\r\nStep 14110, loss: 0.006096096709370613\r\nStep 14111, loss: 0.005075401160866022\r\nStep 14112, loss: 0.0057980772107839584\r\nStep 14113, loss: 0.005073919426649809\r\nStep 14114, loss: 0.006226831581443548\r\nStep 14115, loss: 0.005437750834971666\r\nStep 14116, loss: 0.005633506923913956\r\nStep 14117, loss: 0.005700716748833656\r\nStep 14118, loss: 0.005500595550984144\r\nStep 14119, loss: 0.00533545995131135\r\nStep 14120, loss: 0.005449911113828421\r\nStep 14121, loss: 0.005522486753761768\r\nStep 14122, loss: 0.005895093083381653\r\nStep 14123, loss: 0.005890729371458292\r\nStep 14124, loss: 0.00601503299549222\r\nStep 14125, loss: 0.005567929241806269\r\nStep 14126, loss: 0.005374021362513304\r\nStep 14127, loss: 0.005096671171486378\r\nStep 14128, loss: 0.006076371762901545\r\nStep 14129, loss: 0.005206948611885309\r\nStep 14052, loss: 0.0056651271879673\r\nStep 14053, loss: 0.0057511054910719395\r\nStep 14054, loss: 0.005653360392898321\r\nStep 14055, loss: 0.006465739570558071\r\nStep 14056, loss: 0.005385654512792826\r\nStep 14057, loss: 0.005513523705303669\r\nStep 14058, loss: 0.005006593186408281\r\nStep 14059, loss: 0.006068247836083174\r\nStep 14060, loss: 0.005818826146423817\r\nStep 14061, loss: 0.005489379167556763\r\nStep 14062, loss: 0.0056139761582016945\r\nStep 14063, loss: 0.00583872152492404\r\nStep 14064, loss: 0.00543252844363451\r\nStep 14065, loss: 0.005481886677443981\r\nStep 14066, loss: 0.005980655550956726\r\nStep 14067, loss: 0.005767150782048702\r\nStep 14068, loss: 0.005873392801731825\r\nStep 14069, loss: 0.005222206469625235\r\nStep 14070, loss: 0.005088838282972574\r\nStep 14071, loss: 0.005561526864767075\r\nStep 14072, loss: 0.005540770944207907\r\nStep 14073, loss: 0.005777228157967329\r\nStep 14074, loss: 0.005579386837780476\r\nStep 14075, loss: 0.006171700544655323\r\nStep 14076, loss: 0.005738664418458939\r\nStep 14077, loss: 0.005083790980279446\r\nStep 14130, loss: 0.0055325026623904705\r\nStep 14131, loss: 0.005312476307153702\r\nStep 14132, loss: 0.0060998909175395966\r\nStep 14133, loss: 0.005691619589924812\r\nStep 14134, loss: 0.005320987198501825\r\nStep 14135, loss: 0.0060443882830441\r\nStep 14136, loss: 0.005440847482532263\r\nStep 14137, loss: 0.005523428786545992\r\nStep 14138, loss: 0.005344294011592865\r\nStep 14139, loss: 0.00552268885076046\r\nStep 14140, loss: 0.005881190299987793\r\nStep 14141, loss: 0.005549816880375147\r\nStep 14142, loss: 0.006060194689780474\r\nStep 14143, loss: 0.005829296540468931\r\nStep 14144, loss: 0.005755137652158737\r\nStep 14145, loss: 0.005770351737737656\r\nStep 14146, loss: 0.0050999876111745834\r\nStep 14147, loss: 0.005218647886067629\r\nStep 14148, loss: 0.005474107339978218\r\nStep 14149, loss: 0.006084137596189976\r\nStep 14150, loss: 0.005109443794935942\r\nStep 14151, loss: 0.005723423324525356\r\nStep 14152, loss: 0.005723710171878338\r\nStep 14153, loss: 0.00552727747708559\r\nStep 14154, loss: 0.004998225253075361\r\nStep 14155, loss: 0.005156924482434988\r\nStep 14078, loss: 0.005408609751611948\r\nStep 14079, loss: 0.00533793680369854\r\nStep 14080, loss: 0.005665861535817385\r\nStep 14081, loss: 0.005766092333942652\r\nStep 14082, loss: 0.005746547598391771\r\nStep 14083, loss: 0.005311197135597467\r\nStep 14084, loss: 0.005504858214408159\r\nStep 14085, loss: 0.0054124463349580765\r\nStep 14086, loss: 0.005440876819193363\r\nStep 14087, loss: 0.005958800204098225\r\nStep 14088, loss: 0.00549651961773634\r\nStep 14089, loss: 0.005657386500388384\r\nStep 14090, loss: 0.006051063537597656\r\nStep 14091, loss: 0.005685704294592142\r\nStep 14092, loss: 0.0051639024168252945\r\nStep 14093, loss: 0.005636001005768776\r\nStep 14094, loss: 0.005396587774157524\r\nStep 14095, loss: 0.005838281940668821\r\nStep 14096, loss: 0.005267083644866943\r\nStep 14097, loss: 0.005533750634640455\r\nStep 14098, loss: 0.005617189686745405\r\nStep 14099, loss: 0.005358037538826466\r\nStep 14100, loss: 0.0055032591335475445\r\nStep 14101, loss: 0.005565322004258633\r\nStep 14102, loss: 0.005121121183037758\r\nStep 14103, loss: 0.005191965028643608\r\nStep 14156, loss: 0.005859788041561842\r\nStep 14157, loss: 0.005650458391755819\r\nStep 14158, loss: 0.005723326001316309\r\nStep 14159, loss: 0.00611317390576005\r\nStep 14160, loss: 0.005279841832816601\r\nStep 14161, loss: 0.005833033937960863\r\nStep 14162, loss: 0.005306457169353962\r\nStep 14163, loss: 0.0051093194633722305\r\nStep 14164, loss: 0.005190893076360226\r\nStep 14165, loss: 0.006194702349603176\r\nStep 14166, loss: 0.0054787928238511086\r\nStep 14167, loss: 0.00566356536000967\r\nStep 14168, loss: 0.005384392570704222\r\nStep 14169, loss: 0.0052908617071807384\r\nStep 14170, loss: 0.005024982150644064\r\nStep 14171, loss: 0.0057339356280863285\r\nStep 14172, loss: 0.005511219147592783\r\nStep 14173, loss: 0.005600320175290108\r\nStep 14174, loss: 0.00517815537750721\r\nStep 14175, loss: 0.0052972761914134026\r\nStep 14176, loss: 0.005538758356124163\r\nStep 14177, loss: 0.004929340444505215\r\nStep 14178, loss: 0.005274365656077862\r\nStep 14179, loss: 0.005560077261179686\r\nStep 14180, loss: 0.005800329148769379\r\nStep 14181, loss: 0.005059470888227224\r\nStep 14104, loss: 0.0059740617871284485\r\nStep 14105, loss: 0.005372120067477226\r\nStep 14106, loss: 0.0058049592189490795\r\nStep 14107, loss: 0.0053164539858698845\r\nStep 14108, loss: 0.005402512848377228\r\nStep 14109, loss: 0.006003755610436201\r\nStep 14110, loss: 0.006096096709370613\r\nStep 14111, loss: 0.005075401160866022\r\nStep 14112, loss: 0.0057980772107839584\r\nStep 14113, loss: 0.005073919426649809\r\nStep 14114, loss: 0.006226831581443548\r\nStep 14115, loss: 0.005437750834971666\r\nStep 14116, loss: 0.005633506923913956\r\nStep 14117, loss: 0.005700716748833656\r\nStep 14118, loss: 0.005500595550984144\r\nStep 14119, loss: 0.00533545995131135\r\nStep 14120, loss: 0.005449911113828421\r\nStep 14121, loss: 0.005522486753761768\r\nStep 14122, loss: 0.005895093083381653\r\nStep 14123, loss: 0.005890729371458292\r\nStep 14124, loss: 0.00601503299549222\r\nStep 14125, loss: 0.005567929241806269\r\nStep 14126, loss: 0.005374021362513304\r\nStep 14127, loss: 0.005096671171486378\r\nStep 14128, loss: 0.006076371762901545\r\nStep 14129, loss: 0.005206948611885309\r\nStep 14182, loss: 0.004967685788869858\r\nStep 14183, loss: 0.004787784069776535\r\nStep 14184, loss: 0.005978157743811607\r\nStep 14185, loss: 0.005130748730152845\r\nStep 14186, loss: 0.005504007451236248\r\nStep 14187, loss: 0.005300256423652172\r\nStep 14188, loss: 0.005019632168114185\r\nStep 14189, loss: 0.005699401721358299\r\nStep 14190, loss: 0.005882008001208305\r\nStep 14191, loss: 0.005467270966619253\r\nStep 14192, loss: 0.004971909336745739\r\nStep 14193, loss: 0.005435934290289879\r\nStep 14194, loss: 0.005090567283332348\r\nStep 14195, loss: 0.0049857087433338165\r\nStep 14196, loss: 0.0066054509952664375\r\nStep 14197, loss: 0.005820296239107847\r\nStep 14198, loss: 0.00521948654204607\r\nStep 14199, loss: 0.006478350609540939\r\nStep 14200, loss: 0.006314639933407307\r\nStep 14201, loss: 0.005547133274376392\r\nStep 14202, loss: 0.005370508413761854\r\nStep 14203, loss: 0.00628014886751771\r\nStep 14204, loss: 0.005909750238060951\r\nStep 14205, loss: 0.005536030046641827\r\nStep 14206, loss: 0.006537686102092266\r\nStep 14207, loss: 0.005798934958875179\r\nStep 14130, loss: 0.0055325026623904705\r\nStep 14131, loss: 0.005312476307153702\r\nStep 14132, loss: 0.0060998909175395966\r\nStep 14133, loss: 0.005691619589924812\r\nStep 14134, loss: 0.005320987198501825\r\nStep 14135, loss: 0.0060443882830441\r\nStep 14136, loss: 0.005440847482532263\r\nStep 14137, loss: 0.005523428786545992\r\nStep 14138, loss: 0.005344294011592865\r\nStep 14139, loss: 0.00552268885076046\r\nStep 14140, loss: 0.005881190299987793\r\nStep 14141, loss: 0.005549816880375147\r\nStep 14142, loss: 0.006060194689780474\r\nStep 14143, loss: 0.005829296540468931\r\nStep 14144, loss: 0.005755137652158737\r\nStep 14145, loss: 0.005770351737737656\r\nStep 14146, loss: 0.0050999876111745834\r\nStep 14147, loss: 0.005218647886067629\r\nStep 14148, loss: 0.005474107339978218\r\nStep 14149, loss: 0.006084137596189976\r\nStep 14150, loss: 0.005109443794935942\r\nStep 14151, loss: 0.005723423324525356\r\nStep 14152, loss: 0.005723710171878338\r\nStep 14153, loss: 0.00552727747708559\r\nStep 14154, loss: 0.004998225253075361\r\nStep 14155, loss: 0.005156924482434988\r\nStep 14208, loss: 0.005759541876614094\r\nStep 14156, loss: 0.005859788041561842\r\nStep 14157, loss: 0.005650458391755819\r\nStep 14158, loss: 0.005723326001316309\r\nStep 14159, loss: 0.00611317390576005\r\nStep 14160, loss: 0.005279841832816601\r\nStep 14161, loss: 0.005833033937960863\r\nStep 14162, loss: 0.005306457169353962\r\nStep 14163, loss: 0.0051093194633722305\r\nStep 14164, loss: 0.005190893076360226\r\nStep 14165, loss: 0.006194702349603176\r\nStep 14166, loss: 0.0054787928238511086\r\nStep 14167, loss: 0.00566356536000967\r\nStep 14168, loss: 0.005384392570704222\r\nStep 14169, loss: 0.0052908617071807384\r\nStep 14170, loss: 0.005024982150644064\r\nStep 14171, loss: 0.0057339356280863285\r\nStep 14172, loss: 0.005511219147592783\r\nStep 14173, loss: 0.005600320175290108\r\nStep 14174, loss: 0.00517815537750721\r\nStep 14175, loss: 0.0052972761914134026\r\nStep 14176, loss: 0.005538758356124163\r\nStep 14177, loss: 0.004929340444505215\r\nStep 14178, loss: 0.005274365656077862\r\nStep 14179, loss: 0.005560077261179686\r\nStep 14180, loss: 0.005800329148769379\r\nStep 14181, loss: 0.005059470888227224\r\nStep 14208, loss: 0.005759541876614094\r\nStep 14182, loss: 0.004967685788869858\r\nStep 14183, loss: 0.004787784069776535\r\nStep 14184, loss: 0.005978157743811607\r\nStep 14185, loss: 0.005130748730152845\r\nStep 14186, loss: 0.005504007451236248\r\nStep 14187, loss: 0.005300256423652172\r\nStep 14188, loss: 0.005019632168114185\r\nStep 14189, loss: 0.005699401721358299\r\nStep 14190, loss: 0.005882008001208305\r\nStep 14191, loss: 0.005467270966619253\r\nStep 14192, loss: 0.004971909336745739\r\nStep 14193, loss: 0.005435934290289879\r\nStep 14194, loss: 0.005090567283332348\r\nStep 14195, loss: 0.0049857087433338165\r\nStep 14196, loss: 0.0066054509952664375\r\nStep 14197, loss: 0.005820296239107847\r\nStep 14198, loss: 0.00521948654204607\r\nStep 14199, loss: 0.006478350609540939\r\nStep 14200, loss: 0.006314639933407307\r\nStep 14201, loss: 0.005547133274376392\r\nStep 14202, loss: 0.005370508413761854\r\nStep 14203, loss: 0.00628014886751771\r\nStep 14204, loss: 0.005909750238060951\r\nStep 14205, loss: 0.005536030046641827\r\nStep 14206, loss: 0.006537686102092266\r\nStep 14207, loss: 0.005798934958875179\r\nStep 14208, loss: 0.005759541876614094\r\nStep 14208, loss: 0.005759541876614094\r\n",,terminal_output +7046,20220916,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer]633;D;130",,terminal_output +7047,20232066,"TERMINAL",0,0,"queue",,terminal_command +7048,20232163,"TERMINAL",0,0,"]633;E;2025-08-14 16:14:45 queue;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 16:14:45 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3412401 accelerat train_to tum_cte0 R 1-22:14:44\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-02:25:49\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3423052 accelerat train_to tum_cte0 R\t7:55\t 2 hkn[0714-0715]3422008 accelerat train_dy tum_cte0 R 4:01:35\t 1 hkn07133421964 accelerat train_to tum_cte0 R 4:13:59\t 1 hkn0720",,terminal_output +7049,20233278,"TERMINAL",0,0,"6550664:00",,terminal_output +7050,20234304,"TERMINAL",0,0,"761771",,terminal_output +7051,20235326,"TERMINAL",0,0,"872882",,terminal_output +7052,20236399,"TERMINAL",0,0,"983993",,terminal_output +7053,20236630,"TERMINAL",0,0,"bash",,terminal_focus +7054,20237408,"TERMINAL",0,0,"50948:00404",,terminal_output +7055,20238409,"TERMINAL",0,0,"1505115",,terminal_output +7056,20239452,"TERMINAL",0,0,"216226",,terminal_output +7057,20240603,"TERMINAL",0,0,"327337",,terminal_output +7058,20240616,"TERMINAL",0,0,"scancel 3423052",,terminal_command +7059,20240641,"TERMINAL",0,0,"]633;E;2025-08-14 16:14:53 scancel 3423052;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +7060,20241538,"TERMINAL",0,0,"M43423052 accelerat train_to tum_cte0 CG\t8:03\t 2 hkn[0714-0715]3848",,terminal_output +7061,20241886,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",0,0,"",shellscript,tab +7062,20242596,"TERMINAL",0,0,"54959",,terminal_output +7063,20243690,"TERMINAL",0,0,"656:00610",,terminal_output +7064,20243975,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",0,0,"",shellscript,tab +7065,20244169,"TERMINAL",0,0,"Every 1.0s: squeu... hkn1993.localdomain: Thu Aug 14 16:14:57 2025JOBID PARTITION NAME USER ST\tTIME NODESNODELIST(REASON)3423052 accelerat train_to tum_cte0 CG\t8:03\t 2hkn[0714-0715]3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1(Priority)3412401 accelerat train_to tum_cte0 R 1-22:14:56\t 8hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-02:26:01\t 8hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 4:01:47\t 1hkn07133421964 accelerat train_to tum_cte0 R 4:14:11\t 1hkn0720",,terminal_output +7066,20245244,"TERMINAL",0,0,"87282",,terminal_output +7067,20245594,"TERMINAL",0,0,"watch",,terminal_focus +7068,20246250,"TERMINAL",0,0,"98393",,terminal_output +7069,20247284,"TERMINAL",0,0,"5:0094504",,terminal_output +7070,20247871,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",0,0,"",shellscript,tab +7071,20248416,"TERMINAL",0,0,"15:00515",,terminal_output +7072,20249421,"TERMINAL",0,0,"21626",,terminal_output +7073,20250481,"TERMINAL",0,0,"32737",,terminal_output +7074,20251488,"TERMINAL",0,0,"43848",,terminal_output +7075,20252580,"TERMINAL",0,0,"54959",,terminal_output +7076,20253617,"TERMINAL",0,0,"6510620",,terminal_output +7077,20254309,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",0,0,"",shellscript,tab +7078,20254310,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",461,0,"",shellscript,selection_mouse +7079,20254657,"TERMINAL",0,0,"76171",,terminal_output +7080,20255378,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",460,0,"",shellscript,selection_command +7081,20255707,"TERMINAL",0,0,"87282",,terminal_output +7082,20256375,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",443,29,"",shellscript,content +7083,20256767,"TERMINAL",0,0,"98393",,terminal_output +7084,20257818,"TERMINAL",0,0,"10942:004",,terminal_output +7085,20258303,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",0,0,"",shellscript,tab +7086,20258304,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",466,0,"",shellscript,selection_mouse +7087,20258848,"TERMINAL",0,0,"111626",,terminal_output +7088,20259167,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",443,29,"",shellscript,content +7089,20259883,"TERMINAL",0,0,"32737",,terminal_output +7090,20260948,"TERMINAL",0,0,"43848",,terminal_output +7091,20261972,"TERMINAL",0,0,"54959",,terminal_output +7092,20262999,"TERMINAL",0,0,"6520630",,terminal_output +7093,20264018,"TERMINAL",0,0,"76171",,terminal_output +7094,20265092,"TERMINAL",0,0,"87282",,terminal_output +7095,20266112,"TERMINAL",0,0,"98393",,terminal_output +7096,20267173,"TERMINAL",0,0,"2094104",,terminal_output +7097,20267898,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",0,0,"",shellscript,tab +7098,20267899,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1788,0,"",shellscript,selection_mouse +7099,20268184,"TERMINAL",0,0,"120515",,terminal_output +7100,20269324,"TERMINAL",0,0,"21626",,terminal_output +7101,20270391,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",0,0,"",shellscript,tab +7102,20270392,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1675,0,"",shellscript,selection_mouse +7103,20270426,"TERMINAL",0,0,"32737",,terminal_output +7104,20271206,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1680,0,"\n --wandb_id $slurm_job_id \",shellscript,content +7105,20271259,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1685,0,"",shellscript,selection_command +7106,20271338,"TERMINAL",0,0,"43848",,terminal_output +7107,20271977,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1675,0,"",shellscript,selection_mouse +7108,20272366,"TERMINAL",0,0,"54959",,terminal_output +7109,20272955,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1650,31,"",shellscript,content +7110,20273003,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1654,0,"",shellscript,selection_command +7111,20273408,"TERMINAL",0,0,"6530640",,terminal_output +7112,20274512,"TERMINAL",0,0,"76171",,terminal_output +7113,20275485,"TERMINAL",0,0,"87282",,terminal_output +7114,20275842,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",0,0,"",shellscript,tab +7115,20275843,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1759,0,"",shellscript,selection_mouse +7116,20276523,"TERMINAL",0,0,"98393",,terminal_output +7117,20277620,"TERMINAL",0,0,"3094204",,terminal_output +7118,20278612,"TERMINAL",0,0,"130515",,terminal_output +7119,20279603,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",0,0,"",shellscript,tab +7120,20279604,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1636,0,"",shellscript,selection_mouse +7121,20279667,"TERMINAL",0,0,"21626",,terminal_output +7122,20280304,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1649,0,"\n --restore_ckpt \",shellscript,content +7123,20280316,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1654,0,"",shellscript,selection_command +7124,20280700,"TERMINAL",0,0,"32737",,terminal_output +7125,20281262,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1629,0,"",shellscript,selection_command +7126,20281776,"TERMINAL",0,0,"43848",,terminal_output +7127,20282800,"TERMINAL",0,0,"54959",,terminal_output +7128,20283826,"TERMINAL",0,0,"6641751",,terminal_output +7129,20284557,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1625,25,"",shellscript,content +7130,20284600,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1629,0,"",shellscript,selection_command +7131,20284894,"TERMINAL",0,0,"87282",,terminal_output +7132,20285971,"TERMINAL",0,0,"98393",,terminal_output +7133,20287038,"TERMINAL",0,0,"\r4094304",,terminal_output +7134,20288027,"TERMINAL",0,0,"140515",,terminal_output +7135,20289088,"TERMINAL",0,0,"21626",,terminal_output +7136,20290214,"TERMINAL",0,0,"32737",,terminal_output +7137,20291237,"TERMINAL",0,0,"43848",,terminal_output +7138,20292016,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1873,0,"",shellscript,selection_mouse +7139,20292678,"TERMINAL",0,0,"54959",,terminal_output +7140,20293130,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1873,1,"4",shellscript,content +7141,20293658,"TERMINAL",0,0,"655065:00",,terminal_output +7142,20294695,"TERMINAL",0,0,"76171",,terminal_output +7143,20295747,"TERMINAL",0,0,"87282",,terminal_output +7144,20296784,"TERMINAL",0,0,"98393",,terminal_output +7145,20297857,"TERMINAL",0,0,"50505415",,terminal_output +7146,20298918,"TERMINAL",0,0,"21626",,terminal_output +7147,20299899,"TERMINAL",0,0,"32737",,terminal_output +7148,20300822,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",0,0,"",shellscript,tab +7149,20300823,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",1534,0,"",shellscript,selection_mouse +7150,20300945,"TERMINAL",0,0,"43848",,terminal_output +7151,20301996,"TERMINAL",0,0,"54959",,terminal_output +7152,20303115,"TERMINAL",0,0,"657:00610",,terminal_output +7153,20303724,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",0,0,"",shellscript,tab +7154,20303724,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1420,0,"",shellscript,selection_mouse +7155,20304082,"TERMINAL",0,0,"76171",,terminal_output +7156,20304442,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1529,0,"\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3421973",shellscript,content +7157,20304484,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1530,0,"",shellscript,selection_command +7158,20305301,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1660,0,"\n",shellscript,content +7159,20305382,"TERMINAL",0,0,"87282",,terminal_output +7160,20306212,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1416,0,"",shellscript,selection_mouse +7161,20306426,"TERMINAL",0,0,"98393",,terminal_output +7162,20307501,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1542,0,"",shellscript,selection_command +7163,20307521,"TERMINAL",0,0,"6:0094504",,terminal_output +7164,20308481,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1416,0,"",shellscript,selection_command +7165,20308534,"TERMINAL",0,0,"16:00515",,terminal_output +7166,20309583,"TERMINAL",0,0,"21626",,terminal_output +7167,20309746,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1415,0,"",shellscript,selection_command +7168,20310245,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1414,0,"",shellscript,selection_command +7169,20310284,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1413,0,"",shellscript,selection_command +7170,20310322,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1412,0,"",shellscript,selection_command +7171,20310358,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1411,0,"",shellscript,selection_command +7172,20310394,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1410,0,"",shellscript,selection_command +7173,20310589,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1409,0,"",shellscript,selection_command +7174,20310670,"TERMINAL",0,0,"32737",,terminal_output +7175,20310811,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1408,0,"",shellscript,selection_command +7176,20310995,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1407,0,"",shellscript,selection_command +7177,20311164,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1406,0,"",shellscript,selection_command +7178,20311362,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1405,0,"",shellscript,selection_command +7179,20311530,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1404,0,"",shellscript,selection_command +7180,20311639,"TERMINAL",0,0,"43848",,terminal_output +7181,20312524,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1404,0,"#",shellscript,content +7182,20312525,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1405,0,"",shellscript,selection_keyboard +7183,20312682,"TERMINAL",0,0,"54959",,terminal_output +7184,20312993,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1405,0," ",shellscript,content +7185,20312994,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1406,0,"",shellscript,selection_keyboard +7186,20313721,"TERMINAL",0,0,"6510620",,terminal_output +7187,20314763,"TERMINAL",0,0,"76171",,terminal_output +7188,20315803,"TERMINAL",0,0,"87282",,terminal_output +7189,20316848,"TERMINAL",0,0,"9943:004",,terminal_output +7190,20317886,"TERMINAL",0,0,"1110515",,terminal_output +7191,20318986,"TERMINAL",0,0,"21626",,terminal_output +7192,20319972,"TERMINAL",0,0,"32737",,terminal_output +7193,20321036,"TERMINAL",0,0,"43848",,terminal_output +7194,20322071,"TERMINAL",0,0,"54959",,terminal_output +7195,20323106,"TERMINAL",0,0,"6520630",,terminal_output +7196,20324151,"TERMINAL",0,0,"76171",,terminal_output +7197,20325196,"TERMINAL",0,0,"87282",,terminal_output +7198,20326261,"TERMINAL",0,0,"98393",,terminal_output +7199,20327382,"TERMINAL",0,0,"2094104",,terminal_output +7200,20328339,"TERMINAL",0,0,"120515",,terminal_output +7201,20329392,"TERMINAL",0,0,"21626",,terminal_output +7202,20330468,"TERMINAL",0,0,"32737",,terminal_output +7203,20331482,"TERMINAL",0,0,"43848",,terminal_output +7204,20332527,"TERMINAL",0,0,"54959",,terminal_output +7205,20333633,"TERMINAL",0,0,"6530640",,terminal_output +7206,20334627,"TERMINAL",0,0,"76171",,terminal_output +7207,20335669,"TERMINAL",0,0,"87282",,terminal_output +7208,20336721,"TERMINAL",0,0,"98393",,terminal_output +7209,20337790,"TERMINAL",0,0,"3094204",,terminal_output +7210,20338003,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1662,0,"",shellscript,selection_mouse +7211,20338345,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1661,1,"3",shellscript,selection_mouse +7212,20338429,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1660,2,"73",shellscript,selection_mouse +7213,20338711,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1659,3,"973",shellscript,selection_mouse +7214,20338791,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1658,4,"1973",shellscript,selection_mouse +7215,20338902,"TERMINAL",0,0,"130515",,terminal_output +7216,20338908,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1657,5,"21973",shellscript,selection_mouse +7217,20339197,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1656,6,"421973",shellscript,selection_mouse +7218,20339421,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1655,7,"3421973",shellscript,selection_mouse +7219,20339878,"TERMINAL",0,0,"22737",,terminal_output +7220,20340645,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1655,7,"",shellscript,content +7221,20340976,"TERMINAL",0,0,"43848",,terminal_output +7222,20341053,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1655,0,"3421982",shellscript,content +7223,20342122,"TERMINAL",0,0,"54959",,terminal_output +7224,20343053,"TERMINAL",0,0,"6540650",,terminal_output +7225,20344047,"TERMINAL",0,0,"76171",,terminal_output +7226,20345095,"TERMINAL",0,0,"87282",,terminal_output +7227,20346121,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1688,0,"",shellscript,selection_mouse +7228,20346182,"TERMINAL",0,0,"98393",,terminal_output +7229,20346280,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1663,25,"\nmkdir -p $CHECKPOINT_DIR",shellscript,selection_mouse +7230,20346387,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1688,0,"",shellscript,selection_mouse +7231,20347356,"TERMINAL",0,0,"4094304",,terminal_output +7232,20348228,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1402,0,"",shellscript,selection_mouse +7233,20348280,"TERMINAL",0,0,"140515",,terminal_output +7234,20348440,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1401,1,"D",shellscript,selection_mouse +7235,20348441,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1399,3,"_ID",shellscript,selection_mouse +7236,20348478,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1398,4,"B_ID",shellscript,selection_mouse +7237,20348479,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1397,5,"OB_ID",shellscript,selection_mouse +7238,20348514,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1395,7,"_JOB_ID",shellscript,selection_mouse +7239,20348551,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1394,8,"M_JOB_ID",shellscript,selection_mouse +7240,20348590,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1393,9,"RM_JOB_ID",shellscript,selection_mouse +7241,20348590,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1392,10,"URM_JOB_ID",shellscript,selection_mouse +7242,20348668,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1391,11,"LURM_JOB_ID",shellscript,selection_mouse +7243,20348716,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1390,12,"SLURM_JOB_ID",shellscript,selection_mouse +7244,20348796,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1389,13,"$SLURM_JOB_ID",shellscript,selection_mouse +7245,20349300,"TERMINAL",0,0,"21626",,terminal_output +7246,20350062,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1389,13,"",shellscript,content +7247,20350370,"TERMINAL",0,0,"32737",,terminal_output +7248,20350720,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",1389,0,"3421982",shellscript,content +7249,20351452,"TERMINAL",0,0,"43848",,terminal_output +7250,20352434,"TERMINAL",0,0,"54959",,terminal_output +7251,20353474,"TERMINAL",0,0,"655066:00",,terminal_output +7252,20354524,"TERMINAL",0,0,"76171",,terminal_output +7253,20355600,"TERMINAL",0,0,"87282",,terminal_output +7254,20356791,"TERMINAL",0,0,"98393",,terminal_output +7255,20357677,"TERMINAL",0,0,"5094404",,terminal_output +7256,20358768,"TERMINAL",0,0,"150515",,terminal_output +7257,20359758,"TERMINAL",0,0,"21626",,terminal_output +7258,20360874,"TERMINAL",0,0,"33848",,terminal_output +7259,20361881,"TERMINAL",0,0,"54959",,terminal_output +7260,20362910,"TERMINAL",0,0,"658:00610",,terminal_output +7261,20363988,"TERMINAL",0,0,"76171",,terminal_output +7262,20365069,"TERMINAL",0,0,"87282",,terminal_output +7263,20366063,"TERMINAL",0,0,"98393",,terminal_output +7264,20367112,"TERMINAL",0,0,"7:0094504",,terminal_output +7265,20368161,"TERMINAL",0,0,"17:00515",,terminal_output +7266,20369266,"TERMINAL",0,0,"21626",,terminal_output +7267,20370294,"TERMINAL",0,0,"32737",,terminal_output +7268,20371291,"TERMINAL",0,0,"43848",,terminal_output +7269,20372340,"TERMINAL",0,0,"54959",,terminal_output +7270,20373481,"TERMINAL",0,0,"6510620",,terminal_output +7271,20374490,"TERMINAL",0,0,"76171",,terminal_output +7272,20375529,"TERMINAL",0,0,"87282",,terminal_output +7273,20376524,"TERMINAL",0,0,"98393",,terminal_output +7274,20377669,"TERMINAL",0,0,"10944:004",,terminal_output +7275,20378616,"TERMINAL",0,0,"110515",,terminal_output +7276,20379667,"TERMINAL",0,0,"21626",,terminal_output +7277,20380717,"TERMINAL",0,0,"32737",,terminal_output +7278,20381781,"TERMINAL",0,0,"43848",,terminal_output +7279,20382809,"TERMINAL",0,0,"54959",,terminal_output +7280,20383912,"TERMINAL",0,0,"6621731",,terminal_output +7281,20384910,"TERMINAL",0,0,"87282",,terminal_output +7282,20385965,"TERMINAL",0,0,"98393",,terminal_output +7283,20387008,"TERMINAL",0,0,"2094104",,terminal_output +7284,20388081,"TERMINAL",0,0,"120515",,terminal_output +7285,20389171,"TERMINAL",0,0,"21626",,terminal_output +7286,20390160,"TERMINAL",0,0,"32737",,terminal_output +7287,20391189,"TERMINAL",0,0,"43848",,terminal_output +7288,20392237,"TERMINAL",0,0,"54959",,terminal_output +7289,20393336,"TERMINAL",0,0,"6530640",,terminal_output +7290,20394358,"TERMINAL",0,0,"76171",,terminal_output +7291,20395402,"TERMINAL",0,0,"87282",,terminal_output +7292,20396457,"TERMINAL",0,0,"98393",,terminal_output +7293,20397538,"TERMINAL",0,0,"3094204",,terminal_output +7294,20398557,"TERMINAL",0,0,"130515",,terminal_output +7295,20399620,"TERMINAL",0,0,"21626",,terminal_output +7296,20400705,"TERMINAL",0,0,"32737",,terminal_output +7297,20401653,"TERMINAL",0,0,"43848",,terminal_output +7298,20402707,"TERMINAL",0,0,"54959",,terminal_output +7299,20403880,"TERMINAL",0,0,"6540650",,terminal_output +7300,20404798,"TERMINAL",0,0,"76171",,terminal_output +7301,20405849,"TERMINAL",0,0,"88393",,terminal_output +7302,20406904,"TERMINAL",0,0,"4094304",,terminal_output +7303,20407951,"TERMINAL",0,0,"140515",,terminal_output +7304,20408634,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer]633;D;0",,terminal_output +7305,20423803,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",0,0,"",shellscript,tab +7306,20434586,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",,terminal_command +7307,20434622,"TERMINAL",0,0,"]633;E;2025-08-14 16:18:07 sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;Csbatch: error: Unable to open file slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer]633;D;1",,terminal_output +7308,20439103,"TERMINAL",0,0,"bash",,terminal_focus +7309,20445479,"TERMINAL",0,0,"runner",,terminal_command +7310,20449348,"TERMINAL",0,0,"sync-runner",,terminal_command +7311,20449381,"TERMINAL",0,0,"]633;E;2025-08-14 16:18:22 sync-runner;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Csending incremental file list\r\n",,terminal_output +7312,20451394,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch\r\n",,terminal_output +7313,20451424,"TERMINAL",0,0,"\r\nsent 30,606 bytes received 225 bytes 12,332.40 bytes/sec\r\ntotal size is 128,153,135 speedup is 4,156.63\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +7314,20456931,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch",,terminal_command +7315,20457029,"TERMINAL",0,0,"]633;E;2025-08-14 16:18:30 sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_4nodes_req.sbatch;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CSubmitted batch job 3423080\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +7316,20466245,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch",,terminal_command +7317,20466278,"TERMINAL",0,0,"]633;E;2025-08-14 16:18:39 sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4_2nodes_req.sbatch;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CSubmitted batch job 3423082\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +7318,20468054,"TERMINAL",0,0,"bash",,terminal_focus +7319,20470176,"TERMINAL",0,0,"queue",,terminal_command +7320,20470240,"TERMINAL",0,0,"]633;E;2025-08-14 16:18:43 queue;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h(B[?7hEvery 1.0s: squeu... hkn1993.localdomain: Thu Aug 14 16:18:43 2025JOBID PARTITION NAME USER ST\tTIME NODESNODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1(Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2(Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4(Priority)3412401 accelerat train_to tum_cte0 R 1-22:18:42\t 8hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-02:29:47\t 8hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 4:05:33\t 1hkn07133421964 accelerat train_to tum_cte0 R 4:17:57\t 1hkn0720",,terminal_output +7321,20471302,"TERMINAL",0,0,"43848",,terminal_output +7322,20472349,"TERMINAL",0,0,"54959",,terminal_output +7323,20473124,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 16:18:46 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2 (Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4 (Priority)3412401 accelerat train_to tum_cte0 R 1-22:18:45\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-02:29:50\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 4:05:36\t 1 hkn07133421964 accelerat train_to tum_cte0 R 4:18:00\t 1 hkn0720",,terminal_output +7324,20474394,"TERMINAL",0,0,"76171",,terminal_output +7325,20475470,"TERMINAL",0,0,"87282",,terminal_output +7326,20476266,"TERMINAL",0,0,"98393",,terminal_output +7327,20477404,"TERMINAL",0,0,"5094404",,terminal_output +7328,20478355,"TERMINAL",0,0,"150515",,terminal_output +7329,20479405,"TERMINAL",0,0,"21626",,terminal_output +7330,20480477,"TERMINAL",0,0,"32737",,terminal_output +7331,20481705,"TERMINAL",0,0,"43848",,terminal_output +7332,20482659,"TERMINAL",0,0,"54959",,terminal_output +7333,20483587,"TERMINAL",0,0,"6530:00610",,terminal_output +7334,20484683,"TERMINAL",0,0,"76171",,terminal_output +7335,20485699,"TERMINAL",0,0,"87282",,terminal_output +7336,20486727,"TERMINAL",0,0,"98393",,terminal_output +7337,20487774,"TERMINAL",0,0,"9:0094504",,terminal_output +7338,20488823,"TERMINAL",0,0,"19:01626",,terminal_output +7339,20489861,"TERMINAL",0,0,"32737",,terminal_output +7340,20490918,"TERMINAL",0,0,"43848",,terminal_output +7341,20491952,"TERMINAL",0,0,"54959",,terminal_output +7342,20492997,"TERMINAL",0,0,"6510620",,terminal_output +7343,20494093,"TERMINAL",0,0,"76171",,terminal_output +7344,20495108,"TERMINAL",0,0,"87282",,terminal_output +7345,20496155,"TERMINAL",0,0,"98393",,terminal_output +7346,20497269,"TERMINAL",0,0,"10946:004",,terminal_output +7347,20498297,"TERMINAL",0,0,"110515",,terminal_output +7348,20499319,"TERMINAL",0,0,"21626",,terminal_output +7349,20500341,"TERMINAL",0,0,"32737",,terminal_output +7350,20501427,"TERMINAL",0,0,"43848",,terminal_output +7351,20502415,"TERMINAL",0,0,"54959",,terminal_output +7352,20503648,"TERMINAL",0,0,"6520630",,terminal_output +7353,20504646,"TERMINAL",0,0,"76171",,terminal_output +7354,20505552,"TERMINAL",0,0,"87282",,terminal_output +7355,20506605,"TERMINAL",0,0,"98393",,terminal_output +7356,20507715,"TERMINAL",0,0,"2094104",,terminal_output +7357,20508738,"TERMINAL",0,0,"120515",,terminal_output +7358,20509724,"TERMINAL",0,0,"21626",,terminal_output +7359,20510791,"TERMINAL",0,0,"32737",,terminal_output +7360,20511819,"TERMINAL",0,0,"44959",,terminal_output +7361,20512863,"TERMINAL",0,0,"6530640",,terminal_output +7362,20513911,"TERMINAL",0,0,"76171",,terminal_output +7363,20514984,"TERMINAL",0,0,"87282",,terminal_output +7364,20516007,"TERMINAL",0,0,"98393",,terminal_output +7365,20517045,"TERMINAL",0,0,"3094204",,terminal_output +7366,20518159,"TERMINAL",0,0,"130515",,terminal_output +7367,20519157,"TERMINAL",0,0,"21626",,terminal_output +7368,20520410,"TERMINAL",0,0,"32737",,terminal_output +7369,20521333,"TERMINAL",0,0,"43848",,terminal_output +7370,20522460,"TERMINAL",0,0,"54959",,terminal_output +7371,20523338,"TERMINAL",0,0,"6540650",,terminal_output +7372,20524418,"TERMINAL",0,0,"76171",,terminal_output +7373,20525533,"TERMINAL",0,0,"87282",,terminal_output +7374,20526558,"TERMINAL",0,0,"98393",,terminal_output +7375,20527581,"TERMINAL",0,0,"4094304",,terminal_output +7376,20528603,"TERMINAL",0,0,"140515",,terminal_output +7377,20529610,"TERMINAL",0,0,"21626",,terminal_output +7378,20530858,"TERMINAL",0,0,"32737",,terminal_output +7379,20531715,"TERMINAL",0,0,"43848",,terminal_output +7380,20532754,"TERMINAL",0,0,"54959",,terminal_output +7381,20533790,"TERMINAL",0,0,"655069:00",,terminal_output +7382,20535055,"TERMINAL",0,0,"77282",,terminal_output +7383,20535977,"TERMINAL",0,0,"98393",,terminal_output +7384,20536946,"TERMINAL",0,0,"5094404",,terminal_output +7385,20538023,"TERMINAL",0,0,"150515",,terminal_output +7386,20539032,"TERMINAL",0,0,"21626",,terminal_output +7387,20540091,"TERMINAL",0,0,"32737",,terminal_output +7388,20541207,"TERMINAL",0,0,"43848",,terminal_output +7389,20542235,"TERMINAL",0,0,"54959",,terminal_output +7390,20543247,"TERMINAL",0,0,"651:00610",,terminal_output +7391,20544271,"TERMINAL",0,0,"76171",,terminal_output +7392,20545398,"TERMINAL",0,0,"87282",,terminal_output +7393,20546332,"TERMINAL",0,0,"904228393",,terminal_output +7394,20547373,"TERMINAL",0,0,"20:0094504",,terminal_output +7395,20548470,"TERMINAL",0,0,"120:00515",,terminal_output +7396,20549492,"TERMINAL",0,0,"21626",,terminal_output +7397,20550517,"TERMINAL",0,0,"32737",,terminal_output +7398,20551586,"TERMINAL",0,0,"43848",,terminal_output +7399,20552674,"TERMINAL",0,0,"54959",,terminal_output +7400,20553692,"TERMINAL",0,0,"6510620",,terminal_output +7401,20554665,"TERMINAL",0,0,"76171",,terminal_output +7402,20555710,"TERMINAL",0,0,"87282",,terminal_output +7403,20556788,"TERMINAL",0,0,"98393",,terminal_output +7404,20557792,"TERMINAL",0,0,"10947:004",,terminal_output +7405,20558829,"TERMINAL",0,0,"111626",,terminal_output +7406,20559939,"TERMINAL",0,0,"32737",,terminal_output +7407,20560913,"TERMINAL",0,0,"43848",,terminal_output +7408,20561987,"TERMINAL",0,0,"54959",,terminal_output +7409,20563012,"TERMINAL",0,0,"6520630",,terminal_output +7410,20564137,"TERMINAL",0,0,"76171",,terminal_output +7411,20565110,"TERMINAL",0,0,"87282",,terminal_output +7412,20566185,"TERMINAL",0,0,"98393",,terminal_output +7413,20567209,"TERMINAL",0,0,"2094104",,terminal_output +7414,20568337,"TERMINAL",0,0,"120515",,terminal_output +7415,20569566,"TERMINAL",0,0,"21626",,terminal_output +7416,20570326,"TERMINAL",0,0,"32737",,terminal_output +7417,20571371,"TERMINAL",0,0,"43848",,terminal_output +7418,20572412,"TERMINAL",0,0,"54959",,terminal_output +7419,20573468,"TERMINAL",0,0,"6530640",,terminal_output +7420,20574522,"TERMINAL",0,0,"76171",,terminal_output +7421,20575704,"TERMINAL",0,0,"87282",,terminal_output +7422,20576614,"TERMINAL",0,0,"98393",,terminal_output +7423,20577624,"TERMINAL",0,0,"3094204",,terminal_output +7424,20578712,"TERMINAL",0,0,"130515",,terminal_output +7425,20579718,"TERMINAL",0,0,"21626",,terminal_output +7426,20580829,"TERMINAL",0,0,"32737",,terminal_output +7427,20581852,"TERMINAL",0,0,"43848",,terminal_output +7428,20582856,"TERMINAL",0,0,"5540650",,terminal_output +7429,20583899,"TERMINAL",0,0,"76171",,terminal_output +7430,20585026,"TERMINAL",0,0,"87282",,terminal_output +7431,20586048,"TERMINAL",0,0,"98393",,terminal_output +7432,20587084,"TERMINAL",0,0,"4094304",,terminal_output +7433,20588063,"TERMINAL",0,0,"140515",,terminal_output +7434,20589123,"TERMINAL",0,0,"21626",,terminal_output +7435,20590249,"TERMINAL",0,0,"32737",,terminal_output +7436,20591256,"TERMINAL",0,0,"43848",,terminal_output +7437,20592297,"TERMINAL",0,0,"54959",,terminal_output +7438,20593299,"TERMINAL",0,0,"6550620:00",,terminal_output +7439,20594345,"TERMINAL",0,0,"76171",,terminal_output +7440,20595370,"TERMINAL",0,0,"87282",,terminal_output +7441,20596496,"TERMINAL",0,0,"98393",,terminal_output +7442,20597447,"TERMINAL",0,0,"5094404",,terminal_output +7443,20598853,"TERMINAL",0,0,"150515",,terminal_output +7444,20599751,"TERMINAL",0,0,"21626",,terminal_output +7445,20600796,"TERMINAL",0,0,"32737",,terminal_output +7446,20601835,"TERMINAL",0,0,"44959",,terminal_output +7447,20602878,"TERMINAL",0,0,"652:00610",,terminal_output +7448,20603920,"TERMINAL",0,0,"76171",,terminal_output +7449,20604995,"TERMINAL",0,0,"87282",,terminal_output +7450,20606019,"TERMINAL",0,0,"98393",,terminal_output +7451,20607043,"TERMINAL",0,0,"1:0094504",,terminal_output +7452,20608177,"TERMINAL",0,0,"11:00515",,terminal_output +7453,20609113,"TERMINAL",0,0,"21626",,terminal_output +7454,20610318,"TERMINAL",0,0,"32737",,terminal_output +7455,20611240,"TERMINAL",0,0,"43848",,terminal_output +7456,20612370,"TERMINAL",0,0,"54959",,terminal_output +7457,20613393,"TERMINAL",0,0,"6510620",,terminal_output +7458,20614386,"TERMINAL",0,0,"76171",,terminal_output +7459,20615439,"TERMINAL",0,0,"87282",,terminal_output +7460,20616439,"TERMINAL",0,0,"98393",,terminal_output +7461,20617488,"TERMINAL",0,0,"10948:004",,terminal_output +7462,20618515,"TERMINAL",0,0,"110515",,terminal_output +7463,20619539,"TERMINAL",0,0,"21626",,terminal_output +7464,20620696,"TERMINAL",0,0,"32737",,terminal_output +7465,20621610,"TERMINAL",0,0,"43848",,terminal_output +7466,20622711,"TERMINAL",0,0,"54959",,terminal_output +7467,20623737,"TERMINAL",0,0,"6520630",,terminal_output +7468,20624731,"TERMINAL",0,0,"76171",,terminal_output +7469,20625782,"TERMINAL",0,0,"87282",,terminal_output +7470,20626811,"TERMINAL",0,0,"98393",,terminal_output +7471,20627864,"TERMINAL",0,0,"20205115",,terminal_output +7472,20628894,"TERMINAL",0,0,"21626",,terminal_output +7473,20629979,"TERMINAL",0,0,"32737",,terminal_output +7474,20631016,"TERMINAL",0,0,"43848",,terminal_output +7475,20632148,"TERMINAL",0,0,"54959",,terminal_output +7476,20633155,"TERMINAL",0,0,"6530640",,terminal_output +7477,20634179,"TERMINAL",0,0,"76171",,terminal_output +7478,20635314,"TERMINAL",0,0,"87282",,terminal_output +7479,20636182,"TERMINAL",0,0,"98393",,terminal_output +7480,20637226,"TERMINAL",0,0,"3094204",,terminal_output +7481,20638381,"TERMINAL",0,0,"130515",,terminal_output +7482,20639401,"TERMINAL",0,0,"21626",,terminal_output +7483,20640434,"TERMINAL",0,0,"32737",,terminal_output +7484,20641396,"TERMINAL",0,0,"43848",,terminal_output +7485,20642434,"TERMINAL",0,0,"54959",,terminal_output +7486,20643475,"TERMINAL",0,0,"6540650",,terminal_output +7487,20644522,"TERMINAL",0,0,"76171",,terminal_output +7488,20645549,"TERMINAL",0,0,"87282",,terminal_output +7489,20646677,"TERMINAL",0,0,"98393",,terminal_output +7490,20647697,"TERMINAL",0,0,"4094304",,terminal_output +7491,20648827,"TERMINAL",0,0,"140515",,terminal_output +7492,20649729,"TERMINAL",0,0,"21626",,terminal_output +7493,20650730,"TERMINAL",0,0,"32737",,terminal_output +7494,20651858,"TERMINAL",0,0,"43848",,terminal_output +7495,20652798,"TERMINAL",0,0,"54959",,terminal_output +7496,20653836,"TERMINAL",0,0,"665171:01",,terminal_output +7497,20654883,"TERMINAL",0,0,"87282",,terminal_output +7498,20655920,"TERMINAL",0,0,"98393",,terminal_output +7499,20657015,"TERMINAL",0,0,"5094404",,terminal_output +7500,20657992,"TERMINAL",0,0,"150515",,terminal_output +7501,20659077,"TERMINAL",0,0,"21626",,terminal_output +7502,20660087,"TERMINAL",0,0,"32737",,terminal_output +7503,20661313,"TERMINAL",0,0,"43848",,terminal_output +7504,20662344,"TERMINAL",0,0,"54959",,terminal_output +7505,20663378,"TERMINAL",0,0,"653:00610",,terminal_output +7506,20664389,"TERMINAL",0,0,"76171",,terminal_output +7507,20665413,"TERMINAL",0,0,"87282",,terminal_output +7508,20666435,"TERMINAL",0,0,"98393",,terminal_output +7509,20667459,"TERMINAL",0,0,"2:0094504",,terminal_output +7510,20668484,"TERMINAL",0,0,"12:00515",,terminal_output +7511,20669507,"TERMINAL",0,0,"21626",,terminal_output +7512,20670536,"TERMINAL",0,0,"32737",,terminal_output +7513,20671530,"TERMINAL",0,0,"43848",,terminal_output +7514,20672580,"TERMINAL",0,0,"54959",,terminal_output +7515,20673706,"TERMINAL",0,0,"6510620",,terminal_output +7516,20674647,"TERMINAL",0,0,"76171",,terminal_output +7517,20675794,"TERMINAL",0,0,"87282",,terminal_output +7518,20676774,"TERMINAL",0,0,"98393",,terminal_output +7519,20677804,"TERMINAL",0,0,"10949:004",,terminal_output +7520,20678807,"TERMINAL",0,0,"110515",,terminal_output +7521,20679854,"TERMINAL",0,0,"22737",,terminal_output +7522,20680987,"TERMINAL",0,0,"43848",,terminal_output +7523,20681938,"TERMINAL",0,0,"54959",,terminal_output +7524,20683024,"TERMINAL",0,0,"6520630",,terminal_output +7525,20684027,"TERMINAL",0,0,"76171",,terminal_output +7526,20685076,"TERMINAL",0,0,"87282",,terminal_output +7527,20686200,"TERMINAL",0,0,"98393",,terminal_output +7528,20687325,"TERMINAL",0,0,"2094104",,terminal_output +7529,20688251,"TERMINAL",0,0,"120515",,terminal_output +7530,20689374,"TERMINAL",0,0,"21626",,terminal_output +7531,20690400,"TERMINAL",0,0,"32737",,terminal_output +7532,20691358,"TERMINAL",0,0,"43848",,terminal_output +7533,20692447,"TERMINAL",0,0,"54959",,terminal_output +7534,20693713,"TERMINAL",0,0,"6530640",,terminal_output +7535,20694597,"TERMINAL",0,0,"76171",,terminal_output +7536,20695518,"TERMINAL",0,0,"87282",,terminal_output +7537,20696563,"TERMINAL",0,0,"98393",,terminal_output +7538,20697572,"TERMINAL",0,0,"3094204",,terminal_output +7539,20698692,"TERMINAL",0,0,"130515",,terminal_output +7540,20699717,"TERMINAL",0,0,"21626",,terminal_output +7541,20700848,"TERMINAL",0,0,"32737",,terminal_output +7542,20701870,"TERMINAL",0,0,"43848",,terminal_output +7543,20702801,"TERMINAL",0,0,"54959",,terminal_output +7544,20703811,"TERMINAL",0,0,"6540650",,terminal_output +7545,20704866,"TERMINAL",0,0,"77282",,terminal_output +7546,20705900,"TERMINAL",0,0,"98393",,terminal_output +7547,20706936,"TERMINAL",0,0,"4094304",,terminal_output +7548,20708010,"TERMINAL",0,0,"140515",,terminal_output +7549,20709036,"TERMINAL",0,0,"21626",,terminal_output +7550,20710160,"TERMINAL",0,0,"32737",,terminal_output +7551,20711188,"TERMINAL",0,0,"43848",,terminal_output +7552,20712207,"TERMINAL",0,0,"54959",,terminal_output +7553,20713195,"TERMINAL",0,0,"655062:00",,terminal_output +7554,20714258,"TERMINAL",0,0,"76171",,terminal_output +7555,20715390,"TERMINAL",0,0,"87282",,terminal_output +7556,20716332,"TERMINAL",0,0,"98393",,terminal_output +7557,20717372,"TERMINAL",0,0,"5094404",,terminal_output +7558,20718456,"TERMINAL",0,0,"150515",,terminal_output +7559,20719484,"TERMINAL",0,0,"21626",,terminal_output +7560,20720607,"TERMINAL",0,0,"32737",,terminal_output +7561,20721741,"TERMINAL",0,0,"43848",,terminal_output +7562,20722655,"TERMINAL",0,0,"54959",,terminal_output +7563,20723645,"TERMINAL",0,0,"654:00610",,terminal_output +7564,20724750,"TERMINAL",0,0,"76171",,terminal_output +7565,20725837,"TERMINAL",0,0,"87282",,terminal_output +7566,20726775,"TERMINAL",0,0,"98393",,terminal_output +7567,20727815,"TERMINAL",0,0,"3:0094504",,terminal_output +7568,20728859,"TERMINAL",0,0,"23:01626",,terminal_output +7569,20729906,"TERMINAL",0,0,"32737",,terminal_output +7570,20731052,"TERMINAL",0,0,"43848",,terminal_output +7571,20732076,"TERMINAL",0,0,"54959",,terminal_output +7572,20733102,"TERMINAL",0,0,"6510620",,terminal_output +7573,20734126,"TERMINAL",0,0,"76171",,terminal_output +7574,20735146,"TERMINAL",0,0,"87282",,terminal_output +7575,20736276,"TERMINAL",0,0,"98393",,terminal_output +7576,20737298,"TERMINAL",0,0,"109410:004",,terminal_output +7577,20738321,"TERMINAL",0,0,"110515",,terminal_output +7578,20739347,"TERMINAL",0,0,"21626",,terminal_output +7579,20740367,"TERMINAL",0,0,"32737",,terminal_output +7580,20741398,"TERMINAL",0,0,"43848",,terminal_output +7581,20742459,"TERMINAL",0,0,"54959",,terminal_output +7582,20743646,"TERMINAL",0,0,"6520630",,terminal_output +7583,20744520,"TERMINAL",0,0,"76171",,terminal_output +7584,20745591,"TERMINAL",0,0,"87282",,terminal_output +7585,20746724,"TERMINAL",0,0,"98393",,terminal_output +7586,20747741,"TERMINAL",0,0,"2094104",,terminal_output +7587,20748766,"TERMINAL",0,0,"120515",,terminal_output +7588,20749791,"TERMINAL",0,0,"21626",,terminal_output +7589,20750817,"TERMINAL",0,0,"32737",,terminal_output +7590,20751842,"TERMINAL",0,0,"44959",,terminal_output +7591,20752877,"TERMINAL",0,0,"6530640",,terminal_output +7592,20753917,"TERMINAL",0,0,"76171",,terminal_output +7593,20755013,"TERMINAL",0,0,"87282",,terminal_output +7594,20756007,"TERMINAL",0,0,"98393",,terminal_output +7595,20757060,"TERMINAL",0,0,"3094204",,terminal_output +7596,20758110,"TERMINAL",0,0,"130515",,terminal_output +7597,20759210,"TERMINAL",0,0,"21626",,terminal_output +7598,20760240,"TERMINAL",0,0,"32737",,terminal_output +7599,20761261,"TERMINAL",0,0,"43848",,terminal_output +7600,20762286,"TERMINAL",0,0,"54959",,terminal_output +7601,20763322,"TERMINAL",0,0,"6540650",,terminal_output +7602,20764426,"TERMINAL",0,0,"76171",,terminal_output +7603,20765412,"TERMINAL",0,0,"87282",,terminal_output +7604,20766482,"TERMINAL",0,0,"98393",,terminal_output +7605,20767504,"TERMINAL",0,0,"4094304",,terminal_output +7606,20768634,"TERMINAL",0,0,"140515",,terminal_output +7607,20769583,"TERMINAL",0,0,"21626",,terminal_output +7608,20770628,"TERMINAL",0,0,"32737",,terminal_output +7609,20771668,"TERMINAL",0,0,"43848",,terminal_output +7610,20772770,"TERMINAL",0,0,"54959",,terminal_output +7611,20773763,"TERMINAL",0,0,"655063:00",,terminal_output +7612,20774807,"TERMINAL",0,0,"76171",,terminal_output +7613,20775857,"TERMINAL",0,0,"88393",,terminal_output +7614,20776944,"TERMINAL",0,0,"5094404",,terminal_output +7615,20777940,"TERMINAL",0,0,"150515",,terminal_output +7616,20778977,"TERMINAL",0,0,"21626",,terminal_output +7617,20780097,"TERMINAL",0,0,"32737",,terminal_output +7618,20781062,"TERMINAL",0,0,"43848",,terminal_output +7619,20782109,"TERMINAL",0,0,"54959",,terminal_output +7620,20783159,"TERMINAL",0,0,"655:00610",,terminal_output +7621,20784198,"TERMINAL",0,0,"76171",,terminal_output +7622,20785341,"TERMINAL",0,0,"87282",,terminal_output +7623,20786297,"TERMINAL",0,0,"98393",,terminal_output +7624,20787334,"TERMINAL",0,0,"4:0094504",,terminal_output +7625,20788393,"TERMINAL",0,0,"14:00515",,terminal_output +7626,20789528,"TERMINAL",0,0,"21626",,terminal_output +7627,20790520,"TERMINAL",0,0,"32737",,terminal_output +7628,20791581,"TERMINAL",0,0,"43848",,terminal_output +7629,20792557,"TERMINAL",0,0,"54959",,terminal_output +7630,20793621,"TERMINAL",0,0,"6510620",,terminal_output +7631,20794691,"TERMINAL",0,0,"76171",,terminal_output +7632,20795872,"TERMINAL",0,0,"87282",,terminal_output +7633,20796785,"TERMINAL",0,0,"98393",,terminal_output +7634,20797827,"TERMINAL",0,0,"10941:004",,terminal_output +7635,20798958,"TERMINAL",0,0,"110515",,terminal_output +7636,21206479,"TERMINAL",0,0,"22737",,terminal_output +7637,21206921,"TERMINAL",0,0,"43848549596520630761718728298393209410412051521626327374384854959653064076171872829839330942041305152162632737438485540650761718728298393409430414051521626327374384854959655064:00761718728298393509440415051521626327374384854959656:0061076171983935:009450415:0051521626327374384854959651062076171872829839310942:0041105152162632737438485495965206307617187282983932020511521626327374384854959653064076171872829839330942041305152162632737438485495965406507617187282983934094304140515216263273743848655065:00761718728298393509440415051521626327374384854959657:006107617187282983936:006:00551521626327374384854959651062076171872829839310943:0041105152162632737438485495965206307617188393209410412051521626327374384854959653064076171872829839330942041305152162632737438485495965406507617187282983934094304140515216263384854959655066:00761718728298393509440415051521626327374384854959658:006107617187282983937:009450417:00515216263273743848549596510620761718839310944:004110515216263273743848549596520630761718728298393209410412051521626327374384854959653064076171872829839330942041305152162633848549596540650761718728298393409430414051521626327374384854959655067:00761718728298393509440415051521626327374384854959659:0061076171883938:009450418:0051521626327374384854959651062076171872829839310945:004110515216263273743848549596520630761718728298393209410412051521626327374495965306407617187282983933094204130515216263273743848549596540650761718728298393409430414051521626327374384854959655068:0076171883935094404150515216263273743848549596540:006107617187282983939:009450419:0051521626327374384854959651062076171872829839310946:0041105152162632737449596520630761718728298393209410412051521626327374384854959653064076171872829839330942041305152162632737438485495965406507617187282994304414051521626327374384854959655069:00761718728298393509440415051521626327374384854959651:0061076171872829839330:0094504130:00515216263384854959651062076171872829839310947:0041105152162632737438485495965206307617187282983932094104120515216263273743848549596530640772829839330942041305152162632737438485495965406507617187282983934094304140515216263273743848549596550630:007617187282994404515051521626327374384854959652:00610761718728298393",,terminal_output +7638,21207281,"TERMINAL",0,0,"1:0094504",,terminal_output +7639,21208320,"TERMINAL",0,0,"11:00515",,terminal_output +7640,21209401,"TERMINAL",0,0,"21626",,terminal_output +7641,21210420,"TERMINAL",0,0,"32737",,terminal_output +7642,21211425,"TERMINAL",0,0,"43848",,terminal_output +7643,21212482,"TERMINAL",0,0,"54959",,terminal_output +7644,21213550,"TERMINAL",0,0,"6510620",,terminal_output +7645,21214605,"TERMINAL",0,0,"76171",,terminal_output +7646,21215614,"TERMINAL",0,0,"87282",,terminal_output +7647,21216641,"TERMINAL",0,0,"98393",,terminal_output +7648,21217715,"TERMINAL",0,0,"10948:004",,terminal_output +7649,21218718,"TERMINAL",0,0,"110515",,terminal_output +7650,21219791,"TERMINAL",0,0,"21626",,terminal_output +7651,21220811,"TERMINAL",0,0,"32737",,terminal_output +7652,21221858,"TERMINAL",0,0,"44959",,terminal_output +7653,21222898,"TERMINAL",0,0,"6520630",,terminal_output +7654,21223933,"TERMINAL",0,0,"76171",,terminal_output +7655,21224969,"TERMINAL",0,0,"87282",,terminal_output +7656,21226018,"TERMINAL",0,0,"98393",,terminal_output +7657,21227025,"TERMINAL",0,0,"2094104",,terminal_output +7658,21228058,"TERMINAL",0,0,"120515",,terminal_output +7659,21229094,"TERMINAL",0,0,"21626",,terminal_output +7660,21230141,"TERMINAL",0,0,"32737",,terminal_output +7661,21231195,"TERMINAL",0,0,"43848",,terminal_output +7662,21232254,"TERMINAL",0,0,"54959",,terminal_output +7663,21233273,"TERMINAL",0,0,"6530640",,terminal_output +7664,21234319,"TERMINAL",0,0,"76171",,terminal_output +7665,21235344,"TERMINAL",0,0,"87282",,terminal_output +7666,21236398,"TERMINAL",0,0,"98393",,terminal_output +7667,21237418,"TERMINAL",0,0,"3094204",,terminal_output +7668,21238651,"TERMINAL",0,0,"130515",,terminal_output +7669,21239536,"TERMINAL",0,0,"21626",,terminal_output +7670,21240536,"TERMINAL",0,0,"32737",,terminal_output +7671,21241639,"TERMINAL",0,0,"43848",,terminal_output +7672,21242645,"TERMINAL",0,0,"54959",,terminal_output +7673,21243767,"TERMINAL",0,0,"6540650",,terminal_output +7674,21244726,"TERMINAL",0,0,"76171",,terminal_output +7675,21245821,"TERMINAL",0,0,"87282",,terminal_output +7676,21246847,"TERMINAL",0,0,"98393",,terminal_output +7677,21247885,"TERMINAL",0,0,"40405315",,terminal_output +7678,21248893,"TERMINAL",0,0,"21626",,terminal_output +7679,21250016,"TERMINAL",0,0,"32737",,terminal_output +7680,21251042,"TERMINAL",0,0,"43848",,terminal_output +7681,21252171,"TERMINAL",0,0,"54959",,terminal_output +7682,21253064,"TERMINAL",0,0,"655061:00",,terminal_output +7683,21254089,"TERMINAL",0,0,"76171",,terminal_output +7684,21255123,"TERMINAL",0,0,"87282",,terminal_output +7685,21256167,"TERMINAL",0,0,"98393",,terminal_output +7686,21257289,"TERMINAL",0,0,"5094404",,terminal_output +7687,21258266,"TERMINAL",0,0,"150515",,terminal_output +7688,21259456,"TERMINAL",0,0,"21626",,terminal_output +7689,21260377,"TERMINAL",0,0,"32737",,terminal_output +7690,21261384,"TERMINAL",0,0,"43848",,terminal_output +7691,21261940,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer]633;D;0",,terminal_output +7692,21277307,"train_dynamics.py",0,0,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n darkness_threshold: float = 0.0\n # Tokenizer\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_type: str = ""maskgit"" # supported options: maskgit, causal\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n wandb_id: str = """"\n\n\nargs = tyro.cli(Args)\n\n\ndef dynamics_loss_fn(\n model: Genie, inputs: dict\n) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n """"""Compute masked dynamics loss""""""\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n model.train()\n outputs = model(inputs, training=True)\n mask = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n gt = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]), size=args.num_latent_actions, fill_value=0\n )\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]), size=args.num_patch_latents, fill_value=0\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n psnr=psnr,\n ssim=ssim,\n codebook_usage_lam=codebook_usage_lam,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n\n@nnx.jit\ndef train_step(\n model: Genie, optimizer: nnx.Optimizer, inputs: dict\n) -> tuple[jax.Array, jax.Array, dict]:\n """"""Update state and compute metrics""""""\n\n def loss_fn(model: Genie) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n return dynamics_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(model)\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n decode=False,\n rngs=rngs,\n )\n\n _, params, _ = nnx.split(genie, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n optimizer = nnx.Optimizer(genie, tx)\n del genie\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n darkness_threshold=args.darkness_threshold,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n optimizer = restore_genie_components(optimizer, replicated_sharding, rng, args)\n # NOTE: We have to remove the (unused) tokenizer vq dropout due flax.nnx lazily initializing modules.\n # Specifically, the first dynamics model checkpoint will contain the vq dropout module,\n # but the first full restore will fail due to nnx not initializing the module when\n # dropout is set to 0.0.\n del optimizer.model.tokenizer.vq.drop\n\n # --- TRAIN LOOP ---\n dataloader = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in grain_iterator\n )\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng_mask = jax.random.split(rng, 2)\n inputs = dict(videos=videos, mask_rng=_rng_mask)\n loss, recon, metrics = train_step(optimizer.model, optimizer, inputs)\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n grain_iterator # type: ignore\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +7693,21298766,"train_dynamics.py",3514,0,"",python,selection_mouse +7694,21298931,"train_dynamics.py",3505,16,"index_counts_lam",python,selection_mouse +7695,21307925,"train_dynamics.py",3793,0,"",python,selection_mouse +7696,21308104,"train_dynamics.py",3782,18,"codebook_usage_lam",python,selection_mouse +7697,21316333,"train_dynamics.py",3876,0,"",python,selection_mouse +7698,21316494,"train_dynamics.py",3866,22,"index_counts_tokenizer",python,selection_mouse +7699,21317552,"train_dynamics.py",3853,0,"",python,selection_mouse +7700,21317725,"train_dynamics.py",3838,24,"codebook_usage_tokenizer",python,selection_mouse +7701,21327908,"train_dynamics.py",3872,0,"",python,selection_mouse +7702,21328075,"train_dynamics.py",3866,22,"index_counts_tokenizer",python,selection_mouse +7703,21330660,"train_dynamics.py",3718,0,"",python,selection_mouse +7704,21330832,"train_dynamics.py",3713,12,"video_tokens",python,selection_mouse +7705,21353690,"genie.py",0,0,"from typing import Dict\n\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\nimport orbax.checkpoint as ocp\n\nfrom models.dynamics import DynamicsMaskGIT, DynamicsCausal\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\n\nclass Genie(nnx.Module):\n """"""Genie model""""""\n\n def __init__(\n self,\n in_dim: int,\n tokenizer_dim: int,\n tokenizer_ffn_dim: int,\n latent_patch_dim: int,\n num_patch_latents: int,\n patch_size: int,\n tokenizer_num_blocks: int,\n tokenizer_num_heads: int,\n lam_dim: int,\n lam_ffn_dim: int,\n latent_action_dim: int,\n num_latent_actions: int,\n lam_patch_size: int,\n lam_num_blocks: int,\n lam_num_heads: int,\n lam_co_train: bool,\n dyna_type: str,\n dyna_dim: int,\n dyna_ffn_dim: int,\n dyna_num_blocks: int,\n dyna_num_heads: int,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n dropout: float = 0.0,\n mask_limit: float = 0.0,\n ):\n # --- Tokenizer ---\n self.in_dim = in_dim\n self.tokenizer_dim = tokenizer_dim\n self.tokenizer_ffn_dim = tokenizer_ffn_dim\n self.latent_patch_dim = latent_patch_dim\n self.num_patch_latents = num_patch_latents\n self.patch_size = patch_size\n self.tokenizer_num_blocks = tokenizer_num_blocks\n self.tokenizer_num_heads = tokenizer_num_heads\n # --- LAM ---\n self.lam_dim = lam_dim\n self.lam_ffn_dim = lam_ffn_dim\n self.latent_action_dim = latent_action_dim\n self.num_latent_actions = num_latent_actions\n self.lam_patch_size = lam_patch_size\n self.lam_num_blocks = lam_num_blocks\n self.lam_num_heads = lam_num_heads\n self.lam_co_train = lam_co_train\n # --- Dynamics ---\n self.dyna_type = dyna_type\n self.dyna_dim = dyna_dim\n self.dyna_ffn_dim = dyna_ffn_dim\n self.dyna_num_blocks = dyna_num_blocks\n self.dyna_num_heads = dyna_num_heads\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.dropout = dropout\n self.mask_limit = mask_limit\n\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n ffn_dim=self.tokenizer_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n ffn_dim=self.lam_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_latent_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n if self.dyna_type == ""maskgit"":\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n elif self.dyna_type == ""causal"":\n self.dynamics = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n rngs=rngs,\n )\n else:\n raise ValueError(f""Invalid dynamics type: {self.dyna_type}"")\n\n def __call__(\n self, batch: Dict[str, jax.Array], training: bool = True\n ) -> Dict[str, jax.Array]:\n videos_BTHWC = batch[""videos""]\n tokenizer_outputs = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_indices_BTN = tokenizer_outputs[""indices""]\n lam_outputs = self.lam.vq_encode(videos_BTHWC, training=False)\n z_q_BTm11L = lam_outputs[""z_q""]\n action_indices_E = lam_outputs[""indices""]\n latent_actions_BTm11L = jax.lax.cond(\n self.lam_co_train,\n lambda: z_q_BTm11L,\n lambda: jax.lax.stop_gradient(z_q_BTm11L),\n )\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(token_indices_BTN),\n latent_actions=latent_actions_BTm11L,\n )\n outputs[""mask_rng""] = batch[""mask_rng""]\n dyna_logits_BTNV, dyna_mask = self.dynamics(outputs, training)\n outputs[""token_logits""] = dyna_logits_BTNV\n if dyna_mask is not None:\n outputs[""mask""] = dyna_mask\n mle_indices_BTN = jnp.argmax(outputs[""token_logits""], axis=-1)\n H, W = batch[""videos""].shape[2:4]\n outputs[""recon""] = self.tokenizer.decode(mle_indices_BTN, (H, W))\n outputs[""lam_indices""] = action_indices_E\n return outputs\n\n def sample(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n steps: int = 25,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> jax.Array:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n P: S * N\n """"""\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n latent_actions_E = batch[""latent_actions""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n def maskgit_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array], step: jax.Array\n ) -> tuple[tuple[jax.Array, jax.Array, jax.Array, jax.Array], None]:\n rng, token_idxs_BSN, mask_BSN, action_tokens_EL = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # --- Construct + encode video ---\n vid_embed_BSNM = self.dynamics.patch_embed(token_idxs_BSN)\n mask_token_111M = self.dynamics.mask_token.value\n mask_expanded_BSN1 = mask_BSN[..., None]\n vid_embed_BSNM = jnp.where(mask_expanded_BSN1, mask_token_111M, vid_embed_BSNM)\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = self.dynamics.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1]))\n vid_embed_BSNM += act_embed_BS1M\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (steps * 2))\n step_temp = temperature * (1.0 - unmasked_ratio)\n final_logits_BSNV = self.dynamics.transformer(vid_embed_BSNM) / step_temp\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_BSN = jnp.argmax(final_logits_BSNV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_BSN = jax.random.categorical(_rng, final_logits_BSNV)\n gather_fn = jax.vmap(jax.vmap(jax.vmap(lambda x, y: x[y])))\n final_token_probs_BSN = gather_fn(\n jax.nn.softmax(final_logits_BSNV), sampled_token_idxs_BSN\n )\n final_token_probs_BSN += ~mask_BSN\n # Update masked tokens only\n token_idxs_BSN = jnp.where(mask_BSN, sampled_token_idxs_BSN, token_idxs_BSN)\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n final_token_probs_flat_BP = einops.rearrange(final_token_probs_BSN, ""b s n -> b (s n)"")\n idx_mask_P = jnp.arange(final_token_probs_flat_BP.shape[-1]) <= N - num_unmasked_tokens\n sorted_idxs_BP = jnp.argsort(final_token_probs_flat_BP, axis=-1)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask_P))\n mask_flat_BP = einops.rearrange(mask_BSN, ""b s n -> b (s n)"")\n new_mask_flat_BP = mask_update_fn(mask_flat_BP, sorted_idxs_BP)\n new_mask_BSN = einops.rearrange(new_mask_flat_BP, ""b (s n) -> b s n"", n=N)\n\n new_carry = (rng, token_idxs_BSN, new_mask_BSN, action_tokens_EL)\n return new_carry, None\n\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[tuple[jax.Array, jax.Array], None]:\n rng, current_token_idxs_BSN = carry\n rng, step_rng = jax.random.split(rng)\n\n # Mask current frame (i.e., t == step_t)\n mask_S = jnp.arange(seq_len) == step_t\n mask_BSN = jnp.broadcast_to(mask_S[None, :, None], (B, seq_len, N)).astype(\n bool\n )\n masked_token_idxs_BSN = current_token_idxs_BSN * ~mask_BSN\n\n # --- Initialize and run MaskGIT loop ---\n init_carry_maskgit = (\n step_rng,\n masked_token_idxs_BSN,\n mask_BSN,\n action_tokens_EL,\n )\n final_carry_maskgit, _ = jax.lax.scan(\n maskgit_step_fn, init_carry_maskgit, jnp.arange(steps)\n )\n updated_token_idxs_BSN = final_carry_maskgit[1]\n new_carry = (rng, updated_token_idxs_BSN)\n return new_carry, None\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry, _ = jax.lax.scan(\n generation_step_fn, initial_carry, timesteps_to_scan\n )\n final_token_idxs_BSN = final_carry[1]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC\n\n def sample_causal(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> jax.Array:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n assert isinstance(self.dynamics, DynamicsCausal)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n latent_actions_E = batch[""latent_actions""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n dynamics_causal: DynamicsCausal = self.dynamics\n\n def causal_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array], step_n: jax.Array\n ) -> tuple[tuple[jax.Array, jax.Array, jax.Array, jax.Array], None]:\n rng, token_idxs_BSN, action_tokens_EL, step_t = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_causal.patch_embed(token_idxs_BSN)\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_causal.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1]))\n vid_embed_BSNp1M = jnp.concatenate([act_embed_BS1M, vid_embed_BSNM], axis=2)\n final_logits_BTNp1V = dynamics_causal.transformer(vid_embed_BSNp1M, (step_t, step_n)) / temperature\n final_logits_BV = final_logits_BTNp1V[:, step_t, step_n, :]\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_B = jnp.argmax(final_logits_BV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_B = jax.random.categorical(_rng, final_logits_BV)\n # Update next tokens only\n token_idxs_BSN = token_idxs_BSN.at[:, step_t, step_n].set(sampled_token_idxs_B)\n\n new_carry = (rng, token_idxs_BSN, action_tokens_EL, step_t)\n return new_carry, None\n\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[tuple[jax.Array, jax.Array], None]:\n rng, current_token_idxs_BSN = carry\n rng, step_rng = jax.random.split(rng)\n\n # --- Initialize and run causal loop ---\n init_carry_causal = (\n step_rng,\n current_token_idxs_BSN,\n action_tokens_EL,\n step_t,\n )\n final_carry_causal, _ = jax.lax.scan(\n causal_step_fn, init_carry_causal, jnp.arange(N)\n )\n updated_token_idxs_BSN = final_carry_causal[1]\n new_carry = (rng, updated_token_idxs_BSN)\n return new_carry, None\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry, _ = jax.lax.scan(\n generation_step_fn, initial_carry, timesteps_to_scan\n )\n final_token_idxs_BSN = final_carry[1]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC\n\n def vq_encode(self, batch: Dict[str, jax.Array], training: bool) -> jax.Array:\n # --- Preprocess videos ---\n video_BTHWC = batch[""videos""]\n lam_output = self.lam.vq_encode(video_BTHWC, training=training)\n lam_indices_E = lam_output[""indices""]\n return lam_indices_E\n\n# FIXME (f.srambical): add conversion script for old checkpoints\ndef restore_genie_components(\n optimizer: nnx.Optimizer,\n sharding: jax.sharding.NamedSharding,\n rng: jax.Array,\n args,\n) -> nnx.Optimizer:\n """"""Restore pre-trained Genie components""""""\n rngs = nnx.Rngs(rng)\n\n tx = optimizer.tx\n model = optimizer.model\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n tokenizer_checkpoint_manager = ocp.CheckpointManager(\n directory=args.tokenizer_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n ffn_dim=args.tokenizer_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n dummy_tokenizer_optimizer = nnx.Optimizer(dummy_tokenizer, tx)\n dummy_tokenizer_optimizer_state = nnx.state(dummy_tokenizer_optimizer)\n abstract_sharded_tokenizer_optimizer_state = _create_abstract_sharded_pytree(\n dummy_tokenizer_optimizer_state, sharding\n )\n restored_tokenizer = tokenizer_checkpoint_manager.restore(\n step=tokenizer_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_tokenizer_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_tokenizer_optimizer.model, restored_tokenizer.model)\n model.tokenizer = dummy_tokenizer_optimizer.model\n tokenizer_checkpoint_manager.close()\n\n if args.lam_checkpoint:\n lam_checkpoint_manager = ocp.CheckpointManager(\n directory=args.lam_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n ffn_dim=args.lam_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_latent_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n dummy_lam_optimizer = nnx.Optimizer(dummy_lam, tx)\n dummy_lam_optimizer_state = nnx.state(dummy_lam_optimizer)\n abstract_sharded_lam_optimizer_state = _create_abstract_sharded_pytree(\n dummy_lam_optimizer_state, sharding\n )\n restored_lam_optimizer = lam_checkpoint_manager.restore(\n step=lam_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_lam_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_lam_optimizer.model, restored_lam_optimizer.model)\n model.lam = dummy_lam_optimizer.model\n # Remove the LAM decoder to save memory and avoid unnecessary computation.\n del model.lam.decoder\n lam_checkpoint_manager.close()\n \n # Reinitialize the optimizer states\n optimizer = nnx.Optimizer(model, tx)\n return optimizer\n\n\ndef _create_abstract_sharded_pytree(\n pytree_template: nnx.GraphState, sharding_spec: jax.sharding.NamedSharding\n) -> jax.Array:\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)\n",python,tab +7706,21380036,"genie.py",5948,0,"",python,selection_mouse +7707,21380194,"genie.py",5941,15,"mle_indices_BTN",python,selection_mouse +7708,21385911,"genie.py",6163,0,"",python,selection_mouse +7709,21386079,"genie.py",6153,16,"action_indices_E",python,selection_mouse +7710,21399410,"genie.py",5581,0,"",python,selection_mouse +7711,21399575,"genie.py",5574,12,"video_tokens",python,selection_mouse +7712,21401583,"genie.py",5647,0,"",python,selection_mouse +7713,21401748,"genie.py",5641,14,"latent_actions",python,selection_mouse +7714,21402845,"genie.py",5580,0,"",python,selection_mouse +7715,21403020,"genie.py",5574,12,"video_tokens",python,selection_mouse +7716,21427633,"genie.py",5620,0,"",python,selection_mouse +7717,21429674,"genie.py",5182,0,"",python,selection_mouse +7718,21429872,"genie.py",5174,17,"tokenizer_outputs",python,selection_mouse +7719,21431338,"genie.py",5062,0,"",python,selection_mouse +7720,21431360,"genie.py",5061,0,"",python,selection_command +7721,21431522,"genie.py",5062,0,"",python,selection_mouse +7722,21431528,"genie.py",5061,0,"",python,selection_command +7723,21432163,"genie.py",5104,0,"",python,selection_mouse +7724,21432334,"genie.py",5096,9,"tokenizer",python,selection_mouse +7725,21432937,"genie.py",5062,0,"",python,selection_mouse +7726,21432956,"genie.py",5061,0,"",python,selection_command +7727,21433568,"genie.py",5062,0,"",python,selection_mouse +7728,21433574,"genie.py",5061,0,"",python,selection_command +7729,21434002,"genie.py",5062,0,"",python,selection_mouse +7730,21434014,"genie.py",5061,0,"",python,selection_command +7731,21434238,"genie.py",5062,0,"",python,selection_mouse +7732,21434239,"genie.py",5061,0,"",python,selection_command +7733,21434901,"genie.py",5111,0,"",python,selection_mouse +7734,21435102,"genie.py",5106,9,"vq_encode",python,selection_mouse +7735,21621569,"TERMINAL",0,0,"bash",,terminal_focus +7736,21644155,"TERMINAL",0,0,"queue",,terminal_command +7737,21644231,"TERMINAL",0,0,"]633;E;2025-08-14 16:38:17 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 16:38:17 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4 (Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2 (Priority)3412401 accelerat train_to tum_cte0 R 1-22:38:16\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-02:49:21\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 4:25:07\t 1 hkn07133421964 accelerat train_to tum_cte0 R 4:37:31\t 1 hkn0720",,terminal_output +7738,21645286,"TERMINAL",0,0,"87282",,terminal_output +7739,21646310,"TERMINAL",0,0,"98393",,terminal_output +7740,21647347,"TERMINAL",0,0,"2094104",,terminal_output +7741,21648393,"TERMINAL",0,0,"120515",,terminal_output +7742,21649431,"TERMINAL",0,0,"21626",,terminal_output +7743,21650466,"TERMINAL",0,0,"32737",,terminal_output +7744,21651508,"TERMINAL",0,0,"43848",,terminal_output +7745,21652558,"TERMINAL",0,0,"54959",,terminal_output +7746,21653685,"TERMINAL",0,0,"6530640",,terminal_output +7747,21654634,"TERMINAL",0,0,"76171",,terminal_output +7748,21655731,"TERMINAL",0,0,"87282",,terminal_output +7749,21656705,"TERMINAL",0,0,"98393",,terminal_output +7750,21657742,"TERMINAL",0,0,"3094204",,terminal_output +7751,21658916,"TERMINAL",0,0,"130515",,terminal_output +7752,21659937,"TERMINAL",0,0,"22737",,terminal_output +7753,21660955,"TERMINAL",0,0,"43848",,terminal_output +7754,21661924,"TERMINAL",0,0,"54959",,terminal_output +7755,21662964,"TERMINAL",0,0,"6540650",,terminal_output +7756,21664010,"TERMINAL",0,0,"76171",,terminal_output +7757,21665046,"TERMINAL",0,0,"87282",,terminal_output +7758,21666086,"TERMINAL",0,0,"98393",,terminal_output +7759,21667249,"TERMINAL",0,0,"4094304",,terminal_output +7760,21668172,"TERMINAL",0,0,"140515",,terminal_output +7761,21669253,"TERMINAL",0,0,"21626",,terminal_output +7762,21670384,"TERMINAL",0,0,"32737",,terminal_output +7763,21671400,"TERMINAL",0,0,"43848",,terminal_output +7764,21672422,"TERMINAL",0,0,"54959",,terminal_output +7765,21673449,"TERMINAL",0,0,"655068:00",,terminal_output +7766,21674412,"TERMINAL",0,0,"76171",,terminal_output +7767,21675460,"TERMINAL",0,0,"87282",,terminal_output +7768,21676311,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +7769,21680364,"TERMINAL",0,0,"idling",,terminal_command +7770,21680419,"TERMINAL",0,0,"]633;E;2025-08-14 16:38:53 idling;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Aug 14 16:38:53 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 249 nodes idle\rPartition dev_accelerated:\t 1 nodes idle\rPartition accelerated: 23 nodes idle\rPartition dev_accelerated-h100 :\t 1 nodes idle\rPartition accelerated-h100:\t 1 nodes idle\rPartition large:\t 8 nodes idle",,terminal_output +7771,21681451,"TERMINAL",0,0,"4\t ",,terminal_output +7772,21682048,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +7773,21693470,"TERMINAL",0,0,"queue",,terminal_command +7774,21693527,"TERMINAL",0,0,"]633;E;2025-08-14 16:39:06 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 16:39:06 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4 (Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2 (Priority)3412401 accelerat train_to tum_cte0 R 1-22:39:05\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-02:50:10\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 4:25:56\t 1 hkn07133421964 accelerat train_to tum_cte0 R 4:38:20\t 1 hkn0720",,terminal_output +7775,21694689,"TERMINAL",0,0,"76171",,terminal_output +7776,21695666,"TERMINAL",0,0,"87282",,terminal_output +7777,21696694,"TERMINAL",0,0,"98393",,terminal_output +7778,21697741,"TERMINAL",0,0,"10946:004",,terminal_output +7779,21698738,"TERMINAL",0,0,"110515",,terminal_output +7780,21699864,"TERMINAL",0,0,"21626",,terminal_output +7781,21700891,"TERMINAL",0,0,"33848",,terminal_output +7782,21701913,"TERMINAL",0,0,"54959",,terminal_output +7783,21702938,"TERMINAL",0,0,"6520630",,terminal_output +7784,21703951,"TERMINAL",0,0,"76171",,terminal_output +7785,21705011,"TERMINAL",0,0,"87282",,terminal_output +7786,21706050,"TERMINAL",0,0,"98393",,terminal_output +7787,21707088,"TERMINAL",0,0,"2094104",,terminal_output +7788,21708135,"TERMINAL",0,0,"120515",,terminal_output +7789,21709179,"TERMINAL",0,0,"21626",,terminal_output +7790,21710288,"TERMINAL",0,0,"32737",,terminal_output +7791,21711262,"TERMINAL",0,0,"43848",,terminal_output +7792,21712312,"TERMINAL",0,0,"54959",,terminal_output +7793,21713376,"TERMINAL",0,0,"6530640",,terminal_output +7794,21714426,"TERMINAL",0,0,"76171",,terminal_output +7795,21715436,"TERMINAL",0,0,"87282",,terminal_output +7796,21716486,"TERMINAL",0,0,"98393",,terminal_output +7797,21717514,"TERMINAL",0,0,"3094204",,terminal_output +7798,21718559,"TERMINAL",0,0,"130515",,terminal_output +7799,21719605,"TERMINAL",0,0,"21626",,terminal_output +7800,21720753,"TERMINAL",0,0,"32737",,terminal_output +7801,21721776,"TERMINAL",0,0,"43848",,terminal_output +7802,21722741,"TERMINAL",0,0,"54959",,terminal_output +7803,21723865,"TERMINAL",0,0,"6540650",,terminal_output +7804,21724953,"TERMINAL",0,0,"77282",,terminal_output +7805,21725909,"TERMINAL",0,0,"98393",,terminal_output +7806,21726948,"TERMINAL",0,0,"4094304",,terminal_output +7807,21728013,"TERMINAL",0,0,"140515",,terminal_output +7808,21729022,"TERMINAL",0,0,"21626",,terminal_output +7809,21730082,"TERMINAL",0,0,"32737",,terminal_output +7810,21731107,"TERMINAL",0,0,"43848",,terminal_output +7811,21732149,"TERMINAL",0,0,"54959",,terminal_output +7812,21733361,"TERMINAL",0,0,"655069:00",,terminal_output +7813,21734270,"TERMINAL",0,0,"76171",,terminal_output +7814,21735315,"TERMINAL",0,0,"87282",,terminal_output +7815,21736329,"TERMINAL",0,0,"98393",,terminal_output +7816,21737417,"TERMINAL",0,0,"5094404",,terminal_output +7817,21738424,"TERMINAL",0,0,"150515",,terminal_output +7818,21739496,"TERMINAL",0,0,"21626",,terminal_output +7819,21740518,"TERMINAL",0,0,"32737",,terminal_output +7820,21741562,"TERMINAL",0,0,"43848",,terminal_output +7821,21742670,"TERMINAL",0,0,"54959",,terminal_output +7822,21743693,"TERMINAL",0,0,"651:00610",,terminal_output +7823,21744695,"TERMINAL",0,0,"76171",,terminal_output +7824,21745844,"TERMINAL",0,0,"87282",,terminal_output +7825,21746899,"TERMINAL",0,0,"98393",,terminal_output +7826,21747829,"TERMINAL",0,0,"40:0040:005515",,terminal_output +7827,21748873,"TERMINAL",0,0,"21626",,terminal_output +7828,21749939,"TERMINAL",0,0,"32737",,terminal_output +7829,21750985,"TERMINAL",0,0,"43848",,terminal_output +7830,21752091,"TERMINAL",0,0,"54959",,terminal_output +7831,21753113,"TERMINAL",0,0,"6510620",,terminal_output +7832,21754091,"TERMINAL",0,0,"76171",,terminal_output +7833,21754709,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +7834,21758182,"TERMINAL",0,0,"bash",,terminal_focus +7835,21766346,"killer.sh",0,0,"",shellscript,tab +7836,21801344,"TERMINAL",0,0,"curl https://home.cit.tum.de/~nguyenal/",,terminal_command +7837,21801392,"TERMINAL",0,0,"]633;E;2025-08-14 16:40:54 curl https://home.cit.tum.de/~nguyenal/;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C",,terminal_output +7838,21801547,"TERMINAL",0,0,"\r\n\r\n\r\n\r\n \r\n \r\n Alfred Nguyen\r\n \r\n\r\n\r\n\r\n\r\n

Alfred Nguyen

\r\n

\r\n Hi, I'm Alfred, a researcher in machine learning & foundation models.
\r\n

\r\n\r\n

\r\n currently working at helmholtz munich and p(doom)\r\n

\r\n\r\n

\r\n you can find me here: [\r\n github ·\r\n linkedin ·\r\n email\r\n ]\r\n

\r\n\r\n
\r\n

now

\r\n
    \r\n\r\n
  • \r\n [2025] research assistant at bauer lab, helmholtz munich
    \r\n \r\n
  • \r\n\r\n
  • \r\n [2025] bachelor thesis on scaling test-time training for inference-time adaptation on the arc-agi benchmark
    \r\n \r\n
  • \r\n\r\n
  • \r\n [2025] agi researcher at p(doom)
    \r\n \r\n
  • \r\n\r\n
\r\n
\r\n\r\n
\r\n

selected work

\r\n
    \r\n
  • \r\n hyground: a multi-agent AI system for automated incident response
    \r\n 2024–2025 — developed @ maibornwolff, Python, Multi-agent Systems and more\r\n
  • \r\n
  • \r\n adaptive compute transformer thoughtsblog
    \r\n 2023 — early exploration into dynamic compute scaling\r\n
  • \r\n
  • \r\n matching markets seminar paperpdf
    \r\n 2022 — undergrad research\r\n
  • \r\n
\r\n
\r\n\r\n
\r\n

previously

\r\n
    \r\n
  • \r\n ai engineer @ maibornwolff
    \r\n 2024–2025 — development of Hyground, a multi-agent AI system for automated incident\r\n response\r\n
  • \r\n
  • \r\n software developer @ arrk engineering
    \r\n 2023–2024 — development of an embedded infotainment system (C++, Python and more)\r\n
  • \r\n
  • \r\n intern @ bosch
    \r\n 2017 — designing switching circuitry and programming microcontrollers using Arduino\r\n
  • \r\n
\r\n
\r\n\r\n\r\n\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer]633;D;0",,terminal_output +7839,21804792,"killer.sh",0,0,"",shellscript,tab +7840,21847853,"killer.sh",0,0,"response=$(curl -s https://home.cit.tum.de/~nguyenal/)\n\n",shellscript,content +7841,21847858,"killer.sh",56,0,"\n",shellscript,content +7842,21847949,"killer.sh",56,0,"if echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n",shellscript,content +7843,21848018,"killer.sh",108,0," echo ""killing all jobs""\n",shellscript,content +7844,21848074,"killer.sh",136,0,"else\n",shellscript,content +7845,21848208,"killer.sh",141,0," echo ""continuing jobs""\n",shellscript,content +7846,21848317,"killer.sh",168,0,"fi\n",shellscript,content +7847,21848320,"killer.sh",171,1,"",shellscript,content +7848,21865399,"TERMINAL",0,0,"dev",,terminal_command +7849,21865437,"TERMINAL",0,0,"]633;E;2025-08-14 16:41:58 dev;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +7850,21870035,"TERMINAL",0,0,"sh killer.sh",,terminal_command +7851,21870112,"TERMINAL",0,0,"]633;E;2025-08-14 16:42:03 sh killer.sh ;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C",,terminal_output +7852,21870182,"TERMINAL",0,0,"continuing jobs\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +7853,21872407,"killer.sh",0,0,"",shellscript,tab +7854,21873260,"killer.sh",153,0,"",shellscript,selection_mouse +7855,21873434,"killer.sh",151,10,"continuing",shellscript,selection_mouse +7856,21874870,"killer.sh",151,10,"",shellscript,content +7857,21875631,"killer.sh",151,0,"c",shellscript,content +7858,21875632,"killer.sh",152,0,"",shellscript,selection_keyboard +7859,21875708,"killer.sh",152,0,"o",shellscript,content +7860,21875709,"killer.sh",153,0,"",shellscript,selection_keyboard +7861,21875892,"killer.sh",153,0,"n",shellscript,content +7862,21875893,"killer.sh",154,0,"",shellscript,selection_keyboard +7863,21876161,"killer.sh",154,0,"t",shellscript,content +7864,21876162,"killer.sh",155,0,"",shellscript,selection_keyboard +7865,21876268,"killer.sh",155,0,"i",shellscript,content +7866,21876269,"killer.sh",156,0,"",shellscript,selection_keyboard +7867,21876590,"killer.sh",156,0,"n",shellscript,content +7868,21876591,"killer.sh",157,0,"",shellscript,selection_keyboard +7869,21876726,"killer.sh",157,0,"u",shellscript,content +7870,21876727,"killer.sh",158,0,"",shellscript,selection_keyboard +7871,21876864,"killer.sh",158,0,"e",shellscript,content +7872,21876865,"killer.sh",159,0,"",shellscript,selection_keyboard +7873,21877045,"killer.sh",159,0,"i",shellscript,content +7874,21877045,"killer.sh",160,0,"",shellscript,selection_keyboard +7875,21877153,"killer.sh",160,0,"n",shellscript,content +7876,21877154,"killer.sh",161,0,"",shellscript,selection_keyboard +7877,21877247,"killer.sh",161,0,"g",shellscript,content +7878,21877248,"killer.sh",162,0,"",shellscript,selection_keyboard +7879,21877416,"killer.sh",162,0," ",shellscript,content +7880,21877417,"killer.sh",163,0,"",shellscript,selection_keyboard +7881,21878008,"killer.sh",151,12,"",shellscript,content +7882,21880824,"killer.sh",151,0,"l",shellscript,content +7883,21880825,"killer.sh",152,0,"",shellscript,selection_keyboard +7884,21880999,"killer.sh",152,0,"e",shellscript,content +7885,21881000,"killer.sh",153,0,"",shellscript,selection_keyboard +7886,21881503,"killer.sh",153,0,"a",shellscript,content +7887,21881504,"killer.sh",154,0,"",shellscript,selection_keyboard +7888,21881711,"killer.sh",154,0,"v",shellscript,content +7889,21881711,"killer.sh",155,0,"",shellscript,selection_keyboard +7890,21881805,"killer.sh",155,0,"i",shellscript,content +7891,21881805,"killer.sh",156,0,"",shellscript,selection_keyboard +7892,21881941,"killer.sh",156,0,"n",shellscript,content +7893,21881941,"killer.sh",157,0,"",shellscript,selection_keyboard +7894,21882028,"killer.sh",157,0,"g",shellscript,content +7895,21882029,"killer.sh",158,0,"",shellscript,selection_keyboard +7896,21883476,"killer.sh",163,0," ",shellscript,content +7897,21883477,"killer.sh",164,0,"",shellscript,selection_keyboard +7898,21883601,"killer.sh",164,0,"b",shellscript,content +7899,21883602,"killer.sh",165,0,"",shellscript,selection_keyboard +7900,21883682,"killer.sh",165,0,"e",shellscript,content +7901,21883683,"killer.sh",166,0,"",shellscript,selection_keyboard +7902,21884659,"killer.sh",167,0,"",shellscript,selection_mouse +7903,21885184,"killer.sh",171,0,"",shellscript,selection_mouse +7904,21885370,"killer.sh",170,1,"\n",shellscript,selection_mouse +7905,21885371,"killer.sh",157,14,"g jobs be""\nfi\n",shellscript,selection_mouse +7906,21885388,"killer.sh",140,31,"\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7907,21885413,"killer.sh",116,55," ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7908,21885414,"killer.sh",115,56,"o ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7909,21885433,"killer.sh",60,111,"cho ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7910,21885452,"killer.sh",59,112,"echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7911,21885499,"killer.sh",55,116,"\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7912,21885543,"killer.sh",0,171,"response=$(curl -s https://home.cit.tum.de/~nguyenal/)\n\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7913,21886570,"killer.sh",62,0,"",shellscript,selection_mouse +7914,21887616,"killer.sh",88,0,"",shellscript,selection_mouse +7915,21887751,"killer.sh",87,4,"KILL",shellscript,selection_mouse +7916,21887947,"killer.sh",87,5,"KILL ",shellscript,selection_mouse +7917,21887984,"killer.sh",87,8,"KILL ALL",shellscript,selection_mouse +7918,21888019,"killer.sh",87,9,"KILL ALL ",shellscript,selection_mouse +7919,21888381,"killer.sh",87,13,"KILL ALL JOBS",shellscript,selection_mouse +7920,21903475,"killer.sh",171,0,"",shellscript,selection_mouse +7921,21904643,"killer.sh",89,0,"",shellscript,selection_mouse +7922,21904783,"killer.sh",87,4,"KILL",shellscript,selection_mouse +7923,21904980,"killer.sh",87,8,"KILL ALL",shellscript,selection_mouse +7924,21905020,"killer.sh",87,9,"KILL ALL ",shellscript,selection_mouse +7925,21905057,"killer.sh",87,13,"KILL ALL JOBS",shellscript,selection_mouse +7926,21915416,"killer.sh",171,0,"",shellscript,selection_mouse +7927,21920461,"killer.sh",170,1,"\n",shellscript,selection_mouse +7928,21920488,"killer.sh",145,26,"echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7929,21920525,"killer.sh",140,31,"\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7930,21920563,"killer.sh",112,59,"echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7931,21920686,"killer.sh",59,112,"echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7932,21920724,"killer.sh",55,116,"\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7933,21920863,"killer.sh",0,171,"response=$(curl -s https://home.cit.tum.de/~nguyenal/)\n\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7934,21921247,"killer.sh",2,0,"",shellscript,selection_mouse +7935,21921377,"killer.sh",0,8,"response",shellscript,selection_mouse +7936,21921582,"killer.sh",0,55,"response=$(curl -s https://home.cit.tum.de/~nguyenal/)\n",shellscript,selection_mouse +7937,21921612,"killer.sh",0,63,"response=$(curl -s https://home.cit.tum.de/~nguyenal/)\n\nif echo",shellscript,selection_mouse +7938,21921664,"killer.sh",0,116,"response=$(curl -s https://home.cit.tum.de/~nguyenal/)\n\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo",shellscript,selection_mouse +7939,21921748,"killer.sh",0,140,"response=$(curl -s https://home.cit.tum.de/~nguyenal/)\n\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse",shellscript,selection_mouse +7940,21921749,"killer.sh",0,150,"response=$(curl -s https://home.cit.tum.de/~nguyenal/)\n\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ",shellscript,selection_mouse +7941,21921749,"killer.sh",0,170,"response=$(curl -s https://home.cit.tum.de/~nguyenal/)\n\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi",shellscript,selection_mouse +7942,21921973,"killer.sh",170,0,"",shellscript,selection_mouse +7943,21922167,"killer.sh",168,2,"fi",shellscript,selection_mouse +7944,21922346,"killer.sh",150,20,"""leaving jobs be""\nfi",shellscript,selection_mouse +7945,21922361,"killer.sh",140,30,"\n echo ""leaving jobs be""\nfi",shellscript,selection_mouse +7946,21922381,"killer.sh",112,58,"echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi",shellscript,selection_mouse +7947,21922401,"killer.sh",59,111,"echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi",shellscript,selection_mouse +7948,21922415,"killer.sh",55,115,"\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi",shellscript,selection_mouse +7949,21922498,"killer.sh",0,170,"response=$(curl -s https://home.cit.tum.de/~nguyenal/)\n\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi",shellscript,selection_mouse +7950,21923965,"killer.sh",171,0,"",shellscript,selection_mouse +7951,21926341,"killer.sh",170,1,"\n",shellscript,selection_mouse +7952,21926356,"killer.sh",140,31,"\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7953,21926392,"killer.sh",110,61," echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7954,21926428,"killer.sh",57,114,"f echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7955,21926428,"killer.sh",56,115,"if echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7956,21926429,"killer.sh",55,116,"\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7957,21926512,"killer.sh",0,171,"response=$(curl -s https://home.cit.tum.de/~nguyenal/)\n\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +7958,21927210,"killer.sh",0,0,"",shellscript,selection_command +7959,21934406,"killer.sh",44,0,"",shellscript,selection_mouse +7960,21934903,"killer.sh",49,0,"",shellscript,selection_mouse +7961,21935068,"killer.sh",44,8,"nguyenal",shellscript,selection_mouse +7962,21935729,"killer.sh",170,0,"",shellscript,selection_mouse +7963,22026420,"killer.sh",19,0,"",shellscript,selection_mouse +7964,22026537,"killer.sh",19,5,"https",shellscript,selection_mouse +7965,22026704,"killer.sh",19,6,"https:",shellscript,selection_mouse +7966,22026732,"killer.sh",19,8,"https://",shellscript,selection_mouse +7967,22026733,"killer.sh",19,12,"https://home",shellscript,selection_mouse +7968,22026819,"killer.sh",19,13,"https://home.",shellscript,selection_mouse +7969,22026820,"killer.sh",19,16,"https://home.cit",shellscript,selection_mouse +7970,22026904,"killer.sh",19,17,"https://home.cit.",shellscript,selection_mouse +7971,22026923,"killer.sh",19,20,"https://home.cit.tum",shellscript,selection_mouse +7972,22027024,"killer.sh",19,21,"https://home.cit.tum.",shellscript,selection_mouse +7973,22027025,"killer.sh",19,23,"https://home.cit.tum.de",shellscript,selection_mouse +7974,22027056,"killer.sh",19,24,"https://home.cit.tum.de/",shellscript,selection_mouse +7975,22027073,"killer.sh",19,25,"https://home.cit.tum.de/~",shellscript,selection_mouse +7976,22027102,"killer.sh",19,33,"https://home.cit.tum.de/~nguyenal",shellscript,selection_mouse +7977,22027964,"killer.sh",19,33,"",shellscript,content +7978,22028107,"killer.sh",19,1,"",shellscript,content +7979,22028689,"killer.sh",19,0,"https://home.cit.tum.de/~mahajanm/test.html",shellscript,content +7980,22029477,"killer.sh",61,1,"",shellscript,content +7981,22029628,"killer.sh",60,1,"",shellscript,content +7982,22030664,"killer.sh",60,0,"m",shellscript,content +7983,22030664,"killer.sh",61,0,"",shellscript,selection_keyboard +7984,22030912,"killer.sh",61,0,"l",shellscript,content +7985,22030914,"killer.sh",62,0,"",shellscript,selection_keyboard +7986,22031423,"killer.sh",62,0,"\n",shellscript,content +7987,22032022,"killer.sh",62,1,"",shellscript,content +7988,22035944,"TERMINAL",0,0,"sh killer.sh",,terminal_command +7989,22035992,"TERMINAL",0,0,"]633;E;2025-08-14 16:44:49 sh killer.sh ;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C",,terminal_output +7990,22036143,"TERMINAL",0,0,"leaving jobs be\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +7991,22045473,"TERMINAL",0,0,"watch --help",,terminal_command +7992,22045570,"TERMINAL",0,0,"]633;E;2025-08-14 16:44:58 watch --help;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C\r\nUsage:\r\n watch [options] command\r\n\r\nOptions:\r\n -b, --beep beep if command has a non-zero exit\r\n -c, --color interpret ANSI color and style sequences\r\n -d, --differences[=]\r\n highlight changes between updates\r\n -e, --errexit exit if command has a non-zero exit\r\n -g, --chgexit exit when output from command changes\r\n -n, --interval seconds to wait between updates\r\n -p, --precise attempt run command in precise intervals\r\n -t, --no-title turn off header\r\n -w, --no-wrap turn off line wrapping\r\n -x, --exec pass command to exec instead of ""sh -c""\r\n\r\n -h, --help display this help and exit\r\n -v, --version output version information and exit\r\n\r\nFor more details see watch(1).\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar[?2004h",,terminal_output +7993,22061721,"killer.sh",0,0,"",shellscript,tab +7994,22080293,"TERMINAL",0,0,"watch -n10 ""sh killer.sh""",,terminal_command +7995,22080344,"TERMINAL",0,0,"]633;E;2025-08-14 16:45:33 watch -n10 ""sh killer.sh"";cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C",,terminal_output +7996,22080492,"TERMINAL",0,0,"[?1049h(B[?7hEvery 10.0s: sh killer.shhkn1993.localdomain: Thu Aug 14 16:45:33 2025leaving jobs be",,terminal_output +7997,22090613,"TERMINAL",0,0,"4\t ",,terminal_output +7998,22096403,"killer.sh",0,0,"",shellscript,tab +7999,22098556,"killer.sh",96,0,"",shellscript,selection_mouse +8000,22098698,"killer.sh",96,4,"KILL",shellscript,selection_mouse +8001,22098872,"killer.sh",96,5,"KILL ",shellscript,selection_mouse +8002,22098896,"killer.sh",96,8,"KILL ALL",shellscript,selection_mouse +8003,22098943,"killer.sh",96,9,"KILL ALL ",shellscript,selection_mouse +8004,22099031,"killer.sh",96,13,"KILL ALL JOBS",shellscript,selection_mouse +8005,22100766,"TERMINAL",0,0,"5\t ",,terminal_output +8006,22111111,"TERMINAL",0,0,"6:0\t ",,terminal_output +8007,22121018,"TERMINAL",0,0,"14\t ",,terminal_output +8008,22131252,"TERMINAL",0,0,"2\rkilling all jobs",,terminal_output +8009,22141319,"TERMINAL",0,0,"3\t ",,terminal_output +8010,22141866,"killer.sh",0,0,"",shellscript,tab +8011,22142574,"killer.sh",180,0,"",shellscript,selection_mouse +8012,22143132,"killer.sh",149,0,"",shellscript,selection_mouse +8013,22143776,"killer.sh",64,0,"",shellscript,selection_mouse +8014,22144101,"killer.sh",64,36,"\nif echo ""$response"" | grep -q ""KILL",shellscript,selection_mouse +8015,22144127,"killer.sh",64,37,"\nif echo ""$response"" | grep -q ""KILL ",shellscript,selection_mouse +8016,22144154,"killer.sh",64,40,"\nif echo ""$response"" | grep -q ""KILL ALL",shellscript,selection_mouse +8017,22144551,"killer.sh",104,0,"",shellscript,selection_mouse +8018,22151454,"TERMINAL",0,0,"4\t ",,terminal_output +8019,22153780,"killer.sh",180,0,"",shellscript,selection_mouse +8020,22161570,"TERMINAL",0,0,"5\t ",,terminal_output +8021,22171672,"TERMINAL",0,0,"7:0\rleaving jobs be",,terminal_output +8022,22181794,"TERMINAL",0,0,"1\t ",,terminal_output +8023,22192006,"TERMINAL",0,0,"2\t ",,terminal_output +8024,22202065,"TERMINAL",0,0,"35\t ",,terminal_output +8025,22212194,"TERMINAL",0,0,"4\t ",,terminal_output +8026,22220942,"killer.sh",57,0,"",shellscript,selection_mouse +8027,22221546,"killer.sh",56,1,"",shellscript,content +8028,22221685,"killer.sh",55,1,"",shellscript,content +8029,22221789,"killer.sh",54,1,"",shellscript,content +8030,22221896,"killer.sh",53,1,"",shellscript,content +8031,22222331,"killer.sh",53,0,"k",shellscript,content +8032,22222332,"killer.sh",54,0,"",shellscript,selection_keyboard +8033,22222407,"TERMINAL",0,0,"5\t ",,terminal_output +8034,22222490,"killer.sh",54,0,"i",shellscript,content +8035,22222491,"killer.sh",55,0,"",shellscript,selection_keyboard +8036,22222670,"killer.sh",55,0,"l",shellscript,content +8037,22222671,"killer.sh",56,0,"",shellscript,selection_keyboard +8038,22222779,"killer.sh",56,0,"l",shellscript,content +8039,22222780,"killer.sh",57,0,"",shellscript,selection_keyboard +8040,22222872,"killer.sh",57,0,"e",shellscript,content +8041,22222873,"killer.sh",58,0,"",shellscript,selection_keyboard +8042,22222978,"killer.sh",58,0,"r",shellscript,content +8043,22222979,"killer.sh",59,0,"",shellscript,selection_keyboard +8044,22226187,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +8045,22230624,"TERMINAL",0,0,"sh killer.sh",,terminal_command +8046,22230678,"TERMINAL",0,0,"]633;E;2025-08-14 16:48:03 sh killer.sh;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C",,terminal_output +8047,22230785,"TERMINAL",0,0,"killing all jobs\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +8048,22260529,"killer.sh",0,0,"",shellscript,tab +8049,22261082,"killer.sh",182,0,"",shellscript,selection_mouse +8050,22262490,"killer.sh",52,0,"",shellscript,selection_mouse +8051,22262609,"killer.sh",50,2,"nm",shellscript,selection_mouse +8052,22262625,"killer.sh",49,3,"anm",shellscript,selection_mouse +8053,22262641,"killer.sh",48,4,"janm",shellscript,selection_mouse +8054,22262679,"killer.sh",47,5,"ajanm",shellscript,selection_mouse +8055,22262756,"killer.sh",46,6,"hajanm",shellscript,selection_mouse +8056,22262835,"killer.sh",45,7,"ahajanm",shellscript,selection_mouse +8057,22263233,"killer.sh",44,8,"mahajanm",shellscript,selection_mouse +8058,22263860,"killer.sh",44,8,"n",shellscript,content +8059,22263861,"killer.sh",45,0,"",shellscript,selection_keyboard +8060,22264169,"killer.sh",45,0,"g",shellscript,content +8061,22264170,"killer.sh",46,0,"",shellscript,selection_keyboard +8062,22264283,"killer.sh",46,0,"u",shellscript,content +8063,22264284,"killer.sh",47,0,"",shellscript,selection_keyboard +8064,22264493,"killer.sh",47,0,"y",shellscript,content +8065,22264494,"killer.sh",48,0,"",shellscript,selection_keyboard +8066,22264713,"killer.sh",48,0,"e",shellscript,content +8067,22264713,"killer.sh",49,0,"",shellscript,selection_keyboard +8068,22264786,"killer.sh",49,0,"n",shellscript,content +8069,22264787,"killer.sh",50,0,"",shellscript,selection_keyboard +8070,22264956,"killer.sh",50,0,"a",shellscript,content +8071,22264957,"killer.sh",51,0,"",shellscript,selection_keyboard +8072,22265007,"killer.sh",51,0,"l",shellscript,content +8073,22265008,"killer.sh",52,0,"",shellscript,selection_keyboard +8074,22265804,"killer.sh",178,0,"",shellscript,selection_mouse +8075,22273178,"TERMINAL",0,0,"sh killer.sh",,terminal_command +8076,22273236,"TERMINAL",0,0,"]633;E;2025-08-14 16:48:46 sh killer.sh;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C",,terminal_output +8077,22273382,"TERMINAL",0,0,"leaving jobs be\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +8078,22348376,"killer.sh",0,0,"",shellscript,tab +8079,22350549,"killer.sh",99,0,"",shellscript,selection_mouse +8080,22350692,"killer.sh",98,4,"KILL",shellscript,selection_mouse +8081,22350889,"killer.sh",98,5,"KILL ",shellscript,selection_mouse +8082,22350902,"killer.sh",98,8,"KILL ALL",shellscript,selection_mouse +8083,22350937,"killer.sh",98,9,"KILL ALL ",shellscript,selection_mouse +8084,22350980,"killer.sh",98,13,"KILL ALL JOBS",shellscript,selection_mouse +8085,22432958,"killer.sh",182,0,"",shellscript,selection_mouse +8086,22437487,"TERMINAL",0,0,"bash",,terminal_focus +8087,22492972,"TERMINAL",0,0,"bash",,terminal_focus +8088,22529687,"TERMINAL",0,0,"tmux",,terminal_command +8089,22529775,"TERMINAL",0,0,"]633;E;2025-08-14 16:53:02 tmux;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h[?1h=[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[>c[>q[?25l\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[0] 0:bash* ""hkn1993.localdomain"" 16:53 14-Aug-25(B[?12l[?25h(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[?25l\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[0] 0:bash* ""hkn1993.localdomain"" 16:53 14-Aug-25(B[?12l[?25h",,terminal_output +8090,22529874,"TERMINAL",0,0,"[?7727h",,terminal_output +8091,22531549,"TERMINAL",0,0,"[?2004h(jafar) [tum_cte0515@hkn1993 jafar]$ ",,terminal_output +8092,22534383,"TERMINAL",0,0,"w",,terminal_output +8093,22534668,"TERMINAL",0,0,"a",,terminal_output +8094,22534775,"TERMINAL",0,0,"t",,terminal_output +8095,22534847,"TERMINAL",0,0,"c",,terminal_output +8096,22534942,"TERMINAL",0,0,"h",,terminal_output +8097,22535059,"TERMINAL",0,0," ",,terminal_output +8098,22535300,"TERMINAL",0,0,"-",,terminal_output +8099,22535548,"TERMINAL",0,0,"n",,terminal_output +8100,22535927,"TERMINAL",0,0,"1",,terminal_output +8101,22536824,"TERMINAL",0,0," ",,terminal_output +8102,22539577,"TERMINAL",0,0,"""",,terminal_output +8103,22539899,"TERMINAL",0,0,"s",,terminal_output +8104,22539989,"TERMINAL",0,0,"h",,terminal_output +8105,22540106,"TERMINAL",0,0," ",,terminal_output +8106,22540532,"TERMINAL",0,0,"k",,terminal_output +8107,22540739,"TERMINAL",0,0,"i",,terminal_output +8108,22541046,"TERMINAL",0,0,"l",,terminal_output +8109,22541155,"TERMINAL",0,0,"l",,terminal_output +8110,22541241,"TERMINAL",0,0,"e",,terminal_output +8111,22541419,"TERMINAL",0,0,"r.",,terminal_output +8112,22541662,"TERMINAL",0,0,"s",,terminal_output +8113,22541744,"TERMINAL",0,0,"h",,terminal_output +8114,22542114,"TERMINAL",0,0,"""",,terminal_output +8115,22543473,"TERMINAL",0,0,"",,terminal_output +8116,22543688,"TERMINAL",0,0,"",,terminal_output +8117,22543773,"TERMINAL",0,0,"",,terminal_output +8118,22544102,"TERMINAL",0,0,"",,terminal_output +8119,22544737,"TERMINAL",0,0,"",,terminal_output +8120,22544867,"TERMINAL",0,0,"",,terminal_output +8121,22545295,"TERMINAL",0,0,"",,terminal_output +8122,22545837,"TERMINAL",0,0,"[1@5",,terminal_output +8123,22545902,"TERMINAL",0,0,"[1@0",,terminal_output +8124,22546346,"TERMINAL",0,0,"",,terminal_output +8125,22546414,"TERMINAL",0,0,"",,terminal_output +8126,22547078,"TERMINAL",0,0,"[1@6[1@0",,terminal_output +8127,22562217,"killer.sh",0,0,"",shellscript,tab +8128,22566800,"killer.sh",0,0,"",shellscript,tab +8129,22571478,"TERMINAL",0,0,"\r\n[?2004l",,terminal_output +8130,22571619,"TERMINAL",0,0,"[?25lEvery 60.0s: sh killer.shhkn1993.localdomain: Thu Aug 14 16:53:44 2025\r\nleaving jobs be\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[?12l[?25h",,terminal_output +8131,22574772,"TERMINAL",0,0,"[?25l\r\n[0] 0:watch* ""hkn1993.localdomain"" 16:53 14-Aug-25(B[?12l[?25h",,terminal_output +8132,22576420,"TERMINAL",0,0,"(B[?1l>[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l[?7727l[?1004l[?1049l[detached (from session 0)]\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +8133,22579601,"TERMINAL",0,0,"tmux -q",,terminal_command +8134,22579649,"TERMINAL",0,0,"]633;E;2025-08-14 16:53:52 tmux -q;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h[?1h=[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[>c[>q[?25l\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[1] 0:tmux* ""hkn1993.localdomain"" 16:53 14-Aug-25(B[?12l[?25h(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[?25l\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[1] 0:tmux* ""hkn1993.localdomain"" 16:53 14-Aug-25(B[?12l[?25h",,terminal_output +8135,22579723,"TERMINAL",0,0,"[?7727h",,terminal_output +8136,22581232,"TERMINAL",0,0,"(jafar) [tum_cte0515@hkn1993 jafar]$ [?2004h[?25l[1] 0:bash* ""hkn1993.localdomain"" 16:53 14-Aug-25(B[?12l[?25h",,terminal_output +8137,22583047,"TERMINAL",0,0,"\r\nlogout\r\n[?2004l(B[?1l>[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l[?7727l[?1004l[?1049l[exited]\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +8138,22586306,"TERMINAL",0,0,"tmux a",,terminal_command +8139,22586376,"TERMINAL",0,0,"]633;E;2025-08-14 16:53:59 tmux a;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h[?1h=[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[>c[>q[?25lEvery 60.0s: sh killer.shhkn1993.localdomain: Thu Aug 14 16:53:44 2025\r\nleaving jobs be\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[0] 0:watch* ""hkn1993.localdomain"" 16:53 14-Aug-25(B[?12l[?25h(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[?25lEvery 60.0s: sh killer.shhkn1993.localdomain: Thu Aug 14 16:53:44 2025\r\nleaving jobs be\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[0] 0:watch* ""hkn1993.localdomain"" 16:53 14-Aug-25(B[?12l[?25h",,terminal_output +8140,22586472,"TERMINAL",0,0,"[?7727h",,terminal_output +8141,22589136,"TERMINAL",0,0,"[?25l(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[?12l[?25h[?25l[0] 0:bash* ""hkn1993.localdomain"" 16:54 14-Aug-25(B[?12l[?25h(jafar) [tum_cte0515@hkn1993 jafar]$ [?2004h",,terminal_output +8142,22590410,"killer.sh",0,0,"",shellscript,tab +8143,22591702,"killer.sh",129,0,"",shellscript,selection_mouse +8144,22591838,"killer.sh",129,7,"killing",shellscript,selection_mouse +8145,22592092,"killer.sh",129,11,"killing all",shellscript,selection_mouse +8146,22592128,"killer.sh",129,12,"killing all ",shellscript,selection_mouse +8147,22592140,"killer.sh",129,16,"killing all jobs",shellscript,selection_mouse +8148,22592241,"killer.sh",129,17,"killing all jobs""",shellscript,selection_mouse +8149,22592802,"killer.sh",146,0,"",shellscript,selection_mouse +8150,22593348,"killer.sh",118,0,"",shellscript,selection_mouse +8151,22594405,"killer.sh",118,0,"\n",shellscript,content +8152,22595530,"killer.sh",119,0,"s",shellscript,content +8153,22595531,"killer.sh",120,0,"",shellscript,selection_keyboard +8154,22596027,"killer.sh",119,1,"",shellscript,content +8155,22596177,"killer.sh",119,0," ",shellscript,content +8156,22596797,"killer.sh",123,0,"e",shellscript,content +8157,22596798,"killer.sh",124,0,"",shellscript,selection_keyboard +8158,22596991,"killer.sh",124,0,"c",shellscript,content +8159,22596992,"killer.sh",125,0,"",shellscript,selection_keyboard +8160,22597045,"killer.sh",125,0,"h",shellscript,content +8161,22597046,"killer.sh",126,0,"",shellscript,selection_keyboard +8162,22597210,"killer.sh",126,0,"o",shellscript,content +8163,22597210,"killer.sh",127,0,"",shellscript,selection_keyboard +8164,22597325,"killer.sh",127,0," ",shellscript,content +8165,22597326,"killer.sh",128,0,"",shellscript,selection_keyboard +8166,22597672,"killer.sh",128,0,"""",shellscript,content +8167,22597673,"killer.sh",129,0,"",shellscript,selection_keyboard +8168,22598325,"killer.sh",129,0,"h",shellscript,content +8169,22598325,"killer.sh",130,0,"",shellscript,selection_keyboard +8170,22598711,"killer.sh",130,0,"l",shellscript,content +8171,22598712,"killer.sh",131,0,"",shellscript,selection_keyboard +8172,22599061,"killer.sh",130,1,"",shellscript,content +8173,22599571,"killer.sh",130,0,"e",shellscript,content +8174,22599572,"killer.sh",131,0,"",shellscript,selection_keyboard +8175,22599608,"killer.sh",131,0,"l",shellscript,content +8176,22599609,"killer.sh",132,0,"",shellscript,selection_keyboard +8177,22599753,"killer.sh",132,0,"l",shellscript,content +8178,22599754,"killer.sh",133,0,"",shellscript,selection_keyboard +8179,22599911,"killer.sh",133,0,"o",shellscript,content +8180,22599912,"killer.sh",134,0,"",shellscript,selection_keyboard +8181,22599994,"killer.sh",134,0," ",shellscript,content +8182,22599994,"killer.sh",135,0,"",shellscript,selection_keyboard +8183,22600079,"killer.sh",135,0,"w",shellscript,content +8184,22600080,"killer.sh",136,0,"",shellscript,selection_keyboard +8185,22600161,"killer.sh",136,0,"o",shellscript,content +8186,22600162,"killer.sh",137,0,"",shellscript,selection_keyboard +8187,22600302,"killer.sh",137,0,"r",shellscript,content +8188,22600302,"killer.sh",138,0,"",shellscript,selection_keyboard +8189,22600385,"killer.sh",138,0,"l",shellscript,content +8190,22600386,"killer.sh",139,0,"",shellscript,selection_keyboard +8191,22600517,"killer.sh",139,0,"d",shellscript,content +8192,22600518,"killer.sh",140,0,"",shellscript,selection_keyboard +8193,22600760,"killer.sh",140,0,"""",shellscript,content +8194,22600761,"killer.sh",141,0,"",shellscript,selection_keyboard +8195,22601448,"killer.sh",140,0,"",shellscript,selection_command +8196,22601978,"killer.sh",119,23,"",shellscript,content +8197,22602014,"killer.sh",123,0,"",shellscript,selection_command +8198,22602470,"killer.sh",150,0,"",shellscript,selection_command +8199,22602688,"killer.sh",151,0,"\n echo ""hello world""",shellscript,content +8200,22602694,"killer.sh",156,0,"",shellscript,selection_command +8201,22605522,"TERMINAL",0,0,"watch -n60 ""sh killer.sh""",,terminal_output +8202,22606860,"TERMINAL",0,0,"\r\n[?2004l",,terminal_output +8203,22606997,"TERMINAL",0,0,"[?25lEvery 60.0s: sh killer.shhkn1993.localdomain: Thu Aug 14 16:54:20 2025\r\nhello world\r\nleaving jobs be\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[?12l[?25h",,terminal_output +8204,22609492,"TERMINAL",0,0,"[?25l\r\n[0] 0:watch* ""hkn1993.localdomain"" 16:54 14-Aug-25(B[?12l[?25h[?25l(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[?12l[?25h[?25l[0] 0:bash* ""hkn1993.localdomain"" 16:54 14-Aug-25(B[?12l[?25h[?2004h(jafar) [tum_cte0515@hkn1993 jafar]$ ",,terminal_output +8205,22612066,"killer.sh",0,0,"",shellscript,tab +8206,22613654,"killer.sh",164,0,"",shellscript,selection_mouse +8207,22615064,"killer.sh",162,0,"",shellscript,selection_mouse +8208,22615690,"killer.sh",174,0,"",shellscript,selection_mouse +8209,22615700,"killer.sh",173,0,"",shellscript,selection_command +8210,22616110,"killer.sh",174,0,"",shellscript,selection_command +8211,22616406,"killer.sh",174,0," ",shellscript,content +8212,22616407,"killer.sh",175,0,"",shellscript,selection_keyboard +8213,22617459,"killer.sh",175,0,"$",shellscript,content +8214,22617460,"killer.sh",176,0,"",shellscript,selection_keyboard +8215,22618034,"killer.sh",175,1,"",shellscript,content +8216,22619196,"killer.sh",175,0,"&",shellscript,content +8217,22619197,"killer.sh",176,0,"",shellscript,selection_keyboard +8218,22619351,"killer.sh",176,0,"&",shellscript,content +8219,22619352,"killer.sh",177,0,"",shellscript,selection_keyboard +8220,22619692,"killer.sh",177,0," ",shellscript,content +8221,22619692,"killer.sh",178,0,"",shellscript,selection_keyboard +8222,22619948,"killer.sh",178,0,"\",shellscript,content +8223,22619949,"killer.sh",179,0,"",shellscript,selection_keyboard +8224,22622360,"TERMINAL",0,0,"watch -n60 ""sh killer.sh""",,terminal_output +8225,22622716,"TERMINAL",0,0,"\r\n[?2004l",,terminal_output +8226,22622847,"TERMINAL",0,0,"[?25lEvery 60.0s: sh killer.shhkn1993.localdomain: Thu Aug 14 16:54:35 2025\r\nhello world\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[0] 0:watch* ""hkn1993.localdomain"" 16:54 14-Aug-25(B[?12l[?25hleaving jobs be",,terminal_output +8227,22623954,"TERMINAL",0,0,"[?25l(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[?12l[?25h[?25l[0] 0:bash* ""hkn1993.localdomain"" 16:54 14-Aug-25(B[?12l[?25h(jafar) [tum_cte0515@hkn1993 jafar]$ [?2004h",,terminal_output +8228,22624775,"killer.sh",0,0,"",shellscript,tab +8229,22625608,"killer.sh",167,0,"",shellscript,selection_mouse +8230,22626367,"killer.sh",163,0,"",shellscript,selection_mouse +8231,22626844,"killer.sh",162,0,"",shellscript,selection_command +8232,22627586,"killer.sh",152,28,"",shellscript,content +8233,22627624,"killer.sh",156,0,"",shellscript,selection_command +8234,22627705,"killer.sh",150,0,"",shellscript,selection_command +8235,22627824,"killer.sh",123,0,"",shellscript,selection_command +8236,22628213,"killer.sh",71,0,"",shellscript,selection_command +8237,22628454,"killer.sh",118,0,"\n echo ""hello world"" && \",shellscript,content +8238,22628460,"killer.sh",123,0,"",shellscript,selection_command +8239,22629180,"killer.sh",124,0,"",shellscript,selection_command +8240,22629376,"killer.sh",125,0,"",shellscript,selection_command +8241,22629538,"killer.sh",126,0,"",shellscript,selection_command +8242,22629685,"killer.sh",127,0,"",shellscript,selection_command +8243,22629832,"killer.sh",128,0,"",shellscript,selection_command +8244,22630192,"killer.sh",129,0,"",shellscript,selection_command +8245,22631005,"killer.sh",128,0,"",shellscript,selection_command +8246,22631308,"killer.sh",127,0,"",shellscript,selection_command +8247,22631462,"killer.sh",126,0,"",shellscript,selection_command +8248,22631604,"killer.sh",125,0,"",shellscript,selection_command +8249,22631738,"killer.sh",124,0,"",shellscript,selection_command +8250,22631904,"killer.sh",123,0,"",shellscript,selection_command +8251,22632234,"killer.sh",123,5,"",shellscript,content +8252,22632469,"killer.sh",123,1,"",shellscript,content +8253,22632740,"killer.sh",123,6,"",shellscript,content +8254,22633042,"killer.sh",123,5,"",shellscript,content +8255,22633784,"killer.sh",123,1,"",shellscript,content +8256,22634348,"killer.sh",123,0,"s",shellscript,content +8257,22634349,"killer.sh",124,0,"",shellscript,selection_keyboard +8258,22634556,"killer.sh",124,0,"c",shellscript,content +8259,22634556,"killer.sh",125,0,"",shellscript,selection_keyboard +8260,22634685,"killer.sh",125,0,"a",shellscript,content +8261,22634686,"killer.sh",126,0,"",shellscript,selection_keyboard +8262,22634821,"killer.sh",126,0,"n",shellscript,content +8263,22634822,"killer.sh",127,0,"",shellscript,selection_keyboard +8264,22634934,"killer.sh",127,0,"c",shellscript,content +8265,22634934,"killer.sh",128,0,"",shellscript,selection_keyboard +8266,22635028,"killer.sh",128,0,"e",shellscript,content +8267,22635029,"killer.sh",129,0,"",shellscript,selection_keyboard +8268,22635114,"killer.sh",129,0,"l",shellscript,content +8269,22635114,"killer.sh",130,0,"",shellscript,selection_keyboard +8270,22635305,"killer.sh",130,0," ",shellscript,content +8271,22635305,"killer.sh",131,0,"",shellscript,selection_keyboard +8272,22635449,"killer.sh",131,0,"-",shellscript,content +8273,22635449,"killer.sh",132,0,"",shellscript,selection_keyboard +8274,22635627,"killer.sh",132,0,"-",shellscript,content +8275,22635628,"killer.sh",133,0,"",shellscript,selection_keyboard +8276,22635792,"killer.sh",133,0,"m",shellscript,content +8277,22635793,"killer.sh",134,0,"",shellscript,selection_keyboard +8278,22636352,"killer.sh",134,0,"e",shellscript,content +8279,22636353,"killer.sh",135,0,"",shellscript,selection_keyboard +8280,22636782,"killer.sh",135,0,"^",shellscript,content +8281,22636783,"killer.sh",136,0,"",shellscript,selection_keyboard +8282,22637421,"killer.sh",135,1,"^",shellscript,content +8283,22637436,"killer.sh",135,1,"^",shellscript,content +8284,22637448,"killer.sh",135,1,"",shellscript,content +8285,22637448,"killer.sh",135,0,"",shellscript,selection_keyboard +8286,22637455,"killer.sh",135,0,"^",shellscript,content +8287,22637456,"killer.sh",136,0,"",shellscript,selection_keyboard +8288,22637646,"killer.sh",135,0,"",shellscript,selection_command +8289,22638584,"killer.sh",135,1,"",shellscript,content +8290,22653342,"killer.sh",157,0,"",shellscript,selection_command +8291,22657472,"killer.sh",135,0,"",shellscript,selection_command +8292,22658019,"killer.sh",157,0,"",shellscript,selection_command +8293,22658350,"killer.sh",135,0,"",shellscript,selection_command +8294,22658543,"killer.sh",157,0,"",shellscript,selection_command +8295,22658713,"killer.sh",135,0,"",shellscript,selection_command +8296,22658846,"killer.sh",157,0,"",shellscript,selection_command +8297,22658982,"killer.sh",135,0,"",shellscript,selection_command +8298,22659156,"killer.sh",83,0,"",shellscript,selection_command +8299,22659365,"killer.sh",66,0,"",shellscript,selection_command +8300,22659430,"killer.sh",83,0,"",shellscript,selection_command +8301,22659574,"killer.sh",135,0,"",shellscript,selection_command +8302,22659768,"killer.sh",157,0,"",shellscript,selection_command +8303,22659951,"killer.sh",172,0,"",shellscript,selection_command +8304,22660151,"killer.sh",157,0,"",shellscript,selection_command +8305,22660319,"killer.sh",135,0,"",shellscript,selection_command +8306,22660451,"killer.sh",83,0,"",shellscript,selection_command +8307,22660614,"killer.sh",135,0,"",shellscript,selection_command +8308,22660812,"killer.sh",157,0,"",shellscript,selection_command +8309,22661142,"killer.sh",135,0,"",shellscript,selection_command +8310,22661306,"killer.sh",83,0,"",shellscript,selection_command +8311,22661428,"TERMINAL",0,0,"[?25l[0] 0:bash* ""hkn1993.localdomain"" 16:55 14-Aug-25(B[?12l[?25h",,terminal_output +8312,22661439,"killer.sh",135,0,"",shellscript,selection_command +8313,22661608,"killer.sh",157,0,"",shellscript,selection_command +8314,22661765,"killer.sh",135,0,"",shellscript,selection_command +8315,22661923,"killer.sh",83,0,"",shellscript,selection_command +8316,22662058,"killer.sh",135,0,"",shellscript,selection_command +8317,22662206,"killer.sh",157,0,"",shellscript,selection_command +8318,22662586,"killer.sh",135,0,"",shellscript,selection_command +8319,22669841,"TERMINAL",0,0,"watch -n60 ""sh killer.sh""",,terminal_output +8320,22699555,"TERMINAL",0,0,"\r\n[?2004l",,terminal_output +8321,22699686,"TERMINAL",0,0,"[?25lEvery 60.0s: sh killer.shhkn1993.localdomain: Thu Aug 14 16:55:52 2025\r\nleaving jobs be\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[?12l[?25h",,terminal_output +8322,22703401,"TERMINAL",0,0,"[?25l\r\n[0] 0:watch* ""hkn1993.localdomain"" 16:55 14-Aug-25(B[?12l[?25h",,terminal_output +8323,22703802,"TERMINAL",0,0,"(B[?1l>[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l[?7727l[?1004l[?1049l[detached (from session 0)]\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +8324,22709304,"killer.sh",0,0,"",shellscript,tab +8325,22710244,"killer.sh",204,0,"",shellscript,selection_mouse +8326,22710861,"killer.sh",203,1,"\n",shellscript,selection_mouse +8327,22710899,"killer.sh",182,22," ""leaving jobs be""\nfi\n",shellscript,selection_mouse +8328,22710899,"killer.sh",173,31,"\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +8329,22710900,"killer.sh",146,58,"cho ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +8330,22710900,"killer.sh",145,59,"echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +8331,22710979,"killer.sh",122,82," scancel --me && \\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +8332,22710980,"killer.sh",69,135," echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n scancel --me && \\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +8333,22711027,"killer.sh",66,138,"\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n scancel --me && \\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +8334,22711075,"killer.sh",0,204,"response=$(curl -s https://home.cit.tum.de/~nguyenal/killer.html)\n\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n scancel --me && \\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,selection_mouse +8335,22767206,"killer.sh",204,0,"",shellscript,selection_mouse +8336,23198544,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --job-name=train_dynamics_maskgit_8_node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/holiday/maskgit/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --darkness_threshold=50 \\n --dyna_type=maskgit \\n --num_latent_actions=100 \\n --max_lr=8e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-maskgit-8-node-darkness-filter-$slurm_job_id \\n --tags dynamics maskgit 8-node post-launch-main darkness-filter \\n --entity instant-uv \\n --project jafar \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16 \\n --dyna_ffn_dim=4096 \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +8337,23213674,"TERMINAL",0,0,"bash",,terminal_focus +8338,23215266,"TERMINAL",0,0,"idling",,terminal_command +8339,23215332,"TERMINAL",0,0,"]633;E;2025-08-14 17:04:28 idling;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Aug 14 17:04:28 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 255 nodes idle\rPartition dev_accelerated:\t 1 nodes idle\rPartition accelerated: 32 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 1 nodes idle\rPartition large:\t 8 nodes idle",,terminal_output +8340,23216407,"TERMINAL",0,0,"9\t ",,terminal_output +8341,23216508,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +8342,23226195,"TERMINAL",0,0,"queue",,terminal_command +8343,23226265,"TERMINAL",0,0,"]633;E;2025-08-14 17:04:39 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +8344,23226340,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 17:04:39 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4 (Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2 (Priority)3412401 accelerat train_to tum_cte0 R 1-23:04:38\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-03:15:43\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 4:51:29\t 1 hkn07133421964 accelerat train_to tum_cte0 R 5:03:53\t 1 hkn0720",,terminal_output +8345,23227484,"TERMINAL",0,0,"4094304",,terminal_output +8346,23228389,"TERMINAL",0,0,"140515",,terminal_output +8347,23229530,"TERMINAL",0,0,"21626",,terminal_output +8348,23230480,"TERMINAL",0,0,"32737",,terminal_output +8349,23231526,"TERMINAL",0,0,"43848",,terminal_output +8350,23232552,"TERMINAL",0,0,"54959",,terminal_output +8351,23233596,"TERMINAL",0,0,"655064:00",,terminal_output +8352,23234650,"TERMINAL",0,0,"76171",,terminal_output +8353,23235691,"TERMINAL",0,0,"87282",,terminal_output +8354,23236881,"TERMINAL",0,0,"98393",,terminal_output +8355,23237827,"TERMINAL",0,0,"5094404",,terminal_output +8356,23238850,"TERMINAL",0,0,"151626",,terminal_output +8357,23239978,"TERMINAL",0,0,"32737",,terminal_output +8358,23241000,"TERMINAL",0,0,"43848",,terminal_output +8359,23242234,"TERMINAL",0,0,"54959",,terminal_output +8360,23242705,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +8361,23246471,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",1900,3,"160",shellscript,selection_mouse +8362,23256618,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8363,23265312,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2090,0,"",shellscript,selection_mouse +8364,23266413,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2089,1,"",shellscript,content +8365,23266710,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2089,0,"1",shellscript,content +8366,23266711,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2090,0,"",shellscript,selection_keyboard +8367,23267056,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2090,0,"0",shellscript,content +8368,23267057,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2091,0,"",shellscript,selection_keyboard +8369,23267154,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2091,0,"0",shellscript,content +8370,23267154,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2092,0,"",shellscript,selection_keyboard +8371,23267402,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2091,0,"",shellscript,selection_command +8372,23274539,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",512,0,"",shellscript,selection_mouse +8373,23275167,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",486,0,"",shellscript,selection_mouse +8374,23279962,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8375,23282038,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",443,0,"",shellscript,selection_mouse +8376,23282704,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",471,0,"\n#SBATCH --reservation=llmtum",shellscript,content +8377,23282735,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",472,0,"",shellscript,selection_command +8378,23303655,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",1753,0,"",shellscript,selection_mouse +8379,23303662,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",1752,0,"",shellscript,selection_command +8380,23397084,"TERMINAL",0,0,"bash",,terminal_focus +8381,23398603,"TERMINAL",0,0,"logs",,terminal_command +8382,23398642,"TERMINAL",0,0,"]633;E;2025-08-14 17:07:31 logs;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +8383,23401833,"TERMINAL",0,0,"cd maskgit",,terminal_command +8384,23401862,"TERMINAL",0,0,"]633;E;2025-08-14 17:07:34 cd maskgit;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit]633;D;0",,terminal_output +8385,23402091,"TERMINAL",0,0,"ls",,terminal_command +8386,23402125,"TERMINAL",0,0,"]633;E;2025-08-14 17:07:35 ls;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;Cdynamics-cotraining\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit]633;D;0",,terminal_output +8387,23403533,"TERMINAL",0,0,"cd dynamics-cotraining/",,terminal_command +8388,23403872,"TERMINAL",0,0,"ls",,terminal_command +8389,23403908,"TERMINAL",0,0,"]633;E;2025-08-14 17:07:37 ls;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;Ctrain_dynamics_maskgit_8_node_3412350.log train_dynamics_maskgit_8_node_3417225.log train_dynamics_maskgit_8_node_3418833.log\r\ntrain_dynamics_maskgit_8_node_3412354.log train_dynamics_maskgit_8_node_3417226.log train_dynamics_maskgit_8_node_3418834.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;0",,terminal_output +8390,23406428,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",0,0,"",log,tab +8391,23406745,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",3670,0,"",log,selection_mouse +8392,23406751,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",3669,0,"",log,selection_command +8393,23419614,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8394,23428026,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8395,23436977,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8396,23440581,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8397,23443618,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8398,23444423,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8399,23445994,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",0,0,"",log,tab +8400,23480157,"TERMINAL",0,0,"bash",,terminal_focus +8401,23484126,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8402,23485377,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8403,23492260,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8404,23494354,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",516,0,"",shellscript,selection_mouse +8405,23494880,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",442,0,"",shellscript,selection_mouse +8406,23495640,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",491,0,"",shellscript,selection_mouse +8407,23495795,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",482,11,"reservation",shellscript,selection_mouse +8408,23496179,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",499,0,"",shellscript,selection_mouse +8409,23496310,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",494,6,"llmtum",shellscript,selection_mouse +8410,23501537,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/train_dynamics_maskgit_8_node_3418834.log",0,0,"",log,tab +8411,23503622,"TERMINAL",0,0,"bash",,terminal_focus +8412,23507022,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",,terminal_command +8413,23507058,"TERMINAL",0,0,"]633;E;2025-08-14 17:09:20 sbatch slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;Csbatch: error: Unable to open file slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;1",,terminal_output +8414,23509226,"TERMINAL",0,0,"bash",,terminal_focus +8415,23511440,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",,terminal_command +8416,23511475,"TERMINAL",0,0,"]633;E;2025-08-14 17:09:24 sbatch slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CSubmitted batch job 3423221\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +8417,23516365,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8418,23517697,"TERMINAL",0,0,"bash",,terminal_focus +8419,23518778,"TERMINAL",0,0,"ls",,terminal_command +8420,23525897,"TERMINAL",0,0,"bash",,terminal_focus +8421,23528596,"TERMINAL",0,0,"bash",,terminal_focus +8422,23531555,"TERMINAL",0,0,"tail -f train_dynamics_maskgit_8_node_3423221.log",,terminal_command +8423,23533999,"TERMINAL",0,0,"queue",,terminal_command +8424,23534079,"TERMINAL",0,0,"]633;E;2025-08-14 17:09:47 queue;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 17:09:47 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4 (Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2 (Priority)3423221 accelerat train_dy tum_cte0 PD\t0:00\t 8 (Priority)3412401 accelerat train_to tum_cte0 R 1-23:09:46\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-03:20:51\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 4:56:37\t 1 hkn07133421964 accelerat train_to tum_cte0 R 5:09:01\t 1 hkn0720",,terminal_output +8425,23535301,"TERMINAL",0,0,"87282",,terminal_output +8426,23536219,"TERMINAL",0,0,"98393",,terminal_output +8427,23537246,"TERMINAL",0,0,"5094404",,terminal_output +8428,23538376,"TERMINAL",0,0,"150515",,terminal_output +8429,23539395,"TERMINAL",0,0,"21626",,terminal_output +8430,23540419,"TERMINAL",0,0,"32737",,terminal_output +8431,23541448,"TERMINAL",0,0,"43848",,terminal_output +8432,23542469,"TERMINAL",0,0,"54959",,terminal_output +8433,23543494,"TERMINAL",0,0,"651:00610",,terminal_output +8434,23544517,"TERMINAL",0,0,"76171",,terminal_output +8435,23545531,"TERMINAL",0,0,"87282",,terminal_output +8436,23546123,"TERMINAL",0,0,"bash",,terminal_focus +8437,23546586,"TERMINAL",0,0,"98393",,terminal_output +8438,23547616,"TERMINAL",0,0,"10:0094504",,terminal_output +8439,23548059,"TERMINAL",0,0,"scancel 3423221",,terminal_command +8440,23548110,"TERMINAL",0,0,"]633;E;2025-08-14 17:10:01 scancel 3423221;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +8441,23548666,"TERMINAL",0,0,"\r11240to R 1-23:1hkn[0412,0416,0425,0433,0718,0732,0803,0807]8832dy03:21:0507,0415,0420,0422-0424,0729,0735]22008 4:56:51\t 1 hkn07131964to5:09:1520",,terminal_output +8442,23549464,"TERMINAL",0,0,"watch",,terminal_focus +8443,23549752,"TERMINAL",0,0,"21626",,terminal_output +8444,23550097,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;0",,terminal_output +8445,23552192,"TERMINAL",0,0,"idling",,terminal_command +8446,23552255,"TERMINAL",0,0,"]633;E;2025-08-14 17:10:05 idling;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Aug 14 17:10:05 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 258 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated: 34 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 1 nodes idle\rPartition large:\t 8 nodes idle",,terminal_output +8447,23553356,"TERMINAL",0,0,"6\t ",,terminal_output +8448,23554371,"TERMINAL",0,0,"7\t ",,terminal_output +8449,23555437,"TERMINAL",0,0,"8\t ",,terminal_output +8450,23556415,"TERMINAL",0,0,"9\t ",,terminal_output +8451,23557455,"TERMINAL",0,0,"10\t ",,terminal_output +8452,23558495,"TERMINAL",0,0,"1\t ",,terminal_output +8453,23559569,"TERMINAL",0,0,"2\t ",,terminal_output +8454,23560583,"TERMINAL",0,0,"3\t ",,terminal_output +8455,23561623,"TERMINAL",0,0,"4\t ",,terminal_output +8456,23562669,"TERMINAL",0,0,"5\t ",,terminal_output +8457,23563725,"TERMINAL",0,0,"6\t ",,terminal_output +8458,23564766,"TERMINAL",0,0,"7\t ",,terminal_output +8459,23565861,"TERMINAL",0,0,"8\t ",,terminal_output +8460,23566835,"TERMINAL",0,0,"9\t ",,terminal_output +8461,23567904,"TERMINAL",0,0,"21\t ",,terminal_output +8462,23568922,"TERMINAL",0,0,"2\t ",,terminal_output +8463,23569968,"TERMINAL",0,0,"3\t ",,terminal_output +8464,23571139,"TERMINAL",0,0,"4\t ",,terminal_output +8465,23572163,"TERMINAL",0,0,"5\t ",,terminal_output +8466,23573190,"TERMINAL",0,0,"6\t ",,terminal_output +8467,23574139,"TERMINAL",0,0,"7\t ",,terminal_output +8468,23575238,"TERMINAL",0,0,"8\t ",,terminal_output +8469,23576259,"TERMINAL",0,0,"9\t ",,terminal_output +8470,23577293,"TERMINAL",0,0,"30\t ",,terminal_output +8471,23578309,"TERMINAL",0,0,"1\t ",,terminal_output +8472,23579343,"TERMINAL",0,0,"2\t ",,terminal_output +8473,23580390,"TERMINAL",0,0,"3\t ",,terminal_output +8474,23581432,"TERMINAL",0,0,"4\t ",,terminal_output +8475,23582611,"TERMINAL",0,0,"5\t ",,terminal_output +8476,23583511,"TERMINAL",0,0,"6\t ",,terminal_output +8477,23584548,"TERMINAL",0,0,"7\t ",,terminal_output +8478,23585682,"TERMINAL",0,0,"8\t ",,terminal_output +8479,23586637,"TERMINAL",0,0,"9\t ",,terminal_output +8480,23587729,"TERMINAL",0,0,"40\t ",,terminal_output +8481,23588767,"TERMINAL",0,0,"1\t ",,terminal_output +8482,23589788,"TERMINAL",0,0,"2\t ",,terminal_output +8483,23590800,"TERMINAL",0,0,"3\t ",,terminal_output +8484,23591835,"TERMINAL",0,0,"4\t ",,terminal_output +8485,23592893,"TERMINAL",0,0,"6\t ",,terminal_output +8486,23593977,"TERMINAL",0,0,"7\t ",,terminal_output +8487,23594291,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;0",,terminal_output +8488,23596704,"TERMINAL",0,0,"bash",,terminal_focus +8489,23598620,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",,terminal_command +8490,23598717,"TERMINAL",0,0,"]633;E;2025-08-14 17:10:51 sbatch slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CSubmitted batch job 3423223\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +8491,23600950,"TERMINAL",0,0,"bash",,terminal_focus +8492,23606494,"TERMINAL",0,0,"bash",,terminal_focus +8493,23607941,"TERMINAL",0,0,"queue",,terminal_command +8494,23608003,"TERMINAL",0,0,"]633;E;2025-08-14 17:11:01 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 17:11:01 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4 (Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2 (Priority)3423223 accelerat train_dy tum_cte0 PD\t0:00\t 8 (Priority)3412401 accelerat train_to tum_cte0 R 1-23:11:00\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-03:22:05\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 4:57:51\t 1 hkn07133421964 accelerat train_to tum_cte0 R 5:10:15\t 1 hkn0720",,terminal_output +8495,23609065,"TERMINAL",0,0,"21626",,terminal_output +8496,23610084,"TERMINAL",0,0,"32737",,terminal_output +8497,23610120,"TERMINAL",0,0,"bash",,terminal_focus +8498,23611179,"TERMINAL",0,0,"43848",,terminal_output +8499,23612174,"TERMINAL",0,0,"54959",,terminal_output +8500,23613328,"TERMINAL",0,0,"6510620",,terminal_output +8501,23614267,"TERMINAL",0,0,"76171",,terminal_output +8502,23615309,"TERMINAL",0,0,"87282",,terminal_output +8503,23616403,"TERMINAL",0,0,"98393",,terminal_output +8504,23617446,"TERMINAL",0,0,"10948:004",,terminal_output +8505,23618448,"TERMINAL",0,0,"110515",,terminal_output +8506,23619474,"TERMINAL",0,0,"21626",,terminal_output +8507,23620583,"TERMINAL",0,0,"32737",,terminal_output +8508,23621562,"TERMINAL",0,0,"43848",,terminal_output +8509,23622648,"TERMINAL",0,0,"54959",,terminal_output +8510,23623654,"TERMINAL",0,0,"6520630",,terminal_output +8511,23624730,"TERMINAL",0,0,"76171",,terminal_output +8512,23625822,"TERMINAL",0,0,"87282",,terminal_output +8513,23626845,"TERMINAL",0,0,"98393",,terminal_output +8514,23627871,"TERMINAL",0,0,"20205115",,terminal_output +8515,23628894,"TERMINAL",0,0,"21626",,terminal_output +8516,23630021,"TERMINAL",0,0,"32737",,terminal_output +8517,23631149,"TERMINAL",0,0,"43848",,terminal_output +8518,23632172,"TERMINAL",0,0,"54959",,terminal_output +8519,23633065,"TERMINAL",0,0,"6530640",,terminal_output +8520,23634220,"TERMINAL",0,0,"76171",,terminal_output +8521,23635243,"TERMINAL",0,0,"87282",,terminal_output +8522,23636598,"TERMINAL",0,0,"98393",,terminal_output +8523,23637634,"TERMINAL",0,0,"3094204",,terminal_output +8524,23638726,"TERMINAL",0,0,"130515",,terminal_output +8525,23639731,"TERMINAL",0,0,"21626",,terminal_output +8526,23640773,"TERMINAL",0,0,"32737",,terminal_output +8527,23641801,"TERMINAL",0,0,"43848",,terminal_output +8528,23642922,"TERMINAL",0,0,"5540650",,terminal_output +8529,23643880,"TERMINAL",0,0,"76171",,terminal_output +8530,23644942,"TERMINAL",0,0,"87282",,terminal_output +8531,23645997,"TERMINAL",0,0,"98393",,terminal_output +8532,23647013,"TERMINAL",0,0,"4094304",,terminal_output +8533,23648086,"TERMINAL",0,0,"140515",,terminal_output +8534,23649129,"TERMINAL",0,0,"21626",,terminal_output +8535,23650150,"TERMINAL",0,0,"32737",,terminal_output +8536,23651218,"TERMINAL",0,0,"43848",,terminal_output +8537,23652259,"TERMINAL",0,0,"54959",,terminal_output +8538,23653473,"TERMINAL",0,0,"655061:00",,terminal_output +8539,23654311,"TERMINAL",0,0,"76171",,terminal_output +8540,23655439,"TERMINAL",0,0,"87282",,terminal_output +8541,23656546,"TERMINAL",0,0,"98393",,terminal_output +8542,23657566,"TERMINAL",0,0,"5094404",,terminal_output +8543,23658494,"TERMINAL",0,0,"150515",,terminal_output +8544,23659542,"TERMINAL",0,0,"21626",,terminal_output +8545,23660638,"TERMINAL",0,0,"32737",,terminal_output +8546,23661629,"TERMINAL",0,0,"43848",,terminal_output +8547,23662677,"TERMINAL",0,0,"54959",,terminal_output +8548,23663812,"TERMINAL",0,0,"653:00610",,terminal_output +8549,23664776,"TERMINAL",0,0,"76171",,terminal_output +8550,23665861,"TERMINAL",0,0,"87282",,terminal_output +8551,23666872,"TERMINAL",0,0,"994504",,terminal_output +8552,23667889,"TERMINAL",0,0,"2:012:00515",,terminal_output +8553,23669038,"TERMINAL",0,0,"21626",,terminal_output +8554,23670059,"TERMINAL",0,0,"32737",,terminal_output +8555,23671082,"TERMINAL",0,0,"43848",,terminal_output +8556,23672146,"TERMINAL",0,0,"54959",,terminal_output +8557,23673130,"TERMINAL",0,0,"6510620",,terminal_output +8558,23674246,"TERMINAL",0,0,"76171",,terminal_output +8559,23675224,"TERMINAL",0,0,"87282",,terminal_output +8560,23676260,"TERMINAL",0,0,"98393",,terminal_output +8561,23677328,"TERMINAL",0,0,"10949:004",,terminal_output +8562,23678455,"TERMINAL",0,0,"110515",,terminal_output +8563,23679480,"TERMINAL",0,0,"21626",,terminal_output +8564,23680492,"TERMINAL",0,0,"32737",,terminal_output +8565,23681539,"TERMINAL",0,0,"43848",,terminal_output +8566,23682582,"TERMINAL",0,0,"54959",,terminal_output +8567,23683628,"TERMINAL",0,0,"6520630",,terminal_output +8568,23684672,"TERMINAL",0,0,"76171",,terminal_output +8569,23685734,"TERMINAL",0,0,"87282",,terminal_output +8570,23686853,"TERMINAL",0,0,"98393",,terminal_output +8571,23687790,"TERMINAL",0,0,"2094104",,terminal_output +8572,23688908,"TERMINAL",0,0,"121626",,terminal_output +8573,23689928,"TERMINAL",0,0,"32737",,terminal_output +8574,23690950,"TERMINAL",0,0,"43848",,terminal_output +8575,23692052,"TERMINAL",0,0,"54959",,terminal_output +8576,23693052,"TERMINAL",0,0,"6530640",,terminal_output +8577,23694122,"TERMINAL",0,0,"76171",,terminal_output +8578,23695155,"TERMINAL",0,0,"87282",,terminal_output +8579,23696281,"TERMINAL",0,0,"98393",,terminal_output +8580,23697243,"TERMINAL",0,0,"3094204",,terminal_output +8581,23698332,"TERMINAL",0,0,"130515",,terminal_output +8582,23699371,"TERMINAL",0,0,"21626",,terminal_output +8583,23700407,"TERMINAL",0,0,"32737",,terminal_output +8584,23701494,"TERMINAL",0,0,"43848",,terminal_output +8585,23702466,"TERMINAL",0,0,"54959",,terminal_output +8586,23703489,"TERMINAL",0,0,"6540650",,terminal_output +8587,23704527,"TERMINAL",0,0,"76171",,terminal_output +8588,23705595,"TERMINAL",0,0,"87282",,terminal_output +8589,23706701,"TERMINAL",0,0,"98393",,terminal_output +8590,23707557,"TERMINAL",0,0,"watch",,terminal_focus +8591,23707672,"TERMINAL",0,0,"4094304",,terminal_output +8592,23708744,"TERMINAL",0,0,"140515",,terminal_output +8593,23709766,"TERMINAL",0,0,"21626",,terminal_output +8594,23710815,"TERMINAL",0,0,"32737",,terminal_output +8595,23712044,"TERMINAL",0,0,"44959",,terminal_output +8596,23713067,"TERMINAL",0,0,"655062:00",,terminal_output +8597,23714091,"TERMINAL",0,0,"76171",,terminal_output +8598,23714989,"TERMINAL",0,0,"87282",,terminal_output +8599,23716038,"TERMINAL",0,0,"98393",,terminal_output +8600,23717104,"TERMINAL",0,0,"5094404",,terminal_output +8601,23718195,"TERMINAL",0,0,"150515",,terminal_output +8602,23719185,"TERMINAL",0,0,"21626",,terminal_output +8603,23720218,"TERMINAL",0,0,"32737",,terminal_output +8604,23721262,"TERMINAL",0,0,"43848",,terminal_output +8605,23721538,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +8606,23724988,"TERMINAL",0,0,"sinfo",,terminal_command +8607,23725053,"TERMINAL",0,0,"]633;E;2025-08-14 17:12:58 sinfo ;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Cslurm_load_node: Access/permission denied\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;1",,terminal_output +8608,23733301,"TERMINAL",0,0,"sacct",,terminal_command +8609,23733338,"TERMINAL",0,0,"]633;E;2025-08-14 17:13:06 sacct;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CJobID JobName Partition Account AllocCPUS State ExitCode \r\n------------ ---------- ---------- ---------- ---------- ---------- -------- \r\n3417227 train_tok+ accelerat+ hk-projec+ 48 COMPLETED 0:0 \r\n3417227.bat+ batch hk-projec+ 24 COMPLETED 0:0 \r\n3417227.ext+ extern hk-projec+ 48 COMPLETED 0:0 \r\n3417227.0 python hk-projec+ 40 COMPLETED 0:0 \r\n3412401 train_tok+ accelerat+ hk-projec+ 192 RUNNING 0:0 \r\n3412401.bat+ batch hk-projec+ 24 RUNNING 0:0 \r\n3412401.ext+ extern hk-projec+ 192 RUNNING 0:0 \r\n3412401.0 python hk-projec+ 160 RUNNING 0:0 \r\n3418832 train_dyn+ accelerat+ hk-projec+ 192 RUNNING 0:0 \r\n3418832.bat+ batch hk-projec+ 24 RUNNING 0:0 \r\n3418832.ext+ extern hk-projec+ 192 RUNNING 0:0 \r\n3418832.0 python hk-projec+ 160 RUNNING 0:0 \r\n3415713 train_dyn+ accelerat+ hk-projec+ 0 PENDING 0:0 \r\n3421964 train_tok+ accelerat+ hk-projec+ 24 RUNNING 0:0 \r\n3421964.bat+ batch hk-projec+ 24 RUNNING 0:0 \r\n3421964.ext+ extern hk-projec+ 24 RUNNING 0:0 \r\n3421964.0 python hk-projec+ 20 RUNNING 0:0 \r\n3421973 train_tok+ accelerat+ hk-projec+ 48 CANCELLED+ 0:0 \r\n3421973.bat+ batch hk-projec+ 24 CANCELLED 0:15 \r\n3421973.ext+ extern hk-projec+ 48 COMPLETED 0:0 \r\n3421973.0 python hk-projec+ 40 CANCELLED 0:9 \r\n3421982 train_tok+ accelerat+ hk-projec+ 96 CANCELLED+ 0:0 \r\n3421982.bat+ batch hk-projec+ 24 CANCELLED 0:15 \r\n3421982.ext+ extern hk-projec+ 96 COMPLETED 0:0 \r\n3421982.0 python hk-projec+ 80 CANCELLED 0:9 \r\n3422007 train_dyn+ accelerat+ hk-projec+ 0 CANCELLED+ 0:0 \r\n3422008 train_dyn+ accelerat+ hk-projec+ 24 RUNNING 0:0 \r\n3422008.bat+ batch hk-projec+ 24 RUNNING 0:0 \r\n3422008.ext+ extern hk-projec+ 24 RUNNING 0:0 \r\n3422008.0 python hk-projec+ 20 RUNNING 0:0 \r\n3422101 train_dyn+ accelerat+ hk-projec+ 0 CANCELLED+ 0:0 \r\n3422103 train_dyn+ accelerat+ hk-projec+ 192 CANCELLED+ 0:0 \r\n3422103.bat+ batch hk-projec+ 24 CANCELLED 0:15 \r\n3422103.ext+ extern hk-projec+ 192 COMPLETED 0:0 \r\n3422103.0 python hk-projec+ 160 CANCELLED 0:9 \r\n3423052 train_tok+ accelerat+ hk-projec+ 48 CANCELLED+ 0:0 \r\n3423052.bat+ batch hk-projec+ 24 CANCELLED 0:15 \r\n3423052.ext+ extern hk-projec+ 48 COMPLETED 0:0 \r\n3423052.0 python hk-projec+ 40 CANCELLED 0:9 \r\n3423080 train_tok+ accelerat+ hk-projec+ 0 PENDING 0:0 \r\n3423082 train_tok+ accelerat+ hk-projec+ 0 PENDING 0:0 \r\n3423221 train_dyn+ accelerat+ hk-projec+ 0 CANCELLED+ 0:0 \r\n3423223 train_dyn+ accelerat+ hk-projec+ 0 PENDING 0:0 \r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +8610,23890024,"TERMINAL",0,0,"bash",,terminal_focus +8611,23891197,"TERMINAL",0,0,"bash",,terminal_focus +8612,23892160,"TERMINAL",0,0,"queue",,terminal_command +8613,23892238,"TERMINAL",0,0,"]633;E;2025-08-14 17:15:45 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 17:15:45 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4 (Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2 (Priority)3423223 accelerat train_dy tum_cte0 PD\t0:00\t 8 (Priority)3412401 accelerat train_to tum_cte0 R 1-23:15:44\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-03:26:49\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 5:02:35\t 1 hkn07133421964 accelerat train_to tum_cte0 R 5:14:59\t 1 hkn0720",,terminal_output +8614,23893398,"TERMINAL",0,0,"655065:00",,terminal_output +8615,23894420,"TERMINAL",0,0,"76171",,terminal_output +8616,23895443,"TERMINAL",0,0,"87282",,terminal_output +8617,23896505,"TERMINAL",0,0,"98393",,terminal_output +8618,23897528,"TERMINAL",0,0,"5094404",,terminal_output +8619,23898525,"TERMINAL",0,0,"150515",,terminal_output +8620,23899549,"TERMINAL",0,0,"21626",,terminal_output +8621,23900597,"TERMINAL",0,0,"32737",,terminal_output +8622,23901634,"TERMINAL",0,0,"43848",,terminal_output +8623,23902684,"TERMINAL",0,0,"54959",,terminal_output +8624,23903736,"TERMINAL",0,0,"657:00610",,terminal_output +8625,23904802,"TERMINAL",0,0,"76171",,terminal_output +8626,23905809,"TERMINAL",0,0,"87282",,terminal_output +8627,23906872,"TERMINAL",0,0,"994504",,terminal_output +8628,23907846,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +8629,23910736,"TERMINAL",0,0,"scancel 3423223",,terminal_command +8630,23910772,"TERMINAL",0,0,"]633;E;2025-08-14 17:16:03 scancel 3423223;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +8631,23912215,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8632,23913232,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8633,23914942,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",494,6,"",shellscript,content +8634,23914970,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",493,0,"",shellscript,selection_command +8635,23916927,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",494,0,"llmtum",shellscript,content +8636,23916966,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",494,0,"",shellscript,selection_command +8637,23917817,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",472,29,"",shellscript,content +8638,23919461,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8639,23920850,"TERMINAL",0,0,"bash",,terminal_focus +8640,23922636,"TERMINAL",0,0,"bash",,terminal_focus +8641,23927154,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --job-name=train_dynamics_maskgit_8_node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/holiday/maskgit/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --darkness_threshold=50 \\n --dyna_type=maskgit \\n --num_latent_actions=100 \\n --max_lr=8e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-maskgit-8-node-darkness-filter-$slurm_job_id \\n --tags dynamics maskgit 8-node post-launch-main darkness-filter \\n --entity instant-uv \\n --project jafar \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16 \\n --dyna_ffn_dim=4096 \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +8642,23931088,"TERMINAL",0,0,"bash",,terminal_focus +8643,23932514,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8644,23935295,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +8645,23936413,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",484,0,"servation=llmtum\n#SBATCH --re",shellscript,content +8646,23936436,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",494,0,"",shellscript,selection_command +8647,23938577,"TERMINAL",0,0,"bash",,terminal_focus +8648,23939049,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch^C",,terminal_command +8649,23939074,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]633;E;;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D",,terminal_output +8650,23940428,"TERMINAL",0,0,"dev",,terminal_command +8651,23944679,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",,terminal_command +8652,23944708,"TERMINAL",0,0,"]633;E;2025-08-14 17:16:37 sbatch slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CSubmitted batch job 3423234\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +8653,23945620,"TERMINAL",0,0,"queue",,terminal_command +8654,23945708,"TERMINAL",0,0,"]633;E;2025-08-14 17:16:38 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 17:16:38 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4 (Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2 (Priority)3412401 accelerat train_to tum_cte0 R 1-23:16:37\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-03:27:42\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3423234 accelerat train_dy tum_cte0 R\t0:00\t 8 hkn[0521-0528]3422008 accelerat train_dy tum_cte0 R 5:03:28\t 1 hkn07133421964 accelerat train_to tum_cte0 R 5:15:52\t 1 hkn0720",,terminal_output +8655,23946718,"TERMINAL",0,0,"983193",,terminal_output +8656,23947779,"TERMINAL",0,0,"40942304",,terminal_output +8657,23948421,"TERMINAL",0,0,"bash",,terminal_focus +8658,23948837,"TERMINAL",0,0,"1405315",,terminal_output +8659,23949862,"TERMINAL",0,0,"227537",,terminal_output +8660,23950905,"TERMINAL",0,0,"438648",,terminal_output +8661,23951977,"TERMINAL",0,0,"549759",,terminal_output +8662,23952991,"TERMINAL",0,0,"6550866:00",,terminal_output +8663,23954231,"TERMINAL",0,0,"761971",,terminal_output +8664,23955072,"TERMINAL",0,0,"8721082",,terminal_output +8665,23956271,"TERMINAL",0,0,"983193",,terminal_output +8666,23957193,"TERMINAL",0,0,"50942404",,terminal_output +8667,23958216,"TERMINAL",0,0,"1505315",,terminal_output +8668,23959372,"TERMINAL",0,0,"216426",,terminal_output +8669,23960369,"TERMINAL",0,0,"327537",,terminal_output +8670,23961336,"TERMINAL",0,0,"438648",,terminal_output +8671,23962420,"TERMINAL",0,0,"549759",,terminal_output +8672,23963389,"TERMINAL",0,0,"658:008610",,terminal_output +8673,23964425,"TERMINAL",0,0,"761971",,terminal_output +8674,23965466,"TERMINAL",0,0,"8722082",,terminal_output +8675,23966554,"TERMINAL",0,0,"983193",,terminal_output +8676,23967596,"TERMINAL",0,0,"7:00942504",,terminal_output +8677,23967974,"TERMINAL",0,0,"watch",,terminal_focus +8678,23968645,"TERMINAL",0,0,"17:005315",,terminal_output +8679,23969447,"TERMINAL",0,0,"bash",,terminal_focus +8680,23969649,"TERMINAL",0,0,"216426",,terminal_output +8681,23970683,"TERMINAL",0,0,"327537",,terminal_output +8682,23971731,"TERMINAL",0,0,"438648",,terminal_output +8683,23971809,"TERMINAL",0,0,"tail -f train_dynamics_maskgit_8_node_3423234.log",,terminal_command +8684,23971888,"TERMINAL",0,0,"]633;E;2025-08-14 17:17:04 tail -f train_dynamics_maskgit_8_node_3423234.log;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;Ctail: cannot open 'train_dynamics_maskgit_8_node_3423234.log' for reading: No such file or directory\r\ntail: no files remaining\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;1",,terminal_output +8685,23972782,"TERMINAL",0,0,"549759",,terminal_output +8686,23972817,"TERMINAL",0,0,"tail -f train_dynamics_maskgit_8_node_3423234.log",,terminal_command +8687,23972886,"TERMINAL",0,0,"]633;E;2025-08-14 17:17:05 tail -f train_dynamics_maskgit_8_node_3423234.log;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;CSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0521-0528]\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +8688,23973810,"TERMINAL",0,0,"65108620",,terminal_output +8689,23974857,"TERMINAL",0,0,"7723082",,terminal_output +8690,23976137,"TERMINAL",0,0,"983193",,terminal_output +8691,23976954,"TERMINAL",0,0,"109424:004",,terminal_output +8692,23978186,"TERMINAL",0,0,"1105315",,terminal_output +8693,23979208,"TERMINAL",0,0,"216426",,terminal_output +8694,23980129,"TERMINAL",0,0,"327537",,terminal_output +8695,23981129,"TERMINAL",0,0,"438648",,terminal_output +8696,23981419,"TERMINAL",0,0,"watch",,terminal_focus +8697,23981907,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +8698,23984124,"TERMINAL",0,0,"idling",,terminal_command +8699,23984191,"TERMINAL",0,0,"]633;E;2025-08-14 17:17:17 idling;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Aug 14 17:17:17 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 258 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated: 25 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 1 nodes idle\rPartition large:\t 8 nodes idle",,terminal_output +8700,23985253,"TERMINAL",0,0,"8\t ",,terminal_output +8701,23986273,"TERMINAL",0,0,"9\t ",,terminal_output +8702,23987341,"TERMINAL",0,0,"20\t ",,terminal_output +8703,23988424,"TERMINAL",0,0,"1\t ",,terminal_output +8704,23989655,"TERMINAL",0,0,"2\t ",,terminal_output +8705,23990487,"TERMINAL",0,0,"3\t ",,terminal_output +8706,23991703,"TERMINAL",0,0,"4\t ",,terminal_output +8707,23992502,"TERMINAL",0,0,"5\t ",,terminal_output +8708,23993548,"TERMINAL",0,0,"6\t ",,terminal_output +8709,23994575,"TERMINAL",0,0,"7\t ",,terminal_output +8710,23995619,"TERMINAL",0,0,"8\t ",,terminal_output +8711,23996674,"TERMINAL",0,0,"9\t ",,terminal_output +8712,23997691,"TERMINAL",0,0,"30\t ",,terminal_output +8713,23998768,"TERMINAL",0,0,"1\t ",,terminal_output +8714,23999772,"TERMINAL",0,0,"2\t ",,terminal_output +8715,24000816,"TERMINAL",0,0,"3\t ",,terminal_output +8716,24001851,"TERMINAL",0,0,"47",,terminal_output +8717,24002963,"TERMINAL",0,0,"6\t ",,terminal_output +8718,24003989,"TERMINAL",0,0,"7\t ",,terminal_output +8719,24005013,"TERMINAL",0,0,"8\t ",,terminal_output +8720,24006037,"TERMINAL",0,0,"9\t ",,terminal_output +8721,24006875,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +8722,24007051,"TERMINAL",0,0,"40\t ",,terminal_output +8723,24007882,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.19.11\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250814_171739-3423234\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run dynamics-maskgit-8-node-darkness-filter-3423234\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/3423234\r\n",,terminal_output +8724,24008187,"TERMINAL",0,0,"1\t ",,terminal_output +8725,24009313,"TERMINAL",0,0,"2\t ",,terminal_output +8726,24010235,"TERMINAL",0,0,"3\t ",,terminal_output +8727,24011220,"TERMINAL",0,0,"4\t ",,terminal_output +8728,24012258,"TERMINAL",0,0,"5\t ",,terminal_output +8729,24013307,"TERMINAL",0,0,"6\t ",,terminal_output +8730,24014640,"TERMINAL",0,0,"7\t ",,terminal_output +8731,24015440,"TERMINAL",0,0,"8\t ",,terminal_output +8732,24016487,"TERMINAL",0,0,"9\t ",,terminal_output +8733,24017692,"TERMINAL",0,0,"50\t ",,terminal_output +8734,24018581,"TERMINAL",0,0,"1\t ",,terminal_output +8735,24019553,"TERMINAL",0,0,"2\t ",,terminal_output +8736,24020585,"TERMINAL",0,0,"3\t ",,terminal_output +8737,24021172,"TERMINAL",0,0,"tail",,terminal_focus +8738,24021627,"TERMINAL",0,0,"4\t ",,terminal_output +8739,24022711,"TERMINAL",0,0,"5\t ",,terminal_output +8740,24023708,"TERMINAL",0,0,"6\t ",,terminal_output +8741,24024806,"TERMINAL",0,0,"7\t ",,terminal_output +8742,24025786,"TERMINAL",0,0,"8\t ",,terminal_output +8743,24026826,"TERMINAL",0,0,"9\t ",,terminal_output +8744,24027948,"TERMINAL",0,0,"8:01\t ",,terminal_output +8745,24028975,"TERMINAL",0,0,"2\t ",,terminal_output +8746,24029951,"TERMINAL",0,0,"3\t ",,terminal_output +8747,24030990,"TERMINAL",0,0,"4\t ",,terminal_output +8748,24032030,"TERMINAL",0,0,"5\t ",,terminal_output +8749,24033071,"TERMINAL",0,0,"6\t ",,terminal_output +8750,24034121,"TERMINAL",0,0,"7\t ",,terminal_output +8751,24035153,"TERMINAL",0,0,"8\t ",,terminal_output +8752,24036192,"TERMINAL",0,0,"9\t ",,terminal_output +8753,24037272,"TERMINAL",0,0,"10\t ",,terminal_output +8754,24038277,"TERMINAL",0,0,"1\t ",,terminal_output +8755,24039314,"TERMINAL",0,0,"2\t ",,terminal_output +8756,24040353,"TERMINAL",0,0,"3\t ",,terminal_output +8757,24041398,"TERMINAL",0,0,"4\t ",,terminal_output +8758,24042438,"TERMINAL",0,0,"5\t ",,terminal_output +8759,24043517,"TERMINAL",0,0,"6\t ",,terminal_output +8760,24044539,"TERMINAL",0,0,"7\t ",,terminal_output +8761,24045568,"TERMINAL",0,0,"8\t ",,terminal_output +8762,24046616,"TERMINAL",0,0,"9\t ",,terminal_output +8763,24047655,"TERMINAL",0,0,"20\t ",,terminal_output +8764,24048695,"TERMINAL",0,0,"1\t ",,terminal_output +8765,24049762,"TERMINAL",0,0,"2\t ",,terminal_output +8766,24050835,"TERMINAL",0,0,"3\t ",,terminal_output +8767,24051913,"TERMINAL",0,0,"4\t ",,terminal_output +8768,24052936,"TERMINAL",0,0,"6\t ",,terminal_output +8769,24053961,"TERMINAL",0,0,"7\t ",,terminal_output +8770,24054971,"TERMINAL",0,0,"8\t ",,terminal_output +8771,24056010,"TERMINAL",0,0,"9\t ",,terminal_output +8772,24057143,"TERMINAL",0,0,"30\t ",,terminal_output +8773,24058096,"TERMINAL",0,0,"1\t ",,terminal_output +8774,24059193,"TERMINAL",0,0,"2\t ",,terminal_output +8775,24060183,"TERMINAL",0,0,"3\t ",,terminal_output +8776,24061231,"TERMINAL",0,0,"4\t ",,terminal_output +8777,24062274,"TERMINAL",0,0,"5\t ",,terminal_output +8778,24063396,"TERMINAL",0,0,"6\t ",,terminal_output +8779,24064346,"TERMINAL",0,0,"71",,terminal_output +8780,24065404,"TERMINAL",0,0,"8\t ",,terminal_output +8781,24066457,"TERMINAL",0,0,"9\t ",,terminal_output +8782,24067461,"TERMINAL",0,0,"40\t ",,terminal_output +8783,24068708,"TERMINAL",0,0,"1\t ",,terminal_output +8784,24069732,"TERMINAL",0,0,"2\t ",,terminal_output +8785,24070584,"TERMINAL",0,0,"3\t ",,terminal_output +8786,24071621,"TERMINAL",0,0,"4\t ",,terminal_output +8787,24072667,"TERMINAL",0,0,"5\t ",,terminal_output +8788,24073708,"TERMINAL",0,0,"69",,terminal_output +8789,24074787,"TERMINAL",0,0,"7\t ",,terminal_output +8790,24075873,"TERMINAL",0,0,"8\t ",,terminal_output +8791,24076879,"TERMINAL",0,0,"9\t ",,terminal_output +8792,24077863,"TERMINAL",0,0,"518",,terminal_output +8793,24078911,"TERMINAL",0,0,"2\t ",,terminal_output +8794,24079972,"TERMINAL",0,0,"3\t ",,terminal_output +8795,24081038,"TERMINAL",0,0,"4\t ",,terminal_output +8796,24082123,"TERMINAL",0,0,"5\t ",,terminal_output +8797,24083148,"TERMINAL",0,0,"6\t ",,terminal_output +8798,24084098,"TERMINAL",0,0,"7\t ",,terminal_output +8799,24085192,"TERMINAL",0,0,"8\t ",,terminal_output +8800,24086176,"TERMINAL",0,0,"9\t ",,terminal_output +8801,24087240,"TERMINAL",0,0,"9:00\t ",,terminal_output +8802,24088261,"TERMINAL",0,0,"1\t ",,terminal_output +8803,24088982,"TERMINAL",0,0,"WARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\n",,terminal_output +8804,24089394,"TERMINAL",0,0,"2\t ",,terminal_output +8805,24090416,"TERMINAL",0,0,"3\t ",,terminal_output +8806,24090983,"TERMINAL",0,0,"WARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Dropping 18 examples of 89394 examples (shard 32).\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 120000\r\nWARNING:absl:Missing metrics for step 120000\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nWARNING:absl:Missing metrics for step 200000\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nWARNING:absl:Missing metrics for step 180000\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 180000\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nWARNING:absl:Missing metrics for step 200000\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 289000\r\nWARNING:absl:Missing metrics for step 160000\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 289000\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nWARNING:absl:Missing metrics for step 280000\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 200000\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 180000\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nWARNING:absl:Missing metrics for step 289000\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nWARNING:absl:Missing metrics for step 120000\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 240000\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 240000\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nWARNING:absl:Missing metrics for step 280000\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 180000\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nWARNING:absl:Missing metrics for step 40000\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 200000\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 240000\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 240000\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 120000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 160000\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 120000\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 120000\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 290000\r\nWARNING:absl:Missing metrics for step 160000\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nWARNING:absl:Missing metrics for step 160000\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 20000\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nWARNING:absl:Missing metrics for step 280000\r\nWARNING:absl:Missing metrics for step 120000\r\n",,terminal_output +8807,24091094,"TERMINAL",0,0,"ERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 180000\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 260000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nWARNING:absl:Missing metrics for step 20000\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nWARNING:absl:Missing metrics for step 280000\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nWARNING:absl:Missing metrics for step 289000\r\nWARNING:absl:Missing metrics for step 220000\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 260000\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/260000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 120000\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/120000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 180000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/180000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 290000\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/290000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 140000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/140000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nWARNING:absl:Missing metrics for step 289000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/289000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 200000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/200000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 220000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/220000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/080000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 240000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/240000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 291000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/291000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 100000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/100000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 280000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/280000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 160000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/160000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401/020000/metrics/metrics not found.\r\n",,terminal_output +8808,24091421,"TERMINAL",0,0,"4\t ",,terminal_output +8809,24092061,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1251: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +8810,24092462,"TERMINAL",0,0,"5\t ",,terminal_output +8811,24093461,"TERMINAL",0,0,"6\t ",,terminal_output +8812,24094547,"TERMINAL",0,0,"7\t ",,terminal_output +8813,24095060,"TERMINAL",0,0,"Running on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\nRunning on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\n",,terminal_output +8814,24095577,"TERMINAL",0,0,"8\t ",,terminal_output +8815,24096059,"TERMINAL",0,0,"Running on 32 devices.\r\nCounting all components: ['dynamics', 'lam', 'tokenizer']\r\nParameter counts:\r\n{'dynamics': 271933440, 'lam': 35118240, 'tokenizer': 33750256, 'total': 340801936}\r\nStarting training from step 0...\r\n",,terminal_output +8816,24096588,"TERMINAL",0,0,"99",,terminal_output +8817,24097626,"TERMINAL",0,0,"10\t ",,terminal_output +8818,24098671,"TERMINAL",0,0,"1\t ",,terminal_output +8819,24099714,"TERMINAL",0,0,"2\t ",,terminal_output +8820,24100748,"TERMINAL",0,0,"3\t ",,terminal_output +8821,24101787,"TERMINAL",0,0,"4\t ",,terminal_output +8822,24102908,"TERMINAL",0,0,"5\t ",,terminal_output +8823,24103930,"TERMINAL",0,0,"7\t ",,terminal_output +8824,24104913,"TERMINAL",0,0,"8\t ",,terminal_output +8825,24106133,"TERMINAL",0,0,"98",,terminal_output +8826,24107218,"TERMINAL",0,0,"20\t ",,terminal_output +8827,24108232,"TERMINAL",0,0,"1\t ",,terminal_output +8828,24109261,"TERMINAL",0,0,"2\t ",,terminal_output +8829,24110322,"TERMINAL",0,0,"3\t ",,terminal_output +8830,24111545,"TERMINAL",0,0,"4\t ",,terminal_output +8831,24112573,"TERMINAL",0,0,"5\t ",,terminal_output +8832,24113557,"TERMINAL",0,0,"6\t ",,terminal_output +8833,24114582,"TERMINAL",0,0,"7\t ",,terminal_output +8834,24115507,"TERMINAL",0,0,"8\t ",,terminal_output +8835,24116634,"TERMINAL",0,0,"9\t ",,terminal_output +8836,24117576,"TERMINAL",0,0,"30\t ",,terminal_output +8837,24118681,"TERMINAL",0,0,"1\t ",,terminal_output +8838,24119723,"TERMINAL",0,0,"2\t ",,terminal_output +8839,24120706,"TERMINAL",0,0,"3\t ",,terminal_output +8840,24121765,"TERMINAL",0,0,"4\t ",,terminal_output +8841,24122761,"TERMINAL",0,0,"5\t ",,terminal_output +8842,24123808,"TERMINAL",0,0,"6\t ",,terminal_output +8843,24124850,"TERMINAL",0,0,"7\t ",,terminal_output +8844,24125888,"TERMINAL",0,0,"9\t ",,terminal_output +8845,24126926,"TERMINAL",0,0,"40\t ",,terminal_output +8846,24127070,"TERMINAL",0,0,"2025-08-14 17:19:39.267208: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-08-14 17:19:39.377678: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-08-14 17:19:39.377723: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-08-14 17:19:39.458650: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-08-14 17:19:39.491307: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +8847,24127996,"TERMINAL",0,0,"1\t ",,terminal_output +8848,24129021,"TERMINAL",0,0,"2\t ",,terminal_output +8849,24130049,"TERMINAL",0,0,"3\t ",,terminal_output +8850,24131171,"TERMINAL",0,0,"2025-08-14 17:19:44.016889: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +8851,24131207,"TERMINAL",0,0,"4\t ",,terminal_output +8852,24132097,"TERMINAL",0,0,"2025-08-14 17:19:44.269962: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +8853,24132502,"TERMINAL",0,0,"50",,terminal_output +8854,24133566,"TERMINAL",0,0,"6\t ",,terminal_output +8855,24134580,"TERMINAL",0,0,"7\t ",,terminal_output +8856,24135540,"TERMINAL",0,0,"8\t ",,terminal_output +8857,24136701,"TERMINAL",0,0,"9\t ",,terminal_output +8858,24137727,"TERMINAL",0,0,"50\t ",,terminal_output +8859,24138708,"TERMINAL",0,0,"1\t ",,terminal_output +8860,24139775,"TERMINAL",0,0,"2\t ",,terminal_output +8861,24140796,"TERMINAL",0,0,"3\t ",,terminal_output +8862,24141790,"TERMINAL",0,0,"4\t ",,terminal_output +8863,24142825,"TERMINAL",0,0,"5\t ",,terminal_output +8864,24143871,"TERMINAL",0,0,"7\t ",,terminal_output +8865,24144994,"TERMINAL",0,0,"8\t ",,terminal_output +8866,24145956,"TERMINAL",0,0,"9\t ",,terminal_output +8867,24146993,"TERMINAL",0,0,"20:00\t ",,terminal_output +8868,24148029,"TERMINAL",0,0,"1\t ",,terminal_output +8869,24149068,"TERMINAL",0,0,"2\t ",,terminal_output +8870,24150133,"TERMINAL",0,0,"3\t ",,terminal_output +8871,24151306,"TERMINAL",0,0,"4\t ",,terminal_output +8872,24152264,"TERMINAL",0,0,"5\t ",,terminal_output +8873,24153222,"TERMINAL",0,0,"6\t ",,terminal_output +8874,24154263,"TERMINAL",0,0,"7\t ",,terminal_output +8875,24155293,"TERMINAL",0,0,"8\t ",,terminal_output +8876,24156360,"TERMINAL",0,0,"9\t ",,terminal_output +8877,24157385,"TERMINAL",0,0,"10\t ",,terminal_output +8878,24158614,"TERMINAL",0,0,"1\t ",,terminal_output +8879,24159571,"TERMINAL",0,0,"2\t ",,terminal_output +8880,24160518,"TERMINAL",0,0,"3\t ",,terminal_output +8881,24161614,"TERMINAL",0,0,"4\t ",,terminal_output +8882,24162715,"TERMINAL",0,0,"5\t ",,terminal_output +8883,24163646,"TERMINAL",0,0,"6\t ",,terminal_output +8884,24164765,"TERMINAL",0,0,"7\t ",,terminal_output +8885,24165782,"TERMINAL",0,0,"8\t ",,terminal_output +8886,24166806,"TERMINAL",0,0,"9\t ",,terminal_output +8887,24167794,"TERMINAL",0,0,"20\t ",,terminal_output +8888,24168863,"TERMINAL",0,0,"1\t ",,terminal_output +8889,24169878,"TERMINAL",0,0,"3\t ",,terminal_output +8890,24171009,"TERMINAL",0,0,"4\t ",,terminal_output +8891,24172029,"TERMINAL",0,0,"5\t ",,terminal_output +8892,24173054,"TERMINAL",0,0,"6\t ",,terminal_output +8893,24174077,"TERMINAL",0,0,"7\t ",,terminal_output +8894,24175099,"TERMINAL",0,0,"8\t ",,terminal_output +8895,24176225,"TERMINAL",0,0,"9\t ",,terminal_output +8896,24177250,"TERMINAL",0,0,"30\t ",,terminal_output +8897,24178275,"TERMINAL",0,0,"1\t ",,terminal_output +8898,24179299,"TERMINAL",0,0,"2\t ",,terminal_output +8899,24180283,"TERMINAL",0,0,"3\t ",,terminal_output +8900,24181348,"TERMINAL",0,0,"4\t ",,terminal_output +8901,24182474,"TERMINAL",0,0,"5\t ",,terminal_output +8902,24183631,"TERMINAL",0,0,"6\t ",,terminal_output +8903,24184457,"TERMINAL",0,0,"7\t ",,terminal_output +8904,24185501,"TERMINAL",0,0,"8\t ",,terminal_output +8905,24186679,"TERMINAL",0,0,"9\t ",,terminal_output +8906,24187593,"TERMINAL",0,0,"40\t ",,terminal_output +8907,24188630,"TERMINAL",0,0,"1\t ",,terminal_output +8908,24189747,"TERMINAL",0,0,"2\t ",,terminal_output +8909,24190721,"TERMINAL",0,0,"3\t ",,terminal_output +8910,24191756,"TERMINAL",0,0,"4\t ",,terminal_output +8911,24192815,"TERMINAL",0,0,"5\t ",,terminal_output +8912,24193841,"TERMINAL",0,0,"6\t ",,terminal_output +8913,24194875,"TERMINAL",0,0,"8\t ",,terminal_output +8914,24195991,"TERMINAL",0,0,"9\t ",,terminal_output +8915,24196961,"TERMINAL",0,0,"50\t ",,terminal_output +8916,24197998,"TERMINAL",0,0,"1\t ",,terminal_output +8917,24199037,"TERMINAL",0,0,"2\t ",,terminal_output +8918,24200084,"TERMINAL",0,0,"3\t ",,terminal_output +8919,24201232,"TERMINAL",0,0,"4\t ",,terminal_output +8920,24202238,"TERMINAL",0,0,"5\t ",,terminal_output +8921,24203263,"TERMINAL",0,0,"6\t ",,terminal_output +8922,24204282,"TERMINAL",0,0,"7\t ",,terminal_output +8923,24205288,"TERMINAL",0,0,"8\t ",,terminal_output +8924,24206333,"TERMINAL",0,0,"9\t ",,terminal_output +8925,24207459,"TERMINAL",0,0,"1:00\t ",,terminal_output +8926,24208617,"TERMINAL",0,0,"1\t ",,terminal_output +8927,24209506,"TERMINAL",0,0,"28",,terminal_output +8928,24210487,"TERMINAL",0,0,"3\t ",,terminal_output +8929,24211643,"TERMINAL",0,0,"4\t ",,terminal_output +8930,24212618,"TERMINAL",0,0,"5\t ",,terminal_output +8931,24213617,"TERMINAL",0,0,"6\t ",,terminal_output +8932,24214657,"TERMINAL",0,0,"7\t ",,terminal_output +8933,24215693,"TERMINAL",0,0,"8\t ",,terminal_output +8934,24216778,"TERMINAL",0,0,"9\t ",,terminal_output +8935,24217779,"TERMINAL",0,0,"10\t ",,terminal_output +8936,24218837,"TERMINAL",0,0,"1\t ",,terminal_output +8937,24219856,"TERMINAL",0,0,"3\t ",,terminal_output +8938,24220896,"TERMINAL",0,0,"4\t ",,terminal_output +8939,24221939,"TERMINAL",0,0,"5\t ",,terminal_output +8940,24223019,"TERMINAL",0,0,"6\t ",,terminal_output +8941,24224058,"TERMINAL",0,0,"7\t ",,terminal_output +8942,24225083,"TERMINAL",0,0,"8\t ",,terminal_output +8943,24226199,"TERMINAL",0,0,"9\t ",,terminal_output +8944,24227223,"TERMINAL",0,0,"20\t ",,terminal_output +8945,24228216,"TERMINAL",0,0,"1\t ",,terminal_output +8946,24229270,"TERMINAL",0,0,"2\t ",,terminal_output +8947,24230294,"TERMINAL",0,0,"3\t ",,terminal_output +8948,24231318,"TERMINAL",0,0,"4\t ",,terminal_output +8949,24232546,"TERMINAL",0,0,"56",,terminal_output +8950,24233469,"TERMINAL",0,0,"6\t ",,terminal_output +8951,24234618,"TERMINAL",0,0,"7\t ",,terminal_output +8952,24235471,"TERMINAL",0,0,"8\t ",,terminal_output +8953,24236539,"TERMINAL",0,0,"9\t ",,terminal_output +8954,24237591,"TERMINAL",0,0,"30\t ",,terminal_output +8955,24238692,"TERMINAL",0,0,"1\t ",,terminal_output +8956,24239732,"TERMINAL",0,0,"2\t ",,terminal_output +8957,24240682,"TERMINAL",0,0,"3\t ",,terminal_output +8958,24241776,"TERMINAL",0,0,"4\t ",,terminal_output +8959,24242757,"TERMINAL",0,0,"5\t ",,terminal_output +8960,24243854,"TERMINAL",0,0,"6\t ",,terminal_output +8961,24244836,"TERMINAL",0,0,"7\t ",,terminal_output +8962,24245877,"TERMINAL",0,0,"9\t ",,terminal_output +8963,24246955,"TERMINAL",0,0,"40\t ",,terminal_output +8964,24248009,"TERMINAL",0,0,"1\t ",,terminal_output +8965,24249034,"TERMINAL",0,0,"2\t ",,terminal_output +8966,24250059,"TERMINAL",0,0,"3\t ",,terminal_output +8967,24251082,"TERMINAL",0,0,"4\t ",,terminal_output +8968,24252207,"TERMINAL",0,0,"5\t ",,terminal_output +8969,24253233,"TERMINAL",0,0,"6\t ",,terminal_output +8970,24254258,"TERMINAL",0,0,"7\t ",,terminal_output +8971,24255281,"TERMINAL",0,0,"8\t ",,terminal_output +8972,24256305,"TERMINAL",0,0,"9\t ",,terminal_output +8973,24257337,"TERMINAL",0,0,"50\t ",,terminal_output +8974,24258352,"TERMINAL",0,0,"1\t ",,terminal_output +8975,24259385,"TERMINAL",0,0,"2\t ",,terminal_output +8976,24260410,"TERMINAL",0,0,"3\t ",,terminal_output +8977,24261445,"TERMINAL",0,0,"4\t ",,terminal_output +8978,24262482,"TERMINAL",0,0,"5\t ",,terminal_output +8979,24263609,"TERMINAL",0,0,"6\t ",,terminal_output +8980,24264703,"TERMINAL",0,0,"7\t ",,terminal_output +8981,24265631,"TERMINAL",0,0,"8\t ",,terminal_output +8982,24266715,"TERMINAL",0,0,"9\t ",,terminal_output +8983,24267669,"TERMINAL",0,0,"2:00\t ",,terminal_output +8984,24268703,"TERMINAL",0,0,"1\t ",,terminal_output +8985,24269742,"TERMINAL",0,0,"2\t ",,terminal_output +8986,24270782,"TERMINAL",0,0,"3\t ",,terminal_output +8987,24271828,"TERMINAL",0,0,"4\t ",,terminal_output +8988,24272881,"TERMINAL",0,0,"6\t ",,terminal_output +8989,24273901,"TERMINAL",0,0,"7\t ",,terminal_output +8990,24275046,"TERMINAL",0,0,"8\t ",,terminal_output +8991,24275983,"TERMINAL",0,0,"9\t ",,terminal_output +8992,24277092,"TERMINAL",0,0,"10\t ",,terminal_output +8993,24278116,"TERMINAL",0,0,"1\t ",,terminal_output +8994,24279142,"TERMINAL",0,0,"25",,terminal_output +8995,24280161,"TERMINAL",0,0,"3\t ",,terminal_output +8996,24281188,"TERMINAL",0,0,"4\t ",,terminal_output +8997,24282314,"TERMINAL",0,0,"5\t ",,terminal_output +8998,24283338,"TERMINAL",0,0,"6\t ",,terminal_output +8999,24284298,"TERMINAL",0,0,"7\t ",,terminal_output +9000,24285428,"TERMINAL",0,0,"8\t ",,terminal_output +9001,24286409,"TERMINAL",0,0,"9\t ",,terminal_output +9002,24287447,"TERMINAL",0,0,"20\t ",,terminal_output +9003,24288457,"TERMINAL",0,0,"1\t ",,terminal_output +9004,24289689,"TERMINAL",0,0,"2\t ",,terminal_output +9005,24290527,"TERMINAL",0,0,"3\t ",,terminal_output +9006,24291626,"TERMINAL",0,0,"4\t ",,terminal_output +9007,24292612,"TERMINAL",0,0,"5\t ",,terminal_output +9008,24293403,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;130",,terminal_output +9009,24293635,"TERMINAL",0,0,"6\t ",,terminal_output +9010,24294673,"TERMINAL",0,0,"7\t ",,terminal_output +9011,24295711,"TERMINAL",0,0,"8\t ",,terminal_output +9012,24296745,"TERMINAL",0,0,"9\t ",,terminal_output +9013,24297796,"TERMINAL",0,0,"30\t ",,terminal_output +9014,24298640,"TERMINAL",0,0,"queue",,terminal_command +9015,24298675,"TERMINAL",0,0,"]633;E;2025-08-14 17:22:31 queue;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C",,terminal_output +9016,24298744,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 17:22:31 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4 (Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2 (Priority)3412401 accelerat train_to tum_cte0 R 1-23:22:30\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-03:33:35\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3423234 accelerat train_dy tum_cte0 R\t5:53\t 8 hkn[0521-0528]3422008 accelerat train_dy tum_cte0 R 5:09:21\t 1 hkn07133421964 accelerat train_to tum_cte0 R 5:21:45\t 1 hkn0720",,terminal_output +9017,24298839,"TERMINAL",0,0,"1\t ",,terminal_output +9018,24299804,"TERMINAL",0,0,"216426",,terminal_output +9019,24299873,"TERMINAL",0,0,"3\t ",,terminal_output +9020,24300793,"TERMINAL",0,0,"327537",,terminal_output +9021,24300923,"TERMINAL",0,0,"4\t ",,terminal_output +9022,24301868,"TERMINAL",0,0,"449759",,terminal_output +9023,24301998,"TERMINAL",0,0,"5\t ",,terminal_output +9024,24302324,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;0",,terminal_output +9025,24303073,"TERMINAL",0,0,"6\t ",,terminal_output +9026,24304096,"TERMINAL",0,0,"7\t ",,terminal_output +9027,24305064,"TERMINAL",0,0,"8\t ",,terminal_output +9028,24306110,"TERMINAL",0,0,"9\t ",,terminal_output +9029,24307152,"TERMINAL",0,0,"40\t ",,terminal_output +9030,24308191,"TERMINAL",0,0,"1\t ",,terminal_output +9031,24309235,"TERMINAL",0,0,"2\t ",,terminal_output +9032,24310300,"TERMINAL",0,0,"3\t ",,terminal_output +9033,24310789,"TERMINAL",0,0,"scancel 3423234",,terminal_command +9034,24311308,"TERMINAL",0,0,"433",,terminal_output +9035,24312355,"TERMINAL",0,0,"5\t ",,terminal_output +9036,24312896,"TERMINAL",0,0,"watch",,terminal_focus +9037,24313450,"TERMINAL",0,0,"6\t ",,terminal_output +9038,24313537,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +9039,24318992,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +9040,24320875,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",472,29,"",shellscript,content +9041,24322329,"TERMINAL",0,0,"bash",,terminal_focus +9042,24325366,"TERMINAL",0,0,"bash",,terminal_focus +9043,24331067,"TERMINAL",0,0,"sync-runner",,terminal_command +9044,24331103,"TERMINAL",0,0,"]633;E;2025-08-14 17:23:04 sync-runner;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;Csending incremental file list\r\n",,terminal_output +9045,24331579,"TERMINAL",0,0,"[?25lr[?25h[?25lu[?25h",,terminal_output +9046,24331676,"TERMINAL",0,0,"./\r\nkiller.sh\r\n",,terminal_output +9047,24331792,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +9048,24332036,"TERMINAL",0,0,"[?25ln[?25h[?25le[?25h",,terminal_output +9049,24332089,"TERMINAL",0,0,"[?25lr[?25h",,terminal_output +9050,24332268,"TERMINAL",0,0,"\r\n",,terminal_output +9051,24332432,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch\r\nslurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch\r\n",,terminal_output +9052,24332449,"TERMINAL",0,0,"\r\nsent 29,014 bytes received 229 bytes 19,495.33 bytes/sec\r\ntotal size is 128,153,341 speedup is 4,382.36\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +9053,24332480,"TERMINAL",0,0,"runner",,terminal_command +9054,24332486,"TERMINAL",0,0,"]633;E;2025-08-14 17:23:05 runner;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +9055,24344128,"slurm/jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +9056,24345645,"TERMINAL",0,0,"bash",,terminal_focus +9057,24346815,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +9058,24347990,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",449,0,"",shellscript,selection_mouse +9059,24348132,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",442,29,"train_dynamics_maskgit_8_node",shellscript,selection_mouse +9060,24348266,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",423,49,"#SBATCH --job-name=train_dynamics_maskgit_8_node\n",shellscript,selection_mouse +9061,24349304,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",426,0,"",shellscript,selection_mouse +9062,24349305,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",424,6,"SBATCH",shellscript,selection_mouse +9063,24349538,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",424,7,"SBATCH ",shellscript,selection_mouse +9064,24349539,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",424,12,"SBATCH --job",shellscript,selection_mouse +9065,24349587,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",424,13,"SBATCH --job-",shellscript,selection_mouse +9066,24349588,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",424,17,"SBATCH --job-name",shellscript,selection_mouse +9067,24349651,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",424,47,"SBATCH --job-name=train_dynamics_maskgit_8_node",shellscript,selection_mouse +9068,24350341,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",456,0,"",shellscript,selection_mouse +9069,24350342,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",442,29,"train_dynamics_maskgit_8_node",shellscript,selection_mouse +9070,24350688,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",441,30,"=train_dynamics_maskgit_8_node",shellscript,selection_mouse +9071,24350688,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",437,34,"name=train_dynamics_maskgit_8_node",shellscript,selection_mouse +9072,24350773,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",436,35,"-name=train_dynamics_maskgit_8_node",shellscript,selection_mouse +9073,24350782,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",433,38,"job-name=train_dynamics_maskgit_8_node",shellscript,selection_mouse +9074,24350828,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",432,39,"-job-name=train_dynamics_maskgit_8_node",shellscript,selection_mouse +9075,24350910,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",431,40,"--job-name=train_dynamics_maskgit_8_node",shellscript,selection_mouse +9076,24350910,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",430,41," --job-name=train_dynamics_maskgit_8_node",shellscript,selection_mouse +9077,24350910,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",424,47,"SBATCH --job-name=train_dynamics_maskgit_8_node",shellscript,selection_mouse +9078,24351176,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",423,48,"#SBATCH --job-name=train_dynamics_maskgit_8_node",shellscript,selection_mouse +9079,24351606,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",423,0,"",shellscript,selection_mouse +9080,24351812,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",423,1,"#",shellscript,selection_mouse +9081,24351812,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",423,7,"#SBATCH",shellscript,selection_mouse +9082,24351888,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",423,13,"#SBATCH --job",shellscript,selection_mouse +9083,24351889,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",423,18,"#SBATCH --job-name",shellscript,selection_mouse +9084,24351927,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",423,19,"#SBATCH --job-name=",shellscript,selection_mouse +9085,24351928,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",423,48,"#SBATCH --job-name=train_dynamics_maskgit_8_node",shellscript,selection_mouse +9086,24352906,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",449,0,"",shellscript,selection_mouse +9087,24353530,"TERMINAL",0,0,"bash",,terminal_focus +9088,24357030,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",,terminal_command +9089,24357117,"TERMINAL",0,0,"]633;E;2025-08-14 17:23:30 sbatch slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;CSubmitted batch job 3423250\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +9090,24360195,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",874,0,"",shellscript,selection_mouse +9091,24399267,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",906,0,"",shellscript,selection_mouse +9092,24399736,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",781,0,"",shellscript,selection_mouse +9093,24400759,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",517,0,"",shellscript,selection_mouse +9094,24401271,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",541,0,"",shellscript,selection_mouse +9095,24402815,"TERMINAL",0,0,"bash",,terminal_focus +9096,24404778,"TERMINAL",0,0,"idling",,terminal_command +9097,24404837,"TERMINAL",0,0,"]633;E;2025-08-14 17:24:17 idling;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Aug 14 17:24:17 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 258 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated: 36 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 1 nodes idle\rPartition large:\t 8 nodes idle",,terminal_output +9098,24405911,"TERMINAL",0,0,"9\t ",,terminal_output +9099,24406923,"TERMINAL",0,0,"20\t ",,terminal_output +9100,24408063,"TERMINAL",0,0,"1\t ",,terminal_output +9101,24409087,"TERMINAL",0,0,"2\t ",,terminal_output +9102,24410122,"TERMINAL",0,0,"3\t ",,terminal_output +9103,24411133,"TERMINAL",0,0,"4\t ",,terminal_output +9104,24412159,"TERMINAL",0,0,"5\t ",,terminal_output +9105,24413184,"TERMINAL",0,0,"6\t ",,terminal_output +9106,24414219,"TERMINAL",0,0,"7\t ",,terminal_output +9107,24415332,"TERMINAL",0,0,"8\t ",,terminal_output +9108,24416461,"TERMINAL",0,0,"9\t ",,terminal_output +9109,24417334,"TERMINAL",0,0,"30\t ",,terminal_output +9110,24418405,"TERMINAL",0,0,"1\t ",,terminal_output +9111,24419430,"TERMINAL",0,0,"2\t ",,terminal_output +9112,24420457,"TERMINAL",0,0,"3\t ",,terminal_output +9113,24421581,"TERMINAL",0,0,"4\t ",,terminal_output +9114,24422636,"TERMINAL",0,0,"5\t ",,terminal_output +9115,24423664,"TERMINAL",0,0,"6\t [?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;0",,terminal_output +9116,24432319,"killer.sh",0,0,"",shellscript,tab +9117,24433750,"killer copy.sh",0,0,"response=$(curl -s https://home.cit.tum.de/~nguyenal/killer.html)\n\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n scancel --me && \\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,tab +9118,24439589,"killer_partition.sh",0,0,"response=$(curl -s https://home.cit.tum.de/~nguyenal/killer.html)\n\nif echo ""$response"" | grep -q ""KILL ALL JOBS""; then\n scancel --me && \\n echo ""killing all jobs""\nelse\n echo ""leaving jobs be""\nfi\n",shellscript,tab +9119,24440747,"killer_partition.sh",93,0,"",shellscript,selection_mouse +9120,24441273,"killer_partition.sh",103,0,"",shellscript,selection_mouse +9121,24442669,"killer_partition.sh",108,0,"",shellscript,selection_mouse +9122,24443617,"killer_partition.sh",104,0,"",shellscript,selection_mouse +9123,24444468,"killer_partition.sh",103,0,"",shellscript,selection_mouse +9124,24445217,"killer_partition.sh",103,4,"",shellscript,content +9125,24445644,"killer_partition.sh",103,0,"P",shellscript,content +9126,24445645,"killer_partition.sh",104,0,"",shellscript,selection_keyboard +9127,24446134,"killer_partition.sh",104,0,"A",shellscript,content +9128,24446135,"killer_partition.sh",105,0,"",shellscript,selection_keyboard +9129,24446556,"killer_partition.sh",105,0,"R",shellscript,content +9130,24446556,"killer_partition.sh",106,0,"",shellscript,selection_keyboard +9131,24446829,"killer_partition.sh",106,0,"T",shellscript,content +9132,24446830,"killer_partition.sh",107,0,"",shellscript,selection_keyboard +9133,24446905,"killer_partition.sh",107,0,"I",shellscript,content +9134,24446906,"killer_partition.sh",108,0,"",shellscript,selection_keyboard +9135,24447043,"killer_partition.sh",108,0,"T",shellscript,content +9136,24447044,"killer_partition.sh",109,0,"",shellscript,selection_keyboard +9137,24447219,"killer_partition.sh",109,0,"I",shellscript,content +9138,24447220,"killer_partition.sh",110,0,"",shellscript,selection_keyboard +9139,24447311,"killer_partition.sh",110,0,"O",shellscript,content +9140,24447311,"killer_partition.sh",111,0,"",shellscript,selection_keyboard +9141,24447469,"killer_partition.sh",111,0,"N",shellscript,content +9142,24447470,"killer_partition.sh",112,0,"",shellscript,selection_keyboard +9143,24447846,"killer_partition.sh",112,0," ",shellscript,content +9144,24447846,"killer_partition.sh",113,0,"",shellscript,selection_keyboard +9145,24448819,"killer_partition.sh",112,0,"",shellscript,selection_command +9146,24449753,"killer_partition.sh",100,0,"",shellscript,selection_mouse +9147,24449885,"killer_partition.sh",98,4,"KILL",shellscript,selection_mouse +9148,24450087,"killer_partition.sh",98,5,"KILL ",shellscript,selection_mouse +9149,24450124,"killer_partition.sh",98,14,"KILL PARTITION",shellscript,selection_mouse +9150,24450406,"killer_partition.sh",98,15,"KILL PARTITION ",shellscript,selection_mouse +9151,24450612,"killer_partition.sh",98,19,"KILL PARTITION JOBS",shellscript,selection_mouse +9152,24451546,"killer_partition.sh",170,0,"",shellscript,selection_mouse +9153,24452586,"killer_partition.sh",137,0,"",shellscript,selection_mouse +9154,24452747,"killer_partition.sh",137,1,"-",shellscript,selection_mouse +9155,24452789,"killer_partition.sh",137,2,"--",shellscript,selection_mouse +9156,24452869,"killer_partition.sh",137,3,"--m",shellscript,selection_mouse +9157,24452998,"killer_partition.sh",137,4,"--me",shellscript,selection_mouse +9158,24453524,"killer_partition.sh",137,4,"",shellscript,content +9159,24455350,"TERMINAL",0,0,"bash",,terminal_focus +9160,24456254,"TERMINAL",0,0,"queue",,terminal_command +9161,24456318,"TERMINAL",0,0,"]633;E;2025-08-14 17:25:09 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 17:25:09 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4 (Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2 (Priority)3423250 accelerat train_dy tum_cte0 PD\t0:00\t 8 (Priority)3412401 accelerat train_to tum_cte0 R 1-23:25:08\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-03:36:13\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 5:11:59\t 1 hkn07133421964 accelerat train_to tum_cte0 R 5:24:23\t 1 hkn0720",,terminal_output +9162,24457356,"TERMINAL",0,0,"10942:004",,terminal_output +9163,24458402,"TERMINAL",0,0,"110515",,terminal_output +9164,24459472,"TERMINAL",0,0,"21626",,terminal_output +9165,24460475,"TERMINAL",0,0,"32737",,terminal_output +9166,24460853,"killer_partition.sh",136,0,"",shellscript,selection_mouse +9167,24461543,"TERMINAL",0,0,"43848",,terminal_output +9168,24462200,"killer_partition.sh",137,0,"",shellscript,selection_command +9169,24462465,"killer_partition.sh",137,0,"3421964",shellscript,content +9170,24462597,"TERMINAL",0,0,"54959",,terminal_output +9171,24463196,"killer_partition.sh",144,0," ",shellscript,content +9172,24463197,"killer_partition.sh",145,0,"",shellscript,selection_keyboard +9173,24463660,"TERMINAL",0,0,"6520630",,terminal_output +9174,24464701,"TERMINAL",0,0,"76171",,terminal_output +9175,24465749,"TERMINAL",0,0,"87282",,terminal_output +9176,24466842,"TERMINAL",0,0,"98393",,terminal_output +9177,24467776,"TERMINAL",0,0,"2094104",,terminal_output +9178,24467812,"killer_partition.sh",145,0,"3422008",shellscript,content +9179,24468822,"TERMINAL",0,0,"121626",,terminal_output +9180,24469602,"killer_partition.sh",190,0,"",shellscript,selection_mouse +9181,24469866,"TERMINAL",0,0,"32737",,terminal_output +9182,24470967,"TERMINAL",0,0,"43848",,terminal_output +9183,24471960,"TERMINAL",0,0,"54959",,terminal_output +9184,24473091,"TERMINAL",0,0,"6530640",,terminal_output +9185,24474146,"TERMINAL",0,0,"76171",,terminal_output +9186,24475025,"killer_partition.sh",221,0,"",shellscript,selection_mouse +9187,24475075,"TERMINAL",0,0,"87282",,terminal_output +9188,24476160,"TERMINAL",0,0,"98393",,terminal_output +9189,24477176,"TERMINAL",0,0,"3094204",,terminal_output +9190,24478225,"TERMINAL",0,0,"130515",,terminal_output +9191,24479268,"TERMINAL",0,0,"21626",,terminal_output +9192,24480315,"TERMINAL",0,0,"32737",,terminal_output +9193,24481384,"TERMINAL",0,0,"43848",,terminal_output +9194,24482660,"TERMINAL",0,0,"54959",,terminal_output +9195,24483532,"TERMINAL",0,0,"6540650",,terminal_output +9196,24484558,"TERMINAL",0,0,"76171",,terminal_output +9197,24485535,"TERMINAL",0,0,"87282",,terminal_output +9198,24486604,"TERMINAL",0,0,"98393",,terminal_output +9199,24487729,"TERMINAL",0,0,"4094304",,terminal_output +9200,24488149,"TERMINAL",0,0,"bash",,terminal_focus +9201,24488695,"TERMINAL",0,0,"140515",,terminal_output +9202,24489762,"TERMINAL",0,0,"21626",,terminal_output +9203,24490747,"TERMINAL",0,0,"32737",,terminal_output +9204,24491802,"TERMINAL",0,0,"43848",,terminal_output +9205,24492778,"TERMINAL",0,0,"tmux a",,terminal_command +9206,24492836,"TERMINAL",0,0,"]633;E;2025-08-14 17:25:45 tmux a;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h[?1h=[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[>c[>q[?25lEvery 60.0s: sh killer.shhkn1993.localdomain: Thu Aug 14 17:24:57 2025\r\nleaving jobs be\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n \r\n[0] 0:watch* ""hkn1993.localdomain"" 17:25 14-Aug-25(B[?12l[?25h(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[?25lEvery 60.0s: sh killer.shhkn1993.localdomain: Thu Aug 14 17:24:57 2025\r\nleaving jobs be\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n \r\n[0] 0:watch* ""hkn1993.localdomain"" 17:25 14-Aug-25(B[?12l[?25h",,terminal_output +9207,24492872,"TERMINAL",0,0,"555065:00",,terminal_output +9208,24492904,"TERMINAL",0,0,"[?7727h",,terminal_output +9209,24493976,"TERMINAL",0,0,"76171",,terminal_output +9210,24494943,"TERMINAL",0,0,"87282",,terminal_output +9211,24495009,"TERMINAL",0,0,"[?25l(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[?12l[?25h[?25l[0] 0:bash* ""hkn1993.localdomain"" 17:25 14-Aug-25(B[?12l[?25h(jafar) [tum_cte0515@hkn1993 jafar]$ [?2004h",,terminal_output +9212,24496025,"TERMINAL",0,0,"98393",,terminal_output +9213,24497153,"TERMINAL",0,0,"5094404",,terminal_output +9214,24498074,"TERMINAL",0,0,"150515",,terminal_output +9215,24499199,"TERMINAL",0,0,"21626",,terminal_output +9216,24500225,"TERMINAL",0,0,"32737",,terminal_output +9217,24501216,"TERMINAL",0,0,"43848",,terminal_output +9218,24502249,"TERMINAL",0,0,"54959",,terminal_output +9219,24503445,"TERMINAL",0,0,"657:00610",,terminal_output +9220,24504334,"TERMINAL",0,0,"76171",,terminal_output +9221,24505378,"TERMINAL",0,0,"87282",,terminal_output +9222,24506417,"TERMINAL",0,0,"98393",,terminal_output +9223,24507470,"TERMINAL",0,0,"6:0094504",,terminal_output +9224,24507791,"TERMINAL",0,0,"[?25l[0] 0:bash* ""hkn1993.localdomain"" 17:26 14-Aug-25(B[?12l[?25h",,terminal_output +9225,24508510,"TERMINAL",0,0,"16:00515",,terminal_output +9226,24509628,"TERMINAL",0,0,"21626",,terminal_output +9227,24510603,"TERMINAL",0,0,"32737",,terminal_output +9228,24511061,"TERMINAL",0,0,"[?25l│││││││││││││││││││(B \r\n(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n(jafar) [tum_cte0515@hkn1993 jafar]$ watch -n60 ""sh killer.sh""\r\n(jafar) [tum_cte0515@hkn1993 jafar]$ \n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\r\n[0] 0:bash* ""hkn1993.localdomain"" 17:26 14-Aug-25(B[?12l[?25h[?2004l[?25l│││││││││││││││││││(B\r\n[0] 0:tmux* ""hkn1993.localdomain"" 17:26 14-Aug-25(B[?12l[?25h(jafar) [tum_cte0515@hkn1993 jafar]$ ",,terminal_output +9229,24511702,"TERMINAL",0,0,"43848",,terminal_output +9230,24512716,"TERMINAL",0,0,"54959",,terminal_output +9231,24512725,"TERMINAL",0,0,"[?25l│││││││││││││││││││(B\r\n[0] 0:bash* ""hkn1993.localdomain"" 17:26 14-Aug-25(B[?12l[?25h[?2004h(jafar) [tum_cte0515@hkn1993 jafar]$ ",,terminal_output +9232,24513725,"TERMINAL",0,0,"6510620",,terminal_output +9233,24513934,"TERMINAL",0,0,"[?25l│││││││││││││││││││(B[?12l[?25h",,terminal_output +9234,24514765,"TERMINAL",0,0,"76171",,terminal_output +9235,24514934,"TERMINAL",0,0,"watch -n60 ""sh killer.sh""",,terminal_output +9236,24515809,"TERMINAL",0,0,"87282",,terminal_output +9237,24516864,"TERMINAL",0,0,"9943:004",,terminal_output +9238,24517910,"TERMINAL",0,0,"1110515",,terminal_output +9239,24519064,"TERMINAL",0,0,"21626",,terminal_output +9240,24519997,"TERMINAL",0,0,"32737",,terminal_output +9241,24521122,"TERMINAL",0,0,"43848",,terminal_output +9242,24522088,"TERMINAL",0,0,"54959",,terminal_output +9243,24522491,"TERMINAL",0,0,"[?25l│││││││││││││││││││(B[?12l[?25h",,terminal_output +9244,24523164,"TERMINAL",0,0,"6520630",,terminal_output +9245,24524217,"TERMINAL",0,0,"76171",,terminal_output +9246,24524253,"TERMINAL",0,0,"watch -n60 ""sh killer.sh""(B",,terminal_output +9247,24524802,"TERMINAL",0,0,"watch -n60 ""sh killer.sh""",,terminal_output +9248,24525279,"TERMINAL",0,0,"87282",,terminal_output +9249,24525339,"TERMINAL",0,0,"",,terminal_output +9250,24525675,"TERMINAL",0,0,"",,terminal_output +9251,24525839,"TERMINAL",0,0,"",,terminal_output +9252,24526329,"TERMINAL",0,0,"_.sh""",,terminal_output +9253,24526329,"TERMINAL",0,0,"98393",,terminal_output +9254,24526906,"TERMINAL",0,0,"pa.sh""",,terminal_output +9255,24527043,"TERMINAL",0,0,"r.sh""",,terminal_output +9256,24527212,"TERMINAL",0,0,"t.sh""",,terminal_output +9257,24527297,"TERMINAL",0,0,"(jafar) [tum_cte0515@hkn1993 jafar]$ i.sh"" ",,terminal_output +9258,24527319,"TERMINAL",0,0,"2094104",,terminal_output +9259,24527393,"TERMINAL",0,0,"(jafar) [tum_cte0515@hkn1993 jafar]$ t.sh""",,terminal_output +9260,24527510,"TERMINAL",0,0,"(jafar) [tum_cte0515@hkn1993 jafar]$ i.sh""",,terminal_output +9261,24527675,"TERMINAL",0,0,"(jafar) [tum_cte0515@hkn1993 jafar]$ o.sh""",,terminal_output +9262,24527819,"TERMINAL",0,0,"n.sh""",,terminal_output +9263,24528486,"TERMINAL",0,0,"120515",,terminal_output +9264,24529402,"TERMINAL",0,0,"21626",,terminal_output +9265,24530464,"TERMINAL",0,0,"32737",,terminal_output +9266,24531343,"TERMINAL",0,0,"\n[?2004l",,terminal_output +9267,24531489,"TERMINAL",0,0,"[?25lEvery 60.0s: sh ki... hkn1993.localdomain: Thu Aug 14 17:26:24 2025\nleaving jobs be\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n[?12l[?25h",,terminal_output +9268,24531514,"TERMINAL",0,0,"43848",,terminal_output +9269,24532625,"TERMINAL",0,0,"54959",,terminal_output +9270,24533272,"TERMINAL",0,0,"[?25l│││││││││││││││││││(B\r\n[0] 0:watch* ""hkn1993.localdomain"" 17:26 14-Aug-25(B[?12l[?25h",,terminal_output +9271,24533616,"TERMINAL",0,0,"[?25l│││││││││││││││││││(B[?12l[?25h[?2004h[?25l│││││││││││││││││││(B\r\n[0] 0:bash* ""hkn1993.localdomain"" 17:26 14-Aug-25(B[?12l[?25h",,terminal_output +9272,24533617,"TERMINAL",0,0,"6530640",,terminal_output +9273,24534234,"TERMINAL",0,0,"\r\n[?2004l",,terminal_output +9274,24534362,"TERMINAL",0,0,"[?25lEvery 60.0s: sh kil... hkn1993.localdomain: Thu Aug 14 17:26:27 2025\n\rleaving jobs be\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n[?12l[?25h",,terminal_output +9275,24534752,"TERMINAL",0,0,"76171",,terminal_output +9276,24535761,"TERMINAL",0,0,"87282",,terminal_output +9277,24536744,"TERMINAL",0,0,"98393",,terminal_output +9278,24537017,"TERMINAL",0,0,"[?25l│││││││││││││││││││(B\r\n[0] 0:watch* ""hkn1993.localdomain"" 17:26 14-Aug-25(B[?12l[?25h",,terminal_output +9279,24537301,"TERMINAL",0,0,"(B[?1l>[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l[?7727l[?1004l[?1049l[detached (from session 0)]\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;0",,terminal_output +9280,24537788,"TERMINAL",0,0,"3094204",,terminal_output +9281,24538834,"TERMINAL",0,0,"131626",,terminal_output +9282,24539885,"TERMINAL",0,0,"32737",,terminal_output +9283,24540977,"TERMINAL",0,0,"43848",,terminal_output +9284,24542004,"TERMINAL",0,0,"54959",,terminal_output +9285,24543027,"TERMINAL",0,0,"6540650",,terminal_output +9286,24544068,"TERMINAL",0,0,"76171",,terminal_output +9287,24545113,"TERMINAL",0,0,"87282",,terminal_output +9288,24546165,"TERMINAL",0,0,"98393",,terminal_output +9289,24547224,"TERMINAL",0,0,"4094304",,terminal_output +9290,24548352,"TERMINAL",0,0,"140515",,terminal_output +9291,24549306,"TERMINAL",0,0,"21626",,terminal_output +9292,24550353,"TERMINAL",0,0,"32737",,terminal_output +9293,24551426,"TERMINAL",0,0,"43848",,terminal_output +9294,24552442,"TERMINAL",0,0,"54959",,terminal_output +9295,24553576,"TERMINAL",0,0,"655066:00",,terminal_output +9296,24554772,"TERMINAL",0,0,"76171",,terminal_output +9297,24555585,"TERMINAL",0,0,"87282",,terminal_output +9298,24556686,"TERMINAL",0,0,"98393",,terminal_output +9299,24557774,"TERMINAL",0,0,"5094404",,terminal_output +9300,24558799,"TERMINAL",0,0,"150515",,terminal_output +9301,24559769,"TERMINAL",0,0,"21626",,terminal_output +9302,24560814,"TERMINAL",0,0,"33848",,terminal_output +9303,24561871,"TERMINAL",0,0,"54959",,terminal_output +9304,24562996,"TERMINAL",0,0,"658:00610",,terminal_output +9305,24564014,"TERMINAL",0,0,"76171",,terminal_output +9306,24565043,"TERMINAL",0,0,"87282",,terminal_output +9307,24565636,"killer_partition.sh",0,0,"",shellscript,tab +9308,24566063,"TERMINAL",0,0,"98393",,terminal_output +9309,24567006,"killer_partition.sh",100,0,"",shellscript,selection_mouse +9310,24567099,"TERMINAL",0,0,"7:0094504",,terminal_output +9311,24567133,"killer_partition.sh",98,4,"KILL",shellscript,selection_mouse +9312,24567400,"killer_partition.sh",98,14,"KILL PARTITION",shellscript,selection_mouse +9313,24567556,"killer_partition.sh",98,15,"KILL PARTITION ",shellscript,selection_mouse +9314,24567733,"killer_partition.sh",98,19,"KILL PARTITION JOBS",shellscript,selection_mouse +9315,24568172,"TERMINAL",0,0,"17:00515",,terminal_output +9316,24569248,"TERMINAL",0,0,"21626",,terminal_output +9317,24570430,"TERMINAL",0,0,"32737",,terminal_output +9318,24571320,"TERMINAL",0,0,"43848",,terminal_output +9319,24572365,"TERMINAL",0,0,"54959",,terminal_output +9320,24573416,"TERMINAL",0,0,"6510620",,terminal_output +9321,24574460,"TERMINAL",0,0,"76171",,terminal_output +9322,24575503,"TERMINAL",0,0,"87282",,terminal_output +9323,24576613,"TERMINAL",0,0,"98393",,terminal_output +9324,24577601,"TERMINAL",0,0,"10944:004",,terminal_output +9325,24578660,"TERMINAL",0,0,"110515",,terminal_output +9326,24579797,"TERMINAL",0,0,"21626",,terminal_output +9327,24580738,"TERMINAL",0,0,"32737",,terminal_output +9328,24581786,"TERMINAL",0,0,"43848",,terminal_output +9329,24582836,"TERMINAL",0,0,"5520630",,terminal_output +9330,24583987,"TERMINAL",0,0,"76171",,terminal_output +9331,24584948,"TERMINAL",0,0,"87282",,terminal_output +9332,24586095,"TERMINAL",0,0,"98393",,terminal_output +9333,24587036,"TERMINAL",0,0,"2094104",,terminal_output +9334,24588187,"TERMINAL",0,0,"120515",,terminal_output +9335,24589125,"TERMINAL",0,0,"21626",,terminal_output +9336,24590201,"TERMINAL",0,0,"32737",,terminal_output +9337,24590890,"killer_partition.sh",221,0,"",shellscript,selection_mouse +9338,24591259,"TERMINAL",0,0,"43848",,terminal_output +9339,24592383,"TERMINAL",0,0,"54959",,terminal_output +9340,24593323,"TERMINAL",0,0,"6530640",,terminal_output +9341,24594435,"TERMINAL",0,0,"76171",,terminal_output +9342,24595445,"TERMINAL",0,0,"87282",,terminal_output +9343,24596455,"TERMINAL",0,0,"98393",,terminal_output +9344,24597512,"TERMINAL",0,0,"3094204",,terminal_output +9345,24598734,"TERMINAL",0,0,"130515",,terminal_output +9346,24599694,"TERMINAL",0,0,"21626",,terminal_output +9347,24600714,"TERMINAL",0,0,"32737",,terminal_output +9348,24601712,"TERMINAL",0,0,"43848",,terminal_output +9349,24602753,"TERMINAL",0,0,"54959",,terminal_output +9350,24603857,"TERMINAL",0,0,"6540650",,terminal_output +9351,24604847,"TERMINAL",0,0,"77282",,terminal_output +9352,24605903,"TERMINAL",0,0,"98393",,terminal_output +9353,24606958,"TERMINAL",0,0,"4094304",,terminal_output +9354,24607988,"TERMINAL",0,0,"140515",,terminal_output +9355,24609033,"TERMINAL",0,0,"21626",,terminal_output +9356,24610081,"TERMINAL",0,0,"32737",,terminal_output +9357,24611226,"TERMINAL",0,0,"43848",,terminal_output +9358,24612174,"TERMINAL",0,0,"54959",,terminal_output +9359,24613277,"TERMINAL",0,0,"655067:00",,terminal_output +9360,24614297,"TERMINAL",0,0,"76171",,terminal_output +9361,24615307,"TERMINAL",0,0,"87282",,terminal_output +9362,24616552,"TERMINAL",0,0,"98393",,terminal_output +9363,24617577,"TERMINAL",0,0,"5094404",,terminal_output +9364,24618557,"TERMINAL",0,0,"150515",,terminal_output +9365,24619623,"TERMINAL",0,0,"21626",,terminal_output +9366,24620553,"TERMINAL",0,0,"32737",,terminal_output +9367,24621711,"TERMINAL",0,0,"43848",,terminal_output +9368,24622647,"TERMINAL",0,0,"54959",,terminal_output +9369,24623726,"TERMINAL",0,0,"659:00610",,terminal_output +9370,24624758,"TERMINAL",0,0,"76171",,terminal_output +9371,24625789,"TERMINAL",0,0,"87282",,terminal_output +9372,24626833,"TERMINAL",0,0,"994504",,terminal_output +9373,24627920,"TERMINAL",0,0,"8:018:00515",,terminal_output +9374,24628941,"TERMINAL",0,0,"21626",,terminal_output +9375,24630068,"TERMINAL",0,0,"32737",,terminal_output +9376,24631099,"TERMINAL",0,0,"43848",,terminal_output +9377,24632117,"TERMINAL",0,0,"54959",,terminal_output +9378,24633117,"TERMINAL",0,0,"6510620",,terminal_output +9379,24634167,"TERMINAL",0,0,"76171",,terminal_output +9380,24635300,"TERMINAL",0,0,"87282",,terminal_output +9381,24636316,"TERMINAL",0,0,"98393",,terminal_output +9382,24637339,"TERMINAL",0,0,"10945:004",,terminal_output +9383,24638464,"TERMINAL",0,0,"110515",,terminal_output +9384,24639379,"TERMINAL",0,0,"21626",,terminal_output +9385,24640462,"TERMINAL",0,0,"32737",,terminal_output +9386,24641522,"TERMINAL",0,0,"43848",,terminal_output +9387,24642516,"TERMINAL",0,0,"54959",,terminal_output +9388,24643584,"TERMINAL",0,0,"6520630",,terminal_output +9389,24644605,"TERMINAL",0,0,"76171",,terminal_output +9390,24645707,"TERMINAL",0,0,"87282",,terminal_output +9391,24646762,"TERMINAL",0,0,"98393",,terminal_output +9392,24647739,"TERMINAL",0,0,"2094104",,terminal_output +9393,24648800,"TERMINAL",0,0,"120515",,terminal_output +9394,24649838,"TERMINAL",0,0,"22737",,terminal_output +9395,24650882,"TERMINAL",0,0,"43848",,terminal_output +9396,24651934,"TERMINAL",0,0,"54959",,terminal_output +9397,24652981,"TERMINAL",0,0,"6530640",,terminal_output +9398,24654026,"TERMINAL",0,0,"76171",,terminal_output +9399,24655144,"TERMINAL",0,0,"87282",,terminal_output +9400,24656114,"TERMINAL",0,0,"98393",,terminal_output +9401,24657164,"TERMINAL",0,0,"3094204",,terminal_output +9402,24658233,"TERMINAL",0,0,"130515",,terminal_output +9403,24659354,"TERMINAL",0,0,"21626",,terminal_output +9404,24660380,"TERMINAL",0,0,"32737",,terminal_output +9405,24661348,"TERMINAL",0,0,"43848",,terminal_output +9406,24662428,"TERMINAL",0,0,"54959",,terminal_output +9407,24663453,"TERMINAL",0,0,"6540650",,terminal_output +9408,24664653,"TERMINAL",0,0,"76171",,terminal_output +9409,24665541,"TERMINAL",0,0,"87282",,terminal_output +9410,24666626,"TERMINAL",0,0,"98393",,terminal_output +9411,24667648,"TERMINAL",0,0,"4094304",,terminal_output +9412,24668775,"TERMINAL",0,0,"140515",,terminal_output +9413,24669759,"TERMINAL",0,0,"21626",,terminal_output +9414,24670774,"TERMINAL",0,0,"32737",,terminal_output +9415,24671848,"TERMINAL",0,0,"44959",,terminal_output +9416,24672860,"TERMINAL",0,0,"655068:00",,terminal_output +9417,24673905,"TERMINAL",0,0,"76171",,terminal_output +9418,24674952,"TERMINAL",0,0,"87282",,terminal_output +9419,24676046,"TERMINAL",0,0,"98393",,terminal_output +9420,24677068,"TERMINAL",0,0,"5094404",,terminal_output +9421,24678291,"TERMINAL",0,0,"150515",,terminal_output +9422,24679142,"TERMINAL",0,0,"21626",,terminal_output +9423,24680262,"TERMINAL",0,0,"32737",,terminal_output +9424,24681228,"TERMINAL",0,0,"43848",,terminal_output +9425,24682294,"TERMINAL",0,0,"54959",,terminal_output +9426,24683521,"TERMINAL",0,0,"6540:00610",,terminal_output +9427,24684370,"TERMINAL",0,0,"76171",,terminal_output +9428,24685459,"TERMINAL",0,0,"87282",,terminal_output +9429,24686493,"TERMINAL",0,0,"98393",,terminal_output +9430,24687515,"TERMINAL",0,0,"9:0094504",,terminal_output +9431,24688615,"TERMINAL",0,0,"19:00515",,terminal_output +9432,24689598,"TERMINAL",0,0,"21626",,terminal_output +9433,24690792,"TERMINAL",0,0,"32737",,terminal_output +9434,24691714,"TERMINAL",0,0,"43848",,terminal_output +9435,24692733,"TERMINAL",0,0,"54959",,terminal_output +9436,24693780,"TERMINAL",0,0,"6510620",,terminal_output +9437,24694819,"TERMINAL",0,0,"76171",,terminal_output +9438,24695864,"TERMINAL",0,0,"98393",,terminal_output +9439,24696925,"TERMINAL",0,0,"10946:004",,terminal_output +9440,24698063,"TERMINAL",0,0,"110515",,terminal_output +9441,24699086,"TERMINAL",0,0,"21626",,terminal_output +9442,24700082,"TERMINAL",0,0,"32737",,terminal_output +9443,24701137,"TERMINAL",0,0,"43848",,terminal_output +9444,24702158,"TERMINAL",0,0,"54959",,terminal_output +9445,24703398,"TERMINAL",0,0,"6520630",,terminal_output +9446,24704411,"TERMINAL",0,0,"76171",,terminal_output +9447,24705438,"TERMINAL",0,0,"87282",,terminal_output +9448,24706318,"TERMINAL",0,0,"98393",,terminal_output +9449,24707382,"TERMINAL",0,0,"2094104",,terminal_output +9450,24708403,"TERMINAL",0,0,"120515",,terminal_output +9451,24709531,"TERMINAL",0,0,"21626",,terminal_output +9452,24710483,"TERMINAL",0,0,"32737",,terminal_output +9453,24711579,"TERMINAL",0,0,"43848",,terminal_output +9454,24712611,"TERMINAL",0,0,"54959",,terminal_output +9455,24713631,"TERMINAL",0,0,"6530640",,terminal_output +9456,24714671,"TERMINAL",0,0,"76171",,terminal_output +9457,24715838,"TERMINAL",0,0,"87282",,terminal_output +9458,24716767,"TERMINAL",0,0,"98393",,terminal_output +9459,24717816,"TERMINAL",0,0,"30305215",,terminal_output +9460,24718868,"TERMINAL",0,0,"21626",,terminal_output +9461,24719908,"TERMINAL",0,0,"32737",,terminal_output +9462,24721002,"TERMINAL",0,0,"43848",,terminal_output +9463,24722025,"TERMINAL",0,0,"54959",,terminal_output +9464,24723148,"TERMINAL",0,0,"6540650",,terminal_output +9465,24724210,"TERMINAL",0,0,"76171",,terminal_output +9466,24725196,"TERMINAL",0,0,"87282",,terminal_output +9467,24726222,"TERMINAL",0,0,"98393",,terminal_output +9468,24727245,"TERMINAL",0,0,"4094304",,terminal_output +9469,24728293,"TERMINAL",0,0,"140515",,terminal_output +9470,24729399,"TERMINAL",0,0,"21626",,terminal_output +9471,24730443,"TERMINAL",0,0,"32737",,terminal_output +9472,24731424,"TERMINAL",0,0,"43848",,terminal_output +9473,24732485,"TERMINAL",0,0,"54959",,terminal_output +9474,24733595,"TERMINAL",0,0,"655069:00",,terminal_output +9475,24734618,"TERMINAL",0,0,"76171",,terminal_output +9476,24735641,"TERMINAL",0,0,"87282",,terminal_output +9477,24736739,"TERMINAL",0,0,"98393",,terminal_output +9478,24737794,"TERMINAL",0,0,"5094404",,terminal_output +9479,24738821,"TERMINAL",0,0,"150515",,terminal_output +9480,24739843,"TERMINAL",0,0,"21626",,terminal_output +9481,24740879,"TERMINAL",0,0,"33848",,terminal_output +9482,24741904,"TERMINAL",0,0,"54959",,terminal_output +9483,24742946,"TERMINAL",0,0,"651:00610",,terminal_output +9484,24744042,"TERMINAL",0,0,"76171",,terminal_output +9485,24745070,"TERMINAL",0,0,"87282",,terminal_output +9486,24746106,"TERMINAL",0,0,"98393",,terminal_output +9487,24747226,"TERMINAL",0,0,"30:0094504",,terminal_output +9488,24748209,"TERMINAL",0,0,"130:00515",,terminal_output +9489,24749305,"TERMINAL",0,0,"21626",,terminal_output +9490,24750298,"TERMINAL",0,0,"32737",,terminal_output +9491,24751449,"TERMINAL",0,0,"43848",,terminal_output +9492,24752401,"TERMINAL",0,0,"54959",,terminal_output +9493,24753441,"TERMINAL",0,0,"6510620",,terminal_output +9494,24754483,"TERMINAL",0,0,"76171",,terminal_output +9495,24755546,"TERMINAL",0,0,"87282",,terminal_output +9496,24756634,"TERMINAL",0,0,"98393",,terminal_output +9497,24757632,"TERMINAL",0,0,"10947:004",,terminal_output +9498,24758726,"TERMINAL",0,0,"110515",,terminal_output +9499,24760242,"TERMINAL",0,0,"22737",,terminal_output +9500,24761245,"TERMINAL",0,0,"43848",,terminal_output +9501,24762271,"TERMINAL",0,0,"54959",,terminal_output +9502,24763307,"TERMINAL",0,0,"6520630",,terminal_output +9503,24764419,"TERMINAL",0,0,"76171",,terminal_output +9504,24765426,"TERMINAL",0,0,"87282",,terminal_output +9505,24766467,"TERMINAL",0,0,"98393",,terminal_output +9506,24767491,"TERMINAL",0,0,"2094104",,terminal_output +9507,24768619,"TERMINAL",0,0,"120515",,terminal_output +9508,24769744,"TERMINAL",0,0,"21626",,terminal_output +9509,24770664,"TERMINAL",0,0,"32737",,terminal_output +9510,24771729,"TERMINAL",0,0,"43848",,terminal_output +9511,24772719,"TERMINAL",0,0,"54959",,terminal_output +9512,24773764,"TERMINAL",0,0,"6530640",,terminal_output +9513,24774828,"TERMINAL",0,0,"76171",,terminal_output +9514,24775860,"TERMINAL",0,0,"88393",,terminal_output +9515,24776907,"TERMINAL",0,0,"3094204",,terminal_output +9516,24778036,"TERMINAL",0,0,"130515",,terminal_output +9517,24779068,"TERMINAL",0,0,"21626",,terminal_output +9518,24780085,"TERMINAL",0,0,"32737",,terminal_output +9519,24781108,"TERMINAL",0,0,"43848",,terminal_output +9520,24782121,"TERMINAL",0,0,"54959",,terminal_output +9521,24783165,"TERMINAL",0,0,"6540650",,terminal_output +9522,24784213,"TERMINAL",0,0,"76171",,terminal_output +9523,24785310,"TERMINAL",0,0,"87282",,terminal_output +9524,24786293,"TERMINAL",0,0,"98393",,terminal_output +9525,24787356,"TERMINAL",0,0,"4094304",,terminal_output +9526,24788421,"TERMINAL",0,0,"140515",,terminal_output +9527,24789610,"TERMINAL",0,0,"21626",,terminal_output +9528,24790475,"TERMINAL",0,0,"32737",,terminal_output +9529,24791556,"TERMINAL",0,0,"43848",,terminal_output +9530,24792576,"TERMINAL",0,0,"54959",,terminal_output +9531,24793628,"TERMINAL",0,0,"6550630:00",,terminal_output +9532,24794728,"TERMINAL",0,0,"76171",,terminal_output +9533,24795852,"TERMINAL",0,0,"87282",,terminal_output +9534,24796777,"TERMINAL",0,0,"98393",,terminal_output +9535,24797810,"TERMINAL",0,0,"5094404",,terminal_output +9536,24798852,"TERMINAL",0,0,"151626",,terminal_output +9537,24799906,"TERMINAL",0,0,"32737",,terminal_output +9538,24800982,"TERMINAL",0,0,"43848",,terminal_output +9539,24802000,"TERMINAL",0,0,"54959",,terminal_output +9540,24803040,"TERMINAL",0,0,"652:00610",,terminal_output +9541,24804082,"TERMINAL",0,0,"76171",,terminal_output +9542,24805174,"TERMINAL",0,0,"87282",,terminal_output +9543,24806181,"TERMINAL",0,0,"98393",,terminal_output +9544,24807229,"TERMINAL",0,0,"1:0094504",,terminal_output +9545,24808281,"TERMINAL",0,0,"11:00515",,terminal_output +9546,24809473,"TERMINAL",0,0,"21626",,terminal_output +9547,24810429,"TERMINAL",0,0,"32737",,terminal_output +9548,24811522,"TERMINAL",0,0,"43848",,terminal_output +9549,24812545,"TERMINAL",0,0,"54959",,terminal_output +9550,24813499,"TERMINAL",0,0,"6510620",,terminal_output +9551,24814551,"TERMINAL",0,0,"76171",,terminal_output +9552,24815629,"TERMINAL",0,0,"87282",,terminal_output +9553,24816645,"TERMINAL",0,0,"98393",,terminal_output +9554,24817680,"TERMINAL",0,0,"10948:004",,terminal_output +9555,24818758,"TERMINAL",0,0,"110515",,terminal_output +9556,24819770,"TERMINAL",0,0,"21626",,terminal_output +9557,24820810,"TERMINAL",0,0,"32737",,terminal_output +9558,24821850,"TERMINAL",0,0,"44959",,terminal_output +9559,24822910,"TERMINAL",0,0,"6520630",,terminal_output +9560,24823953,"TERMINAL",0,0,"76171",,terminal_output +9561,24824998,"TERMINAL",0,0,"87282",,terminal_output +9562,24826069,"TERMINAL",0,0,"98393",,terminal_output +9563,24827151,"TERMINAL",0,0,"2094104",,terminal_output +9564,24828317,"TERMINAL",0,0,"120515",,terminal_output +9565,24829172,"TERMINAL",0,0,"21626",,terminal_output +9566,24830226,"TERMINAL",0,0,"32737",,terminal_output +9567,24831389,"TERMINAL",0,0,"43848",,terminal_output +9568,24832413,"TERMINAL",0,0,"54959",,terminal_output +9569,24833388,"TERMINAL",0,0,"6530640",,terminal_output +9570,24834462,"TERMINAL",0,0,"76171",,terminal_output +9571,24835423,"TERMINAL",0,0,"87282",,terminal_output +9572,24836509,"TERMINAL",0,0,"98393",,terminal_output +9573,24837508,"TERMINAL",0,0,"3094204",,terminal_output +9574,24838552,"TERMINAL",0,0,"130515",,terminal_output +9575,24839786,"TERMINAL",0,0,"21626",,terminal_output +9576,24840665,"TERMINAL",0,0,"32737",,terminal_output +9577,24841750,"TERMINAL",0,0,"43848",,terminal_output +9578,24842860,"TERMINAL",0,0,"54959",,terminal_output +9579,24843838,"TERMINAL",0,0,"6540650",,terminal_output +9580,24844849,"TERMINAL",0,0,"77282",,terminal_output +9581,24845900,"TERMINAL",0,0,"98393",,terminal_output +9582,24846950,"TERMINAL",0,0,"4094304",,terminal_output +9583,24848080,"TERMINAL",0,0,"140515",,terminal_output +9584,24849041,"TERMINAL",0,0,"21626",,terminal_output +9585,24850128,"TERMINAL",0,0,"32737",,terminal_output +9586,24851119,"TERMINAL",0,0,"43848",,terminal_output +9587,24852176,"TERMINAL",0,0,"54959",,terminal_output +9588,24853248,"TERMINAL",0,0,"655061:00",,terminal_output +9589,24854264,"TERMINAL",0,0,"76171",,terminal_output +9590,24855348,"TERMINAL",0,0,"87282",,terminal_output +9591,24856373,"TERMINAL",0,0,"98393",,terminal_output +9592,24857416,"TERMINAL",0,0,"5094404",,terminal_output +9593,24858627,"TERMINAL",0,0,"150515",,terminal_output +9594,24859550,"TERMINAL",0,0,"21626",,terminal_output +9595,24860546,"TERMINAL",0,0,"32737",,terminal_output +9596,24861712,"TERMINAL",0,0,"43848",,terminal_output +9597,24862763,"TERMINAL",0,0,"54959",,terminal_output +9598,24863782,"TERMINAL",0,0,"653:00610",,terminal_output +9599,24864773,"TERMINAL",0,0,"76171",,terminal_output +9600,24865794,"TERMINAL",0,0,"87282",,terminal_output +9601,24866818,"TERMINAL",0,0,"98393",,terminal_output +9602,24867863,"TERMINAL",0,0,"2:002:005515",,terminal_output +9603,24868893,"TERMINAL",0,0,"21626",,terminal_output +9604,24869950,"TERMINAL",0,0,"32737",,terminal_output +9605,24870982,"TERMINAL",0,0,"43848",,terminal_output +9606,24872039,"TERMINAL",0,0,"54959",,terminal_output +9607,24873169,"TERMINAL",0,0,"6510620",,terminal_output +9608,24874139,"TERMINAL",0,0,"76171",,terminal_output +9609,24875214,"TERMINAL",0,0,"87282",,terminal_output +9610,24876240,"TERMINAL",0,0,"98393",,terminal_output +9611,24877263,"TERMINAL",0,0,"10949:004",,terminal_output +9612,24878296,"TERMINAL",0,0,"110515",,terminal_output +9613,24879520,"TERMINAL",0,0,"21626",,terminal_output +9614,24880444,"TERMINAL",0,0,"32737",,terminal_output +9615,24881461,"TERMINAL",0,0,"43848",,terminal_output +9616,24882588,"TERMINAL",0,0,"54959",,terminal_output +9617,24883615,"TERMINAL",0,0,"6520630",,terminal_output +9618,24884636,"TERMINAL",0,0,"76171",,terminal_output +9619,24885661,"TERMINAL",0,0,"87282",,terminal_output +9620,24886683,"TERMINAL",0,0,"98393",,terminal_output +9621,24887676,"TERMINAL",0,0,"2094104",,terminal_output +9622,24888769,"TERMINAL",0,0,"120515",,terminal_output +9623,24889860,"TERMINAL",0,0,"21626",,terminal_output +9624,24890801,"TERMINAL",0,0,"32737",,terminal_output +9625,24891837,"TERMINAL",0,0,"44959",,terminal_output +9626,24892884,"TERMINAL",0,0,"6530640",,terminal_output +9627,24893928,"TERMINAL",0,0,"76171",,terminal_output +9628,24894978,"TERMINAL",0,0,"87282",,terminal_output +9629,24896108,"TERMINAL",0,0,"98393",,terminal_output +9630,24897133,"TERMINAL",0,0,"3094204",,terminal_output +9631,24898155,"TERMINAL",0,0,"130515",,terminal_output +9632,24899178,"TERMINAL",0,0,"21626",,terminal_output +9633,24900405,"TERMINAL",0,0,"32737",,terminal_output +9634,24901442,"TERMINAL",0,0,"43848",,terminal_output +9635,24902453,"TERMINAL",0,0,"54959",,terminal_output +9636,24903478,"TERMINAL",0,0,"6540650",,terminal_output +9637,24904406,"TERMINAL",0,0,"76171",,terminal_output +9638,24905451,"TERMINAL",0,0,"87282",,terminal_output +9639,24906497,"TERMINAL",0,0,"98393",,terminal_output +9640,24907576,"TERMINAL",0,0,"4094304",,terminal_output +9641,24908601,"TERMINAL",0,0,"140515",,terminal_output +9642,24909726,"TERMINAL",0,0,"21626",,terminal_output +9643,24910783,"TERMINAL",0,0,"32737",,terminal_output +9644,24911804,"TERMINAL",0,0,"43848",,terminal_output +9645,24912798,"TERMINAL",0,0,"54959",,terminal_output +9646,24913824,"TERMINAL",0,0,"655062:00",,terminal_output +9647,24914852,"TERMINAL",0,0,"77282",,terminal_output +9648,24915898,"TERMINAL",0,0,"98393",,terminal_output +9649,24916941,"TERMINAL",0,0,"5094404",,terminal_output +9650,24917994,"TERMINAL",0,0,"150515",,terminal_output +9651,24919043,"TERMINAL",0,0,"21626",,terminal_output +9652,24920084,"TERMINAL",0,0,"32737",,terminal_output +9653,24921193,"TERMINAL",0,0,"43848",,terminal_output +9654,24922180,"TERMINAL",0,0,"54959",,terminal_output +9655,24923242,"TERMINAL",0,0,"654:00610",,terminal_output +9656,24924266,"TERMINAL",0,0,"76171",,terminal_output +9657,24925428,"TERMINAL",0,0,"87282",,terminal_output +9658,24926408,"TERMINAL",0,0,"98393",,terminal_output +9659,24927441,"TERMINAL",0,0,"3:0094504",,terminal_output +9660,24928461,"TERMINAL",0,0,"13:00515",,terminal_output +9661,24929484,"TERMINAL",0,0,"21626",,terminal_output +9662,24930534,"TERMINAL",0,0,"32737",,terminal_output +9663,24931577,"TERMINAL",0,0,"43848",,terminal_output +9664,24932783,"TERMINAL",0,0,"54959",,terminal_output +9665,24933688,"TERMINAL",0,0,"6510620",,terminal_output +9666,24934722,"TERMINAL",0,0,"76171",,terminal_output +9667,24935766,"TERMINAL",0,0,"87282",,terminal_output +9668,24936808,"TERMINAL",0,0,"98393",,terminal_output +9669,24937861,"TERMINAL",0,0,"1010520:015",,terminal_output +9670,24938910,"TERMINAL",0,0,"21626",,terminal_output +9671,24939958,"TERMINAL",0,0,"32737",,terminal_output +9672,24941061,"TERMINAL",0,0,"43848",,terminal_output +9673,24942051,"TERMINAL",0,0,"54959",,terminal_output +9674,24943092,"TERMINAL",0,0,"6520630",,terminal_output +9675,24944235,"TERMINAL",0,0,"76171",,terminal_output +9676,24945256,"TERMINAL",0,0,"87282",,terminal_output +9677,24946281,"TERMINAL",0,0,"98393",,terminal_output +9678,24947281,"TERMINAL",0,0,"2094104",,terminal_output +9679,24948334,"TERMINAL",0,0,"120515",,terminal_output +9680,24949357,"TERMINAL",0,0,"21626",,terminal_output +9681,24950452,"TERMINAL",0,0,"32737",,terminal_output +9682,24951610,"TERMINAL",0,0,"43848",,terminal_output +9683,24952486,"TERMINAL",0,0,"54959",,terminal_output +9684,24953554,"TERMINAL",0,0,"6530640",,terminal_output +9685,24954575,"TERMINAL",0,0,"76171",,terminal_output +9686,24955613,"TERMINAL",0,0,"87282",,terminal_output +9687,24956715,"TERMINAL",0,0,"98393",,terminal_output +9688,24957689,"TERMINAL",0,0,"3094204",,terminal_output +9689,24958734,"TERMINAL",0,0,"130515",,terminal_output +9690,24960004,"TERMINAL",0,0,"21626",,terminal_output +9691,24960817,"TERMINAL",0,0,"33848",,terminal_output +9692,24961861,"TERMINAL",0,0,"54959",,terminal_output +9693,24962894,"TERMINAL",0,0,"6540650",,terminal_output +9694,24963933,"TERMINAL",0,0,"76171",,terminal_output +9695,24964981,"TERMINAL",0,0,"87282",,terminal_output +9696,24966134,"TERMINAL",0,0,"98393",,terminal_output +9697,24967056,"TERMINAL",0,0,"4094304",,terminal_output +9698,24968195,"TERMINAL",0,0,"140515",,terminal_output +9699,24969221,"TERMINAL",0,0,"21626",,terminal_output +9700,24970243,"TERMINAL",0,0,"32737",,terminal_output +9701,24971266,"TERMINAL",0,0,"43848",,terminal_output +9702,24972294,"TERMINAL",0,0,"54959",,terminal_output +9703,24973327,"TERMINAL",0,0,"655063:00",,terminal_output +9704,24974444,"TERMINAL",0,0,"76171",,terminal_output +9705,24975443,"TERMINAL",0,0,"87282",,terminal_output +9706,24976600,"TERMINAL",0,0,"98393",,terminal_output +9707,24977513,"TERMINAL",0,0,"5094404",,terminal_output +9708,24978538,"TERMINAL",0,0,"150515",,terminal_output +9709,24979664,"TERMINAL",0,0,"21626",,terminal_output +9710,24980689,"TERMINAL",0,0,"32737",,terminal_output +9711,24981712,"TERMINAL",0,0,"43848",,terminal_output +9712,24982736,"TERMINAL",0,0,"54959",,terminal_output +9713,24983864,"TERMINAL",0,0,"655:00610",,terminal_output +9714,24984888,"TERMINAL",0,0,"76171",,terminal_output +9715,24985955,"TERMINAL",0,0,"88393",,terminal_output +9716,24986880,"TERMINAL",0,0,"4:0094504",,terminal_output +9717,24987920,"TERMINAL",0,0,"14:00515",,terminal_output +9718,24988967,"TERMINAL",0,0,"21626",,terminal_output +9719,24990003,"TERMINAL",0,0,"32737",,terminal_output +9720,24991132,"TERMINAL",0,0,"43848",,terminal_output +9721,24992087,"TERMINAL",0,0,"54959",,terminal_output +9722,24993180,"TERMINAL",0,0,"6510620",,terminal_output +9723,24994175,"TERMINAL",0,0,"76171",,terminal_output +9724,24995200,"TERMINAL",0,0,"87282",,terminal_output +9725,24996255,"TERMINAL",0,0,"98393",,terminal_output +9726,24997492,"TERMINAL",0,0,"10941:004",,terminal_output +9727,24998409,"TERMINAL",0,0,"110515",,terminal_output +9728,24999437,"TERMINAL",0,0,"21626",,terminal_output +9729,25000655,"TERMINAL",0,0,"32737",,terminal_output +9730,25001681,"TERMINAL",0,0,"43848",,terminal_output +9731,25002677,"TERMINAL",0,0,"54959",,terminal_output +9732,25003728,"TERMINAL",0,0,"6520630",,terminal_output +9733,25004860,"TERMINAL",0,0,"76171",,terminal_output +9734,25005880,"TERMINAL",0,0,"87282",,terminal_output +9735,25006889,"TERMINAL",0,0,"994104",,terminal_output +9736,25007881,"TERMINAL",0,0,"2120515",,terminal_output +9737,25008922,"TERMINAL",0,0,"21626",,terminal_output +9738,25009982,"TERMINAL",0,0,"32737",,terminal_output +9739,25011001,"TERMINAL",0,0,"43848",,terminal_output +9740,25012047,"TERMINAL",0,0,"54959",,terminal_output +9741,25013077,"TERMINAL",0,0,"6530640",,terminal_output +9742,25014167,"TERMINAL",0,0,"76171",,terminal_output +9743,25015172,"TERMINAL",0,0,"87282",,terminal_output +9744,25016196,"TERMINAL",0,0,"98393",,terminal_output +9745,25017245,"TERMINAL",0,0,"3094204",,terminal_output +9746,25018375,"TERMINAL",0,0,"130515",,terminal_output +9747,25019405,"TERMINAL",0,0,"21626",,terminal_output +9748,25020428,"TERMINAL",0,0,"32737",,terminal_output +9749,25021405,"TERMINAL",0,0,"43848",,terminal_output +9750,25022468,"TERMINAL",0,0,"54959",,terminal_output +9751,25023490,"TERMINAL",0,0,"6540650",,terminal_output +9752,25024617,"TERMINAL",0,0,"76171",,terminal_output +9753,25025571,"TERMINAL",0,0,"87282",,terminal_output +9754,25026656,"TERMINAL",0,0,"98393",,terminal_output +9755,25027658,"TERMINAL",0,0,"4094304",,terminal_output +9756,25028694,"TERMINAL",0,0,"140515",,terminal_output +9757,25029736,"TERMINAL",0,0,"21626",,terminal_output +9758,25030868,"TERMINAL",0,0,"32737",,terminal_output +9759,25031843,"TERMINAL",0,0,"43848",,terminal_output +9760,25032852,"TERMINAL",0,0,"555064:00",,terminal_output +9761,25033893,"TERMINAL",0,0,"76171",,terminal_output +9762,25034968,"TERMINAL",0,0,"87282",,terminal_output +9763,25035976,"TERMINAL",0,0,"98393",,terminal_output +9764,25037017,"TERMINAL",0,0,"5094404",,terminal_output +9765,25038134,"TERMINAL",0,0,"150515",,terminal_output +9766,25039159,"TERMINAL",0,0,"21626",,terminal_output +9767,25040138,"TERMINAL",0,0,"32737",,terminal_output +9768,25041207,"TERMINAL",0,0,"43848",,terminal_output +9769,25042242,"TERMINAL",0,0,"54959",,terminal_output +9770,25043358,"TERMINAL",0,0,"656:00610",,terminal_output +9771,25044383,"TERMINAL",0,0,"76171",,terminal_output +9772,25045405,"TERMINAL",0,0,"87282",,terminal_output +9773,25046430,"TERMINAL",0,0,"98393",,terminal_output +9774,25047518,"TERMINAL",0,0,"5:0094504",,terminal_output +9775,25048580,"TERMINAL",0,0,"15:00515",,terminal_output +9776,25049717,"TERMINAL",0,0,"21626",,terminal_output +9777,25050626,"TERMINAL",0,0,"32737",,terminal_output +9778,25051755,"TERMINAL",0,0,"43848",,terminal_output +9779,25052689,"TERMINAL",0,0,"54959",,terminal_output +9780,25053841,"TERMINAL",0,0,"6510620",,terminal_output +9781,25054754,"TERMINAL",0,0,"76171",,terminal_output +9782,25055829,"TERMINAL",0,0,"87282",,terminal_output +9783,25056835,"TERMINAL",0,0,"9942:004",,terminal_output +9784,25057913,"TERMINAL",0,0,"1110515",,terminal_output +9785,25058935,"TERMINAL",0,0,"21626",,terminal_output +9786,25059282,"TERMINAL",0,0,"watch",,terminal_focus +9787,25059971,"TERMINAL",0,0,"32737",,terminal_output +9788,25061072,"TERMINAL",0,0,"43848",,terminal_output +9789,25061155,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +9790,25071735,"TERMINAL",0,0,"queue",,terminal_command +9791,25071784,"TERMINAL",0,0,"]633;E;2025-08-14 17:35:24 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C",,terminal_output +9792,25071829,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 17:35:24 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4 (Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2 (Priority)3423250 accelerat train_dy tum_cte0 PD\t0:00\t 8 (Priority)3412401 accelerat train_to tum_cte0 R 1-23:35:24\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-03:46:29\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 5:22:15\t 1 hkn07133421964 accelerat train_to tum_cte0 R 5:34:39\t 1 hkn0720",,terminal_output +9793,25072868,"TERMINAL",0,0,"6530640",,terminal_output +9794,25073910,"TERMINAL",0,0,"76171",,terminal_output +9795,25074941,"TERMINAL",0,0,"87282",,terminal_output +9796,25075980,"TERMINAL",0,0,"98393",,terminal_output +9797,25077023,"TERMINAL",0,0,"3094204",,terminal_output +9798,25078123,"TERMINAL",0,0,"130515",,terminal_output +9799,25079107,"TERMINAL",0,0,"21626",,terminal_output +9800,25080221,"TERMINAL",0,0,"32737",,terminal_output +9801,25081182,"TERMINAL",0,0,"43848",,terminal_output +9802,25082269,"TERMINAL",0,0,"54959",,terminal_output +9803,25083273,"TERMINAL",0,0,"6540650",,terminal_output +9804,25084310,"TERMINAL",0,0,"76171",,terminal_output +9805,25085359,"TERMINAL",0,0,"87282",,terminal_output +9806,25086468,"TERMINAL",0,0,"98393",,terminal_output +9807,25087458,"TERMINAL",0,0,"4094304",,terminal_output +9808,25088517,"TERMINAL",0,0,"140515",,terminal_output +9809,25089534,"TERMINAL",0,0,"21626",,terminal_output +9810,25090575,"TERMINAL",0,0,"32737",,terminal_output +9811,25092614,"TERMINAL",0,0,"44959",,terminal_output +9812,25093644,"TERMINAL",0,0,"655065:00",,terminal_output +9813,25094682,"TERMINAL",0,0,"76171",,terminal_output +9814,25095741,"TERMINAL",0,0,"87282",,terminal_output +9815,25096845,"TERMINAL",0,0,"98393",,terminal_output +9816,25097800,"TERMINAL",0,0,"5094404",,terminal_output +9817,25098867,"TERMINAL",0,0,"151626",,terminal_output +9818,25099887,"TERMINAL",0,0,"32737",,terminal_output +9819,25100922,"TERMINAL",0,0,"43848",,terminal_output +9820,25101962,"TERMINAL",0,0,"54959",,terminal_output +9821,25103000,"TERMINAL",0,0,"657:00610",,terminal_output +9822,25104030,"TERMINAL",0,0,"76171",,terminal_output +9823,25105106,"TERMINAL",0,0,"87282",,terminal_output +9824,25106132,"TERMINAL",0,0,"98393",,terminal_output +9825,25107143,"TERMINAL",0,0,"6:0094504",,terminal_output +9826,25108279,"TERMINAL",0,0,"16:00515",,terminal_output +9827,25109221,"TERMINAL",0,0,"21626",,terminal_output +9828,25110327,"TERMINAL",0,0,"32737",,terminal_output +9829,25111319,"TERMINAL",0,0,"43848",,terminal_output +9830,25112406,"TERMINAL",0,0,"54959",,terminal_output +9831,25113432,"TERMINAL",0,0,"6510620",,terminal_output +9832,25114447,"TERMINAL",0,0,"76171",,terminal_output +9833,25114676,"TERMINAL",0,0,"bash",,terminal_focus +9834,25115514,"TERMINAL",0,0,"87282",,terminal_output +9835,25116678,"TERMINAL",0,0,"98393",,terminal_output +9836,25117700,"TERMINAL",0,0,"10943:004",,terminal_output +9837,25118627,"TERMINAL",0,0,"110515",,terminal_output +9838,25119006,"TERMINAL",0,0,"idling",,terminal_command +9839,25119115,"TERMINAL",0,0,"]633;E;2025-08-14 17:36:12 idling;cc3c0b2f-e52c-4c7f-81cf-c8a8bbae24c8]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Aug 14 17:36:12 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 262 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated: 40 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 1 nodes idle\rPartition large:\t 8 nodes idle",,terminal_output +9840,25119683,"TERMINAL",0,0,"21626",,terminal_output +9841,25120098,"TERMINAL",0,0,"3\t ",,terminal_output +9842,25120919,"TERMINAL",0,0,"33848",,terminal_output +9843,25121140,"TERMINAL",0,0,"459",,terminal_output +9844,25121966,"TERMINAL",0,0,"54959",,terminal_output +9845,25122183,"TERMINAL",0,0,"5\t ",,terminal_output +9846,25123004,"TERMINAL",0,0,"6520630",,terminal_output +9847,25123238,"TERMINAL",0,0,"6\t ",,terminal_output +9848,25124049,"TERMINAL",0,0,"76171",,terminal_output +9849,25124265,"TERMINAL",0,0,"7\t ",,terminal_output +9850,25125175,"TERMINAL",0,0,"87282",,terminal_output +9851,25125328,"TERMINAL",0,0,"8\t ",,terminal_output +9852,25126201,"TERMINAL",0,0,"98393",,terminal_output +9853,25126341,"TERMINAL",0,0,"9\t ",,terminal_output +9854,25127197,"TERMINAL",0,0,"2094104",,terminal_output +9855,25127393,"TERMINAL",0,0,"20\t ",,terminal_output +9856,25128250,"TERMINAL",0,0,"120515",,terminal_output +9857,25128466,"TERMINAL",0,0,"1\t ",,terminal_output +9858,25129271,"TERMINAL",0,0,"21626",,terminal_output +9859,25129463,"TERMINAL",0,0,"2\t ",,terminal_output +9860,25130429,"TERMINAL",0,0,"32737",,terminal_output +9861,25130486,"TERMINAL",0,0,"3\t ",,terminal_output +9862,25131424,"TERMINAL",0,0,"43848",,terminal_output +9863,25131557,"TERMINAL",0,0,"4\t ",,terminal_output +9864,25132446,"TERMINAL",0,0,"54959",,terminal_output +9865,25132576,"TERMINAL",0,0,"5\t ",,terminal_output +9866,25133468,"TERMINAL",0,0,"6530640",,terminal_output +9867,25133610,"TERMINAL",0,0,"6\t ",,terminal_output +9868,25134493,"TERMINAL",0,0,"76171",,terminal_output +9869,25134626,"TERMINAL",0,0,"7\t ",,terminal_output +9870,25135484,"TERMINAL",0,0,"87282",,terminal_output +9871,25135673,"TERMINAL",0,0,"8\t ",,terminal_output +9872,25136527,"TERMINAL",0,0,"98393",,terminal_output +9873,25136728,"TERMINAL",0,0,"9\t ",,terminal_output +9874,25137570,"TERMINAL",0,0,"3094204",,terminal_output +9875,25137750,"TERMINAL",0,0,"30\t ",,terminal_output +9876,25138614,"TERMINAL",0,0,"130515",,terminal_output +9877,25138795,"TERMINAL",0,0,"1\t ",,terminal_output +9878,25139841,"TERMINAL",0,0,"21626",,terminal_output +9879,25139876,"TERMINAL",0,0,"2\t ",,terminal_output +9880,25140741,"TERMINAL",0,0,"32737",,terminal_output +9881,25140873,"TERMINAL",0,0,"4\t ",,terminal_output +9882,25141741,"TERMINAL",0,0,"43848",,terminal_output +9883,25141911,"TERMINAL",0,0,"5\t ",,terminal_output +9884,25142780,"TERMINAL",0,0,"54959",,terminal_output +9885,25142954,"TERMINAL",0,0,"6\t ",,terminal_output +9886,25143915,"TERMINAL",0,0,"6641751",,terminal_output +9887,25143995,"TERMINAL",0,0,"7\t ",,terminal_output +9888,25144941,"TERMINAL",0,0,"87282",,terminal_output +9889,25145071,"TERMINAL",0,0,"8\t ",,terminal_output +9890,25145929,"TERMINAL",0,0,"98393",,terminal_output +9891,25146075,"TERMINAL",0,0,"9\t ",,terminal_output +9892,25146949,"TERMINAL",0,0,"4094304",,terminal_output +9893,25147107,"TERMINAL",0,0,"40\t ",,terminal_output +9894,25147989,"TERMINAL",0,0,"140515",,terminal_output +9895,25148149,"TERMINAL",0,0,"1\t ",,terminal_output +9896,25149029,"TERMINAL",0,0,"21626",,terminal_output +9897,25149194,"TERMINAL",0,0,"2\t ",,terminal_output +9898,25150164,"TERMINAL",0,0,"32737",,terminal_output +9899,25150237,"TERMINAL",0,0,"3\t ",,terminal_output +9900,25151147,"TERMINAL",0,0,"43848",,terminal_output +9901,25151291,"TERMINAL",0,0,"4\t ",,terminal_output +9902,25152268,"TERMINAL",0,0,"54959",,terminal_output +9903,25152319,"TERMINAL",0,0,"5\t ",,terminal_output +9904,25153116,"TERMINAL",0,0,"watch",,terminal_focus +9905,25153247,"TERMINAL",0,0,"655066:00",,terminal_output +9906,25153376,"TERMINAL",0,0,"6\t ",,terminal_output +9907,25154280,"TERMINAL",0,0,"76171",,terminal_output +9908,25154375,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +9909,25154431,"TERMINAL",0,0,"7\t ",,terminal_output +9910,25155454,"TERMINAL",0,0,"8\t ",,terminal_output +9911,25156223,"TERMINAL",0,0,"idling",,terminal_command +9912,25156280,"TERMINAL",0,0,"]633;E;2025-08-14 17:36:49 idling;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Aug 14 17:36:49 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 259 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated: 40 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 1 nodes idle\rPartition large:\t 8 nodes idle",,terminal_output +9913,25156474,"TERMINAL",0,0,"9\t ",,terminal_output +9914,25157434,"TERMINAL",0,0,"50\t ",,terminal_output +9915,25157530,"TERMINAL",0,0,"50\t ",,terminal_output +9916,25158457,"TERMINAL",0,0,"1\t ",,terminal_output +9917,25158538,"TERMINAL",0,0,"1\t ",,terminal_output +9918,25159397,"TERMINAL",0,0,"2\t ",,terminal_output +9919,25159577,"TERMINAL",0,0,"2\t ",,terminal_output +9920,25160257,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +9921,25160618,"TERMINAL",0,0,"3\t ",,terminal_output +9922,25161689,"TERMINAL",0,0,"4\t ",,terminal_output +9923,25162661,"TERMINAL",0,0,"queue",,terminal_command +9924,25162732,"TERMINAL",0,0,"5\t ",,terminal_output +9925,25162753,"TERMINAL",0,0,"]633;E;2025-08-14 17:36:55 queue;7bbc766d-cb8c-45e4-8c85-2cc7c12fa702]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Aug 14 17:36:55 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3415713 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3423080 accelerat train_to tum_cte0 PD\t0:00\t 4 (Priority)3423082 accelerat train_to tum_cte0 PD\t0:00\t 2 (Priority)3423250 accelerat train_dy tum_cte0 PD\t0:00\t 8 (Priority)3412401 accelerat train_to tum_cte0 R 1-23:36:54\t 8 hkn[0412,0416,0425,0433,0718,0732,0803,0807]3418832 accelerat train_dy tum_cte0 R 1-03:47:59\t 8 hkn[0407,0415,0420,0422-0424,0729,0735]3422008 accelerat train_dy tum_cte0 R 5:23:45\t 1 hkn07133421964 accelerat train_to tum_cte0 R 5:36:09\t 1 hkn0720",,terminal_output +9926,25163780,"TERMINAL",0,0,"6\t ",,terminal_output +9927,25163780,"TERMINAL",0,0,"658:00610",,terminal_output +9928,25164806,"TERMINAL",0,0,"7\t ",,terminal_output +9929,25164916,"TERMINAL",0,0,"77282",,terminal_output +9930,25165861,"TERMINAL",0,0,"8\t ",,terminal_output +9931,25165897,"TERMINAL",0,0,"98393",,terminal_output +9932,25166865,"TERMINAL",0,0,"7:00\t ",,terminal_output +9933,25166948,"TERMINAL",0,0,"7:0094504",,terminal_output +9934,25167908,"TERMINAL",0,0,"1\t ",,terminal_output +9935,25167966,"TERMINAL",0,0,"17:00515",,terminal_output +9936,25168946,"TERMINAL",0,0,"2\t ",,terminal_output +9937,25169007,"TERMINAL",0,0,"21626",,terminal_output +9938,25169997,"TERMINAL",0,0,"3\t ",,terminal_output +9939,25170083,"TERMINAL",0,0,"32737",,terminal_output +9940,25171025,"TERMINAL",0,0,"4\t ",,terminal_output +9941,25171098,"TERMINAL",0,0,"43848",,terminal_output +9942,25172076,"TERMINAL",0,0,"5\t ",,terminal_output +9943,25172138,"TERMINAL",0,0,"54959",,terminal_output +9944,25173200,"TERMINAL",0,0,"6\t ",,terminal_output +9945,25173201,"TERMINAL",0,0,"6510620",,terminal_output +9946,25174144,"TERMINAL",0,0,"7\t ",,terminal_output +9947,25174253,"TERMINAL",0,0,"76171",,terminal_output +9948,25175251,"TERMINAL",0,0,"8\t ",,terminal_output +9949,25175261,"TERMINAL",0,0,"87282",,terminal_output +9950,25176307,"TERMINAL",0,0,"9\t ",,terminal_output +9951,25176342,"TERMINAL",0,0,"98393",,terminal_output +9952,25177299,"TERMINAL",0,0,"10\t ",,terminal_output +9953,25177357,"TERMINAL",0,0,"10944:004",,terminal_output +9954,25178305,"TERMINAL",0,0,"1\t ",,terminal_output +9955,25178438,"TERMINAL",0,0,"110515",,terminal_output +9956,25179448,"TERMINAL",0,0,"2\t ",,terminal_output +9957,25179449,"TERMINAL",0,0,"21626",,terminal_output +9958,25180473,"TERMINAL",0,0,"3\t ",,terminal_output +9959,25180509,"TERMINAL",0,0,"32737",,terminal_output +9960,25181599,"TERMINAL",0,0,"4\t ",,terminal_output +9961,25181599,"TERMINAL",0,0,"43848",,terminal_output +9962,25182624,"TERMINAL",0,0,"5\t ",,terminal_output +9963,25182625,"TERMINAL",0,0,"54959",,terminal_output +9964,25183556,"TERMINAL",0,0,"6\t ",,terminal_output +9965,25183632,"TERMINAL",0,0,"6520630",,terminal_output +9966,25184557,"TERMINAL",0,0,"7\t ",,terminal_output +9967,25184689,"TERMINAL",0,0,"76171",,terminal_output +9968,25185590,"TERMINAL",0,0,"8\t ",,terminal_output +9969,25185720,"TERMINAL",0,0,"87282",,terminal_output +9970,25186643,"TERMINAL",0,0,"9\t ",,terminal_output +9971,25186773,"TERMINAL",0,0,"98393",,terminal_output +9972,25187750,"TERMINAL",0,0,"20\t ",,terminal_output +9973,25187829,"TERMINAL",0,0,"20205115",,terminal_output +9974,25188784,"TERMINAL",0,0,"1\t ",,terminal_output +9975,25188869,"TERMINAL",0,0,"21626",,terminal_output +9976,25189795,"TERMINAL",0,0,"2\t ",,terminal_output +9977,25189916,"TERMINAL",0,0,"32737",,terminal_output +9978,25190832,"TERMINAL",0,0,"3\t ",,terminal_output +9979,25190968,"TERMINAL",0,0,"43848",,terminal_output +9980,25191906,"TERMINAL",0,0,"4\t ",,terminal_output +9981,25191996,"TERMINAL",0,0,"54959",,terminal_output +9982,25192868,"TERMINAL",0,0,"6\t ",,terminal_output +9983,25193035,"TERMINAL",0,0,"6530640",,terminal_output +9984,25193901,"TERMINAL",0,0,"7\t ",,terminal_output +9985,25194074,"TERMINAL",0,0,"76171",,terminal_output +9986,25194946,"TERMINAL",0,0,"8\t ",,terminal_output +9987,25195114,"TERMINAL",0,0,"87282",,terminal_output +9988,25195981,"TERMINAL",0,0,"9\t ",,terminal_output +9989,25196154,"TERMINAL",0,0,"98393",,terminal_output +9990,25197020,"TERMINAL",0,0,"30\t ",,terminal_output +9991,25197197,"TERMINAL",0,0,"3094204",,terminal_output +9992,25198062,"TERMINAL",0,0,"1\t ",,terminal_output +9993,25198230,"TERMINAL",0,0,"130515",,terminal_output +9994,25199095,"TERMINAL",0,0,"2\t ",,terminal_output +9995,25199280,"TERMINAL",0,0,"21626",,terminal_output +9996,25200119,"TERMINAL",0,0,"3\t ",,terminal_output +9997,25200338,"TERMINAL",0,0,"32737",,terminal_output +9998,25201208,"TERMINAL",0,0,"4\t ",,terminal_output +9999,25201361,"TERMINAL",0,0,"43848",,terminal_output +10000,25201500,"TERMINAL",0,0,"watch",,terminal_focus +10001,25202199,"TERMINAL",0,0,"5\t ",,terminal_output +10002,25202400,"TERMINAL",0,0,"54959",,terminal_output +10003,25203310,"TERMINAL",0,0,"6\t ",,terminal_output +10004,25203441,"TERMINAL",0,0,"6540650",,terminal_output +10005,25204277,"TERMINAL",0,0,"7\t ",,terminal_output +10006,25204483,"TERMINAL",0,0,"76171",,terminal_output +10007,25205324,"TERMINAL",0,0,"8\t ",,terminal_output +10008,25205530,"TERMINAL",0,0,"87282",,terminal_output +10009,25206361,"TERMINAL",0,0,"9\t ",,terminal_output +10010,25206572,"TERMINAL",0,0,"98393",,terminal_output +10011,25207404,"TERMINAL",0,0,"40\t ",,terminal_output +10012,25207615,"TERMINAL",0,0,"4094304",,terminal_output +10013,25208531,"TERMINAL",0,0,"1\t ",,terminal_output +10014,25208663,"TERMINAL",0,0,"140515",,terminal_output +10015,25209689,"TERMINAL",0,0,"2\t ",,terminal_output +10016,25209725,"TERMINAL",0,0,"21626",,terminal_output +10017,25210522,"TERMINAL",0,0,"3\t ",,terminal_output +10018,25210741,"TERMINAL",0,0,"32737",,terminal_output +10019,25211563,"TERMINAL",0,0,"4\t ",,terminal_output +10020,25211792,"TERMINAL",0,0,"43848",,terminal_output +10021,25212624,"TERMINAL",0,0,"5\t ",,terminal_output +10022,25212822,"TERMINAL",0,0,"555067:00",,terminal_output +10023,25213650,"TERMINAL",0,0,"6\t ",,terminal_output +10024,25213883,"TERMINAL",0,0,"76171",,terminal_output +10025,25214783,"TERMINAL",0,0,"7\t ",,terminal_output +10026,25214956,"TERMINAL",0,0,"87282",,terminal_output +10027,25215801,"TERMINAL",0,0,"8\t ",,terminal_output +10028,25215959,"TERMINAL",0,0,"98393",,terminal_output +10029,25216824,"TERMINAL",0,0,"9\t ",,terminal_output +10030,25217003,"TERMINAL",0,0,"5094404",,terminal_output +10031,25217798,"TERMINAL",0,0,"50\t ",,terminal_output +10032,25218050,"TERMINAL",0,0,"150515",,terminal_output +10033,25218975,"TERMINAL",0,0,"13",,terminal_output +10034,25219108,"TERMINAL",0,0,"21626",,terminal_output +10035,25219936,"TERMINAL",0,0,"3\t ",,terminal_output +10036,25220131,"TERMINAL",0,0,"32737",,terminal_output +10037,25221025,"TERMINAL",0,0,"4\t ",,terminal_output +10038,25221173,"TERMINAL",0,0,"43848",,terminal_output +10039,25222016,"TERMINAL",0,0,"5\t ",,terminal_output +10040,25222208,"TERMINAL",0,0,"54959",,terminal_output +10041,25223058,"TERMINAL",0,0,"6\t ",,terminal_output +10042,25223257,"TERMINAL",0,0,"659:00610",,terminal_output +10043,25224100,"TERMINAL",0,0,"7\t ",,terminal_output +10044,25224294,"TERMINAL",0,0,"76171",,terminal_output +10045,25225145,"TERMINAL",0,0,"8\t ",,terminal_output +10046,25225339,"TERMINAL",0,0,"87282",,terminal_output +10047,25226183,"TERMINAL",0,0,"9\t ",,terminal_output +10048,25226382,"TERMINAL",0,0,"98393",,terminal_output +10049,25227226,"TERMINAL",0,0,"8:00\t ",,terminal_output +10050,25227426,"TERMINAL",0,0,"8:0094504",,terminal_output +10051,25228294,"TERMINAL",0,0,"1\t ",,terminal_output +10052,25228471,"TERMINAL",0,0,"18:00515",,terminal_output +10053,25229318,"TERMINAL",0,0,"2\t ",,terminal_output +10054,25229508,"TERMINAL",0,0,"21626",,terminal_output +10055,25230340,"TERMINAL",0,0,"3\t ",,terminal_output +10056,25230548,"TERMINAL",0,0,"32737",,terminal_output +10057,25231365,"TERMINAL",0,0,"4\t ",,terminal_output +10058,25231592,"TERMINAL",0,0,"43848",,terminal_output +10059,25232597,"TERMINAL",0,0,"5\t ",,terminal_output +10060,25232669,"TERMINAL",0,0,"54959",,terminal_output +10061,25233438,"TERMINAL",0,0,"6\t ",,terminal_output +10062,25233674,"TERMINAL",0,0,"6510620",,terminal_output +10063,25234462,"TERMINAL",0,0,"7\t ",,terminal_output +10064,25234746,"TERMINAL",0,0,"76171",,terminal_output +10065,25235500,"TERMINAL",0,0,"8\t ",,terminal_output +10066,25235769,"TERMINAL",0,0,"87282",,terminal_output +10067,25236589,"TERMINAL",0,0,"9\t ",,terminal_output +10068,25236797,"TERMINAL",0,0,"98393",,terminal_output +10069,25237716,"TERMINAL",0,0,"10\t ",,terminal_output +10070,25237846,"TERMINAL",0,0,"101055:015",,terminal_output +10071,25238739,"TERMINAL",0,0,"1\t ",,terminal_output +10072,25238885,"TERMINAL",0,0,"21626",,terminal_output +10073,25239763,"TERMINAL",0,0,"2\t ",,terminal_output +10074,25239957,"TERMINAL",0,0,"32737",,terminal_output +10075,25240789,"TERMINAL",0,0,"3\t ",,terminal_output +10076,25241196,"TERMINAL",0,0,"43848",,terminal_output +10077,25241813,"TERMINAL",0,0,"4\t ",,terminal_output +10078,25242193,"TERMINAL",0,0,"54959",,terminal_output +10079,25242836,"TERMINAL",0,0,"5\t ",,terminal_output +10080,25243228,"TERMINAL",0,0,"6520630",,terminal_output +10081,25243819,"TERMINAL",0,0,"6\t ",,terminal_output +10082,25244268,"TERMINAL",0,0,"76171",,terminal_output +10083,25244883,"TERMINAL",0,0,"7\t ",,terminal_output +10084,25245405,"TERMINAL",0,0,"87282",,terminal_output +10085,25245887,"TERMINAL",0,0,"9\t ",,terminal_output +10086,25246419,"TERMINAL",0,0,"98393",,terminal_output +10087,25246932,"TERMINAL",0,0,"20\t ",,terminal_output +10088,25247549,"TERMINAL",0,0,"2094104",,terminal_output +10089,25247970,"TERMINAL",0,0,"1\t ",,terminal_output +10090,25248477,"TERMINAL",0,0,"120515",,terminal_output +10091,25249013,"TERMINAL",0,0,"2\t ",,terminal_output +10092,25249490,"TERMINAL",0,0,"21626",,terminal_output +10093,25250049,"TERMINAL",0,0,"3\t ",,terminal_output +10094,25250502,"TERMINAL",0,0,"32737",,terminal_output +10095,25251089,"TERMINAL",0,0,"4\t ",,terminal_output +10096,25251642,"TERMINAL",0,0,"43848",,terminal_output +10097,25252129,"TERMINAL",0,0,"5\t ",,terminal_output +10098,25252584,"TERMINAL",0,0,"54959",,terminal_output +10099,25253178,"TERMINAL",0,0,"6\t ",,terminal_output +10100,25253692,"TERMINAL",0,0,"6530640",,terminal_output +10101,25254211,"TERMINAL",0,0,"7\t ",,terminal_output +10102,25254714,"TERMINAL",0,0,"76171",,terminal_output +10103,25255283,"TERMINAL",0,0,"8\t ",,terminal_output +10104,25255710,"TERMINAL",0,0,"87282",,terminal_output +10105,25256454,"TERMINAL",0,0,"9\t ",,terminal_output +10106,25256745,"TERMINAL",0,0,"98393",,terminal_output +10107,25257386,"TERMINAL",0,0,"30\t ",,terminal_output +10108,25257789,"TERMINAL",0,0,"3094204",,terminal_output +10109,25258372,"TERMINAL",0,0,"1\t ",,terminal_output +10110,25258822,"TERMINAL",0,0,"131626",,terminal_output +10111,25259423,"TERMINAL",0,0,"2\t ",,terminal_output +10112,25259865,"TERMINAL",0,0,"32737",,terminal_output +10113,25260445,"TERMINAL",0,0,"3\t ",,terminal_output +10114,25260962,"TERMINAL",0,0,"43848",,terminal_output +10115,25261678,"TERMINAL",0,0,"4\t ",,terminal_output +10116,25261987,"TERMINAL",0,0,"54959",,terminal_output +10117,25262535,"TERMINAL",0,0,"5\t ",,terminal_output +10118,25263013,"TERMINAL",0,0,"6540650",,terminal_output +10119,25263681,"TERMINAL",0,0,"6\t ",,terminal_output +10120,25264053,"TERMINAL",0,0,"76171",,terminal_output +10121,25264488,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining]633;D;0",,terminal_output +10122,25265110,"TERMINAL",0,0,"87282",,terminal_output +10123,25266212,"TERMINAL",0,0,"watch",,terminal_focus +10124,25266217,"TERMINAL",0,0,"98393",,terminal_output diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-0538b58f-47c4-46c6-9641-bb5ac5ffb10f1759593994343-2025_10_04-18.07.24.830/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-0538b58f-47c4-46c6-9641-bb5ac5ffb10f1759593994343-2025_10_04-18.07.24.830/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..55a24e90aad677c951a5c1e3aed75ef760969832 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-0538b58f-47c4-46c6-9641-bb5ac5ffb10f1759593994343-2025_10_04-18.07.24.830/source.csv @@ -0,0 +1,654 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +2,388,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"6:07:24 PM [info] Activating crowd-code\n6:07:24 PM [info] Recording started\n6:07:24 PM [info] Initializing git provider using file system watchers...\n",Log,tab +3,616,"extension-output-pdoom-org.crowd-code-#1-crowd-code",150,0,"6:07:25 PM [info] Git repository found\n6:07:25 PM [info] Git provider initialized successfully\n6:07:25 PM [info] Initial git state: [object Object]\n",Log,content +4,4644,"TERMINAL",0,0,"bash",,terminal_focus +5,6178,"TERMINAL",0,0,"queue",,terminal_command +6,6218,"TERMINAL",0,0,"]633;C",,terminal_output +7,6303,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1990.localdomain: Sat Oct 4 18:07:30 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3541783 accelerat preproce tum_cte0 R 1:21:01\t 1 hkn0517",,terminal_output +8,7354,"TERMINAL",0,0,"13",,terminal_output +9,9048,"TERMINAL",0,0,"34",,terminal_output +10,9444,"TERMINAL",0,0,"45",,terminal_output +11,10479,"TERMINAL",0,0,"56",,terminal_output +12,11534,"TERMINAL",0,0,"67",,terminal_output +13,12599,"TERMINAL",0,0,"78",,terminal_output +14,13011,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1990:~/Projects/jasmine",,terminal_output +15,15087,"TERMINAL",0,0,"logs",,terminal_command +16,15114,"TERMINAL",0,0,"]633;C]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir",,terminal_output +17,15649,"TERMINAL",0,0,"ls",,terminal_command +18,15690,"TERMINAL",0,0,"]633;C",,terminal_output +19,16007,"TERMINAL",0,0,"atari\r\nbig_run\r\nbig-runs\r\nbreakout\r\ncausal\r\ncoinrun\r\ndata_coinrun\r\njafar_og_reproduction\r\nlam\r\nmaskgit\r\nmaskgit-maskprob-fix\r\npreprocess\r\ntrain_dyn_causal_180M_3372931.log\r\ntrain_dyn_causal_180M_3372963.log\r\ntrain_dyn_causal_180M_3372969.log\r\ntrain_dyn_causal_180M_3373107.log\r\ntrain_dyn_causal_255M_3372932.log\r\ntrain_dyn_causal_255M_3372970.log\r\ntrain_dyn_causal_255M_3373108.log\r\ntrain_dyn_causal_356M_3372934.log\r\ntrain_dyn_causal_356M_3372971.log\r\ntrain_dyn_causal_356M_3373109.log\r\ntrain_dyn_causal_500M_3372936.log\r\ntrain_dyn_causal_500M_3372972.log\r\ntrain_dyn_causal_500M_3373110.log\r\ntrain_dyn_new_arch-bugfixed-spatial-shift_3359343.log\r\ntrain_dyn_new_arch-bugfixed-temporal-shift_3359349.log\r\ntrain_dyn_yolorun_3333026.log\r\ntrain_dyn_yolorun_3333448.log\r\ntrain_dyn_yolorun_3335345.log\r\ntrain_dyn_yolorun_3335362.log\r\ntrain_dyn_yolorun_3348592.log\r\ntrain_dyn_yolorun_new_arch_3351743.log\r\ntrain_dyn_yolorun_new_arch_3352103.log\r\ntrain_dyn_yolorun_new_arch_3352115.log\r\ntrain_dyn_yolorun_new_arch_3358457.log\r\ntrain_lam_action_space_scaling_10_3320179.log\r\ntrain_lam_action_space_scaling_10_3321529.log\r\ntrain_lam_action_space_scaling_10_3329786.log\r\ntrain_lam_action_space_scaling_10_3329801.log\r\ntrain_lam_action_space_scaling_10_3331283.log\r\ntrain_lam_action_space_scaling_12_3318546.log\r\ntrain_lam_action_space_scaling_12_3320177.log\r\ntrain_lam_action_space_scaling_12_3321527.log\r\ntrain_lam_action_space_scaling_12_3329787.log\r\ntrain_lam_action_space_scaling_12_3329802.log\r\ntrain_lam_action_space_scaling_12_3331284.log\r\ntrain_lam_action_space_scaling_20_3318547.log\r\ntrain_lam_action_space_scaling_20_3329788.log\r\ntrain_lam_action_space_scaling_20_3329803.log\r\ntrain_lam_action_space_scaling_20_3331285.log\r\ntrain_lam_action_space_scaling_50_3320180.log\r\ntrain_lam_action_space_scaling_50_3329789.log\r\ntrain_lam_action_space_scaling_50_3329804.log\r\ntrain_lam_action_space_scaling_50_3331286.log\r\ntrain_lam_action_space_scaling_6_3318549.log\r\ntrain_lam_action_space_scaling_6_3320178.log\r\ntrain_lam_action_space_scaling_6_3321528.log\r\ntrain_lam_action_space_scaling_6_3329790.log\r\ntrain_lam_action_space_scaling_6_3329805.log\r\ntrain_lam_action_space_scaling_6_3331287.log\r\ntrain_lam_action_space_scaling_8_3318550.log\r\ntrain_lam_action_space_scaling_8_3329791.log\r\ntrain_lam_action_space_scaling_8_3329806.log\r\ntrain_lam_action_space_scaling_8_3331288.log\r\ntrain_lam_batch_size_scaling_sqrt_lr_8_nodes_3528968.log\r\ntrain_lam_minecraft_overfit_sample_3309655.log\r\ntrain_lam_model_size_scaling_38M_3317098.log\r\ntrain_lam_model_size_scaling_38M_3317115.log\r\ntrain_lam_model_size_scaling_38M_3317231.log\r\ntrain_tokenizer_batch_size_scaling_16_node_3321526.log\r\ntrain_tokenizer_batch_size_scaling_1_node_3318551.log\r\ntrain_tokenizer_batch_size_scaling_2_node_3318552.log\r\ntrain_tokenizer_batch_size_scaling_2_node_3330806.log\r\ntrain_tokenizer_batch_size_scaling_2_node_3330848.log\r\ntrain_tokenizer_batch_size_scaling_2_node_3331282.log\r\ntrain_tokenizer_batch_size_scaling_4_node_3318553.log\r\ntrain_tokenizer_batch_size_scaling_4_node_3320175.log\r\ntrain_tokenizer_batch_size_scaling_4_node_3321524.log\r\ntrain_tokenizer_batch_size_scaling_8_node_3320176.log\r\ntrain_tokenizer_batch_size_scaling_8_node_3321525.log\r\ntrain_tokenizer_batch_size_scaling_8_node_3528969.log\r\ntrain_tokenizer_minecraft_overfit_sample_3309656.log\r\ntrain_tokenizer_model_size_scaling_127M_3317233.log\r\ntrain_tokenizer_model_size_scaling_127M_3318554.log\r\ntrain_tokenizer_model_size_scaling_140M_3313562.log\r\ntrain_tokenizer_model_size_scaling_140M_3316019.log\r\ntrain_tokenizer_model_size_scaling_200M_3313563.log\r\ntrain_tokenizer_model_size_scaling_200M_3316020.log\r\ntrain_tokenizer_model_size_scaling_227M_3317234.log\r\ntrain_tokenizer_model_size_scaling_227M_3318555.log\r\ntrain_tokenizer_model_size_scaling_227M_3320173.log\r\ntrain_tokenizer_model_size_scaling_227M_3321523.log\r\ntrain_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_tokenizer_model_size_scaling_80M_3316026.log\r\nyoloruns\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir",,terminal_output +20,18899,"TERMINAL",0,0,"cd preprocess/",,terminal_command +21,19210,"TERMINAL",0,0,"ls",,terminal_command +22,19283,"TERMINAL",0,0,"]633;C",,terminal_output +23,19331,"TERMINAL",0,0,"atari breakout coinrun doom minecraft\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/preprocess",,terminal_output +24,22268,"TERMINAL",0,0,"cd doom/",,terminal_command +25,22560,"TERMINAL",0,0,"ls",,terminal_command +26,27813,"TERMINAL",0,0,"queue",,terminal_command +27,27855,"TERMINAL",0,0,"]633;C",,terminal_output +28,27899,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1990.localdomain: Sat Oct 4 18:07:52 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3541783 accelerat preproce tum_cte0 R 1:21:23\t 1 hkn0517",,terminal_output +29,28952,"TERMINAL",0,0,"34",,terminal_output +30,29406,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/preprocess/doom",,terminal_output +31,32015,"TERMINAL",0,0,"cd ..",,terminal_command +32,33459,"TERMINAL",0,0,"cd coinrun/",,terminal_command +33,33723,"TERMINAL",0,0,"ls",,terminal_command +34,33742,"TERMINAL",0,0,"]633;C",,terminal_output +35,33849,"TERMINAL",0,0,"generate_coinrun_chunked_500m_3533473.log preprocess_coinrun_chunked_3500818.log\r\ngenerate_coinrun_chunked_500m_3533684.log preprocess_coinrun_chunked_3533689.log\r\ngenerate_coinrun_chunked_500m_3533688.log preprocess_coinrun_chunked_3533695.log\r\ngenerate_coinrun_chunked_500m_3533696.log preprocess_coinrun_chunked_3533706.log\r\ngenerate_coinrun_chunked_500m_3533705.log preprocess_coinrun_chunked_3541775.log\r\npreprocess_coinrun_chunked_3476960.log preprocess_coinrun_chunked_3541783.log\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/preprocess/coinrun",,terminal_output +36,45296,"TERMINAL",0,0,"tail -f preprocess_coinrun_chunked_3541783.log",,terminal_command +37,45340,"TERMINAL",0,0,"]633;C warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/data/.venv/lib/python3.10/site-packages/stable_baselines3/common/save_util.py:167: UserWarning: Could not deserialize object clip_range. Consider using `custom_objects` argument to replace this object.\r\nException: code expected at most 16 arguments, got 18\r\n warnings.warn(\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/data/.venv/lib/python3.10/site-packages/stable_baselines3/common/save_util.py:167: UserWarning: Could not deserialize object lr_schedule. Consider using `custom_objects` argument to replace this object.\r\nException: code expected at most 16 arguments, got 18\r\n warnings.warn(\r\nAL lib: (WW) alc_initconfig: Failed to initialize backend ""pulse""\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/data/.venv/lib/python3.10/site-packages/stable_baselines3/common/vec_env/patch_gym.py:49: UserWarning: You provided an OpenAI Gym environment. We strongly recommend transitioning to Gymnasium environments. Stable-Baselines3 is automatically wrapping your environments in a compatibility layer, which could potentially cause issues.\r\n warnings.warn(\r\n",,terminal_output +38,149567,"TERMINAL",0,0,"bash",,terminal_focus +39,161968,"data/jasmine_data/ViZDoomPPO/common/envs.py",0,0,"import typing as t\n\nimport cv2\nimport numpy as np\nimport vizdoom\nfrom gym import Env\nfrom gym import spaces\nfrom stable_baselines3.common import vec_env\nfrom stable_baselines3.common.callbacks import EvalCallback\nfrom stable_baselines3.ppo import ppo, policies\nfrom vizdoom import GameVariable\n\nfrom common.models import init_model\nfrom common.utils import get_available_actions\n\nfrom tqdm import tqdm\nimport time\nfrom stable_baselines3.common.callbacks import BaseCallback\nfrom stable_baselines3.common.vec_env import SubprocVecEnv, VecTransposeImage\n\nimport torch\nimport os \nimport sys\n\nFrame = np.ndarray\n\nDOOM_ENV_WITH_BOTS_ARGS = """"""\n -host 1 \n -deathmatch \n +viz_nocheat 0 \n +cl_run 1 \n +name AGENT \n +colorset 0 \n +sv_forcerespawn 1 \n +sv_respawnprotect 1 \n +sv_nocrouch 1 \n +sv_noexit 1\n """"""\n\n\nclass DoomEnv(Env):\n """"""Wrapper environment following OpenAI's gym interface for a VizDoom game instance.""""""\n\n def __init__(self,\n game: vizdoom.DoomGame,\n frame_processor: t.Callable,\n frame_skip: int = 4):\n super().__init__()\n\n # Determine action space\n self.action_space = spaces.Discrete(game.get_available_buttons_size())\n\n # Determine observation space\n h, w, c = game.get_screen_height(), game.get_screen_width(), game.get_screen_channels()\n new_h, new_w, new_c = frame_processor(np.zeros((h, w, c))).shape\n self.observation_space = spaces.Box(low=0, high=255, shape=(new_h, new_w, new_c), dtype=np.uint8)\n\n # Assign other variables\n self.game = game\n self.possible_actions = np.eye(self.action_space.n).tolist() # VizDoom needs a list of buttons states.\n self.frame_skip = frame_skip\n self.frame_processor = frame_processor\n\n self.empty_frame = np.zeros(self.observation_space.shape, dtype=np.uint8)\n self.state = self.empty_frame\n\n def step(self, action: int) -> t.Tuple[Frame, int, bool, t.Dict]:\n """"""Apply an action to the environment.\n\n Args:\n action:\n\n Returns:\n A tuple containing:\n - A numpy ndarray containing the current environment state.\n - The reward obtained by applying the provided action.\n - A boolean flag indicating whether the episode has ended.\n - An empty info dict.\n """"""\n reward = self.game.make_action(self.possible_actions[action], self.frame_skip)\n done = self.game.is_episode_finished()\n self.state = self._get_frame(done)\n\n return self.state, reward, done, {}\n\n def reset(self) -> Frame:\n """"""Resets the environment.\n\n Returns:\n The initial state of the new environment.\n """"""\n self.game.new_episode()\n self.state = self._get_frame()\n\n return self.state\n\n def close(self) -> None:\n self.game.close()\n\n def render(self, mode='human'):\n pass\n\n def _get_frame(self, done: bool = False) -> Frame:\n return self.frame_processor(\n self.game.get_state().screen_buffer) if not done else self.empty_frame\n\n\nclass DoomWithBots(DoomEnv):\n\n def __init__(self, game, frame_processor, frame_skip, n_bots):\n super().__init__(game, frame_processor, frame_skip)\n self.n_bots = n_bots\n self.last_frags = 0\n self._reset_bots()\n\n # Redefine the action space using combinations.\n self.possible_actions = get_available_actions(np.array(game.get_available_buttons()))\n self.action_space = spaces.Discrete(len(self.possible_actions))\n\n def step(self, action):\n self.game.make_action(self.possible_actions[action], self.frame_skip)\n\n # Compute rewards.\n frags = self.game.get_game_variable(GameVariable.FRAGCOUNT)\n reward = frags - self.last_frags\n self.last_frags = frags\n\n # Check for episode end.\n self._respawn_if_dead()\n done = self.game.is_episode_finished()\n self.state = self._get_frame(done)\n\n return self.state, reward, done, {}\n\n def reset(self):\n self._reset_bots()\n self.last_frags = 0\n\n return super().reset()\n\n def _respawn_if_dead(self):\n if not self.game.is_episode_finished():\n if self.game.is_player_dead():\n self.game.respawn_player()\n\n def _reset_bots(self):\n # Make sure you have the bots.cfg file next to the program entry point.\n self.game.send_game_command('removebots')\n for i in range(self.n_bots):\n self.game.send_game_command('addbot')\n\n def _print_state(self):\n server_state = self.game.get_server_state()\n player_scores = list(zip(\n server_state.players_names,\n server_state.players_frags,\n server_state.players_in_game))\n player_scores = sorted(player_scores, key=lambda tup: tup[1])\n\n # print('*** DEATHMATCH RESULTS ***')\n # for player_name, player_score, player_ingame in player_scores:\n # if player_ingame:\n # print(f' - {player_name}: {player_score}')\n\n\ndef default_frame_processor(frame: Frame) -> Frame:\n return cv2.resize(frame[40:, 4:-4], None, fx=.5, fy=.5, interpolation=cv2.INTER_AREA)\n\n\ndef create_env(scenario: str, **kwargs) -> DoomEnv:\n # Create a VizDoom instance.\n game = vizdoom.DoomGame()\n game.load_config(f'scenarios/{scenario}.cfg')\n game.set_window_visible(False) \n game.init()\n\n # Wrap the game with the Gym adapter.\n return DoomEnv(game, **kwargs)\n\n\ndef create_env_with_bots(scenario: str, **kwargs) -> DoomEnv:\n # Create a VizDoom instance.\n game = vizdoom.DoomGame()\n game.load_config(f'scenarios/{scenario}.cfg')\n game.add_game_args(DOOM_ENV_WITH_BOTS_ARGS)\n game.set_window_visible(False) \n game.init()\n\n return DoomWithBots(game, **kwargs)\n\ndef create_vec_env(n_envs: int = 1, **kwargs) -> VecTransposeImage:\n return VecTransposeImage(SubprocVecEnv([lambda: create_env(**kwargs) for _ in range(n_envs)]))\n\ndef vec_env_with_bots(n_envs: int = 1, **kwargs) -> VecTransposeImage:\n return VecTransposeImage(SubprocVecEnv([lambda: create_env_with_bots(**kwargs) for _ in range(n_envs)]))\n\n\ndef create_eval_vec_env(**kwargs) -> vec_env.VecTransposeImage:\n return create_vec_env(n_envs=1, **kwargs)\n\n\n\ndef solve_env(env: vec_env.VecTransposeImage, eval_env: vec_env.VecTransposeImage, scenario: str, agent_args: t.Dict, resume: bool = False, load_path: str = None):\n device = torch.device(""cuda"" if torch.cuda.is_available() else ""cpu"")\n \n if resume:\n # Load the existing model\n if load_path != """":\n agent = ppo.PPO.load(load_path, env=env, tensorboard_log='logs/tensorboard', **agent_args)\n print(f""Resumed training from {load_path}"")\n else:\n print(""Resume selected but no path provided"")\n sys.exit()\n else:\n # Create a new agent\n agent = ppo.PPO(policies.ActorCriticCnnPolicy, env, tensorboard_log='logs/tensorboard', seed=0, **agent_args)\n init_model(agent)\n\n agent.policy.to(device)\n\n # Create callbacks.\n eval_callback = EvalCallback(\n eval_env,\n n_eval_episodes=5,\n eval_freq=4000,\n log_path=f'logs/evaluations/{scenario}',\n best_model_save_path=f'logs/models/{scenario}'\n )\n\n # Set up progress bar\n total_timesteps = 10_000_000\n pbar = tqdm(total=total_timesteps, desc=""Training Progress"")\n class ProgressBarCallback(BaseCallback):\n def __init__(self, pbar):\n super().__init__()\n self.pbar = pbar\n self.last_time = time.time()\n self.last_timesteps = 0\n \n def _on_step(self):\n current_timesteps = self.num_timesteps - self.last_timesteps\n self.pbar.update(current_timesteps)\n self.last_timesteps = self.num_timesteps\n \n current_time = time.time()\n steps_per_second = current_timesteps / (current_time - self.last_time)\n self.pbar.set_postfix({""steps/s"": f""{steps_per_second:.2f}""})\n self.last_time = current_time\n \n return True\n \n def on_training_end(self):\n self.pbar.close()\n\n progress_callback = ProgressBarCallback(pbar)\n\n # Start the training process.\n try:\n agent.learn(\n total_timesteps=10_000_000, \n tb_log_name=scenario, \n callback=[eval_callback, progress_callback],\n reset_num_timesteps=not resume # Don't reset timesteps if resuming\n )\n finally:\n pbar.close()\n env.close()\n eval_env.close()\n\n return agent\n\ndef save_model(agent: ppo.PPO, scenario: str):\n """"""Save the trained model.""""""\n save_path = f'logs/models/{scenario}/final_model'\n if not os.path.exists(os.path.dirname(save_path)):\n os.makedirs(os.path.dirname(save_path))\n agent.save(save_path)\n print(f""Model saved to {save_path}"")\n\n\ndef load_model(load_path: str, env: vec_env.VecTransposeImage) -> ppo.PPO:\n """"""Load a trained model.""""""\n if not os.path.exists(os.path.dirname(load_path)):\n os.makedirs(os.path.dirname(load_path))\n agent = ppo.PPO.load(load_path, env=env)\n print(f""Model loaded from {load_path}"")\n return agent",python,tab +40,161971,"data/jasmine_data/ViZDoomPPO/common/envs.py",8130,0,"",python,selection_mouse +41,162562,"data/jasmine_data/ViZDoomPPO/common/envs.py",8038,0,"",python,selection_mouse +42,162726,"data/jasmine_data/ViZDoomPPO/common/envs.py",8025,17,"current_timesteps",python,selection_mouse +43,164958,"data/jasmine_data/ViZDoomPPO/common/envs.py",8057,0,"",python,selection_mouse +44,165150,"data/jasmine_data/ViZDoomPPO/common/envs.py",8046,12,"current_time",python,selection_mouse +45,165272,"data/jasmine_data/ViZDoomPPO/common/envs.py",7994,83," steps_per_second = current_timesteps / (current_time - self.last_time)\n",python,selection_mouse +46,166089,"data/jasmine_data/ViZDoomPPO/common/envs.py",8150,0,"",python,selection_mouse +47,166097,"data/jasmine_data/ViZDoomPPO/common/envs.py",8149,0,"",python,selection_command +48,166257,"data/jasmine_data/ViZDoomPPO/common/envs.py",8150,0,"",python,selection_mouse +49,166258,"data/jasmine_data/ViZDoomPPO/common/envs.py",8149,0,"",python,selection_command +50,167023,"data/jasmine_data/ViZDoomPPO/common/envs.py",8045,0,"",python,selection_mouse +51,167847,"data/jasmine_data/ViZDoomPPO/common/envs.py",8051,0,"",python,selection_mouse +52,168012,"data/jasmine_data/ViZDoomPPO/common/envs.py",8046,12,"current_time",python,selection_mouse +53,172549,"TERMINAL",0,0,"tail",,terminal_focus +54,179667,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/preprocess/coinrun",,terminal_output +55,187693,"TERMINAL",0,0,"cd $ws_dir",,terminal_command +56,188301,"TERMINAL",0,0,"ls",,terminal_command +57,188344,"TERMINAL",0,0,"]633;C",,terminal_output +58,188450,"TERMINAL",0,0,"checkpoints data_breakout data_new scripts\r\ncount_items.sh data_coinrun huggingface\r\ndata data_doom logs\r\ndata_atari data_minecraft possibly_corrupt_files_in_this_workspace.txt\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared",,terminal_output +59,191700,"TERMINAL",0,0,"cd data_doom/",,terminal_command +60,192016,"TERMINAL",0,0,"ls",,terminal_command +61,194593,"TERMINAL",0,0,"cd doom_episodes_50m/",,terminal_command +62,195716,"TERMINAL",0,0,"ls",,terminal_command +63,197536,"TERMINAL",0,0,"cd train/",,terminal_command +64,197802,"TERMINAL",0,0,"ls",,terminal_command +65,197851,"TERMINAL",0,0,"]633;C",,terminal_output +66,198219,"TERMINAL",0,0,"data_0000.array_record data_0063.array_record data_0126.array_record\r\ndata_0001.array_record data_0064.array_record data_0127.array_record\r\ndata_0002.array_record data_0065.array_record data_0128.array_record\r\ndata_0003.array_record data_0066.array_record data_0129.array_record\r\ndata_0004.array_record data_0067.array_record data_0130.array_record\r\ndata_0005.array_record data_0068.array_record data_0131.array_record\r\ndata_0006.array_record data_0069.array_record data_0132.array_record\r\ndata_0007.array_record data_0070.array_record data_0133.array_record\r\ndata_0008.array_record data_0071.array_record data_0134.array_record\r\ndata_0009.array_record data_0072.array_record data_0135.array_record\r\ndata_0010.array_record data_0073.array_record data_0136.array_record\r\ndata_0011.array_record data_0074.array_record data_0137.array_record\r\ndata_0012.array_record data_0075.array_record data_0138.array_record\r\ndata_0013.array_record data_0076.array_record data_0139.array_record\r\ndata_0014.array_record data_0077.array_record data_0140.array_record\r\ndata_0015.array_record data_0078.array_record data_0141.array_record\r\ndata_0016.array_record data_0079.array_record data_0142.array_record\r\ndata_0017.array_record data_0080.array_record data_0143.array_record\r\ndata_0018.array_record data_0081.array_record data_0144.array_record\r\ndata_0019.array_record data_0082.array_record data_0145.array_record\r\ndata_0020.array_record data_0083.array_record data_0146.array_record\r\ndata_0021.array_record data_0084.array_record data_0147.array_record\r\ndata_0022.array_record data_0085.array_record data_0148.array_record\r\ndata_0023.array_record data_0086.array_record data_0149.array_record\r\ndata_0024.array_record data_0087.array_record data_0150.array_record\r\ndata_0025.array_record data_0088.array_record data_0151.array_record\r\ndata_0026.array_record data_0089.array_record data_0152.array_record\r\ndata_0027.array_record data_0090.array_record data_0153.array_record\r\ndata_0028.array_record data_0091.array_record data_0154.array_record\r\ndata_0029.array_record data_0092.array_record data_0155.array_record\r\ndata_0030.array_record data_0093.array_record data_0156.array_record\r\ndata_0031.array_record data_0094.array_record data_0157.array_record\r\ndata_0032.array_record data_0095.array_record data_0158.array_record\r\ndata_0033.array_record data_0096.array_record data_0159.array_record\r\ndata_0034.array_record data_0097.array_record data_0160.array_record\r\ndata_0035.array_record data_0098.array_record data_0161.array_record\r\ndata_0036.array_record data_0099.array_record data_0162.array_record\r\ndata_0037.array_record data_0100.array_record data_0163.array_record\r\ndata_0038.array_record data_0101.array_record data_0164.array_record\r\ndata_0039.array_record data_0102.array_record data_0165.array_record\r\ndata_0040.array_record data_0103.array_record data_0166.array_record\r\ndata_0041.array_record data_0104.array_record data_0167.array_record\r\ndata_0042.array_record data_0105.array_record data_0168.array_record\r\ndata_0043.array_record data_0106.array_record data_0169.array_record\r\ndata_0044.array_record data_0107.array_record data_0170.array_record\r\ndata_0045.array_record data_0108.array_record data_0171.array_record\r\ndata_0046.array_record data_0109.array_record data_0172.array_record\r\ndata_0047.array_record data_0110.array_record data_0173.array_record\r\ndata_0048.array_record data_0111.array_record data_0174.array_record\r\ndata_0049.array_record data_0112.array_record data_0175.array_record\r\ndata_0050.array_record data_0113.array_record data_0176.array_record\r\ndata_0051.array_record data_0114.array_record data_0177.array_record\r\ndata_0052.array_record data_0115.array_record data_0178.array_record\r\ndata_0053.array_record data_0116.array_record data_0179.array_record\r\ndata_0054.array_record data_0117.array_record data_0180.array_record\r\ndata_0055.array_record data_0118.array_record data_0181.array_record\r\ndata_0056.array_record data_0119.array_record data_0182.array_record\r\ndata_0057.array_record data_0120.array_record data_0183.array_record\r\ndata_0058.array_record data_0121.array_record data_0184.array_record\r\ndata_0059.array_record data_0122.array_record data_0185.array_record\r\ndata_0060.array_record data_0123.array_record data_0186.array_record\r\ndata_0061.array_record data_0124.array_record\r\ndata_0062.array_record data_0125.array_record\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_doom/doom_episodes_50m/train",,terminal_output +67,210825,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",0,0,"# # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # \n# /$$ /$$ /$$ /$$$$$$$ /$$$$$$$ /$$$$$$$ /$$$$$$ #\n# | $$ | $$|__/ | $$__ $$ | $$__ $$| $$__ $$ /$$__ $$ #\n# | $$ | $$ /$$ /$$$$$$$$| $$ \ $$ /$$$$$$ /$$$$$$ /$$$$$$/$$$$ | $$ \ $$| $$ \ $$| $$ \ $$ #\n# | $$ / $$/| $$|____ /$$/| $$ | $$ /$$__ $$ /$$__ $$| $$_ $$_ $$ | $$$$$$$/| $$$$$$$/| $$ | $$ #\n# \ $$ $$/ | $$ /$$$$/ | $$ | $$| $$ \ $$| $$ \ $$| $$ \ $$ \ $$ | $$____/ | $$____/ | $$ | $$ #\n# \ $$$/ | $$ /$$__/ | $$ | $$| $$ | $$| $$ | $$| $$ | $$ | $$ | $$ | $$ | $$ | $$ #\n# \ $/ | $$ /$$$$$$$$| $$$$$$$/| $$$$$$/| $$$$$$/| $$ | $$ | $$ | $$ | $$ | $$$$$$/ #\n# \_/ |__/|________/|_______/ \______/ \______/ |__/ |__/ |__/ |__/ |__/ \______/ #\n# #\n# # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # \n \n# FORK OF LEANDRO KIELIGER'S DOOM PPO TUTORIAL: https://lkieliger.medium.com/deep-reinforcement-learning-in-practice-by-playing-doom-part-1-getting-started-618c99075c77 \n\n# SCRIPT TO TRAIN A PPO AGENT ON THE ViZDOOM DEATHMATCH ENVIRONMENT. \n\nimport numpy as np\nimport os\nimport vizdoom\nfrom common import envs\nfrom common.models import CustomCNN\nfrom stable_baselines3.common.vec_env import VecTransposeImage, SubprocVecEnv\nimport random\nimport string\nfrom vizdoom.vizdoom import GameVariable\nfrom collections import deque\nimport torch\n\ndevice = torch.device(""cuda"" if torch.cuda.is_available() else ""cpu"")\n# device=torch.device(""mps"")\nprint(f""Using device: {device}"")\n\n# Rewards\n# 1 per kill\nreward_factor_frag = 1.0\nreward_factor_damage = 0.01\n\n# Player can move at ~16.66 units per tick\nreward_factor_distance = 5e-4\npenalty_factor_distance = -2.5e-3\nreward_threshold_distance = 3.0\n\n# Pistol clips have 10 bullets\nreward_factor_ammo_increment = 0.02\nreward_factor_ammo_decrement = -0.01\n\n# Player starts at 100 health\nreward_factor_health_increment = 0.02\nreward_factor_health_decrement = -0.01\nreward_factor_armor_increment = 0.01\n\n\n# List of game variables storing ammunition information. Used for keeping track of ammunition-related rewards.\nAMMO_VARIABLES = [\n GameVariable.AMMO0,\n GameVariable.AMMO1,\n GameVariable.AMMO2,\n GameVariable.AMMO3,\n GameVariable.AMMO4,\n GameVariable.AMMO5,\n GameVariable.AMMO6,\n GameVariable.AMMO7,\n GameVariable.AMMO8,\n GameVariable.AMMO9,\n]\n\n# List of game variables storing weapon information. Used for keeping track of ammunition-related rewards.\nWEAPON_VARIABLES = [\n GameVariable.WEAPON0,\n GameVariable.WEAPON1,\n GameVariable.WEAPON2,\n GameVariable.WEAPON3,\n GameVariable.WEAPON4,\n GameVariable.WEAPON5,\n GameVariable.WEAPON6,\n GameVariable.WEAPON7,\n GameVariable.WEAPON8,\n GameVariable.WEAPON9,\n]\n\n\nclass DoomWithBotsShaped(envs.DoomWithBots):\n """"""An environment wrapper for a Doom deathmatch game with bots.\n\n Rewards are shaped according to the multipliers defined in the notebook.\n """"""\n\n def __init__(self, game, frame_processor, frame_skip, n_bots, shaping):\n super().__init__(game, frame_processor, frame_skip, n_bots)\n\n # Give a random two-letter name to the agent for identifying instances in parallel learning.\n self.name = """".join(random.choices(string.ascii_uppercase + string.digits, k=2))\n self.shaping = shaping\n\n # Internal states\n self.last_health = 100\n self.last_x, self.last_y = self._get_player_pos()\n self.ammo_state = self._get_ammo_state()\n self.weapon_state = self._get_weapon_state()\n self.total_rew = self.last_damage_dealt = self.deaths = self.last_frags = (\n self.last_armor\n ) = 0\n\n # Store individual reward contributions for logging purposes\n self.rewards_stats = {\n ""frag"": 0,\n ""damage"": 0,\n ""ammo"": 0,\n ""health"": 0,\n ""armor"": 0,\n ""distance"": 0,\n }\n\n def step(self, action, array=False):\n # Perform the action as usual\n state, reward, done, info = super().step(action)\n\n self._log_reward_stat(""frag"", reward)\n\n # Adjust the reward based on the shaping table\n if self.shaping:\n shaped_reward = reward + self.shape_rewards()\n else:\n shaped_reward = reward\n\n self.total_rew += shaped_reward\n\n return state, shaped_reward, done, info\n\n def reset(self):\n self._print_state()\n\n state = super().reset()\n\n self.last_health = 100\n self.last_x, self.last_y = self._get_player_pos()\n self.last_armor = self.last_frags = self.total_rew = self.deaths = 0\n\n # Damage count is not cleared when starting a new episode: https://github.com/mwydmuch/ViZDoom/issues/399\n # self.last_damage_dealt = 0\n\n # Reset reward stats\n for k in self.rewards_stats.keys():\n self.rewards_stats[k] = 0\n\n return state\n\n def seed(self, seed=None):\n if seed is not None:\n random.seed(seed)\n np.random.seed(seed)\n return [seed]\n\n def shape_rewards(self):\n reward_contributions = [\n self._compute_damage_reward(),\n self._compute_ammo_reward(),\n self._compute_health_reward(),\n self._compute_armor_reward(),\n self._compute_distance_reward(*self._get_player_pos()),\n ]\n\n return sum(reward_contributions)\n\n def _respawn_if_dead(self):\n if not self.game.is_episode_finished():\n # Check if player is dead\n if self.game.is_player_dead():\n self.deaths += 1\n self._reset_player()\n\n def _compute_distance_reward(self, x, y):\n """"""Computes a reward/penalty based on the distance travelled since last update.""""""\n dx = self.last_x - x\n dy = self.last_y - y\n\n distance = np.sqrt(dx**2 + dy**2)\n\n if distance - reward_threshold_distance > 0:\n reward = reward_factor_distance\n else:\n reward = -reward_factor_distance\n\n self.last_x = x\n self.last_y = y\n self._log_reward_stat(""distance"", reward)\n\n return reward\n\n def _compute_damage_reward(self):\n """"""Computes a reward based on total damage inflicted to enemies since last update.""""""\n damage_dealt = self.game.get_game_variable(GameVariable.DAMAGECOUNT)\n reward = reward_factor_damage * (damage_dealt - self.last_damage_dealt)\n\n self.last_damage_dealt = damage_dealt\n self._log_reward_stat(""damage"", reward)\n\n return reward\n\n def _compute_health_reward(self):\n """"""Computes a reward/penalty based on total health change since last update.""""""\n # When the player is dead, the health game variable can be -999900\n health = max(self.game.get_game_variable(GameVariable.HEALTH), 0)\n\n health_reward = reward_factor_health_increment * max(\n 0, health - self.last_health\n )\n health_penalty = reward_factor_health_decrement * min(\n 0, health - self.last_health\n )\n reward = health_reward - health_penalty\n\n self.last_health = health\n self._log_reward_stat(""health"", reward)\n\n return reward\n\n def _compute_armor_reward(self):\n """"""Computes a reward/penalty based on total armor change since last update.""""""\n armor = self.game.get_game_variable(GameVariable.ARMOR)\n reward = reward_factor_armor_increment * max(0, armor - self.last_armor)\n\n self.last_armor = armor\n self._log_reward_stat(""armor"", reward)\n\n return reward\n\n def _compute_ammo_reward(self):\n """"""Computes a reward/penalty based on total ammunition change since last update.""""""\n self.weapon_state = self._get_weapon_state()\n\n new_ammo_state = self._get_ammo_state()\n ammo_diffs = (new_ammo_state - self.ammo_state) * self.weapon_state\n ammo_reward = reward_factor_ammo_increment * max(0, np.sum(ammo_diffs))\n ammo_penalty = reward_factor_ammo_decrement * min(0, np.sum(ammo_diffs))\n reward = ammo_reward - ammo_penalty\n\n self.ammo_state = new_ammo_state\n self._log_reward_stat(""ammo"", reward)\n\n return reward\n\n def _get_player_pos(self):\n """"""Returns the player X- and Y- coordinates.""""""\n return self.game.get_game_variable(\n GameVariable.POSITION_X\n ), self.game.get_game_variable(GameVariable.POSITION_Y)\n\n def _get_ammo_state(self):\n """"""Returns the total available ammunition per weapon slot.""""""\n ammo_state = np.zeros(10)\n\n for i in range(10):\n ammo_state[i] = self.game.get_game_variable(AMMO_VARIABLES[i])\n\n return ammo_state\n\n def _get_weapon_state(self):\n """"""Returns which weapon slots can be used. Available weapons are encoded as ones.""""""\n weapon_state = np.zeros(10)\n\n for i in range(10):\n weapon_state[i] = self.game.get_game_variable(WEAPON_VARIABLES[i])\n\n return weapon_state\n\n def _log_reward_stat(self, kind: str, reward: float):\n self.rewards_stats[kind] += reward\n\n def _reset_player(self):\n self.last_health = 100\n self.last_armor = 0\n self.game.respawn_player()\n self.last_x, self.last_y = self._get_player_pos()\n self.ammo_state = self._get_ammo_state()\n\n def _print_state(self):\n super()._print_state()\n\n\nREWARD_THRESHOLDS = [5, 10, 15, 20, 25, 25]\n\n# Curriculum based learning\nclass DoomWithBotsCurriculum(DoomWithBotsShaped):\n def __init__(\n self,\n game,\n frame_processor,\n frame_skip,\n n_bots,\n shaping,\n initial_level=0,\n max_level=5,\n rolling_mean_length=10,\n ):\n super().__init__(game, frame_processor, frame_skip, n_bots, shaping)\n\n # Initialize ACS script difficulty level\n game.send_game_command(""pukename change_difficulty 0"")\n\n # Internal state\n self.level = initial_level\n self.max_level = max_level\n self.rolling_mean_length = rolling_mean_length\n self.last_rewards = deque(maxlen=rolling_mean_length)\n\n def step(self, action, array=False):\n # Perform action step as usual\n state, reward, done, infos = super().step(action, array)\n\n # After an episode, check whether difficulty should be increased.\n if done:\n self.last_rewards.append(self.total_rew)\n run_mean = np.mean(self.last_rewards)\n print(\n ""Avg. last 10 runs of {}: {:.2f}. Current difficulty level: {}"".format(\n self.name, run_mean, self.level\n )\n )\n if (\n run_mean > REWARD_THRESHOLDS[self.level]\n and len(self.last_rewards) >= self.rolling_mean_length\n ):\n self._change_difficulty()\n\n return state, reward, done, infos\n\n def reset(self):\n state = super().reset()\n self.game.send_game_command(f""pukename change_difficulty {self.level}"")\n\n return state\n\n def _change_difficulty(self):\n """"""Adjusts the difficulty by setting the difficulty level in the ACS script.""""""\n if self.level < self.max_level:\n self.level += 1\n print(f""Changing difficulty for {self.name} to {self.level}"")\n self.game.send_game_command(f""pukename change_difficulty {self.level}"")\n self.last_rewards = deque(maxlen=self.rolling_mean_length)\n else:\n print(f""{self.name} already at max level!"")\n\n\ndef game_instance(scenario):\n """"""Creates a Doom game instance.""""""\n game = vizdoom.DoomGame()\n dir_path = os.path.dirname(os.path.realpath(__file__))\n cfg_pth = os.path.join(dir_path, ""scenarios"", f""{scenario}.cfg"")\n game.load_config(cfg_pth)\n game.add_game_args(envs.DOOM_ENV_WITH_BOTS_ARGS)\n game.set_window_visible(False)\n game.init()\n\n return game\n\n\ndef env_with_bots_shaped(scenario, **kwargs) -> envs.DoomEnv:\n """"""Wraps a Doom game instance in an environment with shaped rewards.""""""\n game = game_instance(scenario)\n return DoomWithBotsShaped(game, **kwargs)\n\n\ndef vec_env_with_bots_shaped(n_envs=1, **kwargs) -> VecTransposeImage:\n """"""Wraps Doom game instances in a vectorized environment with shaped rewards using true parallelism.""""""\n return VecTransposeImage(\n SubprocVecEnv([lambda: env_with_bots_shaped(**kwargs) for _ in range(n_envs)])\n )\n\n\ndef env_with_bots_curriculum(scenario, **kwargs) -> envs.DoomEnv:\n """"""Wraps a Doom game instance in an environment with shaped rewards and curriculum.""""""\n game = game_instance(scenario)\n return DoomWithBotsCurriculum(game, **kwargs)\n\n\ndef vec_env_with_bots_curriculum(n_envs=1, **kwargs) -> VecTransposeImage:\n """"""Wraps a Doom game instance in a vectorized environment with shaped rewards and curriculum.""""""\n return VecTransposeImage(\n SubprocVecEnv([lambda: env_with_bots_curriculum(**kwargs) for _ in range(n_envs)])\n )\n\n\nif __name__ == ""__main__"":\n scenario = ""deathmatch_simple""\n\n # Agent parameters.\n agent_args = {\n ""n_steps"": 4096,\n ""learning_rate"": 1e-4,\n ""batch_size"": 64,\n ""policy_kwargs"": {""features_extractor_class"": CustomCNN},\n ""device"": device,\n }\n\n # Environment parameters.\n env_args = {\n ""scenario"": scenario,\n ""frame_skip"": 4,\n ""frame_processor"": envs.default_frame_processor,\n ""n_bots"": 6,\n ""shaping"": True,\n ""initial_level"": 1,\n ""max_level"": 5,\n }\n\n # In the evaluation environment we measure frags only.\n eval_env_args = dict(env_args)\n eval_env_args[""shaping""] = False\n\n n_envs = 32\n # Create environments with bots and shaping.\n env = vec_env_with_bots_curriculum(\n n_envs, **env_args\n ) # You can increase the number of parallel environments\n eval_env = vec_env_with_bots_curriculum(n_envs, **eval_env_args)\n\n agent = envs.solve_env(\n env,\n eval_env,\n scenario,\n agent_args,\n resume=False,\n )\n envs.save_model(agent, ""agent_test"")\n",python,tab +68,219592,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",0,0,"# # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # #\n# /$$ /$$ /$$ /$$$$$$$ /$$$$$$$ /$$$$$$$ /$$$$$$ #\n# | $$ | $$|__/ | $$__ $$ | $$__ $$| $$__ $$ /$$__ $$ #\n# | $$ | $$ /$$ /$$$$$$$$| $$ \ $$ /$$$$$$ /$$$$$$ /$$$$$$/$$$$ | $$ \ $$| $$ \ $$| $$ \ $$ #\n# | $$ / $$/| $$|____ /$$/| $$ | $$ /$$__ $$ /$$__ $$| $$_ $$_ $$ | $$$$$$$/| $$$$$$$/| $$ | $$ #\n# \ $$ $$/ | $$ /$$$$/ | $$ | $$| $$ \ $$| $$ \ $$| $$ \ $$ \ $$ | $$____/ | $$____/ | $$ | $$ #\n# \ $$$/ | $$ /$$__/ | $$ | $$| $$ | $$| $$ | $$| $$ | $$ | $$ | $$ | $$ | $$ | $$ #\n# \ $/ | $$ /$$$$$$$$| $$$$$$$/| $$$$$$/| $$$$$$/| $$ | $$ | $$ | $$ | $$ | $$$$$$/ #\n# \_/ |__/|________/|_______/ \______/ \______/ |__/ |__/ |__/ |__/ |__/ \______/ #\n# #\n# # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # \n \n# FORK OF LEANDRO KIELIGER'S DOOM PPO TUTORIAL: https://lkieliger.medium.com/deep-reinforcement-learning-in-practice-by-playing-doom-part-1-getting-started-618c99075c77 \n\n# SCRIPT TO RUN PPO AGENT AND GENERATE DATASET FOR DOOM ENVIRONMENT. \n\nfrom dataclasses import dataclass\nimport imageio\nfrom common import envs\nimport torch\nfrom vizdoom.vizdoom import GameVariable\n\n\nimport os\n\nimport numpy as np\nfrom train_ppo_parallel import DoomWithBotsCurriculum, game_instance\nfrom stable_baselines3.common.vec_env import (\n VecTransposeImage,\n DummyVecEnv\n)\nfrom loguru import logger\nimport tyro\nfrom jasmine_data.utils import save_chunks\n\n# To replicate frame_skip in the environment\nACTION_REPEAT = 4\n\n@dataclass\nclass Args:\n num_episodes_train: int = 1000\n num_episodes_val: int = 100\n num_episodes_test: int = 100\n min_episode_length: int = 100\n max_episode_length: int = 1000\n chunk_size: int = 160\n chunks_per_file: int = 100\n agent_path: str = """"\n seed: int = 0\n output_dir: str = ""data/vizdoom_episodes""\n generate_gif: bool = False\n\nargs = tyro.cli(Args)\ndevice = torch.device(""cuda"" if torch.cuda.is_available() else ""cpu"")\nlogger.info(f""Using device: {device}"")\n\ndef dummy_vec_env_with_bots_curriculum(n_envs=1, **kwargs) -> VecTransposeImage:\n """"""Wraps a Doom game instance in a vectorized environment with shaped rewards and curriculum.""""""\n scenario = kwargs.pop(""scenario"") # Remove 'scenario' from kwargs\n return VecTransposeImage(\n DummyVecEnv(\n [lambda: DoomWithBotsCurriculum(game_instance(scenario), **kwargs)] * n_envs\n )\n )\n\ndef make_gif(agent, eval_env_args):\n """"""Generate a GIF by running the agent in the environment.\n\n Args:\n agent: The trained PPO agent.\n file_path (str): Path to save the generated GIF.\n eval_env_args (dict): Arguments for the evaluation environment.\n num_episodes (int): Number of episodes to run.\n \n Returns:\n list: Collected health values for analysis.\n """"""\n # Set frame_skip to 1 to capture all frames\n eval_env_args['frame_skip'] = 1\n env = dummy_vec_env_with_bots_curriculum(1, **eval_env_args)\n\n images = []\n actions = []\n health_values = []\n current_action = None\n frame_counter = 0\n\n obs = env.reset()\n\n done = False\n while not done and frame_counter < args.max_episode_length:\n if frame_counter % ACTION_REPEAT == 0:\n current_action, _ = agent.predict(obs)\n \n obs, _, done, _ = env.step(current_action)\n\n # Get the raw screen buffer from the Doom game instance\n screen = env.venv.envs[0].game.get_state().screen_buffer\n\n # Get the current health value\n health = env.venv.envs[0].game.get_game_variable(GameVariable.HEALTH)\n health_values.append(health) # Store the health value\n\n actions.append(current_action)\n images.append(screen)\n\n frame_counter += 1\n\n print(""Health values:"", health_values)\n print(""Number of health values:"", len(health_values))\n print(""Number of actions:"", len(actions))\n print(""Number of images:"", len(images))\n\n # Save only the first 1000 frames to avoid large file size\n imageio.mimsave(args.output_dir, images, fps=20)\n env.close()\n logger.info(f""GIF saved to {args.output_dir}"")\n \n return health_values\n\ndef make_array_records_dataset(agent, eval_env_args, num_episodes, split):\n """"""Generate a dataset by running the agent in the environment and saving the data as array record files.\n\n Args:\n agent: The trained PPO agent.\n output_dir (str): Directory to save the array record files.\n eval_env_args (dict): Arguments for the evaluation environment.\n num_episodes (int): Number of episodes to run.\n """"""\n # Set frame_skip to 1 to capture all frames\n eval_env_args['frame_skip'] = 1\n env = dummy_vec_env_with_bots_curriculum(1, **eval_env_args)\n\n current_action = None\n episode_idx = 0\n episode_metadata = []\n obs_chunks = []\n act_chunks = []\n file_idx = 0\n output_dir_split = os.path.join(args.output_dir, split)\n os.makedirs(output_dir_split, exist_ok=True)\n\n while episode_idx < num_episodes:\n obs = env.reset()\n done = False\n frame_counter = 0\n\n observations_seq = []\n actions_seq = []\n health_values_seq = []\n episode_obs_chunks = []\n episode_act_chunks = []\n\n # --- Run episode ---\n while not done and frame_counter < args.max_episode_length:\n screen = env.venv.envs[0].game.get_state().screen_buffer\n health = env.venv.envs[0].game.get_game_variable(GameVariable.HEALTH)\n if frame_counter % ACTION_REPEAT == 0:\n current_action, _ = agent.predict(obs)\n\n obs, _, done, _ = env.step(current_action)\n\n observations_seq.append(screen)\n actions_seq.append(int(current_action.item()))\n health_values_seq.append(int(health))\n\n if len(observations_seq) == args.chunk_size:\n episode_obs_chunks.append(observations_seq)\n episode_act_chunks.append(actions_seq)\n observations_seq = []\n actions_seq = []\n\n frame_counter += 1\n\n # --- Save episode ---\n if frame_counter >= args.min_episode_length:\n if observations_seq:\n if len(observations_seq) < args.chunk_size:\n print(\n f""Warning: Inconsistent chunk_sizes. Episode has {len(observations_seq)} frames, ""\n f""which is smaller than the requested chunk_size: {args.chunk_size}. ""\n ""This might lead to performance degradation during training.""\n )\n episode_obs_chunks.append(observations_seq)\n episode_act_chunks.append(actions_seq)\n\n obs_chunks_data = [\n np.stack(seq).astype(np.uint8)\n for seq in episode_obs_chunks\n ]\n act_chunks_data = [\n np.stack(act).astype(np.uint8) for act in episode_act_chunks\n ]\n obs_chunks.extend(obs_chunks_data)\n act_chunks.extend(act_chunks_data)\n\n ep_metadata, file_idx, obs_chunks, act_chunks = save_chunks(\n file_idx, args.chunks_per_file, output_dir_split, obs_chunks, act_chunks\n )\n episode_metadata.extend(ep_metadata)\n\n print(f""Episode {episode_idx} completed, length: {frame_counter}."")\n episode_idx += 1\n else:\n print(f""Episode too short ({frame_counter}), resampling..."")\n env.close()\n\ndef main():\n scenario = ""deathmatch_simple""\n\n env_args = {\n ""scenario"": scenario,\n ""frame_skip"": 1,\n ""frame_processor"": envs.default_frame_processor,\n ""n_bots"": 8,\n ""shaping"": True,\n ""initial_level"": 5,\n ""max_level"": 5,\n ""rolling_mean_length"": 10,\n }\n\n eval_env_args = dict(env_args)\n new_env = dummy_vec_env_with_bots_curriculum(1, **env_args)\n agent = envs.load_model(\n args.agent_path,\n new_env,\n )\n\n if args.generate_gif:\n make_gif(agent, eval_env_args)\n else:\n make_array_records_dataset(agent, num_episodes=args.num_episodes_train, eval_env_args=eval_env_args, split=""train"")\n make_array_records_dataset(agent, num_episodes=args.num_episodes_val, eval_env_args=eval_env_args, split=""val"")\n make_array_records_dataset(agent, num_episodes=args.num_episodes_test, eval_env_args=eval_env_args, split=""test"")\n\nif __name__ == ""__main__"":\n main()\n",python,tab +69,226400,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",0,0,"""""""\nGenerates a dataset of random-action CoinRun episodes.\nEpisodes are saved individually as memory-mapped files for efficient loading.\n""""""\n\nfrom dataclasses import dataclass\n\nfrom gym3 import types_np\nimport numpy as np\nfrom procgen import ProcgenGym3Env\nimport tyro\nimport json\nimport os\nfrom jasmine_data.utils import save_chunks\n\n\n@dataclass\nclass Args:\n num_episodes_train: int = 10000\n num_episodes_val: int = 500\n num_episodes_test: int = 500\n output_dir: str = ""data/coinrun_episodes""\n min_episode_length: int = 1000\n max_episode_length: int = 1000\n chunk_size: int = 160\n chunks_per_file: int = 100\n seed: int = 0\n\n\nargs = tyro.cli(Args)\nassert (\n args.max_episode_length >= args.min_episode_length\n), ""Maximum episode length must be greater than or equal to minimum episode length.""\n\nif args.min_episode_length < args.chunk_size:\n print(\n ""Warning: Minimum episode length is smaller than chunk size. Note that episodes shorter than the chunk size will be discarded.""\n )\n\n\n# --- Generate episodes ---\ndef generate_episodes(num_episodes, split):\n episode_idx = 0\n episode_metadata = []\n obs_chunks = []\n act_chunks = []\n file_idx = 0\n output_dir_split = os.path.join(args.output_dir, split)\n while episode_idx < num_episodes:\n seed = np.random.randint(0, 10000)\n env = ProcgenGym3Env(num=1, env_name=""coinrun"", start_level=seed)\n\n observations_seq = []\n actions_seq = []\n episode_obs_chunks = []\n episode_act_chunks = []\n\n # --- Run episode ---\n step_t = 0\n first_obs = True\n for step_t in range(args.max_episode_length):\n _, obs, first = env.observe()\n action = types_np.sample(env.ac_space, bshape=(env.num,))\n env.act(action)\n observations_seq.append(obs[""rgb""])\n actions_seq.append(action)\n if len(observations_seq) == args.chunk_size:\n episode_obs_chunks.append(observations_seq)\n episode_act_chunks.append(actions_seq)\n observations_seq = []\n actions_seq = []\n if first and not first_obs:\n break\n first_obs = False\n\n # --- Save episode ---\n if step_t + 1 >= args.min_episode_length:\n if observations_seq:\n if len(observations_seq) < args.chunk_size:\n print(\n f""Warning: Inconsistent chunk_sizes. Episode has {len(observations_seq)} frames, ""\n f""which is smaller than the requested chunk_size: {args.chunk_size}. ""\n ""This might lead to performance degradation during training.""\n )\n episode_obs_chunks.append(observations_seq)\n episode_act_chunks.append(actions_seq)\n\n obs_chunks_data = [\n np.concatenate(seq, axis=0).astype(np.uint8)\n for seq in episode_obs_chunks\n ]\n act_chunks_data = [\n np.concatenate(act, axis=0) for act in episode_act_chunks\n ]\n obs_chunks.extend(obs_chunks_data)\n act_chunks.extend(act_chunks_data)\n\n ep_metadata, file_idx, obs_chunks, act_chunks = save_chunks(\n file_idx, args.chunks_per_file, output_dir_split, obs_chunks, act_chunks\n )\n episode_metadata.extend(ep_metadata)\n\n print(f""Episode {episode_idx} completed, length: {step_t + 1}."")\n episode_idx += 1\n else:\n print(f""Episode too short ({step_t + 1}), resampling..."")\n\n if len(obs_chunks) > 0:\n print(\n f""Warning: Dropping {len(obs_chunks)} chunks for consistent number of chunks per file."",\n ""Consider changing the chunk_size and chunks_per_file parameters to prevent data-loss."",\n )\n\n print(f""Done generating {split} split"")\n return episode_metadata\n\n\ndef get_action_space():\n env = ProcgenGym3Env(num=1, env_name=""coinrun"", start_level=0)\n return env.ac_space.eltype.n\n\n\ndef main():\n # Set random seed and create dataset directories\n np.random.seed(args.seed)\n # --- Generate episodes ---\n train_episode_metadata = generate_episodes(args.num_episodes_train, ""train"")\n val_episode_metadata = generate_episodes(args.num_episodes_val, ""val"")\n test_episode_metadata = generate_episodes(args.num_episodes_test, ""test"")\n\n # --- Save metadata ---\n metadata = {\n ""env"": ""coinrun"",\n ""num_actions"": get_action_space(),\n ""num_episodes_train"": args.num_episodes_train,\n ""num_episodes_val"": args.num_episodes_val,\n ""num_episodes_test"": args.num_episodes_test,\n ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")\n\n\nif __name__ == ""__main__"":\n main()\n",python,tab +70,247266,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",0,0,"",python,tab +71,247268,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8034,0,"",python,selection_mouse +72,247900,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8026,16,"episode_metadata",python,selection_mouse +73,257800,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8275,0,"",python,selection_mouse +74,257808,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8274,0,"",python,selection_command +75,258447,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8275,0,"\n ",python,content +76,259213,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8280,0,"r",python,content +77,259215,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8281,0,"",python,selection_keyboard +78,259376,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8281,0,"e",python,content +79,259377,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8282,0,"",python,selection_keyboard +80,259462,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8282,0,"t",python,content +81,259464,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8283,0,"",python,selection_keyboard +82,259607,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8283,0,"u",python,content +83,259609,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8284,0,"",python,selection_keyboard +84,259732,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8284,0,"r",python,content +85,259734,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8285,0,"",python,selection_keyboard +86,259850,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8285,0,"n",python,content +87,259852,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8286,0,"",python,selection_keyboard +88,259911,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8286,0," ",python,content +89,259913,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8287,0,"",python,selection_keyboard +90,260111,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8287,0,"e",python,content +91,260112,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8288,0,"",python,selection_keyboard +92,260190,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8288,0,"p",python,content +93,260192,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8289,0,"",python,selection_keyboard +94,261305,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8289,0,"i",python,content +95,261307,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8290,0,"",python,selection_keyboard +96,261491,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8290,0,"s",python,content +97,261492,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8291,0,"",python,selection_keyboard +98,261569,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8291,0,"o",python,content +99,261570,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8292,0,"",python,selection_keyboard +100,261696,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8292,0,"d",python,content +101,261698,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8293,0,"",python,selection_keyboard +102,261809,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8293,0,"e",python,content +103,261811,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8294,0,"",python,selection_keyboard +104,262037,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8294,0,"_",python,content +105,262038,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8295,0,"",python,selection_keyboard +106,262335,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8295,0,"m",python,content +107,262336,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8296,0,"",python,selection_keyboard +108,262463,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8296,0,"e",python,content +109,262464,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8297,0,"",python,selection_keyboard +110,262963,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8287,10,"episode_metadata",python,content +111,266314,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8882,0,"",python,selection_mouse +112,268313,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8882,0,"m",python,content +113,268314,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8883,0,"",python,selection_keyboard +114,268717,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8883,0,"e",python,content +115,268717,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8884,0,"",python,selection_keyboard +116,268878,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8884,0,"t",python,content +117,268879,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8885,0,"",python,selection_keyboard +118,269057,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8885,0,"a",python,content +119,269058,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8886,0,"",python,selection_keyboard +120,269659,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8886,0,"_",python,content +121,269660,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8887,0,"",python,selection_keyboard +122,269990,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8887,0,"d",python,content +123,269991,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8888,0,"",python,selection_keyboard +124,270254,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8888,0,"a",python,content +125,270255,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8889,0,"",python,selection_keyboard +126,270286,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8889,0,"t",python,content +127,270287,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8890,0,"",python,selection_keyboard +128,270416,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8890,0,"a",python,content +129,270417,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8891,0,"",python,selection_keyboard +130,271152,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8882,9,"",python,content +131,280455,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",0,0,"",python,tab +132,280456,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4228,0,"",python,selection_mouse +133,281104,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4227,0,"",python,selection_mouse +134,288820,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",0,0,"",python,tab +135,290861,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8882,0,"t",python,content +136,290863,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8883,0,"",python,selection_keyboard +137,290976,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8883,0,"r",python,content +138,290977,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8884,0,"",python,selection_keyboard +139,291200,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8884,0,"a",python,content +140,291201,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8885,0,"",python,selection_keyboard +141,291229,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8885,0,"i",python,content +142,291230,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8886,0,"",python,selection_keyboard +143,291354,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8886,0,"n",python,content +144,291355,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8887,0,"",python,selection_keyboard +145,291774,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8887,0,"_",python,content +146,291775,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8888,0,"",python,selection_keyboard +147,292086,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8888,0,"e",python,content +148,292087,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8889,0,"",python,selection_keyboard +149,292173,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8889,0,"p",python,content +150,292174,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8890,0,"",python,selection_keyboard +151,292388,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8890,0,"i",python,content +152,292389,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8891,0,"",python,selection_keyboard +153,292475,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8891,0,"s",python,content +154,292477,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8892,0,"",python,selection_keyboard +155,292606,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8892,0,"o",python,content +156,292607,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8893,0,"",python,selection_keyboard +157,292666,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8893,0,"d",python,content +158,292667,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8894,0,"",python,selection_keyboard +159,292774,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8894,0,"e",python,content +160,292775,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8895,0,"",python,selection_keyboard +161,292934,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8895,0,"_",python,content +162,292935,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8896,0,"",python,selection_keyboard +163,293223,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8896,0,",",python,content +164,293224,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8897,0,"",python,selection_keyboard +165,293368,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8897,0,"e",python,content +166,293369,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8898,0,"",python,selection_keyboard +167,293767,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8897,1,"",python,content +168,293902,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8896,1,"",python,content +169,294084,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8896,0,"m",python,content +170,294085,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8897,0,"",python,selection_keyboard +171,294198,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8897,0,"e",python,content +172,294199,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8898,0,"",python,selection_keyboard +173,294440,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8898,0,"t",python,content +174,294441,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8899,0,"",python,selection_keyboard +175,294796,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8899,0,"a",python,content +176,294797,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8900,0,"",python,selection_keyboard +177,295210,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8900,0,"d",python,content +178,295211,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8901,0,"",python,selection_keyboard +179,295387,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8901,0,"a",python,content +180,295388,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8902,0,"",python,selection_keyboard +181,295433,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8902,0,"t",python,content +182,295434,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8903,0,"",python,selection_keyboard +183,295589,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8903,0,"a",python,content +184,295590,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8904,0,"",python,selection_keyboard +185,296672,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8904,0," ",python,content +186,296673,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8905,0,"",python,selection_keyboard +187,296990,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8905,0,"=",python,content +188,296991,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8906,0,"",python,selection_keyboard +189,297085,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8906,0," ",python,content +190,297088,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8907,0,"",python,selection_keyboard +191,297762,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8906,0,"",python,selection_command +192,297877,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9055,0,"",python,selection_command +193,298217,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9031,0,"",python,selection_command +194,299776,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9031,0,"v",python,content +195,299777,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9032,0,"",python,selection_keyboard +196,299960,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9032,0,"a",python,content +197,299961,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9033,0,"",python,selection_keyboard +198,300294,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9033,0,"l",python,content +199,300295,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9034,0,"",python,selection_keyboard +200,300591,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9034,0,"_",python,content +201,300592,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9035,0,"",python,selection_keyboard +202,300884,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9035,0,"e",python,content +203,300885,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9036,0,"",python,selection_keyboard +204,300970,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9036,0,"p",python,content +205,300971,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9037,0,"",python,selection_keyboard +206,301176,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9037,0,"i",python,content +207,301177,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9038,0,"",python,selection_keyboard +208,301278,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9038,0,"s",python,content +209,301279,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9039,0,"",python,selection_keyboard +210,301385,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9039,0,"o",python,content +211,301386,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9040,0,"",python,selection_keyboard +212,301487,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9040,0,"d",python,content +213,301488,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9041,0,"",python,selection_keyboard +214,301589,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9041,0,"e",python,content +215,301591,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9042,0,"",python,selection_keyboard +216,301833,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9042,0,"_",python,content +217,301834,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9043,0,"",python,selection_keyboard +218,302130,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9043,0,"m",python,content +219,302131,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9044,0,"",python,selection_keyboard +220,302295,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9044,0,"e",python,content +221,302296,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9045,0,"",python,selection_keyboard +222,302539,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9045,0,"t",python,content +223,302540,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9046,0,"",python,selection_keyboard +224,302748,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9046,0,"a",python,content +225,302749,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9047,0,"",python,selection_keyboard +226,303002,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9047,0,"d",python,content +227,303003,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9048,0,"",python,selection_keyboard +228,303136,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9048,0,"a",python,content +229,303137,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9049,0,"",python,selection_keyboard +230,303215,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9049,0,"t",python,content +231,303216,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9050,0,"",python,selection_keyboard +232,303317,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9050,0,"a",python,content +233,303318,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9051,0,"",python,selection_keyboard +234,303761,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9051,0," ",python,content +235,303762,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9052,0,"",python,selection_keyboard +236,303932,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9052,0,"=",python,content +237,303933,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9053,0,"",python,selection_keyboard +238,304022,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9053,0," ",python,content +239,304023,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9054,0,"",python,selection_keyboard +240,304630,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9053,0,"",python,selection_command +241,304746,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9196,0,"",python,selection_command +242,305197,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9174,0,"",python,selection_command +243,305844,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9174,0,"t",python,content +244,305845,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9175,0,"",python,selection_keyboard +245,306072,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9175,0,"e",python,content +246,306073,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9176,0,"",python,selection_keyboard +247,306253,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9176,0,"s",python,content +248,306254,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9177,0,"",python,selection_keyboard +249,306347,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9177,0,"t",python,content +250,306348,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9178,0,"",python,selection_keyboard +251,307788,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9178,0,"_",python,content +252,307789,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9179,0,"",python,selection_keyboard +253,308123,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9179,0,"e",python,content +254,308124,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9180,0,"",python,selection_keyboard +255,308250,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9180,0,"p",python,content +256,308251,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9181,0,"",python,selection_keyboard +257,308434,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9181,0,"s",python,content +258,308435,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9182,0,"",python,selection_keyboard +259,308467,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9182,0,"i",python,content +260,308468,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9183,0,"",python,selection_keyboard +261,309031,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9182,1,"",python,content +262,309181,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9181,1,"",python,content +263,309558,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9181,0,"i",python,content +264,309558,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9182,0,"",python,selection_keyboard +265,309663,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9182,0,"s",python,content +266,309664,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9183,0,"",python,selection_keyboard +267,309771,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9183,0,"o",python,content +268,309772,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9184,0,"",python,selection_keyboard +269,309887,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9184,0,"d",python,content +270,309888,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9185,0,"",python,selection_keyboard +271,310002,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9185,0,"e",python,content +272,310002,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9186,0,"",python,selection_keyboard +273,310234,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9186,0,"_",python,content +274,310235,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9187,0,"",python,selection_keyboard +275,310594,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9187,0,"m",python,content +276,310595,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9188,0,"",python,selection_keyboard +277,310764,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9188,0,"e",python,content +278,310764,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9189,0,"",python,selection_keyboard +279,310931,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9189,0,"t",python,content +280,310932,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9190,0,"",python,selection_keyboard +281,310972,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9190,0,"a",python,content +282,310973,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9191,0,"",python,selection_keyboard +283,311151,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9191,0,"d",python,content +284,311152,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9192,0,"",python,selection_keyboard +285,311286,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9192,0,"a",python,content +286,311287,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9193,0,"",python,selection_keyboard +287,311359,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9193,0,"t",python,content +288,311360,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9194,0,"",python,selection_keyboard +289,311463,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9194,0,"a",python,content +290,311464,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9195,0,"",python,selection_keyboard +291,311734,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9195,0," ",python,content +292,311735,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9196,0,"",python,selection_keyboard +293,311953,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9196,0,"=",python,content +294,311954,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9197,0,"",python,selection_keyboard +295,311999,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9197,0," ",python,content +296,312000,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9198,0,"",python,selection_keyboard +297,312628,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9197,0,"",python,selection_command +298,314589,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9050,0,"",python,selection_mouse +299,316800,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",0,0,"",python,tab +300,316801,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5397,0,"",python,selection_mouse +301,317210,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5381,0,"",python,selection_mouse +302,317737,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5358,38," print(f""Done generating dataset."")",python,selection_command +303,317937,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5357,39,"\n print(f""Done generating dataset."")",python,selection_command +304,318461,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5326,70," json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +305,318477,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5253,143," with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +306,318520,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5247,149," }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +307,318543,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5191,205," ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +308,318578,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5137,259," ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +309,318763,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5079,317," ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +310,318764,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5068,328," ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +311,318764,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5004,392," [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +312,318765,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4963,433," ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +313,318765,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4952,444," ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +314,318765,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4889,507," [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +315,318766,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4849,547," ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +316,318787,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4838,558," ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +317,318827,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4773,623," [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +318,318851,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4731,665," ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +319,318892,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4678,718," ""num_episodes_test"": args.num_episodes_test,\n ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +320,318936,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4627,769," ""num_episodes_val"": args.num_episodes_val,\n ""num_episodes_test"": args.num_episodes_test,\n ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +321,318963,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4572,824," ""num_episodes_train"": args.num_episodes_train,\n ""num_episodes_val"": args.num_episodes_val,\n ""num_episodes_test"": args.num_episodes_test,\n ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +322,318979,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4529,867," ""num_actions"": get_action_space(),\n ""num_episodes_train"": args.num_episodes_train,\n ""num_episodes_val"": args.num_episodes_val,\n ""num_episodes_test"": args.num_episodes_test,\n ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +323,319307,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4503,893," ""env"": ""coinrun"",\n ""num_actions"": get_action_space(),\n ""num_episodes_train"": args.num_episodes_train,\n ""num_episodes_val"": args.num_episodes_val,\n ""num_episodes_test"": args.num_episodes_test,\n ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +324,319475,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4486,910," metadata = {\n ""env"": ""coinrun"",\n ""num_actions"": get_action_space(),\n ""num_episodes_train"": args.num_episodes_train,\n ""num_episodes_val"": args.num_episodes_val,\n ""num_episodes_test"": args.num_episodes_test,\n ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +325,319611,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4458,938," # --- Save metadata ---\n metadata = {\n ""env"": ""coinrun"",\n ""num_actions"": get_action_space(),\n ""num_episodes_train"": args.num_episodes_train,\n ""num_episodes_val"": args.num_episodes_val,\n ""num_episodes_test"": args.num_episodes_test,\n ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,selection_command +326,319819,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",4458,0,"",python,selection_command +327,321737,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",0,0,"",python,tab +328,321738,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9181,0,"",python,selection_mouse +329,322358,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9311,0,"\n # --- Save metadata ---\n metadata = {\n ""env"": ""coinrun"",\n ""num_actions"": get_action_space(),\n ""num_episodes_train"": args.num_episodes_train,\n ""num_episodes_val"": args.num_episodes_val,\n ""num_episodes_test"": args.num_episodes_test,\n ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")",python,content +330,322398,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9316,0,"",python,selection_command +331,324957,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9312,27," # --- Save metadata ---",python,selection_command +332,326121,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9316,0,"",python,selection_command +333,326312,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9170,0,"",python,selection_command +334,326471,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9027,0,"",python,selection_command +335,326611,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8878,0,"",python,selection_command +336,327130,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8874,148," train_episode_metadata = make_array_records_dataset(agent, num_episodes=args.num_episodes_train, eval_env_args=eval_env_args, split=""train"")",python,selection_command +337,327320,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8874,291," train_episode_metadata = make_array_records_dataset(agent, num_episodes=args.num_episodes_train, eval_env_args=eval_env_args, split=""train"")\n val_episode_metadata = make_array_records_dataset(agent, num_episodes=args.num_episodes_val, eval_env_args=eval_env_args, split=""val"")",python,selection_command +338,327472,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8874,437," train_episode_metadata = make_array_records_dataset(agent, num_episodes=args.num_episodes_train, eval_env_args=eval_env_args, split=""train"")\n val_episode_metadata = make_array_records_dataset(agent, num_episodes=args.num_episodes_val, eval_env_args=eval_env_args, split=""val"")\n test_episode_metadata = make_array_records_dataset(agent, num_episodes=args.num_episodes_test, eval_env_args=eval_env_args, split=""test"")",python,selection_command +339,327740,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8882,0,"",python,selection_command +340,327973,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9170,4,"",python,content +341,327973,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9027,4,"",python,content +342,327973,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8878,4,"",python,content +343,328140,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8877,0,"",python,selection_command +344,328341,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8867,0,"",python,selection_command +345,329039,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8828,0,"",python,selection_command +346,329424,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8867,0,"",python,selection_command +347,329782,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8864,10,"",python,content +348,329794,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8868,0,"",python,selection_command +349,329863,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8829,0,"",python,selection_command +350,330401,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8863,0,"\n ",python,content +351,331047,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8872,0,"r",python,content +352,331048,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8873,0,"",python,selection_keyboard +353,331215,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8873,0,"e",python,content +354,331216,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8874,0,"",python,selection_keyboard +355,331259,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8874,0,"t",python,content +356,331260,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8875,0,"",python,selection_keyboard +357,331460,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8875,0,"u",python,content +358,331461,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8876,0,"",python,selection_keyboard +359,331580,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8876,0,"r",python,content +360,331581,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8877,0,"",python,selection_keyboard +361,331605,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8877,0,"n",python,content +362,331606,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8878,0,"",python,selection_keyboard +363,331757,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8878,0," ",python,content +364,331758,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8879,0,"",python,selection_keyboard +365,332297,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8879,0,"\n ",python,content +366,332584,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8880,8,"",python,content +367,333586,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9452,0,"",python,selection_mouse +368,334163,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9377,0,"",python,selection_mouse +369,334176,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9376,0,"",python,selection_command +370,334757,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9437,0,"",python,selection_mouse +371,335294,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9398,0,"",python,selection_mouse +372,340264,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9406,0,"",python,selection_mouse +373,340379,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9401,16,"get_action_space",python,selection_mouse +374,342761,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9406,0,"",python,selection_mouse +375,344134,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8632,0,"",python,selection_mouse +376,344254,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8626,13,"eval_env_args",python,selection_mouse +377,345689,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8742,0,"",python,selection_mouse +378,349591,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",8440,0,"",python,selection_mouse +379,352619,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9397,0,"",python,selection_mouse +380,354935,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9400,0,"",python,selection_mouse +381,355700,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9400,1,"",python,content +382,356073,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9400,16,"",python,content +383,356816,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9400,3,"",python,content +384,356857,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9399,0,"",python,selection_command +385,357719,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9400,0,"",python,selection_command +386,358063,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9400,0," ",python,content +387,358064,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9401,0,"",python,selection_keyboard +388,358365,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9401,0,"0",python,content +389,358365,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9402,0,"",python,selection_keyboard +390,358584,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9402,0," ",python,content +391,358585,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9403,0,"",python,selection_keyboard +392,359239,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9402,1,"",python,content +393,359510,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9402,0,",",python,content +394,359511,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9403,0,"",python,selection_keyboard +395,359601,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9403,0," ",python,content +396,359602,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9404,0,"",python,selection_keyboard +397,360310,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9404,0,"#",python,content +398,360311,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9405,0,"",python,selection_keyboard +399,360427,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9405,0," ",python,content +400,360428,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9406,0,"",python,selection_keyboard +401,361271,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9406,0,"T",python,content +402,361272,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9407,0,"",python,selection_keyboard +403,361468,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9407,0,"O",python,content +404,361469,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9408,0,"",python,selection_keyboard +405,361554,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9408,0,"D",python,content +406,361555,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9409,0,"",python,selection_keyboard +407,361621,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9409,0,"O",python,content +408,361623,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9410,0,"",python,selection_keyboard +409,361729,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9410,0," ",python,content +410,361729,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9411,0,"",python,selection_keyboard +411,361882,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9411,0,"m",python,content +412,361883,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9412,0,"",python,selection_keyboard +413,362044,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9412,0,"i",python,content +414,362045,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9413,0,"",python,selection_keyboard +415,362143,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9413,0,"h",python,content +416,362144,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9414,0,"",python,selection_keyboard +417,362235,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9414,0,"i",python,content +418,362236,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9415,0,"",python,selection_keyboard +419,362317,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9415,0,"r",python,content +420,362318,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9416,0,"",python,selection_keyboard +421,362820,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9415,0,"",python,selection_command +422,364836,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",0,0,"# # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # \n# /$$ /$$ /$$ /$$$$$$$ /$$$$$$$ /$$$$$$$ /$$$$$$ #\n# | $$ | $$|__/ | $$__ $$ | $$__ $$| $$__ $$ /$$__ $$ #\n# | $$ | $$ /$$ /$$$$$$$$| $$ \ $$ /$$$$$$ /$$$$$$ /$$$$$$/$$$$ | $$ \ $$| $$ \ $$| $$ \ $$ #\n# | $$ / $$/| $$|____ /$$/| $$ | $$ /$$__ $$ /$$__ $$| $$_ $$_ $$ | $$$$$$$/| $$$$$$$/| $$ | $$ #\n# \ $$ $$/ | $$ /$$$$/ | $$ | $$| $$ \ $$| $$ \ $$| $$ \ $$ \ $$ | $$____/ | $$____/ | $$ | $$ #\n# \ $$$/ | $$ /$$__/ | $$ | $$| $$ | $$| $$ | $$| $$ | $$ | $$ | $$ | $$ | $$ | $$ #\n# \ $/ | $$ /$$$$$$$$| $$$$$$$/| $$$$$$/| $$$$$$/| $$ | $$ | $$ | $$ | $$ | $$$$$$/ #\n# \_/ |__/|________/|_______/ \______/ \______/ |__/ |__/ |__/ |__/ |__/ \______/ #\n# #\n# # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # \n \n# FORK OF LEANDRO KIELIGER'S DOOM PPO TUTORIAL: https://lkieliger.medium.com/deep-reinforcement-learning-in-practice-by-playing-doom-part-1-getting-started-618c99075c77 \n\n# SCRIPT TO TRAIN A PPO AGENT ON THE ViZDOOM DEATHMATCH ENVIRONMENT. \n\nimport numpy as np\nimport os\nimport vizdoom\nfrom common import envs\nfrom common.models import CustomCNN\nfrom stable_baselines3.common.vec_env import VecTransposeImage, SubprocVecEnv\nimport random\nimport string\nfrom vizdoom.vizdoom import GameVariable\nfrom collections import deque\nimport torch\n\ndevice = torch.device(""cuda"" if torch.cuda.is_available() else ""cpu"")\n# device=torch.device(""mps"")\nprint(f""Using device: {device}"")\n\n# Rewards\n# 1 per kill\nreward_factor_frag = 1.0\nreward_factor_damage = 0.01\n\n# Player can move at ~16.66 units per tick\nreward_factor_distance = 5e-4\npenalty_factor_distance = -2.5e-3\nreward_threshold_distance = 3.0\n\n# Pistol clips have 10 bullets\nreward_factor_ammo_increment = 0.02\nreward_factor_ammo_decrement = -0.01\n\n# Player starts at 100 health\nreward_factor_health_increment = 0.02\nreward_factor_health_decrement = -0.01\nreward_factor_armor_increment = 0.01\n\n\n# List of game variables storing ammunition information. Used for keeping track of ammunition-related rewards.\nAMMO_VARIABLES = [\n GameVariable.AMMO0,\n GameVariable.AMMO1,\n GameVariable.AMMO2,\n GameVariable.AMMO3,\n GameVariable.AMMO4,\n GameVariable.AMMO5,\n GameVariable.AMMO6,\n GameVariable.AMMO7,\n GameVariable.AMMO8,\n GameVariable.AMMO9,\n]\n\n# List of game variables storing weapon information. Used for keeping track of ammunition-related rewards.\nWEAPON_VARIABLES = [\n GameVariable.WEAPON0,\n GameVariable.WEAPON1,\n GameVariable.WEAPON2,\n GameVariable.WEAPON3,\n GameVariable.WEAPON4,\n GameVariable.WEAPON5,\n GameVariable.WEAPON6,\n GameVariable.WEAPON7,\n GameVariable.WEAPON8,\n GameVariable.WEAPON9,\n]\n\n\nclass DoomWithBotsShaped(envs.DoomWithBots):\n """"""An environment wrapper for a Doom deathmatch game with bots.\n\n Rewards are shaped according to the multipliers defined in the notebook.\n """"""\n\n def __init__(self, game, frame_processor, frame_skip, n_bots, shaping):\n super().__init__(game, frame_processor, frame_skip, n_bots)\n\n # Give a random two-letter name to the agent for identifying instances in parallel learning.\n self.name = """".join(random.choices(string.ascii_uppercase + string.digits, k=2))\n self.shaping = shaping\n\n # Internal states\n self.last_health = 100\n self.last_x, self.last_y = self._get_player_pos()\n self.ammo_state = self._get_ammo_state()\n self.weapon_state = self._get_weapon_state()\n self.total_rew = self.last_damage_dealt = self.deaths = self.last_frags = (\n self.last_armor\n ) = 0\n\n # Store individual reward contributions for logging purposes\n self.rewards_stats = {\n ""frag"": 0,\n ""damage"": 0,\n ""ammo"": 0,\n ""health"": 0,\n ""armor"": 0,\n ""distance"": 0,\n }\n\n def step(self, action, array=False):\n # Perform the action as usual\n state, reward, done, info = super().step(action)\n\n self._log_reward_stat(""frag"", reward)\n\n # Adjust the reward based on the shaping table\n if self.shaping:\n shaped_reward = reward + self.shape_rewards()\n else:\n shaped_reward = reward\n\n self.total_rew += shaped_reward\n\n return state, shaped_reward, done, info\n\n def reset(self):\n self._print_state()\n\n state = super().reset()\n\n self.last_health = 100\n self.last_x, self.last_y = self._get_player_pos()\n self.last_armor = self.last_frags = self.total_rew = self.deaths = 0\n\n # Damage count is not cleared when starting a new episode: https://github.com/mwydmuch/ViZDoom/issues/399\n # self.last_damage_dealt = 0\n\n # Reset reward stats\n for k in self.rewards_stats.keys():\n self.rewards_stats[k] = 0\n\n return state\n\n def seed(self, seed=None):\n if seed is not None:\n random.seed(seed)\n np.random.seed(seed)\n return [seed]\n\n def shape_rewards(self):\n reward_contributions = [\n self._compute_damage_reward(),\n self._compute_ammo_reward(),\n self._compute_health_reward(),\n self._compute_armor_reward(),\n self._compute_distance_reward(*self._get_player_pos()),\n ]\n\n return sum(reward_contributions)\n\n def _respawn_if_dead(self):\n if not self.game.is_episode_finished():\n # Check if player is dead\n if self.game.is_player_dead():\n self.deaths += 1\n self._reset_player()\n\n def _compute_distance_reward(self, x, y):\n """"""Computes a reward/penalty based on the distance travelled since last update.""""""\n dx = self.last_x - x\n dy = self.last_y - y\n\n distance = np.sqrt(dx**2 + dy**2)\n\n if distance - reward_threshold_distance > 0:\n reward = reward_factor_distance\n else:\n reward = -reward_factor_distance\n\n self.last_x = x\n self.last_y = y\n self._log_reward_stat(""distance"", reward)\n\n return reward\n\n def _compute_damage_reward(self):\n """"""Computes a reward based on total damage inflicted to enemies since last update.""""""\n damage_dealt = self.game.get_game_variable(GameVariable.DAMAGECOUNT)\n reward = reward_factor_damage * (damage_dealt - self.last_damage_dealt)\n\n self.last_damage_dealt = damage_dealt\n self._log_reward_stat(""damage"", reward)\n\n return reward\n\n def _compute_health_reward(self):\n """"""Computes a reward/penalty based on total health change since last update.""""""\n # When the player is dead, the health game variable can be -999900\n health = max(self.game.get_game_variable(GameVariable.HEALTH), 0)\n\n health_reward = reward_factor_health_increment * max(\n 0, health - self.last_health\n )\n health_penalty = reward_factor_health_decrement * min(\n 0, health - self.last_health\n )\n reward = health_reward - health_penalty\n\n self.last_health = health\n self._log_reward_stat(""health"", reward)\n\n return reward\n\n def _compute_armor_reward(self):\n """"""Computes a reward/penalty based on total armor change since last update.""""""\n armor = self.game.get_game_variable(GameVariable.ARMOR)\n reward = reward_factor_armor_increment * max(0, armor - self.last_armor)\n\n self.last_armor = armor\n self._log_reward_stat(""armor"", reward)\n\n return reward\n\n def _compute_ammo_reward(self):\n """"""Computes a reward/penalty based on total ammunition change since last update.""""""\n self.weapon_state = self._get_weapon_state()\n\n new_ammo_state = self._get_ammo_state()\n ammo_diffs = (new_ammo_state - self.ammo_state) * self.weapon_state\n ammo_reward = reward_factor_ammo_increment * max(0, np.sum(ammo_diffs))\n ammo_penalty = reward_factor_ammo_decrement * min(0, np.sum(ammo_diffs))\n reward = ammo_reward - ammo_penalty\n\n self.ammo_state = new_ammo_state\n self._log_reward_stat(""ammo"", reward)\n\n return reward\n\n def _get_player_pos(self):\n """"""Returns the player X- and Y- coordinates.""""""\n return self.game.get_game_variable(\n GameVariable.POSITION_X\n ), self.game.get_game_variable(GameVariable.POSITION_Y)\n\n def _get_ammo_state(self):\n """"""Returns the total available ammunition per weapon slot.""""""\n ammo_state = np.zeros(10)\n\n for i in range(10):\n ammo_state[i] = self.game.get_game_variable(AMMO_VARIABLES[i])\n\n return ammo_state\n\n def _get_weapon_state(self):\n """"""Returns which weapon slots can be used. Available weapons are encoded as ones.""""""\n weapon_state = np.zeros(10)\n\n for i in range(10):\n weapon_state[i] = self.game.get_game_variable(WEAPON_VARIABLES[i])\n\n return weapon_state\n\n def _log_reward_stat(self, kind: str, reward: float):\n self.rewards_stats[kind] += reward\n\n def _reset_player(self):\n self.last_health = 100\n self.last_armor = 0\n self.game.respawn_player()\n self.last_x, self.last_y = self._get_player_pos()\n self.ammo_state = self._get_ammo_state()\n\n def _print_state(self):\n super()._print_state()\n\n\nREWARD_THRESHOLDS = [5, 10, 15, 20, 25, 25]\n\n# Curriculum based learning\nclass DoomWithBotsCurriculum(DoomWithBotsShaped):\n def __init__(\n self,\n game,\n frame_processor,\n frame_skip,\n n_bots,\n shaping,\n initial_level=0,\n max_level=5,\n rolling_mean_length=10,\n ):\n super().__init__(game, frame_processor, frame_skip, n_bots, shaping)\n\n # Initialize ACS script difficulty level\n game.send_game_command(""pukename change_difficulty 0"")\n\n # Internal state\n self.level = initial_level\n self.max_level = max_level\n self.rolling_mean_length = rolling_mean_length\n self.last_rewards = deque(maxlen=rolling_mean_length)\n\n def step(self, action, array=False):\n # Perform action step as usual\n state, reward, done, infos = super().step(action, array)\n\n # After an episode, check whether difficulty should be increased.\n if done:\n self.last_rewards.append(self.total_rew)\n run_mean = np.mean(self.last_rewards)\n print(\n ""Avg. last 10 runs of {}: {:.2f}. Current difficulty level: {}"".format(\n self.name, run_mean, self.level\n )\n )\n if (\n run_mean > REWARD_THRESHOLDS[self.level]\n and len(self.last_rewards) >= self.rolling_mean_length\n ):\n self._change_difficulty()\n\n return state, reward, done, infos\n\n def reset(self):\n state = super().reset()\n self.game.send_game_command(f""pukename change_difficulty {self.level}"")\n\n return state\n\n def _change_difficulty(self):\n """"""Adjusts the difficulty by setting the difficulty level in the ACS script.""""""\n if self.level < self.max_level:\n self.level += 1\n print(f""Changing difficulty for {self.name} to {self.level}"")\n self.game.send_game_command(f""pukename change_difficulty {self.level}"")\n self.last_rewards = deque(maxlen=self.rolling_mean_length)\n else:\n print(f""{self.name} already at max level!"")\n\n\ndef game_instance(scenario):\n """"""Creates a Doom game instance.""""""\n game = vizdoom.DoomGame()\n dir_path = os.path.dirname(os.path.realpath(__file__))\n cfg_pth = os.path.join(dir_path, ""scenarios"", f""{scenario}.cfg"")\n game.load_config(cfg_pth)\n game.add_game_args(envs.DOOM_ENV_WITH_BOTS_ARGS)\n game.set_window_visible(False)\n game.init()\n\n return game\n\n\ndef env_with_bots_shaped(scenario, **kwargs) -> envs.DoomEnv:\n """"""Wraps a Doom game instance in an environment with shaped rewards.""""""\n game = game_instance(scenario)\n return DoomWithBotsShaped(game, **kwargs)\n\n\ndef vec_env_with_bots_shaped(n_envs=1, **kwargs) -> VecTransposeImage:\n """"""Wraps Doom game instances in a vectorized environment with shaped rewards using true parallelism.""""""\n return VecTransposeImage(\n SubprocVecEnv([lambda: env_with_bots_shaped(**kwargs) for _ in range(n_envs)])\n )\n\n\ndef env_with_bots_curriculum(scenario, **kwargs) -> envs.DoomEnv:\n """"""Wraps a Doom game instance in an environment with shaped rewards and curriculum.""""""\n game = game_instance(scenario)\n return DoomWithBotsCurriculum(game, **kwargs)\n\n\ndef vec_env_with_bots_curriculum(n_envs=1, **kwargs) -> VecTransposeImage:\n """"""Wraps a Doom game instance in a vectorized environment with shaped rewards and curriculum.""""""\n return VecTransposeImage(\n SubprocVecEnv([lambda: env_with_bots_curriculum(**kwargs) for _ in range(n_envs)])\n )\n\n\nif __name__ == ""__main__"":\n scenario = ""deathmatch_simple""\n\n # Agent parameters.\n agent_args = {\n ""n_steps"": 4096,\n ""learning_rate"": 1e-4,\n ""batch_size"": 64,\n ""policy_kwargs"": {""features_extractor_class"": CustomCNN},\n ""device"": device,\n }\n\n # Environment parameters.\n env_args = {\n ""scenario"": scenario,\n ""frame_skip"": 4,\n ""frame_processor"": envs.default_frame_processor,\n ""n_bots"": 6,\n ""shaping"": True,\n ""initial_level"": 1,\n ""max_level"": 5,\n }\n\n # In the evaluation environment we measure frags only.\n eval_env_args = dict(env_args)\n eval_env_args[""shaping""] = False\n\n n_envs = 32\n # Create environments with bots and shaping.\n env = vec_env_with_bots_curriculum(\n n_envs, **env_args\n ) # You can increase the number of parallel environments\n eval_env = vec_env_with_bots_curriculum(n_envs, **eval_env_args)\n\n agent = envs.solve_env(\n env,\n eval_env,\n scenario,\n agent_args,\n resume=False,\n )\n envs.save_model(agent, ""agent_test"")\n",python,tab +423,365585,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",4384,0,"",python,selection_mouse +424,369913,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",4637,0,"",python,selection_command +425,370881,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",4681,0,"",python,selection_command +426,371274,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",4746,0,"",python,selection_command +427,372422,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",10875,0,"",python,selection_command +428,374580,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",10915,0,"",python,selection_command +429,374933,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",10986,0,"",python,selection_command +430,375302,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",4637,0,"",python,selection_command +431,376517,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",10986,0,"",python,selection_command +432,377415,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",10915,0,"",python,selection_command +433,377773,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",10875,0,"",python,selection_command +434,378169,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",4746,0,"",python,selection_command +435,379103,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",4681,0,"",python,selection_command +436,379914,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",4637,0,"",python,selection_command +437,380293,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",10986,0,"",python,selection_command +438,381160,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",10915,0,"",python,selection_command +439,381393,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",10875,0,"",python,selection_command +440,381652,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",4746,0,"",python,selection_command +441,382190,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",4681,0,"",python,selection_command +442,382445,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",4637,0,"",python,selection_command +443,382653,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",10986,0,"",python,selection_command +444,406280,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",3486,0,"",python,selection_mouse +445,406602,"data/jasmine_data/ViZDoomPPO/common/envs.py",0,0,"",python,tab +446,406603,"data/jasmine_data/ViZDoomPPO/common/envs.py",3162,0,"",python,selection_command +447,419203,"data/jasmine_data/ViZDoomPPO/common/envs.py",5501,0,"",python,selection_mouse +448,419220,"data/jasmine_data/ViZDoomPPO/common/envs.py",5500,0,"",python,selection_command +449,421352,"data/jasmine_data/ViZDoomPPO/common/envs.py",5562,0,"",python,selection_mouse +450,421483,"data/jasmine_data/ViZDoomPPO/common/envs.py",5556,7,"DoomEnv",python,selection_mouse +451,437850,"data/jasmine_data/ViZDoomPPO/train_ppo_parallel.py",0,0,"",python,tab +452,440614,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",0,0,"",python,tab +453,442375,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",0,0,"",python,tab +454,442377,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5078,0,"",python,selection_mouse +455,442380,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",5077,0,"",python,selection_command +456,444053,"slurm/jobs/mihir/horeka/preprocessing/doom_chunked.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --gres=gpu:1\n#SBATCH --cpus-per-task=16\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/preprocess/doom/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/preprocess/doom/%x_%j.log\n#SBATCH --job-name=preprocess_coinrun_chunked\n\n# export PYTHONUNBUFFERED=1\n\n\n# source .venv/bin/activate\n\nsource .venv/bin/activate\n\npython jasmine_data/ViZDoomPPO/load_model_generate_dataset.py \\n --output_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_doom/doom_episodes_50m \\n --agent_path /home/hk-project-p0023960/tum_cte0515/Projects/jasmine/data/jasmine_data/ViZDoomPPO/logs/models/deathmatch_simple_bak/best_model.zip \\n --num_episodes_train 50_000 \\n \n\n",shellscript,tab +457,461940,"TERMINAL",0,0,"bash",,terminal_focus +458,463242,"TERMINAL",0,0,"python",,terminal_command +459,463289,"TERMINAL",0,0,"]633;C",,terminal_output +460,463354,"TERMINAL",0,0,"Python 3.9.18 (main, Sep 4 2025, 00:00:00) \r\n[GCC 11.4.1 20231218 (Red Hat 11.4.1-3)] on linux\r\nType ""help"", ""copyright"", ""credits"" or ""license"" for more information.\r\n>>> ",,terminal_output +461,472291,"TERMINAL",0,0,"1",,terminal_output +462,472388,"TERMINAL",0,0,"0",,terminal_output +463,472735,"TERMINAL",0,0,"0",,terminal_output +464,472938,"TERMINAL",0,0,"0",,terminal_output +465,473540,"TERMINAL",0,0,"",,terminal_output +466,473761,"TERMINAL",0,0,"",,terminal_output +467,473857,"TERMINAL",0,0,"",,terminal_output +468,473961,"TERMINAL",0,0,"",,terminal_output +469,488292,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",0,0,"",python,tab +470,488293,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9807,0,"",python,selection_mouse +471,488334,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",9806,0,"",python,selection_command +472,488893,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",10281,0,"",python,selection_command +473,489639,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",0,0,"",python,selection_command +474,489983,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",114,0,"",python,selection_command +475,490205,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",228,0,"",python,selection_command +476,490691,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",342,0,"",python,selection_command +477,490713,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",456,0,"",python,selection_command +478,490758,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",570,0,"",python,selection_command +479,490801,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",684,0,"",python,selection_command +480,490843,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",798,0,"",python,selection_command +481,490844,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",912,0,"",python,selection_command +482,490878,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1026,0,"",python,selection_command +483,490923,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1140,0,"",python,selection_command +484,490963,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1350,0,"",python,selection_command +485,490964,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1456,0,"",python,selection_command +486,491080,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1728,0,"",python,selection_command +487,491239,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1729,0,"",python,selection_command +488,491394,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1799,0,"",python,selection_command +489,491530,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1800,0,"",python,selection_command +490,491660,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1834,0,"",python,selection_command +491,491952,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1849,0,"",python,selection_command +492,492132,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1873,0,"",python,selection_command +493,492345,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1885,0,"\n",python,content +494,492561,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1886,0,"i",python,content +495,492562,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1887,0,"",python,selection_keyboard +496,492639,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1887,0,"m",python,content +497,492641,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1888,0,"",python,selection_keyboard +498,492823,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1888,0,"p",python,content +499,492825,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1889,0,"",python,selection_keyboard +500,492867,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1889,0,"o",python,content +501,492868,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1890,0,"",python,selection_keyboard +502,493034,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1890,0,"r",python,content +503,493035,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1891,0,"",python,selection_keyboard +504,493233,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1891,0,"t",python,content +505,493233,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1892,0,"",python,selection_keyboard +506,493444,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1892,0,"j",python,content +507,493446,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1893,0,"",python,selection_keyboard +508,493839,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1892,1,"",python,content +509,494029,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1892,0," ",python,content +510,494030,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1893,0,"",python,selection_keyboard +511,494159,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1893,0,"j",python,content +512,494160,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1894,0,"",python,selection_keyboard +513,494356,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1894,0,"s",python,content +514,494358,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1895,0,"",python,selection_keyboard +515,494405,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1895,0,"o",python,content +516,494406,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1896,0,"",python,selection_keyboard +517,494534,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1896,0,"n",python,content +518,494535,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1897,0,"",python,selection_keyboard +519,494983,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1896,0,"",python,selection_command +520,496010,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1906,0,"",python,selection_mouse +521,496590,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1940,0,"",python,selection_mouse +522,497178,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1939,2,"",python,content +523,497584,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1949,0,"",python,selection_command +524,500244,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1960,0,"",python,selection_mouse +525,500392,"data/jasmine_data/ViZDoomPPO/load_model_generate_dataset.py",1957,5,"numpy",python,selection_mouse +526,502151,"TERMINAL",0,0,"bash",,terminal_focus +527,505606,"TERMINAL",0,0,"python",,terminal_focus +528,511140,"TERMINAL",0,0,"1",,terminal_output +529,511870,"TERMINAL",0,0,"6",,terminal_output +530,512569,"TERMINAL",0,0,"0",,terminal_output +531,513080,"TERMINAL",0,0," ",,terminal_output +532,513676,"TERMINAL",0,0,"*",,terminal_output +533,513773,"TERMINAL",0,0," ",,terminal_output +534,516738,"TERMINAL",0,0,"1",,terminal_output +535,516796,"TERMINAL",0,0,"0",,terminal_output +536,517027,"TERMINAL",0,0,"0",,terminal_output +537,517721,"TERMINAL",0,0,"\r\n16000\r\n>>> ",,terminal_output +538,521193,"TERMINAL",0,0,"16000",,terminal_output +539,521823,"TERMINAL",0,0," ",,terminal_output +540,531354,"TERMINAL",0,0,"bash",,terminal_focus +541,532880,"TERMINAL",0,0,"python",,terminal_focus +542,533895,"TERMINAL",0,0,"*",,terminal_output +543,534247,"TERMINAL",0,0," ",,terminal_output +544,535569,"TERMINAL",0,0,"1",,terminal_output +545,537408,"TERMINAL",0,0,"8",,terminal_output +546,537796,"TERMINAL",0,0,"6",,terminal_output +547,538780,"TERMINAL",0,0,"\r\n2976000\r\n>>> ",,terminal_output +548,583065,"slurm/jobs/mihir/horeka/preprocessing/doom_chunked.sh",0,0,"",shellscript,tab +549,588819,"TERMINAL",0,0,"bash",,terminal_focus +550,592013,"TERMINAL",0,0,"ls",,terminal_command +551,592096,"TERMINAL",0,0,"]633;C",,terminal_output +552,592207,"TERMINAL",0,0,"data_0000.array_record data_0068.array_record data_0136.array_record\r\ndata_0001.array_record data_0069.array_record data_0137.array_record\r\ndata_0002.array_record data_0070.array_record data_0138.array_record\r\ndata_0003.array_record data_0071.array_record data_0139.array_record\r\ndata_0004.array_record data_0072.array_record data_0140.array_record\r\ndata_0005.array_record data_0073.array_record data_0141.array_record\r\ndata_0006.array_record data_0074.array_record data_0142.array_record\r\ndata_0007.array_record data_0075.array_record data_0143.array_record\r\ndata_0008.array_record data_0076.array_record data_0144.array_record\r\ndata_0009.array_record data_0077.array_record data_0145.array_record\r\ndata_0010.array_record data_0078.array_record data_0146.array_record\r\ndata_0011.array_record data_0079.array_record data_0147.array_record\r\ndata_0012.array_record data_0080.array_record data_0148.array_record\r\ndata_0013.array_record data_0081.array_record data_0149.array_record\r\ndata_0014.array_record data_0082.array_record data_0150.array_record\r\ndata_0015.array_record data_0083.array_record data_0151.array_record\r\ndata_0016.array_record data_0084.array_record data_0152.array_record\r\ndata_0017.array_record data_0085.array_record data_0153.array_record\r\ndata_0018.array_record data_0086.array_record data_0154.array_record\r\ndata_0019.array_record data_0087.array_record data_0155.array_record\r\ndata_0020.array_record data_0088.array_record data_0156.array_record\r\ndata_0021.array_record data_0089.array_record data_0157.array_record\r\ndata_0022.array_record data_0090.array_record data_0158.array_record\r\ndata_0023.array_record data_0091.array_record data_0159.array_record\r\ndata_0024.array_record data_0092.array_record data_0160.array_record\r\ndata_0025.array_record data_0093.array_record data_0161.array_record\r\ndata_0026.array_record data_0094.array_record data_0162.array_record\r\ndata_0027.array_record data_0095.array_record data_0163.array_record\r\ndata_0028.array_record data_0096.array_record data_0164.array_record\r\ndata_0029.array_record data_0097.array_record data_0165.array_record\r\ndata_0030.array_record data_0098.array_record data_0166.array_record\r\ndata_0031.array_record data_0099.array_record data_0167.array_record\r\ndata_0032.array_record data_0100.array_record data_0168.array_record\r\ndata_0033.array_record data_0101.array_record data_0169.array_record\r\ndata_0034.array_record data_0102.array_record data_0170.array_record\r\ndata_0035.array_record data_0103.array_record data_0171.array_record\r\ndata_0036.array_record data_0104.array_record data_0172.array_record\r\ndata_0037.array_record data_0105.array_record data_0173.array_record\r\ndata_0038.array_record data_0106.array_record data_0174.array_record\r\ndata_0039.array_record data_0107.array_record data_0175.array_record\r\ndata_0040.array_record data_0108.array_record data_0176.array_record\r\ndata_0041.array_record data_0109.array_record data_0177.array_record\r\ndata_0042.array_record data_0110.array_record data_0178.array_record\r\ndata_0043.array_record data_0111.array_record data_0179.array_record\r\ndata_0044.array_record data_0112.array_record data_0180.array_record\r\ndata_0045.array_record data_0113.array_record data_0181.array_record\r\ndata_0046.array_record data_0114.array_record data_0182.array_record\r\ndata_0047.array_record data_0115.array_record data_0183.array_record\r\ndata_0048.array_record data_0116.array_record data_0184.array_record\r\ndata_0049.array_record data_0117.array_record data_0185.array_record\r\ndata_0050.array_record data_0118.array_record data_0186.array_record\r\ndata_0051.array_record data_0119.array_record data_0187.array_record\r\ndata_0052.array_record data_0120.array_record data_0188.array_record\r\ndata_0053.array_record data_0121.array_record data_0189.array_record\r\ndata_0054.array_record data_0122.array_record data_0190.array_record\r\ndata_0055.array_record data_0123.array_record data_0191.array_record\r\ndata_0056.array_record data_0124.array_record data_0192.array_record\r\ndata_0057.array_record data_0125.array_record data_0193.array_record\r\ndata_0058.array_record data_0126.array_record data_0194.array_record\r\ndata_0059.array_record data_0127.array_record data_0195.array_record\r\ndata_0060.array_record data_0128.array_record data_0196.array_record\r\ndata_0061.array_record data_0129.array_record data_0197.array_record\r\ndata_0062.array_record data_0130.array_record data_0198.array_record\r\ndata_0063.array_record data_0131.array_record data_0199.array_record\r\ndata_0064.array_record data_0132.array_record data_0200.array_record\r\ndata_0065.array_record data_0133.array_record data_0201.array_record\r\ndata_0066.array_record data_0134.array_record\r\ndata_0067.array_record data_0135.array_record\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_doom/doom_episodes_50m/train",,terminal_output +553,594583,"TERMINAL",0,0,"python",,terminal_focus +554,595629,"TERMINAL",0,0,"",,terminal_output +555,596085,"TERMINAL",0,0,"16000 * 186",,terminal_output +556,596356,"TERMINAL",0,0," * 100",,terminal_output +557,597754,"TERMINAL",0,0,"",,terminal_output +558,598187,"TERMINAL",0,0,"\r",,terminal_output +559,599200,"TERMINAL",0,0,"",,terminal_output +560,599570,"TERMINAL",0,0,"",,terminal_output +561,603397,"TERMINAL",0,0," ",,terminal_output +562,603594,"TERMINAL",0,0,"*",,terminal_output +563,603736,"TERMINAL",0,0," ",,terminal_output +564,604304,"TERMINAL",0,0,"2",,terminal_output +565,604487,"TERMINAL",0,0,"0",,terminal_output +566,604636,"TERMINAL",0,0,"0",,terminal_output +567,606624,"TERMINAL",0,0,"\r\n3200000\r\n>>> ",,terminal_output +568,616526,"TERMINAL",0,0,"bash",,terminal_focus +569,617806,"TERMINAL",0,0,"queue",,terminal_command +570,617873,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1990.localdomain: Sat Oct 4 18:17:42 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3541783 accelerat preproce tum_cte0 R 1:31:13\t 1 hkn0517",,terminal_output +571,618979,"TERMINAL",0,0,"34",,terminal_output +572,619990,"TERMINAL",0,0,"45",,terminal_output +573,621027,"TERMINAL",0,0,"56",,terminal_output +574,622022,"TERMINAL",0,0,"67",,terminal_output +575,623078,"TERMINAL",0,0,"78",,terminal_output +576,624208,"TERMINAL",0,0,"89",,terminal_output +577,625131,"TERMINAL",0,0,"920",,terminal_output +578,626166,"TERMINAL",0,0,"501",,terminal_output +579,627273,"TERMINAL",0,0,"12",,terminal_output +580,628295,"TERMINAL",0,0,"23",,terminal_output +581,629320,"TERMINAL",0,0,"35",,terminal_output +582,630446,"TERMINAL",0,0,"56",,terminal_output +583,631369,"TERMINAL",0,0,"67",,terminal_output +584,632395,"TERMINAL",0,0,"78",,terminal_output +585,633435,"TERMINAL",0,0,"89",,terminal_output +586,634473,"TERMINAL",0,0,"930",,terminal_output +587,635509,"TERMINAL",0,0,"8:001",,terminal_output +588,636590,"TERMINAL",0,0,"12",,terminal_output +589,637582,"TERMINAL",0,0,"23",,terminal_output +590,638614,"TERMINAL",0,0,"34",,terminal_output +591,639654,"TERMINAL",0,0,"45",,terminal_output +592,640690,"TERMINAL",0,0,"56",,terminal_output +593,641788,"TERMINAL",0,0,"67",,terminal_output +594,642946,"dataset_duplicates.ipynb",0,0,"\nprint(f""\nFound {len(duplicates)} sets of duplicate videos."")\nif duplicates:\n # Print the first 5 duplicate sets as an example\n for i, (h, paths) in enumerate(duplicates.items()):\n print(f"" - Hash: {h[:10]}... ({len(paths)} files)"")\n for path in paths:\n print(f"" - {os.path.basename(path)}"")\n\n",python,tab +595,643037,"TERMINAL",0,0,"78",,terminal_output +596,643832,"TERMINAL",0,0,"89",,terminal_output +597,644076,"TERMINAL",0,0,"python",,terminal_focus +598,644940,"TERMINAL",0,0,"940",,terminal_output +599,645067,"TERMINAL",0,0,"\r\nKeyboardInterrupt\r\n>>> ",,terminal_output +600,645886,"TERMINAL",0,0,"\r\n]0;tum_cte0515@hkn1990:~/Projects/jasmine",,terminal_output +601,645927,"TERMINAL",0,0,"101",,terminal_output +602,647053,"TERMINAL",0,0,"watch",,terminal_focus +603,647056,"TERMINAL",0,0,"12",,terminal_output +604,647727,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_doom/doom_episodes_50m/train",,terminal_output +605,648847,"TERMINAL",0,0,"pwd",,terminal_command +606,650439,"TERMINAL",0,0,"cd ..",,terminal_command +607,650949,"TERMINAL",0,0,"pwd",,terminal_command +608,654076,"dataset_duplicates.ipynb",0,0,"",python,tab +609,658821,"dataset_duplicates.ipynb",0,0,"base = ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_doom/dev""\n\ntrain_dir = os.path.join(base, ""train"")\ntest_dir = os.path.join(base, ""test"")\nval_dir = os.path.join(base, ""val"")\n\ntrain_array_record_files = get_array_record_files(train_dir)\ntest_array_record_files = get_array_record_files(test_dir)\nval_array_record_files = get_array_record_files(val_dir)\n\narray_record_files = train_array_record_files + test_array_record_files + val_array_record_files\nprint(f""Found {len(array_record_files)} files to process."")\n\nnum_processes = multiprocessing.cpu_count()\nprint(f""Using {num_processes} worker processes."")\n",python,tab +610,658822,"dataset_duplicates.ipynb",8,0,"",python,selection_mouse +611,658969,"dataset_duplicates.ipynb",7,2,"""/",python,selection_mouse +612,659019,"dataset_duplicates.ipynb",7,6,"""/hkfs",python,selection_mouse +613,659059,"dataset_duplicates.ipynb",7,11,"""/hkfs/work",python,selection_mouse +614,659060,"dataset_duplicates.ipynb",7,12,"""/hkfs/work/",python,selection_mouse +615,659068,"dataset_duplicates.ipynb",7,21,"""/hkfs/work/workspace",python,selection_mouse +616,659115,"dataset_duplicates.ipynb",7,29,"""/hkfs/work/workspace/scratch",python,selection_mouse +617,659133,"dataset_duplicates.ipynb",7,30,"""/hkfs/work/workspace/scratch/",python,selection_mouse +618,659160,"dataset_duplicates.ipynb",7,41,"""/hkfs/work/workspace/scratch/tum_ind3695",python,selection_mouse +619,659226,"dataset_duplicates.ipynb",7,72,"""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_doom/dev""\n",python,selection_mouse +620,659756,"dataset_duplicates.ipynb",7,42,"""/hkfs/work/workspace/scratch/tum_ind3695-",python,selection_mouse +621,659780,"dataset_duplicates.ipynb",7,41,"""/hkfs/work/workspace/scratch/tum_ind3695",python,selection_mouse +622,659825,"dataset_duplicates.ipynb",0,9,"base = ""/",python,selection_mouse +623,660553,"dataset_duplicates.ipynb",7,71,"""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_doom/dev""",python,selection_mouse +624,661415,"dataset_duplicates.ipynb",7,70,"""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_doom/dev",python,selection_mouse +625,662213,"dataset_duplicates.ipynb",7,70,"",python,content +626,662646,"dataset_duplicates.ipynb",7,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_doom/doom_episodes_50m",python,content +627,665373,"dataset_duplicates.ipynb",7,0,"",python,selection_mouse +628,666191,"dataset_duplicates.ipynb",7,0,"""",python,content +629,666193,"dataset_duplicates.ipynb",8,0,"",python,selection_keyboard +630,667178,"dataset_duplicates.ipynb",137,0,"",python,selection_mouse +631,668968,"dataset_duplicates.ipynb",291,0,"",python,selection_mouse +632,670755,"dataset_duplicates.ipynb",432,0,"",python,selection_mouse +633,671897,"dataset_duplicates.ipynb",432,0,"#",python,content +634,671899,"dataset_duplicates.ipynb",433,0,"",python,selection_keyboard +635,677266,"dataset_duplicates.ipynb",0,0,"",python,tab +636,678978,"dataset_duplicates.ipynb",0,0,"import os\nimport pickle\nimport hashlib\nimport numpy as np\nfrom collections import defaultdict\nfrom array_record.python.array_record_module import ArrayRecordReader\nimport multiprocessing\nfrom tqdm import tqdm # Using tqdm for a nice progress bar\n",python,tab +637,680407,"dataset_duplicates.ipynb",0,0,"# --- Helper Functions ---\n\ndef hash_byte_data(bytes_data: bytes) -> str:\n """"""Calculates the SHA256 hash of a byte string.""""""\n return hashlib.sha256(bytes_data).hexdigest()\n\n\ndef hash_numpy_frame(frame: np.ndarray) -> str:\n """"""Calculates the SHA256 hash of a numpy array.""""""\n return hashlib.sha256(np.ascontiguousarray(frame)).hexdigest()\n\n\ndef get_episode_level_hash(video_path: str) -> tuple[str, str] | None:\n """"""\n Reads a single array_record file, extracts the video, hashes it,\n and returns the (hash, video_path) tuple.\n """"""\n try:\n reader = ArrayRecordReader(video_path)\n record_data = reader.read()\n record_unpickled = pickle.loads(record_data)\n video_bytes = record_unpickled[""raw_video""]\n video_hash = hash_byte_data(video_bytes)\n return video_hash, video_path\n except Exception as e:\n print(f""Error processing file {video_path}: {e}"")\n return None\n\n\ndef get_frame_level_hashes(video_path: str) -> tuple[list[str], str] | None:\n """"""\n Reads a single array_record file, extracts the frames, hashes each frame,\n and returns the (list of hashes, video_path) tuple.\n """"""\n try:\n reader = ArrayRecordReader(video_path)\n record_data = pickle.loads(reader.read())\n\n # video shape (seq_len, 64, 64, 3)\n video_shape = ( record_data[""sequence_length""], 64, 64, 3)\n episode_tensor = np.frombuffer(record_data[""raw_video""], dtype=np.uint8)\n episode_tensor = episode_tensor.reshape(video_shape) \n\n frame_hashes = [hash_numpy_frame(frame) for frame in episode_tensor]\n\n return frame_hashes, video_path\n except Exception as e:\n print(f""Error processing file {video_path}: {e}"")\n return None\n\ndef get_array_record_files(dir):\n return [\n os.path.join(dir, x)\n for x in os.listdir(dir)\n if x.endswith("".array_record"")\n ]\n",python,tab +638,684141,"dataset_duplicates.ipynb",0,0,"# Find episode level duplicates\nduplicate_episode = defaultdict(list)\n\n# The 'with' statement ensures the pool is properly closed\nwith multiprocessing.Pool(processes=num_processes) as pool:\n # Use pool.imap_unordered for efficiency. It processes items as they are submitted\n # and returns results as they complete, which is perfect for progress bars.\n # We wrap the result iterator with tqdm to show progress.\n results = pool.imap_unordered(get_episode_level_hash, array_record_files)\n \n print(""Processing files and calculating hashes..."")\n for result in tqdm(results, total=len(array_record_files)):\n if result: # Ensure the worker didn't return None due to an error\n video_hash, video_path = result\n duplicate_episode[video_hash].append(video_path)\n",python,tab +639,688637,"dataset_duplicates.ipynb",0,0,"",python,tab +640,700131,"dataset_duplicates.ipynb",365,0,"",python,selection_mouse +641,701262,"dataset_duplicates.ipynb",347,0,"",python,selection_mouse +642,701994,"dataset_duplicates.ipynb",353,0,"",python,selection_mouse +643,702600,"dataset_duplicates.ipynb",352,0,"",python,selection_command +644,703232,"dataset_duplicates.ipynb",329,56,"val_array_record_files = get_array_record_files(val_dir)",python,selection_command +645,703424,"dataset_duplicates.ipynb",270,115,"test_array_record_files = get_array_record_files(test_dir)\nval_array_record_files = get_array_record_files(val_dir)",python,selection_command +646,703672,"dataset_duplicates.ipynb",270,0,"",python,selection_command +647,704314,"dataset_duplicates.ipynb",329,0,"#",python,content +648,704314,"dataset_duplicates.ipynb",270,0,"#",python,content +649,704316,"dataset_duplicates.ipynb",271,0,"",python,selection_keyboard +650,704358,"dataset_duplicates.ipynb",331,0," ",python,content +651,704358,"dataset_duplicates.ipynb",271,0," ",python,content +652,704360,"dataset_duplicates.ipynb",272,0,"",python,selection_keyboard +653,704749,"dataset_duplicates.ipynb",271,0,"",python,selection_command +654,709671,"dataset_duplicates.ipynb",0,0,"",python,tab diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-1103382d-4338-4569-ad1a-fd99664b131a1756899628806-2025_09_03-13.40.59.36/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-1103382d-4338-4569-ad1a-fd99664b131a1756899628806-2025_09_03-13.40.59.36/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..b7ae549a3324aeafc27e0420174e23f34d81fb09 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-1103382d-4338-4569-ad1a-fd99664b131a1756899628806-2025_09_03-13.40.59.36/source.csv @@ -0,0 +1,551 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +1,4,"models/dynamics.py",0,0,"from typing import Dict\n\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\nimport einops\n\nfrom utils.nn import STTransformer, Transformer\n\n\nclass DynamicsMaskGIT(nnx.Module):\n """"""\n MaskGIT dynamics model\n \n Dimension keys:\n B: batch size\n T: sequence length\n N: number of patches per frame\n L: latent dimension\n V: vocabulary size (number of latents)\n """"""\n\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_latents: int,\n latent_action_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n mask_limit: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_latents = num_latents\n self.latent_action_dim = latent_action_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.mask_limit = mask_limit\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.transformer = STTransformer(\n self.model_dim,\n self.model_dim,\n self.ffn_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.patch_embed = nnx.Embed(self.num_latents, self.model_dim, rngs=rngs)\n self.mask_token = nnx.Param(\n nnx.initializers.lecun_uniform()(rngs.params(), (1, 1, 1, self.model_dim))\n )\n self.action_up = nnx.Linear(\n self.latent_action_dim,\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(\n self, batch: Dict[str, jax.Array], training: bool = True\n ) -> tuple[jax.Array, jax.Array | None]:\n # --- Mask videos ---\n video_tokens_BTN = batch[""video_tokens""]\n latent_actions_BTm11L = batch[""latent_actions""]\n vid_embed_BTNM = self.patch_embed(video_tokens_BTN)\n if training:\n batch_size = vid_embed_BTNM.shape[0]\n _rng_prob, *_rngs_mask = jax.random.split(batch[""mask_rng""], batch_size + 1)\n mask_prob = jax.random.uniform(\n _rng_prob, shape=(batch_size,), minval=self.mask_limit\n )\n per_sample_shape = vid_embed_BTNM.shape[1:-1]\n mask = jax.vmap(\n lambda rng, prob: jax.random.bernoulli(rng, prob, per_sample_shape),\n in_axes=(0, 0),\n )(jnp.asarray(_rngs_mask), mask_prob)\n mask = mask.at[:, 0].set(False)\n vid_embed_BTNM = jnp.where(\n jnp.expand_dims(mask, -1), self.mask_token.value, vid_embed_BTNM\n )\n else:\n mask = None\n\n # --- Predict transition ---\n act_embed_BTm11M = self.action_up(latent_actions_BTm11L)\n padded_act_embed_BT1M = jnp.pad(act_embed_BTm11M, ((0, 0), (1, 0), (0, 0), (0, 0)))\n padded_act_embed_BTNM = jnp.broadcast_to(padded_act_embed_BT1M, vid_embed_BTNM.shape)\n vid_embed_BTNM += padded_act_embed_BTNM\n logits_BTNV = self.transformer(vid_embed_BTNM)\n return logits_BTNV, mask\n\nclass DynamicsCausal(nnx.Module):\n """"""Causal dynamics model""""""\n\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_latents: int,\n latent_action_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_latents = num_latents\n self.latent_action_dim = latent_action_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.transformer = Transformer(\n self.model_dim,\n self.model_dim,\n self.ffn_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n rngs=rngs,\n )\n self.patch_embed = nnx.Embed(self.num_latents, self.model_dim, rngs=rngs)\n self.action_up = nnx.Linear(\n self.latent_action_dim,\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(\n self, batch: Dict[str, jax.Array], training: bool = True\n ) -> tuple[jax.Array, jax.Array | None]:\n video_tokens_BTN = batch[""video_tokens""]\n latent_actions_BTm11L = batch[""latent_actions""]\n\n vid_embed_BTNM = self.patch_embed(video_tokens_BTN)\n act_embed_BTm11M = self.action_up(latent_actions_BTm11L)\n padded_act_embed_BT1M = jnp.pad(act_embed_BTm11M, ((0, 0), (1, 0), (0, 0), (0, 0)))\n vid_embed_BTNp1M = jnp.concatenate([padded_act_embed_BT1M, vid_embed_BTNM], axis=2)\n\n logits_BTNp1V = self.transformer(vid_embed_BTNp1M)\n logits_BTNV = logits_BTNp1V[:, :, :-1]\n\n return logits_BTNV, jnp.ones_like(video_tokens_BTN)\n",python,tab +2,312,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"1:40:58 PM [info] Activating crowd-code\n1:40:59 PM [info] Recording started\n1:40:59 PM [info] Initializing git provider using file system watchers...\n1:40:59 PM [info] Git repository found\n1:40:59 PM [info] Git provider initialized successfully\n",Log,tab +3,487,"extension-output-pdoom-org.crowd-code-#1-crowd-code",245,0,"1:40:59 PM [info] Initial git state: [object Object]\n",Log,content +4,9066,"models/dynamics.py",0,0,"",python,tab +5,10669,"train_dynamics.py",0,0,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n darkness_threshold: float = 0.0\n # Tokenizer\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_type: str = ""maskgit"" # supported options: maskgit, causal\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n wandb_id: str = """"\n\n\nargs = tyro.cli(Args)\n\n\ndef dynamics_loss_fn(\n model: Genie, inputs: dict\n) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n """"""Compute masked dynamics loss""""""\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n model.train()\n outputs = model(inputs, training=True)\n mask = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n gt = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]), size=args.num_latent_actions, fill_value=0\n )\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]), size=args.num_patch_latents, fill_value=0\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n psnr=psnr,\n ssim=ssim,\n codebook_usage_lam=codebook_usage_lam,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n\n@nnx.jit\ndef train_step(\n model: Genie, optimizer: nnx.Optimizer, inputs: dict\n) -> tuple[jax.Array, jax.Array, dict]:\n """"""Update state and compute metrics""""""\n\n def loss_fn(model: Genie) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n return dynamics_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(model)\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n decode=False,\n rngs=rngs,\n )\n\n _, params, _ = nnx.split(genie, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n optimizer = nnx.Optimizer(genie, tx)\n del genie\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n darkness_threshold=args.darkness_threshold,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n optimizer = restore_genie_components(optimizer, replicated_sharding, rng, args)\n # NOTE: We have to remove the (unused) tokenizer vq dropout due flax.nnx lazily initializing modules.\n # Specifically, the first dynamics model checkpoint will contain the vq dropout module,\n # but the first full restore will fail due to nnx not initializing the module when\n # dropout is set to 0.0.\n del optimizer.model.tokenizer.vq.drop\n\n # --- TRAIN LOOP ---\n dataloader = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in grain_iterator\n )\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng_mask = jax.random.split(rng, 2)\n inputs = dict(videos=videos, mask_rng=_rng_mask)\n loss, recon, metrics = train_step(optimizer.model, optimizer, inputs)\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n grain_iterator # type: ignore\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +6,15765,"genie.py",0,0,"from typing import Dict\n\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\nimport orbax.checkpoint as ocp\n\nfrom models.dynamics import DynamicsMaskGIT, DynamicsCausal\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\n\nclass Genie(nnx.Module):\n """"""Genie model""""""\n\n def __init__(\n self,\n in_dim: int,\n tokenizer_dim: int,\n tokenizer_ffn_dim: int,\n latent_patch_dim: int,\n num_patch_latents: int,\n patch_size: int,\n tokenizer_num_blocks: int,\n tokenizer_num_heads: int,\n lam_dim: int,\n lam_ffn_dim: int,\n latent_action_dim: int,\n num_latent_actions: int,\n lam_patch_size: int,\n lam_num_blocks: int,\n lam_num_heads: int,\n lam_co_train: bool,\n dyna_type: str,\n dyna_dim: int,\n dyna_ffn_dim: int,\n dyna_num_blocks: int,\n dyna_num_heads: int,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n dropout: float = 0.0,\n mask_limit: float = 0.0,\n ):\n # --- Tokenizer ---\n self.in_dim = in_dim\n self.tokenizer_dim = tokenizer_dim\n self.tokenizer_ffn_dim = tokenizer_ffn_dim\n self.latent_patch_dim = latent_patch_dim\n self.num_patch_latents = num_patch_latents\n self.patch_size = patch_size\n self.tokenizer_num_blocks = tokenizer_num_blocks\n self.tokenizer_num_heads = tokenizer_num_heads\n # --- LAM ---\n self.lam_dim = lam_dim\n self.lam_ffn_dim = lam_ffn_dim\n self.latent_action_dim = latent_action_dim\n self.num_latent_actions = num_latent_actions\n self.lam_patch_size = lam_patch_size\n self.lam_num_blocks = lam_num_blocks\n self.lam_num_heads = lam_num_heads\n self.lam_co_train = lam_co_train\n # --- Dynamics ---\n self.dyna_type = dyna_type\n self.dyna_dim = dyna_dim\n self.dyna_ffn_dim = dyna_ffn_dim\n self.dyna_num_blocks = dyna_num_blocks\n self.dyna_num_heads = dyna_num_heads\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.dropout = dropout\n self.mask_limit = mask_limit\n\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n ffn_dim=self.tokenizer_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n ffn_dim=self.lam_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_latent_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n if self.dyna_type == ""maskgit"":\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n elif self.dyna_type == ""causal"":\n self.dynamics = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n rngs=rngs,\n )\n else:\n raise ValueError(f""Invalid dynamics type: {self.dyna_type}"")\n\n def __call__(\n self, batch: Dict[str, jax.Array], training: bool = True\n ) -> Dict[str, jax.Array]:\n videos_BTHWC = batch[""videos""]\n tokenizer_outputs = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_indices_BTN = tokenizer_outputs[""indices""]\n lam_outputs = self.lam.vq_encode(videos_BTHWC, training=False)\n z_q_BTm11L = lam_outputs[""z_q""]\n action_indices_E = lam_outputs[""indices""]\n latent_actions_BTm11L = jax.lax.cond(\n self.lam_co_train,\n lambda: z_q_BTm11L,\n lambda: jax.lax.stop_gradient(z_q_BTm11L),\n )\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(token_indices_BTN),\n latent_actions=latent_actions_BTm11L,\n )\n outputs[""mask_rng""] = batch[""mask_rng""]\n dyna_logits_BTNV, dyna_mask = self.dynamics(outputs, training)\n outputs[""token_logits""] = dyna_logits_BTNV\n if dyna_mask is not None:\n outputs[""mask""] = dyna_mask\n mle_indices_BTN = jnp.argmax(outputs[""token_logits""], axis=-1)\n H, W = batch[""videos""].shape[2:4]\n outputs[""recon""] = self.tokenizer.decode(mle_indices_BTN, (H, W))\n outputs[""lam_indices""] = action_indices_E\n return outputs\n\n def sample(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n steps: int = 25,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> jax.Array:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n P: S * N\n """"""\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n latent_actions_E = batch[""latent_actions""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n def maskgit_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array], step: jax.Array\n ) -> tuple[tuple[jax.Array, jax.Array, jax.Array, jax.Array], None]:\n rng, token_idxs_BSN, mask_BSN, action_tokens_EL = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # --- Construct + encode video ---\n vid_embed_BSNM = self.dynamics.patch_embed(token_idxs_BSN)\n mask_token_111M = self.dynamics.mask_token.value\n mask_expanded_BSN1 = mask_BSN[..., None]\n vid_embed_BSNM = jnp.where(mask_expanded_BSN1, mask_token_111M, vid_embed_BSNM)\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = self.dynamics.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1]))\n vid_embed_BSNM += act_embed_BS1M\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (steps * 2))\n step_temp = temperature * (1.0 - unmasked_ratio)\n final_logits_BSNV = self.dynamics.transformer(vid_embed_BSNM) / step_temp\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_BSN = jnp.argmax(final_logits_BSNV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_BSN = jax.random.categorical(_rng, final_logits_BSNV)\n gather_fn = jax.vmap(jax.vmap(jax.vmap(lambda x, y: x[y])))\n final_token_probs_BSN = gather_fn(\n jax.nn.softmax(final_logits_BSNV), sampled_token_idxs_BSN\n )\n final_token_probs_BSN += ~mask_BSN\n # Update masked tokens only\n token_idxs_BSN = jnp.where(mask_BSN, sampled_token_idxs_BSN, token_idxs_BSN)\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n final_token_probs_flat_BP = einops.rearrange(final_token_probs_BSN, ""b s n -> b (s n)"")\n idx_mask_P = jnp.arange(final_token_probs_flat_BP.shape[-1]) <= N - num_unmasked_tokens\n sorted_idxs_BP = jnp.argsort(final_token_probs_flat_BP, axis=-1)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask_P))\n mask_flat_BP = einops.rearrange(mask_BSN, ""b s n -> b (s n)"")\n new_mask_flat_BP = mask_update_fn(mask_flat_BP, sorted_idxs_BP)\n new_mask_BSN = einops.rearrange(new_mask_flat_BP, ""b (s n) -> b s n"", n=N)\n\n new_carry = (rng, token_idxs_BSN, new_mask_BSN, action_tokens_EL)\n return new_carry, None\n\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[tuple[jax.Array, jax.Array], None]:\n rng, current_token_idxs_BSN = carry\n rng, step_rng = jax.random.split(rng)\n\n # Mask current frame (i.e., t == step_t)\n mask_S = jnp.arange(seq_len) == step_t\n mask_BSN = jnp.broadcast_to(mask_S[None, :, None], (B, seq_len, N)).astype(\n bool\n )\n masked_token_idxs_BSN = current_token_idxs_BSN * ~mask_BSN\n\n # --- Initialize and run MaskGIT loop ---\n init_carry_maskgit = (\n step_rng,\n masked_token_idxs_BSN,\n mask_BSN,\n action_tokens_EL,\n )\n final_carry_maskgit, _ = jax.lax.scan(\n maskgit_step_fn, init_carry_maskgit, jnp.arange(steps)\n )\n updated_token_idxs_BSN = final_carry_maskgit[1]\n new_carry = (rng, updated_token_idxs_BSN)\n return new_carry, None\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry, _ = jax.lax.scan(\n generation_step_fn, initial_carry, timesteps_to_scan\n )\n final_token_idxs_BSN = final_carry[1]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC\n\n def sample_causal(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> jax.Array:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n assert isinstance(self.dynamics, DynamicsCausal)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n latent_actions_E = batch[""latent_actions""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n dynamics_causal: DynamicsCausal = self.dynamics\n\n def causal_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array], step_n: jax.Array\n ) -> tuple[tuple[jax.Array, jax.Array, jax.Array, jax.Array], None]:\n rng, token_idxs_BSN, action_tokens_EL, step_t = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_causal.patch_embed(token_idxs_BSN)\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_causal.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1]))\n vid_embed_BSNp1M = jnp.concatenate([act_embed_BS1M, vid_embed_BSNM], axis=2)\n final_logits_BTNp1V = dynamics_causal.transformer(vid_embed_BSNp1M, (step_t, step_n)) / temperature\n final_logits_BV = final_logits_BTNp1V[:, step_t, step_n, :]\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_B = jnp.argmax(final_logits_BV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_B = jax.random.categorical(_rng, final_logits_BV)\n # Update next tokens only\n token_idxs_BSN = token_idxs_BSN.at[:, step_t, step_n].set(sampled_token_idxs_B)\n\n new_carry = (rng, token_idxs_BSN, action_tokens_EL, step_t)\n return new_carry, None\n\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[tuple[jax.Array, jax.Array], None]:\n rng, current_token_idxs_BSN = carry\n rng, step_rng = jax.random.split(rng)\n\n # --- Initialize and run causal loop ---\n init_carry_causal = (\n step_rng,\n current_token_idxs_BSN,\n action_tokens_EL,\n step_t,\n )\n final_carry_causal, _ = jax.lax.scan(\n causal_step_fn, init_carry_causal, jnp.arange(N)\n )\n updated_token_idxs_BSN = final_carry_causal[1]\n new_carry = (rng, updated_token_idxs_BSN)\n return new_carry, None\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry, _ = jax.lax.scan(\n generation_step_fn, initial_carry, timesteps_to_scan\n )\n final_token_idxs_BSN = final_carry[1]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC\n\n def vq_encode(self, batch: Dict[str, jax.Array], training: bool) -> jax.Array:\n # --- Preprocess videos ---\n video_BTHWC = batch[""videos""]\n lam_output = self.lam.vq_encode(video_BTHWC, training=training)\n lam_indices_E = lam_output[""indices""]\n return lam_indices_E\n\n# FIXME (f.srambical): add conversion script for old checkpoints\ndef restore_genie_components(\n optimizer: nnx.Optimizer,\n sharding: jax.sharding.NamedSharding,\n rng: jax.Array,\n args,\n) -> nnx.Optimizer:\n """"""Restore pre-trained Genie components""""""\n rngs = nnx.Rngs(rng)\n\n tx = optimizer.tx\n model = optimizer.model\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n tokenizer_checkpoint_manager = ocp.CheckpointManager(\n directory=args.tokenizer_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n ffn_dim=args.tokenizer_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n dummy_tokenizer_optimizer = nnx.Optimizer(dummy_tokenizer, tx)\n dummy_tokenizer_optimizer_state = nnx.state(dummy_tokenizer_optimizer)\n abstract_sharded_tokenizer_optimizer_state = _create_abstract_sharded_pytree(\n dummy_tokenizer_optimizer_state, sharding\n )\n restored_tokenizer = tokenizer_checkpoint_manager.restore(\n step=tokenizer_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_tokenizer_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_tokenizer_optimizer.model, restored_tokenizer.model)\n model.tokenizer = dummy_tokenizer_optimizer.model\n tokenizer_checkpoint_manager.close()\n\n if args.lam_checkpoint:\n lam_checkpoint_manager = ocp.CheckpointManager(\n directory=args.lam_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n ffn_dim=args.lam_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_latent_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n dummy_lam_optimizer = nnx.Optimizer(dummy_lam, tx)\n dummy_lam_optimizer_state = nnx.state(dummy_lam_optimizer)\n abstract_sharded_lam_optimizer_state = _create_abstract_sharded_pytree(\n dummy_lam_optimizer_state, sharding\n )\n restored_lam_optimizer = lam_checkpoint_manager.restore(\n step=lam_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_lam_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_lam_optimizer.model, restored_lam_optimizer.model)\n model.lam = dummy_lam_optimizer.model\n # Remove the LAM decoder to save memory and avoid unnecessary computation.\n del model.lam.decoder\n lam_checkpoint_manager.close()\n \n # Reinitialize the optimizer states\n optimizer = nnx.Optimizer(model, tx)\n return optimizer\n\n\ndef _create_abstract_sharded_pytree(\n pytree_template: nnx.GraphState, sharding_spec: jax.sharding.NamedSharding\n) -> jax.Array:\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)\n",python,tab +7,21363,"utils/nn.py",0,0,"import math\nfrom typing import Tuple, Callable, List\n\nfrom flax import nnx\nimport jax\nimport jax.numpy as jnp\nimport einops\n\n\nclass SpatioTemporalPositionalEncoding(nnx.Module):\n """"""\n Applies separate sinusoidal positional encodings to the temporal and spatial dimensions.\n """"""\n def __init__(self, d_model: int, max_len: int = 5000):\n self.d_model = d_model\n self.max_len = max_len\n\n pe = jnp.zeros((self.max_len, self.d_model))\n position = jnp.arange(0, self.max_len, dtype=jnp.float32)[:, None]\n div_term = jnp.exp(\n jnp.arange(0, self.d_model, 2) * (-math.log(10000.0) / self.d_model)\n )\n pe = pe.at[:, 0::2].set(jnp.sin(position * div_term))\n pe = pe.at[:, 1::2].set(jnp.cos(position * div_term))\n self.pe = nnx.Variable(pe)\n\n def __call__(self, x: jax.Array) -> jax.Array:\n """"""\n Args:\n x: The input tensor of shape (Batch, Time, Space, Dimension).\n\n Returns:\n The input tensor with positional encodings added.\n """"""\n assert x.ndim == 4, f""Input must be 4-dimensional, but got shape {x.shape}""\n\n num_timesteps = x.shape[1]\n num_spatial_patches = x.shape[2]\n\n # Temporal positional encoding: (1, T, 1, D)\n temporal_pe = self.pe.value[None, :num_timesteps, None, :]\n x = x + temporal_pe\n\n # Spatial positional encoding: (1, 1, S, D)\n spatial_pe = self.pe.value[None, None, :num_spatial_patches, :]\n x = x + spatial_pe\n\n return x\n\n\nclass STBlock(nnx.Module):\n def __init__(\n self,\n dim: int,\n ffn_dim: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n ):\n self.dim = dim\n self.ffn_dim = ffn_dim\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.spatial_norm = nnx.LayerNorm(\n num_features=self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.spatial_attention = nnx.MultiHeadAttention(\n num_heads=self.num_heads,\n in_features=self.dim,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=False\n ),\n rngs=rngs,\n decode=False,\n )\n\n self.temporal_norm = nnx.LayerNorm(\n num_features=self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.temporal_attention = nnx.MultiHeadAttention(\n num_heads=self.num_heads,\n in_features=self.dim,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=True\n ),\n rngs=rngs,\n decode=False,\n )\n\n self.ffn_norm = nnx.LayerNorm(\n num_features=self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.ffn_dense1 = nnx.Linear(\n in_features=self.dim,\n out_features=self.ffn_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.ffn_dense2 = nnx.Linear(\n in_features=self.ffn_dim,\n out_features=self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n @nnx.remat\n def __call__(self, x_BTNM: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z_BTNM = self.spatial_norm(x_BTNM)\n z_BTNM = self.spatial_attention(z_BTNM)\n x_BTNM = x_BTNM + z_BTNM\n\n # --- Temporal attention ---\n x_BNTM = x_BTNM.swapaxes(1, 2)\n z_BNTM = self.temporal_norm(x_BNTM)\n z_BNTM = self.temporal_attention(z_BNTM)\n x_BNTM = x_BNTM + z_BNTM\n x_BTNM = x_BNTM.swapaxes(1, 2)\n\n # --- Feedforward ---\n z_BTNM = self.ffn_norm(x_BTNM)\n z_BTND = self.ffn_dense1(z_BTNM)\n z_BTND = jax.nn.gelu(z_BTND)\n z_BTNM = self.ffn_dense2(z_BTND)\n x_BTNM = x_BTNM + z_BTNM\n\n return x_BTNM\n\n\nclass STTransformer(nnx.Module):\n """"""\n Dimension keys:\n B: batch size\n T: number of frames\n N: number of patches per frame\n I: number of input features\n M: model dimension\n D: FFN dimension\n O: number of output features\n """"""\n def __init__(\n self,\n input_dim: int,\n model_dim: int,\n ffn_dim: int,\n out_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n max_len: int = 5000,\n ):\n self.input_dim = input_dim\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.out_dim = out_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.input_norm1 = nnx.LayerNorm(\n num_features=self.input_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.input_dense = nnx.Linear(\n in_features=self.input_dim,\n out_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.input_norm2 = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n self.pos_enc = SpatioTemporalPositionalEncoding(self.model_dim, max_len=max_len)\n\n self.blocks = []\n for _ in range(self.num_blocks):\n self.blocks.append(\n STBlock(\n dim=self.model_dim,\n ffn_dim=self.ffn_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n )\n\n self.output_dense = nnx.Linear(\n in_features=self.model_dim,\n out_features=self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(self, x_BTNI: jax.Array) -> jax.Array:\n x_BTNI = self.input_norm1(x_BTNI)\n x_BTNM = self.input_dense(x_BTNI)\n x_BTNM = self.input_norm2(x_BTNM)\n x_BTNM = self.pos_enc(x_BTNM)\n for block in self.blocks:\n x_BTNM = block(x_BTNM)\n\n x_BTNO = self.output_dense(x_BTNM)\n return x_BTNO\n\nclass TransformerBlock(nnx.Module):\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.decode = decode\n\n self.temporal_norm = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.spatial_norm = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.ffn_norm = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.temporal_attention = nnx.MultiHeadAttention(\n num_heads=self.num_heads,\n in_features=self.model_dim,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=True\n ),\n rngs=rngs,\n decode=self.decode,\n )\n self.spatial_attention = nnx.MultiHeadAttention(\n num_heads=self.num_heads,\n in_features=self.model_dim,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=True\n ),\n rngs=rngs,\n decode=self.decode,\n )\n self.ffn_dense1 = nnx.Linear(\n in_features=self.model_dim,\n out_features=self.ffn_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.ffn_dense2 = nnx.Linear(\n in_features=self.ffn_dim,\n out_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n @nnx.remat\n def __call__(self, x_BTNM: jax.Array, pos_index: Tuple[jax.Array, jax.Array] | None = None) -> jax.Array:\n # --- Spatial attention ---\n B, T, N, M = x_BTNM.shape\n z_FNM = einops.rearrange(x_BTNM, ""b t n m -> (b t) n m"")\n z_FNM = self.spatial_norm(z_FNM)\n z_FNM = self.spatial_attention(z_FNM)\n z_BTNM = einops.rearrange(z_FNM, ""(b t) n m -> b t n m"", t=T)\n x_BTNM = x_BTNM + z_BTNM\n # --- Temporal attention ---\n z_PTM = einops.rearrange(x_BTNM, ""b t n m -> (b n) t m"")\n z_PTM = self.temporal_norm(z_PTM)\n z_PTM = self.temporal_attention(z_PTM)\n z_BTNM = einops.rearrange(z_PTM, ""(b n) t m -> b t n m"", n=N)\n x_BTNM = x_BTNM + z_BTNM\n # --- Feedforward ---\n z_BTNM = self.ffn_norm(x_BTNM)\n z_BTND = self.ffn_dense1(z_BTNM)\n z_BTND = jax.nn.gelu(z_BTND)\n z_BTNM = self.ffn_dense2(z_BTND)\n x_BTNM = x_BTNM + z_BTNM\n\n return x_BTNM\n\nclass Transformer(nnx.Module):\n """"""\n Dimension keys:\n B: batch size\n T: number of frames\n N: number of patches per frame\n I: number of input features\n M: model dimension\n D: FFN dimension\n O: number of output features\n F: number of frames in batch\n P: number of patch positions in batch\n """"""\n def __init__(\n self,\n input_dim: int,\n model_dim: int,\n ffn_dim: int,\n out_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n max_len: int = 5000,\n ):\n self.input_dim = input_dim\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.out_dim = out_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.input_norm1 = nnx.LayerNorm(\n num_features=self.input_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.input_dense = nnx.Linear(\n in_features=self.input_dim,\n out_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.input_norm2 = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n self.pos_enc = SpatioTemporalPositionalEncoding(self.model_dim, max_len=max_len)\n\n self.blocks: List[TransformerBlock] = []\n for _ in range(self.num_blocks):\n self.blocks.append(\n TransformerBlock(\n model_dim=self.model_dim,\n ffn_dim=self.ffn_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n rngs=rngs,\n )\n )\n self.output_dense = nnx.Linear(\n in_features=self.model_dim,\n out_features=self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(self, x_BTNI: jax.Array, pos_index: Tuple[jax.Array, jax.Array] | None = None) -> jax.Array:\n x_BTNI = self.input_norm1(x_BTNI)\n x_BTNM = self.input_dense(x_BTNI)\n x_BTNM = self.input_norm2(x_BTNM)\n x_BTNM = self.pos_enc(x_BTNM)\n for block in self.blocks:\n x_BTNM = block(x_BTNM, pos_index)\n\n x_BTNV = self.output_dense(x_BTNM)\n return x_BTNV\n\ndef normalize(x: jax.Array) -> jax.Array:\n return x / (jnp.linalg.norm(x, ord=2, axis=-1, keepdims=True) + 1e-8)\n\n\nclass VectorQuantizer(nnx.Module):\n """"""\n Dimension keys:\n D: B * T * N\n K: number of latents\n L: latent dimension\n """"""\n def __init__(\n self, latent_dim: int, num_latents: int, dropout: float, rngs: nnx.Rngs\n ):\n self.latent_dim = latent_dim\n self.num_latents = num_latents\n self.dropout = dropout\n\n self.codebook = nnx.Param(\n normalize(\n nnx.initializers.lecun_uniform()(\n rngs.params(), (self.num_latents, self.latent_dim)\n )\n )\n )\n self.drop = nnx.Dropout(self.dropout, rngs=rngs)\n\n def __call__(\n self, x_DL: jax.Array, training: bool\n ) -> Tuple[jax.Array, jax.Array, jax.Array, jax.Array]:\n # --- Compute distances ---\n x_DL = normalize(x_DL)\n normalized_codebook_KL = normalize(self.codebook.value)\n distance_DK = -jnp.matmul(x_DL, normalized_codebook_KL.T)\n if training:\n distance_DK = self.drop(distance_DK)\n\n # --- Get indices and embeddings ---\n indices_D = jnp.argmin(distance_DK, axis=-1)\n z_DL = self.codebook[indices_D]\n\n # --- Straight through estimator ---\n z_q_DL = x_DL + jax.lax.stop_gradient(z_DL - x_DL)\n return z_q_DL, z_DL, x_DL, indices_D\n\n def get_codes(self, indices_E: jax.Array) -> jax.Array:\n return self.codebook[indices_E]\n\n\ndef _create_flash_attention_fn(use_flash_attention: bool, is_causal: bool) -> Callable:\n """"""\n Create an attention function that uses flash attention if enabled.\n\n flax.nnx.MultiHeadAttention provides tensors with shape (batch..., length, num_heads, head_dim),\n but jax.nn.dot_product_attention expects (batch, length, num_heads, head_dim). We reshape to\n ensure compatibility. cuDNN's flash attention additionally requires a sequence length that\n is a multiple of 4. We pad the sequence length to the nearest multiple of 4 and mask\n accordingly. Note that cuDNN requires the mask to be broadcast before calling the attention\n function due to strict shape checking.\n """"""\n\n def attention_fn(query_BTHD, key_BSHD, value_BSHD, bias=None, mask_B111=None, **kwargs):\n implementation = ""cudnn"" if use_flash_attention else None\n\n def _merge_batch_dims(x):\n return einops.rearrange(x, ""... l h k -> (...) l h k"")\n\n def _pad(x, pad_size):\n return jnp.pad(x, ((0, 0), (0, pad_size), (0, 0), (0, 0)))\n\n original_shape = query_BTHD.shape\n T = query_BTHD.shape[-3]\n S = key_BSHD.shape[-3]\n\n # Pad to nearest multiple of 4\n Q = ((T + 3) // 4) * 4\n pad_size_Q = Q - T\n K = ((S + 3) // 4) * 4\n pad_size_K = K - S\n\n query_BQHD = _pad(_merge_batch_dims(query_BTHD), pad_size_Q)\n key_BKHD = _pad(_merge_batch_dims(key_BSHD), pad_size_K)\n value_BKHD = _pad(_merge_batch_dims(value_BSHD), pad_size_K)\n\n attention_mask = jnp.ones((Q, K), dtype=jnp.bool_)\n attention_mask = attention_mask.at[T:, :].set(False)\n attention_mask = attention_mask.at[:, S:].set(False)\n\n mask_11TS = attention_mask[jnp.newaxis, jnp.newaxis, :, :]\n\n bias_4d = jnp.pad(_merge_batch_dims(bias), ((0, 0), (0, 0), (0, pad_size_Q), (0, pad_size_K))) if bias is not None else None\n\n # NOTE: jax.nn.dot_product_attention does not support dropout\n output_4d = jax.nn.dot_product_attention(\n query=query_BQHD,\n key=key_BKHD,\n value=value_BKHD,\n bias=bias_4d,\n mask=mask_11TS,\n implementation=implementation,\n is_causal=is_causal,\n )\n return output_4d[..., :T, :, :].reshape(original_shape)\n\n return attention_fn\n",python,tab +8,84155,"utils/nn.py",16785,0,"",python,selection_mouse +9,138465,"utils/nn.py",16789,0,"",python,selection_mouse +10,138480,"utils/nn.py",16788,0,"",python,selection_command +11,139099,"utils/nn.py",16718,0,"",python,selection_mouse +12,139109,"utils/nn.py",16717,0,"",python,selection_command +13,140678,"utils/nn.py",16874,0,"",python,selection_mouse +14,140693,"utils/nn.py",16873,0,"",python,selection_command +15,146771,"utils/nn.py",16647,0,"",python,selection_mouse +16,147465,"utils/nn.py",16719,0,"",python,selection_mouse +17,149012,"utils/nn.py",16758,0,"",python,selection_mouse +18,149017,"utils/nn.py",16757,0,"",python,selection_command +19,151452,"utils/nn.py",16719,0,"",python,selection_command +20,151603,"utils/nn.py",16717,0,"",python,selection_command +21,151965,"utils/nn.py",16719,0,"",python,selection_command +22,152096,"utils/nn.py",16757,0,"",python,selection_command +23,152242,"utils/nn.py",16788,0,"",python,selection_command +24,152388,"utils/nn.py",16815,0,"",python,selection_command +25,825613,"TERMINAL",0,0,"neofetch",,terminal_command +26,825662,"TERMINAL",0,0,"]633;E;2025-09-03 13:54:44 neofetch;6a06e7a8-9391-4d17-9f3a-a329f1fb7609]633;Cbash: neofetch: command not found...\r\n",,terminal_output +27,826759,"TERMINAL",0,0,"]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;127",,terminal_output +28,827175,"TERMINAL",0,0,"^C",,terminal_command +29,830682,"TERMINAL",0,0,"clear",,terminal_command +30,830700,"TERMINAL",0,0,"]633;E;2025-09-03 13:54:49 clear;6a06e7a8-9391-4d17-9f3a-a329f1fb7609]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +31,927507,"utils/nn.py",16509,0,"",python,selection_mouse +32,1174031,"utils/nn.py",15860,0,"",python,selection_mouse +33,1210327,"utils/nn.py",15366,0,"",python,selection_mouse +34,1224242,"utils/nn.py",15859,0,"",python,selection_mouse +35,1579802,"utils/nn.py",16437,0,"",python,selection_mouse +36,1580754,"utils/nn.py",16441,0,"",python,selection_mouse +37,1580757,"utils/nn.py",16440,0,"",python,selection_command +38,1581460,"utils/nn.py",16407,0,"",python,selection_mouse +39,2723196,"train_tokenizer.py",0,0,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n init_lr: float = 0.0\n max_lr: float = 3e-4\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 20000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n warmup_steps: int = 10000\n darkness_threshold: float = 0.0\n # Tokenizer\n model_dim: int = 512\n ffn_dim: int = 2048\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 4\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\n\ndef tokenizer_loss_fn(\n model: TokenizerVQVAE, inputs: dict\n) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n # --- Compute loss ---\n # FIXME (f.srambical): Can we even do native int8 training without casting the video at all?\n # FIXME (f.srambical): If the tokenizer is the reason for the dynamics model being memory-bound,\n # should we at least train the tokenizer natively in int8?\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n model.train()\n outputs = model(inputs, training=True)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n mse = jnp.square(gt - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n\n@nnx.jit\ndef train_step(\n tokenizer: TokenizerVQVAE, optimizer: nnx.Optimizer, inputs: dict\n) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(model: TokenizerVQVAE) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n return tokenizer_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n tokenizer\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n ffn_dim=args.ffn_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n\n _, params, _ = nnx.split(tokenizer, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n optimizer = nnx.Optimizer(tokenizer, tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n darkness_threshold=args.darkness_threshold,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in grain_iterator\n )\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n inputs = dict(videos=videos)\n loss, recon, metrics = train_step(tokenizer, optimizer, inputs)\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n grain_iterator # type: ignore\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +40,2728858,"train_tokenizer.py",1380,0,"",python,selection_mouse +41,2729005,"train_tokenizer.py",1373,10,"patch_size",python,selection_mouse +42,2746713,"train_lam.py",0,0,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom models.lam import LatentActionModel\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n vq_beta: float = 0.25\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n vq_reset_thresh: int = 50\n darkness_threshold: float = 0.0\n # LAM\n model_dim: int = 512\n ffn_dim: int = 2048\n latent_dim: int = 32\n num_latents: int = 6\n patch_size: int = 16\n num_blocks: int = 4\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.0\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_lam""\n tags: list[str] = field(default_factory=lambda: [""lam""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\n\ndef lam_loss_fn(\n model: LatentActionModel, inputs: dict\n) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n # --- Compute loss ---\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n model.train()\n outputs = model(inputs, training=True)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=(index_counts != 0).mean(),\n )\n return loss, (outputs[""recon""], index_counts, metrics)\n\n\n@nnx.jit\ndef train_step(\n lam: LatentActionModel,\n optimizer: nnx.Optimizer,\n inputs: dict,\n action_last_active: jax.Array,\n rng: jax.Array,\n) -> tuple[jax.Array, jax.Array, jax.Array, dict]:\n def loss_fn(\n model: LatentActionModel,\n ) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n return lam_loss_fn(model, inputs)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)\n\n # --- Reset inactive latent actions ---\n codebook = lam.vq.codebook\n num_codes = len(codebook)\n active_codes = idx_counts != 0.0\n action_last_active = jnp.where(active_codes, 0, action_last_active + 1)\n p_code = active_codes / active_codes.sum()\n reset_idxs = jax.random.choice(rng, num_codes, shape=(num_codes,), p=p_code)\n do_reset = action_last_active >= args.vq_reset_thresh\n new_codebook = jnp.where(\n jnp.expand_dims(do_reset, -1), codebook[reset_idxs], codebook.value\n )\n lam.vq.codebook.value = new_codebook\n action_last_active = jnp.where(do_reset, 0, action_last_active)\n return loss, recon, action_last_active, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n ffn_dim=args.ffn_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n\n # Count parameters\n _, params, _ = nnx.split(lam, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n optimizer = nnx.Optimizer(lam, tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n darkness_threshold=args.darkness_threshold,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in grain_iterator\n )\n print(f""Starting training from step {step}..."")\n action_last_active = jnp.zeros(args.num_latents, dtype=jnp.int32)\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng = jax.random.split(rng)\n\n inputs = dict(videos=videos, rng=_rng)\n rng, _rng = jax.random.split(rng)\n loss, recon, action_last_active, metrics = train_step(\n lam, optimizer, inputs, action_last_active, _rng\n )\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0, 1:].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n grain_iterator # type: ignore\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +43,2753802,"train_lam.py",1406,0,"",python,selection_mouse +44,5002152,"train_lam.py",1423,0,"",python,selection_mouse +45,5007085,"train_tokenizer.py",0,0,"",python,tab +46,5012806,"train_tokenizer.py",1403,0,"",python,selection_mouse +47,11335024,"TERMINAL",0,0,"",,terminal_focus +48,11337086,"TERMINAL",0,0,"source /home/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/bin/activate",,terminal_command +49,11337136,"TERMINAL",0,0,"]633;E;2025-09-03 16:49:56 source /home/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/bin/activate;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +50,11338099,"TERMINAL",0,0,"queue",,terminal_command +51,11338144,"TERMINAL",0,0,"]633;E;2025-09-03 16:49:57 queue;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C",,terminal_output +52,11338216,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Wed Sep 3 16:49:57 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3457966 accelerat train_la tum_cte0 PD\t0:00\t 8 (Resources)3457967 accelerat train_la tum_cte0 PD\t0:00\t 8 (Priority)3457968 accelerat train_la tum_cte0 PD\t0:00\t 8 (Priority)",,terminal_output +53,11339316,"TERMINAL",0,0,"8",,terminal_output +54,11339485,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +55,11342233,"TERMINAL",0,0,"idling",,terminal_command +56,11342301,"TERMINAL",0,0,"]633;E;2025-09-03 16:50:01 idling;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Wed Sep 3 16:50:01 2025Partition dev_cpuonly:\t 8 nodes idle\rPartition cpuonly: 19 nodes idle\rPartition dev_accelerated:\t 1 nodes idle\rPartition accelerated:\t 0 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 0 nodes idle\rPartition accelerated-h200:\t 0 nodes idle",,terminal_output +57,11343412,"TERMINAL",0,0,"2",,terminal_output +58,11344436,"TERMINAL",0,0,"3",,terminal_output +59,11345136,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +60,11348557,"TERMINAL",0,0,"idling",,terminal_command +61,11348642,"TERMINAL",0,0,"]633;E;2025-09-03 16:50:07 idling;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Wed Sep 3 16:50:07 2025Partition dev_cpuonly:\t 8 nodes idle\rPartition cpuonly: 19 nodes idle\rPartition dev_accelerated:\t 1 nodes idle\rPartition accelerated:\t 0 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 0 nodes idle\rPartition accelerated-h200:\t 0 nodes idle",,terminal_output +62,11349761,"TERMINAL",0,0,"8",,terminal_output +63,11350591,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +64,13842370,"TERMINAL",0,0,"queue",,terminal_command +65,13842462,"TERMINAL",0,0,"]633;E;2025-09-03 17:31:41 queue;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Wed Sep 3 17:31:41 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3457966 accelerat train_la tum_cte0 PD\t0:00\t 8 (Resources)3457967 accelerat train_la tum_cte0 PD\t0:00\t 8 (Priority)3457968 accelerat train_la tum_cte0 PD\t0:00\t 8 (Priority)",,terminal_output +66,13843118,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +67,13884392,"TERMINAL",0,0,"queue",,terminal_command +68,13884443,"TERMINAL",0,0,"]633;E;2025-09-03 17:32:23 queue;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C",,terminal_output +69,13884514,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Wed Sep 3 17:32:23 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3457966 accelerat train_la tum_cte0 PD\t0:00\t 8 (Resources)3457967 accelerat train_la tum_cte0 PD\t0:00\t 8 (Priority)3457968 accelerat train_la tum_cte0 PD\t0:00\t 8 (Priority)",,terminal_output +70,13885540,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +71,13888862,"TERMINAL",0,0,"fsacct_week",,terminal_command +72,13888895,"TERMINAL",0,0,"]633;E;2025-09-03 17:32:27 fsacct_week;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C JobID JobName Partition All State Elapsed Timelimit \r\n--------------- ------------------------------ ---------------- --- ------------ ---------- ---------- \r\n 3454890 train_lam_minecraft_1node_dev accelerated 24 REQUEUED 00:04:42 00:10:00 \r\n 3454917 train_lam_minecraft_1node_dev accelerated 24 REQUEUED 00:04:36 00:10:00 \r\n 3457966 train_lam_minecraft_8node_dar+ accelerated 0 PENDING 00:00:00 2-00:00:00 \r\n 3457967 train_lam_minecraft_8node_dar+ accelerated 0 PENDING 00:00:00 2-00:00:00 \r\n 3457968 train_lam_minecraft_8node_dar+ accelerated 0 PENDING 00:00:00 2-00:00:00 \r\n 3457969 train_lam_minecraft_8node_dar+ accelerated 192 FAILED 00:05:59 2-00:00:00 \r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +73,13893779,"TERMINAL",0,0,"logs",,terminal_command +74,13893806,"TERMINAL",0,0,"]633;E;2025-09-03 17:32:32 logs;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +75,13898899,"TERMINAL",0,0,"cd coinrun/",,terminal_command +76,13899210,"TERMINAL",0,0,"ls",,terminal_command +77,13899233,"TERMINAL",0,0,"]633;E;2025-09-03 17:32:38 ls;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;Cdynamics tokenizer\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun]633;D;0",,terminal_output +78,13903031,"TERMINAL",0,0,"cd ..",,terminal_command +79,13903047,"TERMINAL",0,0,"]633;E;2025-09-03 17:32:41 cd ..;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +80,13903314,"TERMINAL",0,0,"ls",,terminal_command +81,13903395,"TERMINAL",0,0,"]633;E;2025-09-03 17:32:42 ls;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;Catari train_dyn_causal_500M_3372972.log train_lam_action_space_scaling_12_3321527.log train_lam_action_space_scaling_8_3329806.log train_tokenizer_model_size_scaling_140M_3313562.log\r\nbig_run train_dyn_causal_500M_3373110.log train_lam_action_space_scaling_12_3329787.log train_lam_action_space_scaling_8_3331288.log train_tokenizer_model_size_scaling_140M_3316019.log\r\nbig-runs train_dyn_new_arch-bugfixed-spatial-shift_3359343.log train_lam_action_space_scaling_12_3329802.log train_lam_minecraft_overfit_sample_3309655.log train_tokenizer_model_size_scaling_200M_3313563.log\r\ncausal train_dyn_new_arch-bugfixed-temporal-shift_3359349.log train_lam_action_space_scaling_12_3331284.log train_lam_model_size_scaling_38M_3317098.log train_tokenizer_model_size_scaling_200M_3316020.log\r\ncoinrun train_dyn_yolorun_3333026.log train_lam_action_space_scaling_20_3318547.log train_lam_model_size_scaling_38M_3317115.log train_tokenizer_model_size_scaling_227M_3317234.log\r\nlam train_dyn_yolorun_3333448.log train_lam_action_space_scaling_20_3329788.log train_lam_model_size_scaling_38M_3317231.log train_tokenizer_model_size_scaling_227M_3318555.log\r\nmaskgit train_dyn_yolorun_3335345.log train_lam_action_space_scaling_20_3329803.log train_tokenizer_batch_size_scaling_16_node_3321526.log train_tokenizer_model_size_scaling_227M_3320173.log\r\nmaskgit-maskprob-fix train_dyn_yolorun_3335362.log train_lam_action_space_scaling_20_3331285.log train_tokenizer_batch_size_scaling_1_node_3318551.log train_tokenizer_model_size_scaling_227M_3321523.log\r\npreprocess train_dyn_yolorun_3348592.log train_lam_action_space_scaling_50_3320180.log train_tokenizer_batch_size_scaling_2_node_3318552.log train_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_dyn_causal_180M_3372931.log train_dyn_yolorun_new_arch_3351743.log train_lam_action_space_scaling_50_3329789.log train_tokenizer_batch_size_scaling_2_node_3330806.log train_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_dyn_causal_180M_3372963.log train_dyn_yolorun_new_arch_3352103.log train_lam_action_space_scaling_50_3329804.log train_tokenizer_batch_size_scaling_2_node_3330848.log train_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_dyn_causal_180M_3372969.log train_dyn_yolorun_new_arch_3352115.log train_lam_action_space_scaling_50_3331286.log train_tokenizer_batch_size_scaling_2_node_3331282.log train_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_dyn_causal_180M_3373107.log train_dyn_yolorun_new_arch_3358457.log train_lam_action_space_scaling_6_3318549.log train_tokenizer_batch_size_scaling_4_node_3318553.log train_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_dyn_causal_255M_3372932.log train_lam_action_space_scaling_10_3320179.log train_lam_action_space_scaling_6_3320178.log train_tokenizer_batch_size_scaling_4_node_3320175.log train_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_dyn_causal_255M_3372970.log train_lam_action_space_scaling_10_3321529.log train_lam_action_space_scaling_6_3321528.log train_tokenizer_batch_size_scaling_4_node_3321524.log train_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_dyn_causal_255M_3373108.log train_lam_action_space_scaling_10_3329786.log train_lam_action_space_scaling_6_3329790.log train_tokenizer_batch_size_scaling_8_node_3320176.log train_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_dyn_causal_356M_3372934.log train_lam_action_space_scaling_10_3329801.log train_lam_action_space_scaling_6_3329805.log train_tokenizer_batch_size_scaling_8_node_3321525.log train_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_dyn_causal_356M_3372971.log train_lam_action_space_scaling_10_3331283.log train_lam_action_space_scaling_6_3331287.log train_tokenizer_minecraft_overfit_sample_3309656.log train_tokenizer_model_size_scaling_80M_3316026.log\r\ntrain_dyn_causal_356M_3373109.log train_lam_action_space_scaling_12_3318546.log train_lam_action_space_scaling_8_3318550.log train_tokenizer_model_size_scaling_127M_3317233.log yoloruns\r\ntrain_dyn_causal_500M_3372936.log train_lam_action_space_scaling_12_3320177.log train_lam_action_space_scaling_8_3329791.log train_tokenizer_model_size_scaling_127M_3318554.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +82,13905306,"TERMINAL",0,0,"cd lam/",,terminal_command +83,13905646,"TERMINAL",0,0,"ls",,terminal_command +84,13905684,"TERMINAL",0,0,"]633;E;2025-09-03 17:32:44 ls;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C",,terminal_output +85,13905977,"TERMINAL",0,0,"train_lam_minecraft_8node_3431870.log train_lam_minecraft_8node_3431885.log train_lam_minecraft_8node_3454917.log train_lam_minecraft_8node_3454948.log train_lam_minecraft_8node_darkness_filter_37M_3454953.log\r\ntrain_lam_minecraft_8node_3431875.log train_lam_minecraft_8node_3431895.log train_lam_minecraft_8node_3454941.log train_lam_minecraft_8node_darkness_filter_133M_3454956.log train_lam_minecraft_8node_darkness_filter_37M_3457969.log\r\ntrain_lam_minecraft_8node_3431876.log train_lam_minecraft_8node_3454890.log train_lam_minecraft_8node_3454944.log train_lam_minecraft_8node_darkness_filter_311M_3454955.log train_lam_minecraft_8node_darkness_filter_400M_3454954.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +86,13931971,"TERMINAL",0,0,"ls -l | grep 3457969",,terminal_command +87,13932008,"TERMINAL",0,0,"]633;E;2025-09-03 17:33:10 ls -l | grep 3457969;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C-rw-rw----+ 1 tum_cte0515 hk-project-p0023960 58963 Sep 2 16:56 train_lam_minecraft_8node_darkness_filter_37M_3457969.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +88,13935482,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/train_lam_minecraft_8node_darkness_filter_37M_3457969.log",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node_darkness_filter_37M\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\n# slurm_job_id=$SLURM_JOB_ID\nslurm_job_id=3454953\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n --restore_ckpt \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --darkness_threshold=50 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-darkness-filter-37M-$slurm_job_id \\n --tags lam minecraft 8-node darkness-filter 37M \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n/var/spool/slurmd/job3457969/slurm_script: line 38: .venv/bin/activate: No such file or directory\nSLURM_JOB_USER=tum_cte0515\nSLURM_TASKS_PER_NODE=4(x8)\nSLURM_JOB_UID=999226\nSLURM_TASK_PID=1440498\nSLURM_JOB_GPUS=0,1,2,3\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs\nSLURMD_NODENAME=hkn0404\nSLURM_JOB_START_TIME=1756824635\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1756997435\nSLURM_CPUS_ON_NODE=24\nSLURM_JOB_CPUS_PER_NODE=24(x8)\nSLURM_GPUS_ON_NODE=4\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=accelerated\nSLURM_TRES_PER_TASK=cpu=5\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=8\nSLURM_JOBID=3457969\nSLURM_JOB_QOS=normal\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=5\nSLURM_NTASKS=32\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0404\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_NODELIST=hkn[0404,0407-0408,0417,0503,0725,0729,0732]\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=32\nSLURM_NNODES=8\nSLURM_SUBMIT_HOST=hkn1990.localdomain\nSLURM_JOB_ID=3457969\nSLURM_NODEID=0\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_lam_minecraft_8node_darkness_filter_37M\nSLURM_NTASKS_PER_NODE=4\nSLURM_JOB_GID=502226\nSLURM_JOB_NODELIST=hkn[0404,0407-0408,0417,0503,0725,0729,0732]\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\n2025-09-02 16:51:32.979826: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\n2025-09-02 16:51:32.979716: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\n2025-09-02 16:51:32.979963: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\n2025-09-02 16:51:32.979958: E external/xla/xla/stream_executor/cuda/cuda_platform.cc:51] failed call to cuInit: INTERNAL: CUDA error: Failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 742, in backends\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = _init_backend(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 828, in _init_backend\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n backend = registration.factory()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 528, in factory\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jaxlib/xla_client.py"", line 153, in make_c_api_client\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_lam.py"", line 153, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_lam.py"", line 153, in \n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_lam.py"", line 153, in \n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return _xla.get_c_api_client(plugin_name, options, distributed_client)\njaxlib._jax.XlaRuntimeError: FAILED_PRECONDITION: No visible GPU devices.\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/train_lam.py"", line 153, in \n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n num_devices = jax.device_count()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 907, in device_count\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return int(get_backend(backend).device_count())\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 876, in get_backend\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n return _get_backend_uncached(platform)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 855, in _get_backend_uncached\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n bs = backends()\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 758, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': FAILED_PRECONDITION: No visible GPU devices. (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n2025-09-02 16:56:33.847340: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1798] Shutdown barrier in coordination service has failed:\nDEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\nThis suggests that the workers are out of sync. Either at least one worker (a) crashed early due to program error or scheduler events (e.g. preemption, eviction), (b) was too fast in its execution, or (c) too slow / hanging. Check the logs (both the program and scheduler events) for an earlier error to identify the root cause.\n2025-09-02 16:56:33.847379: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1839] Use error polling to propagate the following error to all tasks: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.847562: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.847711: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.847756: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.847765: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.847814: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.847865: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.848049: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848309: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.848088: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848204: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848026: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.848100: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848254: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848027: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848062: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848225: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848211: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848202: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848085: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848183: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848217: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848280: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848218: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848263: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848329: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848260: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848395: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848441: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848796: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.848399: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848611: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.848439: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848749: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848760: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.848398: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.848833: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849068: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.849200: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849066: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.849198: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.849003: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849035: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.849080: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849229: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.849107: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.849359: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.849266: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849228: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.849209: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.849270: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849223: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\n2025-09-02 16:56:33.849264: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849408: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849271: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849221: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849411: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849434: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849435: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.849434: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::2947760685903380055::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 4/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:1. Some timed out task names:\n/job:jax_worker/replica:0/task:13\n/job:jax_worker/replica:0/task:20\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1dShutdown::2947760685903380055']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-09-02 16:56:33.931086: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.931108: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1756824993.930994068"",""description"":""Error received from peer ipv4:10.0.1.36:62385"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\n2025-09-02 16:56:33.931120: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:429] Polled an error from coordination service (this can be an error from this or another task).\n2025-09-02 16:56:33.931151: F external/xla/xla/pjrt/distributed/client.h:88] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: Failed to send RPC to coordination service. Either the leader task was preempted/died/restarted unexpectedly or this task is experiencing network issues. Check earlier logs from 1) this task, 2) the leader (usually slice 0 task 0), and 3) cluster scheduler to debug further.\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/PollForError:\n:{""created"":""@1756824993.931007203"",""description"":""Error received from peer ipv4:10.0.1.36:62385"",""file"":""external/com_github_grpc_grpc/src/core/lib/surface/call.cc"",""file_line"":1056,""grpc_message"":""Socket closed"",""grpc_status"":14}\nsrun: error: hkn0404: tasks 0-3: Aborted (core dumped)\nsrun: error: hkn0732: tasks 29-30: Aborted (core dumped)\nsrun: error: hkn0407: tasks 4,7: Aborted (core dumped)\nsrun: error: hkn0725: tasks 21,23: Aborted (core dumped)\nsrun: error: hkn0408: tasks 9-10: Aborted (core dumped)\nsrun: error: hkn0729: tasks 24-25: Aborted (core dumped)\nsrun: error: hkn0417: tasks 14-15: Aborted (core dumped)\nsrun: error: hkn0503: tasks 16,18: Aborted (core dumped)\nsrun: error: hkn0407: task 6: Aborted (core dumped)\nsrun: error: hkn0732: task 31: Aborted (core dumped)\nsrun: error: hkn0725: task 22: Aborted (core dumped)\nsrun: error: hkn0729: task 26: Aborted (core dumped)\nsrun: error: hkn0503: task 17: Aborted (core dumped)\nsrun: error: hkn0408: task 11: Aborted (core dumped)\nsrun: error: hkn0417: task 12: Aborted (core dumped)\nsrun: error: hkn0732: task 28: Aborted (core dumped)\nsrun: error: hkn0725: task 20: Aborted (core dumped)\nsrun: error: hkn0407: task 5: Aborted (core dumped)\nsrun: error: hkn0408: task 8: Aborted (core dumped)\nsrun: error: hkn0417: task 13: Aborted (core dumped)\nsrun: error: hkn0729: task 27: Aborted (core dumped)\nsrun: error: hkn0503: task 19: Aborted (core dumped)\n\n============================= JOB FEEDBACK =============================\n\nJob ID: 3457969\nCluster: hk\nUser/Group: tum_cte0515/hk-project-p0023960\nAccount: hk-project-p0023960\nState: FAILED (exit code 134)\nPartition: accelerated\nNodes: 8\nCores per node: 24\nNodelist: hkn[0404,0407-0408,0417,0503,0725,0729,0732]\nCPU Utilized: 00:02:37\nCPU Efficiency: 0.23% of 19:08:48 core-walltime\nJob Wall-clock time: 00:05:59\nStarttime: Tue Sep 2 16:50:35 2025\nEndtime: Tue Sep 2 16:56:34 2025\nMemory Utilized: 12.17 GB (estimated maximum)\nMemory Efficiency: 0.00% of 0.00 MB (0.00 MB/node)\nEnergy Consumed: 3157333 Joule / 877.036944444444 Watthours\nAverage node power draw: 8794.79944289694 Watt\n",log,tab +89,13935894,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/train_lam_minecraft_8node_darkness_filter_37M_3457969.log",508,0,"",log,selection_mouse +90,13935910,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/train_lam_minecraft_8node_darkness_filter_37M_3457969.log",507,0,"",log,selection_command +91,13936594,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/train_lam_minecraft_8node_darkness_filter_37M_3457969.log",58963,0,"",log,selection_command +92,13967121,"TERMINAL",0,0,"runner",,terminal_command +93,13979199,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-^Crkness-filter-base.sbatch",,terminal_command +94,13979221,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]633;E;;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D",,terminal_output +95,13987679,"TERMINAL",0,0,"cat slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",,terminal_command +96,13987708,"TERMINAL",0,0,"]633;E;2025-09-03 17:34:06 cat slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=8\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=48:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:4\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\r\n#SBATCH --job-name=train_lam_minecraft_8node_darkness_filter_37M\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# --- signal trap to requeue job before timeout ---\r\nrequeue_job() {\r\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\r\n scontrol requeue $SLURM_JOB_ID\r\n exit 0\r\n}\r\n\r\ntrap requeue_job sigusr1\r\n\r\n# set checkpoint flag based on restart count\r\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\r\n\r\nif [ $restart_count -eq 0 ]; then\r\n restore_ckpt_flag=""--no-restore-ckpt""\r\nelse\r\n restore_ckpt_flag=""--restore-ckpt""\r\nfi\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\n\r\njob_name=$SLURM_JOB_NAME\r\n# slurm_job_id=$SLURM_JOB_ID\r\nslurm_job_id=3454953\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nenv | grep SLURM\r\n\r\nsrun python train_lam.py \\r\n --save_ckpt \\r\n --restore_ckpt \\r\n --wandb_id $slurm_job_id \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=160 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --darkness_threshold=50 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --log_checkpoint_interval=1000 \\r\n --name=lam-minecraft-8-node-darkness-filter-37M-$slurm_job_id \\r\n --tags lam minecraft 8-node darkness-filter 37M \\r\n --entity instant-uv \\r\n --project jafar \\r\n --num_latents=100 \\r\n --data_dir $array_records_dir &\r\n\r\nchild_pid=$!\r\n\r\nwait $child_pid\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +97,13996678,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",,terminal_command +98,13996728,"TERMINAL",0,0,"]633;E;2025-09-03 17:34:15 sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C",,terminal_output +99,13996951,"TERMINAL",0,0,"Submitted batch job 3463210\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +100,13999265,"TERMINAL",0,0,"queue",,terminal_command +101,13999316,"TERMINAL",0,0,"]633;E;2025-09-03 17:34:18 queue;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C",,terminal_output +102,13999389,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Wed Sep 3 17:34:18 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3457966 accelerat train_la tum_cte0 PD\t0:00\t 8 (Resources)3457967 accelerat train_la tum_cte0 PD\t0:00\t 8 (Priority)3457968 accelerat train_la tum_cte0 PD\t0:00\t 8 (Priority)3463210 accelerat train_la tum_cte0 PD\t0:00\t 8 (Priority)",,terminal_output +103,14000077,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +104,14215026,"TERMINAL",0,0,"cd ..",,terminal_command +105,14215299,"TERMINAL",0,0,"ls",,terminal_command +106,14215337,"TERMINAL",0,0,"]633;E;2025-09-03 17:37:54 ls;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;Ccheckpoints jafar jafar_jobs jafar_jobs_2 jafar_jobs_3 tmp\r\n]0;tum_cte0515@hkn1993:~/Projects]633;D;0",,terminal_output +107,14222012,"TERMINAL",0,0,"cd ls",,terminal_command +108,14223194,"TERMINAL",0,0,"ls",,terminal_command +109,14227747,"TERMINAL",0,0,"mv jafar jasmine",,terminal_command +110,14227770,"TERMINAL",0,0,"]633;E;2025-09-03 17:38:06 mv jafar jasmine;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C]0;tum_cte0515@hkn1993:~/Projects]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects",,terminal_output +111,14228805,"TERMINAL",0,0,"ls",,terminal_command +112,14228820,"TERMINAL",0,0,"]633;E;2025-09-03 17:38:07 ls;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;Ccheckpoints jafar_jobs jafar_jobs_2 jafar_jobs_3 jasmine tmp\r\n]0;tum_cte0515@hkn1993:~/Projects]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects",,terminal_output +113,14241517,"TERMINAL",0,0,"mv jafar_jobs jasmine_jobs",,terminal_command +114,14241539,"TERMINAL",0,0,"]633;E;2025-09-03 17:38:20 mv jafar_jobs jasmine_jobs;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C]0;tum_cte0515@hkn1993:~/Projects]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects",,terminal_output +115,14247458,"TERMINAL",0,0,"mv jafar_jobs_2 jasmine_jobs_2",,terminal_command +116,14247470,"TERMINAL",0,0,"]633;E;2025-09-03 17:38:26 mv jafar_jobs_2 jasmine_jobs_2;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C]0;tum_cte0515@hkn1993:~/Projects]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects",,terminal_output +117,14252310,"TERMINAL",0,0,"mv jafar_jobs_3 jasmine_jobs_3",,terminal_command +118,14252339,"TERMINAL",0,0,"]633;E;2025-09-03 17:38:31 mv jafar_jobs_3 jasmine_jobs_3;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C]0;tum_cte0515@hkn1993:~/Projects]633;D;0",,terminal_output +119,14269977,"TERMINAL",0,0,"vim ~/.bashrc",,terminal_command +120,14270165,"TERMINAL",0,0,"]633;E;2025-09-03 17:38:48 vim ~/.bashrc;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C[?1049h[>4;2m[?1h=[?2004h[?1004h[?12h[?12l[?25l""~/.bashrc"" 45L, 2218B▽ Pzz\[0%m [>c]10;?]11;?# .bashrc# Source global definitionsif [ -f /etc/bashrc ]; then\r\n . /etc/bashrc\r\nfi\r\n\r\n# User specific environment\r\nif ! [[ ""$PATH"" =~ ""$HOME/.local/bin:$HOME/bin:"" ]]\r\nthen\r\n PATH=""$HOME/.local/bin:$HOME/bin:$PATH""\r\nfi\r\nexport PATH\r\n\r\nalias idle='sinfo_t_idle'\r\nexport ws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\nalias salloc_node='salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8'\r\nalias salloc_cpu='salloc --time=01:00:00 --partition=dev_cpuonly --nodes=1 --cpus-per-task=128'\r\nalias sync-runner=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jafar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""\r\nalias sync-runner-2=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jafar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""\r\nalias sync-runner-3=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jafar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""\r\nalias runner=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""\r\nalias runner-2=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""\r\nalias runner-3=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""\r\nalias dev=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar/""\r\nalias smi=""watch -n1 nvidia-smi""\r\nalias idling=""watch -n1 sinfo_t_idle""\r\nalias queue='watch -n1 squeue --me'\r\nalias fqueue='watch -n 1 ""squeue -o \""%.10i %.16P %.30j %.8u %.8T %.10M %.9l %.6D %R\""""'\r\nalias fsacct_week='sacct --format=""JobID%15,JobName%30,Partition%16,AllocCPUS%3,State%12,Elapsed%10,Timelimit%10"" --starttime $(date -d ""last week"" +%Y-%m-%d) | grep -vE ""*.batch|*.extern|*.inter|bash|python|CANCELLED|echo""'\r\nalias logs=""cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir""overlapper() {\r\n if [ -z ""$1"" ]; thenecho ""Usage: overlap ""21,87Top[?25hP+q436f\P+q6b75\P+q6b64\P+q6b72\P+q6b6c\P+q2332\P+q2334\P+q2569\P+q2a37\P+q6b31\[?12$p[?25l/3333/3333 [?25h[?25l/f6f6/e3e3 [?25h",,terminal_output +121,14270693,"TERMINAL",0,0,"[?25lj 2,76[?25h",,terminal_output +122,14271318,"TERMINAL",0,0,"[?25lk 1,87[?25h",,terminal_output +123,14271444,"TERMINAL",0,0,"[?25lk 0[?25h",,terminal_output +124,14271632,"TERMINAL",0,0,"[?25lk 19[?25h",,terminal_output +125,14272005,"TERMINAL",0,0,"[?25lj 20[?25h",,terminal_output +126,14272312,"TERMINAL",0,0,"[?25lk 19[?25h",,terminal_output +127,14272487,"TERMINAL",0,0,"[?25lh 6[?25h",,terminal_output +128,14272999,"TERMINAL",0,0,"[?25lh 5[?25h",,terminal_output +129,14273089,"TERMINAL",0,0,"[?25lh 4[?25h[?25lh 3[?25h[?25lh 2[?25h",,terminal_output +130,14273337,"TERMINAL",0,0,"[?25lh 1[?25h",,terminal_output +131,14273579,"TERMINAL",0,0,"[?25lh 0[?25h",,terminal_output +132,14273744,"TERMINAL",0,0,"[?25lh 79[?25h",,terminal_output +133,14273916,"TERMINAL",0,0,"[?25lh 8[?25h",,terminal_output +134,14274134,"TERMINAL",0,0,"[?25ll 9[?25h",,terminal_output +135,14274615,"TERMINAL",0,0,"[?25l80[?25h[?25ll 1[?25h",,terminal_output +136,14274788,"TERMINAL",0,0,"[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h",,terminal_output +137,14274900,"TERMINAL",0,0,"[?25lh 5[?25h",,terminal_output +138,14275390,"TERMINAL",0,0,"[?25lh 4[?25h[?25lh 3[?25h",,terminal_output +139,14275466,"TERMINAL",0,0,"[?25lh 2[?25h",,terminal_output +140,14275533,"TERMINAL",0,0,"[?25ll 3[?25h",,terminal_output +141,14276029,"TERMINAL",0,0,"[?25ll 4[?25h",,terminal_output +142,14276493,"TERMINAL",0,0,"[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 90[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h",,terminal_output +143,14277163,"TERMINAL",0,0,"[?25ll 100[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 10[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 20[?25h[?25ll 1[?25h",,terminal_output +144,14277342,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +145,14277515,"TERMINAL",0,0,"[?25ll 3[?25h",,terminal_output +146,14277645,"TERMINAL",0,0,"[?25ll 4[?25h",,terminal_output +147,14277808,"TERMINAL",0,0,"[?25ll 5[?25h",,terminal_output +148,14278061,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +149,14278146,"TERMINAL",0,0,"[?25lw /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""[?25h",,terminal_output +150,14278283,"TERMINAL",0,0,"[?25li -- INSERT --19,125Top[?25h",,terminal_output +151,14278835,"TERMINAL",0,0,"[?25ls/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""6[?25h",,terminal_output +152,14279315,"TERMINAL",0,0,"[?25lm/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""7[?25h",,terminal_output +153,14279438,"TERMINAL",0,0,"[?25li/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""8[?25h",,terminal_output +154,14279506,"TERMINAL",0,0,"[?25ln/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""9[?25h",,terminal_output +155,14279618,"TERMINAL",0,0,"[?25le/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""30[?25h",,terminal_output +156,14280680,"TERMINAL",0,0,"[?25l^[",,terminal_output +157,14280808,"TERMINAL",0,0," 19,129Top[?25h",,terminal_output +158,14280874,"TERMINAL",0,0,"[?25lj 20[?25h",,terminal_output +159,14281422,"TERMINAL",0,0,"[?25lh 8[?25h",,terminal_output +160,14281608,"TERMINAL",0,0,"[?25lh 7[?25h",,terminal_output +161,14281769,"TERMINAL",0,0,"[?25lh 6[?25h",,terminal_output +162,14282347,"TERMINAL",0,0,"[?25ll 7[?25h",,terminal_output +163,14283891,"TERMINAL",0,0,"[?25l^V1x1-- VISUAL BLOCK --1x120,127Top[?25h",,terminal_output +164,14284136,"TERMINAL",0,0,"[?25lj 2x1f1[?25h",,terminal_output +165,14285649,"TERMINAL",0,0,"[?25lI f20,127Top-- INSERT --20,127Top[?25h",,terminal_output +166,14286567,"TERMINAL",0,0,"[?25lar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""[?25h",,terminal_output +167,14286733,"TERMINAL",0,0,"[?25lr /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""[?25h",,terminal_output +168,14287232,"TERMINAL",0,0,"[?25l /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""[?25h",,terminal_output +169,14287921,"TERMINAL",0,0,"[?25ls /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""8[?25h[?25lm /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""9[?25h",,terminal_output +170,14288158,"TERMINAL",0,0,"[?25li /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""30[?25h[?25ln /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""1[?25h",,terminal_output +171,14288272,"TERMINAL",0,0,"[?25le /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""2[?25h",,terminal_output +172,14288679,"TERMINAL",0,0,"[?25l^[",,terminal_output +173,14288745,"TERMINAL",0,0," smfar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""20,127Top[?25h[?25lj 1[?25h",,terminal_output +174,14290420,"TERMINAL",0,0,"[?25li -- INSERT --21,127Top[?25h",,terminal_output +175,14293082,"TERMINAL",0,0,"[?25l^[",,terminal_output +176,14293171,"TERMINAL",0,0," 21,126Top[?25h",,terminal_output +177,14293738,"TERMINAL",0,0,"[?25lu2 changes; before #3 7 seconds agofar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""far /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""20,127Top20,127Top[?25h",,terminal_output +178,14295113,"TERMINAL",0,0,"[?25l^V1x1-- VISUAL BLOCK --1x120,127Top[?25h",,terminal_output +179,14295324,"TERMINAL",0,0,"[?25lj 2x1f1[?25h",,terminal_output +180,14295635,"TERMINAL",0,0,"[?25lI f20,127Top-- INSERT --20,127Top[?25h",,terminal_output +181,14299317,"TERMINAL",0,0,"[?25lar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""[?25h",,terminal_output +182,14299536,"TERMINAL",0,0,"[?25lr /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""[?25h",,terminal_output +183,14299649,"TERMINAL",0,0,"[?25l /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""[?25h",,terminal_output +184,14300458,"TERMINAL",0,0,"[?25lj /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""8[?25h",,terminal_output +185,14300531,"TERMINAL",0,0,"[?25la /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""9[?25h",,terminal_output +186,14300657,"TERMINAL",0,0,"[?25ls /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""30[?25h",,terminal_output +187,14300733,"TERMINAL",0,0,"[?25lm /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""1[?25h",,terminal_output +188,14300919,"TERMINAL",0,0,"[?25li /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""2[?25h[?25ln /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""3[?25h",,terminal_output +189,14301040,"TERMINAL",0,0,"[?25le /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""4[?25h",,terminal_output +190,14303213,"TERMINAL",0,0,"[?25l /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""3[?25h",,terminal_output +191,14303392,"TERMINAL",0,0,"[?25l /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""2[?25h",,terminal_output +192,14303574,"TERMINAL",0,0,"[?25l /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""1[?25h",,terminal_output +193,14303738,"TERMINAL",0,0,"[?25l /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""0[?25h",,terminal_output +194,14303958,"TERMINAL",0,0,"[?25l /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""29[?25h",,terminal_output +195,14304160,"TERMINAL",0,0,"[?25l /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""8[?25h",,terminal_output +196,14304373,"TERMINAL",0,0,"[?25l /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""7[?25h",,terminal_output +197,14304620,"TERMINAL",0,0,"[?25ls /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""8[?25h",,terminal_output +198,14304686,"TERMINAL",0,0,"[?25lm /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""9[?25h",,terminal_output +199,14304938,"TERMINAL",0,0,"[?25li /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""30[?25h[?25ln /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""1[?25h",,terminal_output +200,14305078,"TERMINAL",0,0,"[?25le /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_2/""2[?25h",,terminal_output +201,14305499,"TERMINAL",0,0,"[?25l^[",,terminal_output +202,14305609,"TERMINAL",0,0," smfar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""20,127Top[?25h",,terminal_output +203,14306761,"TERMINAL",0,0,"[?25lj 1[?25h",,terminal_output +204,14309241,"TERMINAL",0,0,"[?25lx dl mfar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""[?25h",,terminal_output +205,14309567,"TERMINAL",0,0,"[?25lx dl far /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""[?25h",,terminal_output +206,14309816,"TERMINAL",0,0,"[?25lx dl ar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""[?25h",,terminal_output +207,14309954,"TERMINAL",0,0,"[?25lx dl r /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""[?25h",,terminal_output +208,14310383,"TERMINAL",0,0,"[?25lx dl  /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""[?25h",,terminal_output +209,14310529,"TERMINAL",0,0,"[?25li -- INSERT --21,127Top[?25h",,terminal_output +210,14310749,"TERMINAL",0,0,"[?25ls /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""8[?25h",,terminal_output +211,14311364,"TERMINAL",0,0,"[?25lm /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""9[?25h",,terminal_output +212,14311535,"TERMINAL",0,0,"[?25li /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""30[?25h[?25ln /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""1[?25h",,terminal_output +213,14311657,"TERMINAL",0,0,"[?25le /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs_3/""2[?25h",,terminal_output +214,14312073,"TERMINAL",0,0,"[?25l^[",,terminal_output +215,14312164,"TERMINAL",0,0," 21,131Top[?25h",,terminal_output +216,14312303,"TERMINAL",0,0,"[?25lk 0[?25h",,terminal_output +217,14312799,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +218,14313464,"TERMINAL",0,0,"[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h",,terminal_output +219,14314318,"TERMINAL",0,0,"[?25ll 9[?25h[?25ll 40[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 50[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 60[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h",,terminal_output +220,14314572,"TERMINAL",0,0,"[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 70[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h",,terminal_output +221,14314640,"TERMINAL",0,0,"[?25ll 6[?25h",,terminal_output +222,14314765,"TERMINAL",0,0,"[?25ll 7[?25h",,terminal_output +223,14314928,"TERMINAL",0,0,"[?25ll 8[?25h",,terminal_output +224,14315033,"TERMINAL",0,0,"[?25ll 9[?25h",,terminal_output +225,14315258,"TERMINAL",0,0,"[?25ll 80[?25h",,terminal_output +226,14315372,"TERMINAL",0,0,"[?25ll 1[?25h",,terminal_output +227,14315784,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +228,14316136,"TERMINAL",0,0,"[?25lx dl ar_jobs_2/""[?25h",,terminal_output +229,14316289,"TERMINAL",0,0,"[?25lx dl r_jobs_2/""[?25h",,terminal_output +230,14316480,"TERMINAL",0,0,"[?25lx dl _jobs_2/""[?25h",,terminal_output +231,14316621,"TERMINAL",0,0,"[?25li -- INSERT --20,182Top[?25h",,terminal_output +232,14316748,"TERMINAL",0,0,"[?25ls_jobs_2/""3[?25h",,terminal_output +233,14317302,"TERMINAL",0,0,"[?25lm_jobs_2/""4[?25h",,terminal_output +234,14317425,"TERMINAL",0,0,"[?25li_jobs_2/""5[?25h",,terminal_output +235,14317478,"TERMINAL",0,0,"[?25ln_jobs_2/""6[?25h",,terminal_output +236,14317546,"TERMINAL",0,0,"[?25le_jobs_2/""7[?25h",,terminal_output +237,14317868,"TERMINAL",0,0,"[?25l^[",,terminal_output +238,14317953,"TERMINAL",0,0," 20,186Top[?25h",,terminal_output +239,14318006,"TERMINAL",0,0,"[?25lj 1[?25h",,terminal_output +240,14318238,"TERMINAL",0,0,"[?25lh 5[?25h",,terminal_output +241,14318419,"TERMINAL",0,0,"[?25lh 4[?25h",,terminal_output +242,14318540,"TERMINAL",0,0,"[?25lh 3[?25h",,terminal_output +243,14318735,"TERMINAL",0,0,"[?25lh 2[?25h",,terminal_output +244,14318899,"TERMINAL",0,0,"[?25lx dl ar_jobs_3/""[?25h",,terminal_output +245,14319050,"TERMINAL",0,0,"[?25lx dl r_jobs_3/""[?25h",,terminal_output +246,14319259,"TERMINAL",0,0,"[?25lx dl _jobs_3/""[?25h",,terminal_output +247,14319385,"TERMINAL",0,0,"[?25li -- INSERT --21,182Top[?25h",,terminal_output +248,14319580,"TERMINAL",0,0,"[?25ls_jobs_3/""3[?25h",,terminal_output +249,14319905,"TERMINAL",0,0,"[?25lm_jobs_3/""4[?25h",,terminal_output +250,14320126,"TERMINAL",0,0,"[?25li_jobs_3/""5[?25h[?25ln_jobs_3/""6[?25h",,terminal_output +251,14320191,"TERMINAL",0,0,"[?25le_jobs_3/""7[?25h",,terminal_output +252,14320470,"TERMINAL",0,0,"[?25l^[",,terminal_output +253,14320538,"TERMINAL",0,0," 21,186Top[?25h",,terminal_output +254,14323643,"TERMINAL",0,0,"[?25lk 0[?25h",,terminal_output +255,14323848,"TERMINAL",0,0,"[?25lk 19[?25h",,terminal_output +256,14325131,"TERMINAL",0,0,"[?25ly[?25h",,terminal_output +257,14325285,"TERMINAL",0,0,"[?25ly [?25h",,terminal_output +258,14325353,"TERMINAL",0,0,"[?25lp alias sync-runner=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""20,1Top[?25h",,terminal_output +259,14326223,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +260,14326909,"TERMINAL",0,0,"[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 10[?25h",,terminal_output +261,14327067,"TERMINAL",0,0,"[?25ll 1[?25h",,terminal_output +262,14327236,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +263,14327360,"TERMINAL",0,0,"[?25ll 3[?25h",,terminal_output +264,14327680,"TERMINAL",0,0,"[?25ll 4[?25h",,terminal_output +265,14327852,"TERMINAL",0,0,"[?25ll 5[?25h",,terminal_output +266,14327918,"TERMINAL",0,0,"[?25ll 6[?25h",,terminal_output +267,14328104,"TERMINAL",0,0,"[?25ll 7[?25h",,terminal_output +268,14328397,"TERMINAL",0,0,"[?25lh 6[?25h",,terminal_output +269,14328576,"TERMINAL",0,0,"[?25lh 5[?25h",,terminal_output +270,14328722,"TERMINAL",0,0,"[?25lh 4[?25h",,terminal_output +271,14328891,"TERMINAL",0,0,"[?25lh 3[?25h",,terminal_output +272,14329015,"TERMINAL",0,0,"[?25lh 2[?25h",,terminal_output +273,14329378,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +274,14329496,"TERMINAL",0,0,"[?25lw =""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""[?25h",,terminal_output +275,14329575,"TERMINAL",0,0,"[?25li -- INSERT --20,12Top[?25h",,terminal_output +276,14330152,"TERMINAL",0,0,"[?25lj=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""3[?25h",,terminal_output +277,14330645,"TERMINAL",0,0,"[?25la=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""4[?25h",,terminal_output +278,14330772,"TERMINAL",0,0,"[?25lf=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""5[?25h",,terminal_output +279,14330889,"TERMINAL",0,0,"[?25la=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""6[?25h",,terminal_output +280,14331074,"TERMINAL",0,0,"[?25lr=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""7[?25h",,terminal_output +281,14331620,"TERMINAL",0,0,"[?25l^[",,terminal_output +282,14331729,"TERMINAL",0,0," 20,16Top[?25h",,terminal_output +283,14332039,"TERMINAL",0,0,"[?25ll 7[?25h",,terminal_output +284,14332785,"TERMINAL",0,0,"[?25ll 8[?25h[?25ll 9[?25h[?25ll 20[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h",,terminal_output +285,14332949,"TERMINAL",0,0,"[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25l30[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h",,terminal_output +286,14333116,"TERMINAL",0,0,"[?25ll 4[?25h",,terminal_output +287,14333247,"TERMINAL",0,0,"[?25ll 5[?25h",,terminal_output +288,14333472,"TERMINAL",0,0,"[?25ll 6[?25h",,terminal_output +289,14333876,"TERMINAL",0,0,"[?25ll 7[?25h",,terminal_output +290,14334940,"TERMINAL",0,0,"[?25ll 8[?25h[?25ll 9[?25h[?25ll 40[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 50[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 60[?25h",,terminal_output +291,14335203,"TERMINAL",0,0,"[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h",,terminal_output +292,14336335,"TERMINAL",0,0,"[?25ll 8[?25h[?25ll 9[?25h[?25ll 70[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 80[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 90[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 100[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h",,terminal_output +293,14336401,"TERMINAL",0,0,"[?25ll 7[?25h",,terminal_output +294,14336514,"TERMINAL",0,0,"[?25ll 8[?25h",,terminal_output +295,14336729,"TERMINAL",0,0,"[?25ll 9[?25h",,terminal_output +296,14336985,"TERMINAL",0,0,"[?25ll 10[?25h",,terminal_output +297,14337643,"TERMINAL",0,0,"[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h",,terminal_output +298,14337831,"TERMINAL",0,0,"[?25ll 7[?25h",,terminal_output +299,14338169,"TERMINAL",0,0,"[?25ll 8[?25h",,terminal_output +300,14338268,"TERMINAL",0,0,"[?25ll 9[?25h",,terminal_output +301,14338424,"TERMINAL",0,0,"[?25ll 20[?25h",,terminal_output +302,14338608,"TERMINAL",0,0,"[?25ll 1[?25h",,terminal_output +303,14338729,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +304,14340308,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +305,14340361,"TERMINAL",0,0,"[?25lw /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""[?25h",,terminal_output +306,14340663,"TERMINAL",0,0,"[?25li -- INSERT --20,122Top[?25h",,terminal_output +307,14341085,"TERMINAL",0,0,"[?25lj/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""3[?25h",,terminal_output +308,14341255,"TERMINAL",0,0,"[?25la/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""4[?25h",,terminal_output +309,14341441,"TERMINAL",0,0,"[?25lf/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""5[?25h",,terminal_output +310,14341522,"TERMINAL",0,0,"[?25la/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""6[?25h",,terminal_output +311,14341770,"TERMINAL",0,0,"[?25lr/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""7[?25h",,terminal_output +312,14342368,"TERMINAL",0,0,"[?25l /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""8[?25h",,terminal_output +313,14342597,"TERMINAL",0,0,"[?25l^[",,terminal_output +314,14342699,"TERMINAL",0,0," 20,127Top[?25h[?25lk 19[?25h",,terminal_output +315,14343080,"TERMINAL",0,0,"[?25ll 8[?25h",,terminal_output +316,14343253,"TERMINAL",0,0,"[?25ll 9[?25h",,terminal_output +317,14343411,"TERMINAL",0,0,"[?25ll 30[?25h",,terminal_output +318,14343855,"TERMINAL",0,0,"[?25li -- INSERT --19,130Top[?25h",,terminal_output +319,14343966,"TERMINAL",0,0,"[?25l /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""1[?25h",,terminal_output +320,14344095,"TERMINAL",0,0,"[?25l^[",,terminal_output +321,14344184,"TERMINAL",0,0," 19,130Top[?25h",,terminal_output +322,14344364,"TERMINAL",0,0,"[?25lj 20[?25h",,terminal_output +323,14344492,"TERMINAL",0,0,"[?25ll 1[?25h",,terminal_output +324,14345682,"TERMINAL",0,0,"[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 40[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 50[?25h[?25ll 1[?25h[?25ll 2[?25h",,terminal_output +325,14346366,"TERMINAL",0,0,"[?25ll 3[?25h[?25ll 4[?25h[?25l5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 60[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 70[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h",,terminal_output +326,14346429,"TERMINAL",0,0,"[?25ll 8[?25h",,terminal_output +327,14346873,"TERMINAL",0,0,"[?25ll 9[?25h",,terminal_output +328,14347186,"TERMINAL",0,0,"[?25lk 19[?25h",,terminal_output +329,14347505,"TERMINAL",0,0,"[?25lh 8[?25h",,terminal_output +330,14347652,"TERMINAL",0,0,"[?25ll 9[?25h",,terminal_output +331,14347851,"TERMINAL",0,0,"[?25ll 80[?25h",,terminal_output +332,14348095,"TERMINAL",0,0,"[?25lx dl ar_jobs/""[?25h",,terminal_output +333,14348289,"TERMINAL",0,0,"[?25lx dl r_jobs/""[?25h",,terminal_output +334,14348491,"TERMINAL",0,0,"[?25lx dl _jobs/""[?25h",,terminal_output +335,14348557,"TERMINAL",0,0,"[?25li -- INSERT --19,180Top[?25h",,terminal_output +336,14348770,"TERMINAL",0,0,"[?25ls_jobs/""1[?25h",,terminal_output +337,14348918,"TERMINAL",0,0,"[?25lm_jobs/""2[?25h",,terminal_output +338,14349071,"TERMINAL",0,0,"[?25li_jobs/""3[?25h",,terminal_output +339,14349139,"TERMINAL",0,0,"[?25ln_jobs/""4[?25h",,terminal_output +340,14349207,"TERMINAL",0,0,"[?25le_jobs/""5[?25h",,terminal_output +341,14349575,"TERMINAL",0,0,"[?25l^[",,terminal_output +342,14349684,"TERMINAL",0,0," 19,184Top[?25h",,terminal_output +343,14350016,"TERMINAL",0,0,"[?25l^[",,terminal_output +344,14350144,"TERMINAL",0,0," ^[ [?25h",,terminal_output +345,14350257,"TERMINAL",0,0,"[?25l^[",,terminal_output +346,14350321,"TERMINAL",0,0," ^[ [?25h",,terminal_output +347,14354112,"TERMINAL",0,0,"[?25lj 20[?25h",,terminal_output +348,14354318,"TERMINAL",0,0,"[?25lj 1[?25h",,terminal_output +349,14354707,"TERMINAL",0,0,"[?25lh 3[?25h",,terminal_output +350,14355361,"TERMINAL",0,0,"[?25lh 2[?25h[?25lh 1[?25h[?25lh 0[?25h[?25lh 79[?25h[?25lh 8[?25h[?25lh 7[?25h",,terminal_output +351,14355578,"TERMINAL",0,0,"[?25lj 2[?25h",,terminal_output +352,14356116,"TERMINAL",0,0,"[?25lh 6[?25h",,terminal_output +353,14356454,"TERMINAL",0,0,"[?25lj 3,76 [?25h",,terminal_output +354,14356972,"TERMINAL",0,0,"[?25lh 5[?25h",,terminal_output +355,14357678,"TERMINAL",0,0,"[?25lh 4[?25h[?25lh 3[?25h[?25lh 2[?25h[?25lh 1[?25h[?25lh 0[?25h[?25lh 69[?25h[?25lh 8[?25h[?25lh 7[?25h[?25lh 6[?25h",,terminal_output +356,14357954,"TERMINAL",0,0,"[?25ll 7[?25h",,terminal_output +357,14358177,"TERMINAL",0,0,"[?25ll 8[?25h",,terminal_output +358,14358321,"TERMINAL",0,0,"[?25lx dl r_jobs/""[?25h",,terminal_output +359,14358729,"TERMINAL",0,0,"[?25lx dl _jobs/""[?25h",,terminal_output +360,14358978,"TERMINAL",0,0,"[?25lh 7[?25h",,terminal_output +361,14359336,"TERMINAL",0,0,"[?25lx dl _jobs/""[?25h",,terminal_output +362,14359507,"TERMINAL",0,0,"[?25li -- INSERT --23,67Top[?25h",,terminal_output +363,14359944,"TERMINAL",0,0,"[?25ls_jobs/""8[?25h",,terminal_output +364,14360011,"TERMINAL",0,0,"[?25lm_jobs/""9[?25h",,terminal_output +365,14360142,"TERMINAL",0,0,"[?25li_jobs/""70[?25h",,terminal_output +366,14360196,"TERMINAL",0,0,"[?25ln_jobs/""1[?25h",,terminal_output +367,14360294,"TERMINAL",0,0,"[?25le_jobs/""2[?25h",,terminal_output +368,14360547,"TERMINAL",0,0,"[?25l^[",,terminal_output +369,14360600,"TERMINAL",0,0," 23,71Top[?25h[?25lj 4[?25h",,terminal_output +370,14360929,"TERMINAL",0,0,"[?25lh 0[?25h",,terminal_output +371,14361047,"TERMINAL",0,0,"[?25lh 69[?25h",,terminal_output +372,14361315,"TERMINAL",0,0,"[?25lx dl ar_jobs_2/""[?25h",,terminal_output +373,14361497,"TERMINAL",0,0,"[?25lx dl r_jobs_2/""[?25h",,terminal_output +374,14361656,"TERMINAL",0,0,"[?25lx dl _jobs_2/""[?25h",,terminal_output +375,14361721,"TERMINAL",0,0,"[?25li -- INSERT --24,69Top[?25h",,terminal_output +376,14361885,"TERMINAL",0,0,"[?25ls_jobs_2/""70[?25h",,terminal_output +377,14361999,"TERMINAL",0,0,"[?25li_jobs_2/""1[?25h",,terminal_output +378,14362065,"TERMINAL",0,0,"[?25lm_jobs_2/""2[?25h",,terminal_output +379,14362675,"TERMINAL",0,0,"[?25l_jobs_2/""1[?25h",,terminal_output +380,14362801,"TERMINAL",0,0,"[?25l_jobs_2/""0[?25h",,terminal_output +381,14363016,"TERMINAL",0,0,"[?25lm_jobs_2/""1[?25h",,terminal_output +382,14363215,"TERMINAL",0,0,"[?25li_jobs_2/""2[?25h[?25ln_jobs_2/""3[?25h",,terminal_output +383,14363312,"TERMINAL",0,0,"[?25le_jobs_2/""4[?25h",,terminal_output +384,14363555,"TERMINAL",0,0,"[?25l^[",,terminal_output +385,14363681,"TERMINAL",0,0," 24,73Top[?25h[?25lj 5[?25h",,terminal_output +386,14364011,"TERMINAL",0,0,"[?25lh 2[?25h",,terminal_output +387,14364151,"TERMINAL",0,0,"[?25lh 1[?25h",,terminal_output +388,14364322,"TERMINAL",0,0,"[?25lh 0[?25h",,terminal_output +389,14364448,"TERMINAL",0,0,"[?25lh 69[?25h",,terminal_output +390,14364598,"TERMINAL",0,0,"[?25lx dl ar_jobs_3/""[?25h",,terminal_output +391,14364762,"TERMINAL",0,0,"[?25lx dl r_jobs_3/""[?25h",,terminal_output +392,14365255,"TERMINAL",0,0,"[?25lx dl _jobs_3/""[?25h",,terminal_output +393,14365499,"TERMINAL",0,0,"[?25li -- INSERT --25,69Top[?25h",,terminal_output +394,14366160,"TERMINAL",0,0,"[?25ls_jobs_3/""70[?25h",,terminal_output +395,14366274,"TERMINAL",0,0,"[?25lm_jobs_3/""1[?25h",,terminal_output +396,14366500,"TERMINAL",0,0,"[?25lin_jobs_3/""3[?25h",,terminal_output +397,14366609,"TERMINAL",0,0,"[?25le_jobs_3/""4[?25h",,terminal_output +398,14366795,"TERMINAL",0,0,"[?25l^[",,terminal_output +399,14366905,"TERMINAL",0,0," 25,73Top[?25h",,terminal_output +400,14368260,"TERMINAL",0,0,"[?25ly[?25h",,terminal_output +401,14368457,"TERMINAL",0,0,"[?25ly [?25h[?25lp alias runner-3=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""26,1Top[?25h",,terminal_output +402,14368880,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +403,14369363,"TERMINAL",0,0,"[?25ll 3[?25h[?25ll 4[?25h",,terminal_output +404,14369437,"TERMINAL",0,0,"[?25ll 5[?25h",,terminal_output +405,14369539,"TERMINAL",0,0,"[?25ll 6[?25h",,terminal_output +406,14369702,"TERMINAL",0,0,"[?25ll 7[?25h",,terminal_output +407,14370121,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +408,14370175,"TERMINAL",0,0,"[?25lw  -3=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""[?25h",,terminal_output +409,14370837,"TERMINAL",0,0,"[?25lx dl 3=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""[?25h",,terminal_output +410,14371017,"TERMINAL",0,0,"[?25lx dl =""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""[?25h",,terminal_output +411,14371624,"TERMINAL",0,0,"[?25la -- INSERT --26,8Top[?25h",,terminal_output +412,14372139,"TERMINAL",0,0,"[?25l""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""7[?25h",,terminal_output +413,14374402,"TERMINAL",0,0,"[?25lj""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""8[?25h",,terminal_output +414,14374563,"TERMINAL",0,0,"[?25l ja""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""9[?25h",,terminal_output +415,14374785,"TERMINAL",0,0,"[?25lf""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""10[?25h",,terminal_output +416,14374997,"TERMINAL",0,0,"[?25la""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""1[?25h",,terminal_output +417,14375191,"TERMINAL",0,0,"[?25lr""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""2[?25h",,terminal_output +418,14375664,"TERMINAL",0,0,"[?25l=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""3[?25h",,terminal_output +419,14376706,"TERMINAL",0,0,"[?25l^[",,terminal_output +420,14376817,"TERMINAL",0,0," 26,12Top[?25h",,terminal_output +421,14377315,"TERMINAL",0,0,"[?25ll 3[?25h",,terminal_output +422,14377837,"TERMINAL",0,0,"[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h",,terminal_output +423,14378119,"TERMINAL",0,0,"[?25l7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 20[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h",,terminal_output +424,14378204,"TERMINAL",0,0,"[?25ll 7[?25h",,terminal_output +425,14379170,"TERMINAL",0,0,"[?25l8[?25h[?25ll 9[?25h[?25ll 30[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 40[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 50[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 60[?25h[?25ll 1[?25h",,terminal_output +426,14379837,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +427,14380375,"TERMINAL",0,0,"[?25ll 3[?25h",,terminal_output +428,14380885,"TERMINAL",0,0,"[?25ll 4[?25h",,terminal_output +429,14381075,"TERMINAL",0,0,"[?25ll 5[?25h",,terminal_output +430,14381804,"TERMINAL",0,0,"[?25lh 4[?25h",,terminal_output +431,14382320,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +432,14382373,"TERMINAL",0,0,"[?25lw /""[?25h",,terminal_output +433,14383784,"TERMINAL",0,0,"[?25li -- INSERT --26,64Top[?25h",,terminal_output +434,14385957,"TERMINAL",0,0,"[?25lj/""5[?25h",,terminal_output +435,14386023,"TERMINAL",0,0,"[?25la/""6[?25h",,terminal_output +436,14386265,"TERMINAL",0,0,"[?25lf/""7[?25h",,terminal_output +437,14386376,"TERMINAL",0,0,"[?25la/""8[?25h",,terminal_output +438,14386554,"TERMINAL",0,0,"[?25lr/""9[?25h",,terminal_output +439,14386992,"TERMINAL",0,0,"[?25l^[",,terminal_output +440,14387159,"TERMINAL",0,0," 26,68Top[?25h[?25lj 7[?25h",,terminal_output +441,14388038,"TERMINAL",0,0,"[?25lh 7[?25h",,terminal_output +442,14388159,"TERMINAL",0,0,"[?25lh 6[?25h",,terminal_output +443,14388359,"TERMINAL",0,0,"[?25lh 5[?25h",,terminal_output +444,14388492,"TERMINAL",0,0,"[?25lh 4[?25h",,terminal_output +445,14388687,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +446,14388884,"TERMINAL",0,0,"[?25lw /""[?25h",,terminal_output +447,14389064,"TERMINAL",0,0,"[?25li -- INSERT --27,64Top[?25h",,terminal_output +448,14389327,"TERMINAL",0,0,"[?25ls/""5[?25h",,terminal_output +449,14389566,"TERMINAL",0,0,"[?25lm/""6[?25h",,terminal_output +450,14389691,"TERMINAL",0,0,"[?25li/""7[?25h[?25ln/""8[?25h",,terminal_output +451,14389763,"TERMINAL",0,0,"[?25le/""9[?25h",,terminal_output +452,14390013,"TERMINAL",0,0,"[?25l^[",,terminal_output +453,14390143,"TERMINAL",0,0," 27,68Top[?25h",,terminal_output +454,14391050,"TERMINAL",0,0,"[?25lk 6[?25h",,terminal_output +455,14392009,"TERMINAL",0,0,"[?25ly[?25h",,terminal_output +456,14392134,"TERMINAL",0,0,"[?25ly [?25h",,terminal_output +457,14392190,"TERMINAL",0,0,"[?25lp alias jafar=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar/""27,1Top[?25h",,terminal_output +458,14392707,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +459,14393401,"TERMINAL",0,0,"[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 10[?25h[?25ll 1[?25h",,terminal_output +460,14393467,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +461,14393954,"TERMINAL",0,0,"[?25li -- INSERT --27,12Top[?25h",,terminal_output +462,14394589,"TERMINAL",0,0,"[?25l-=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar/""3[?25h",,terminal_output +463,14394959,"TERMINAL",0,0,"[?25lr=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar/""4[?25h[?25lu=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar/""5[?25h",,terminal_output +464,14395051,"TERMINAL",0,0,"[?25ln=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar/""6[?25h",,terminal_output +465,14395163,"TERMINAL",0,0,"[?25ln=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar/""7[?25h[?25le=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar/""8[?25h",,terminal_output +466,14395695,"TERMINAL",0,0,"[?25lr=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar/""9[?25h",,terminal_output +467,14396218,"TERMINAL",0,0,"[?25l^[",,terminal_output +468,14396328,"TERMINAL",0,0," 27,18Top[?25h",,terminal_output +469,14396693,"TERMINAL",0,0,"[?25ll 9[?25h",,terminal_output +470,14397176,"TERMINAL",0,0,"[?25ll 20[?25h[?25ll 1[?25h",,terminal_output +471,14397257,"TERMINAL",0,0,"[?25ll 2[?25h[?25ll 3[?25h",,terminal_output +472,14397533,"TERMINAL",0,0,"[?25ll 4[?25h[?25l5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 30[?25h",,terminal_output +473,14397544,"TERMINAL",0,0,"[?25ll 1[?25h[?25ll 2[?25h",,terminal_output +474,14398045,"TERMINAL",0,0,"[?25ll 3[?25h[?25l4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 40[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 50[?25h",,terminal_output +475,14398477,"TERMINAL",0,0,"[?25lh 49[?25h",,terminal_output +476,14398645,"TERMINAL",0,0,"[?25lh 8[?25h",,terminal_output +477,14398860,"TERMINAL",0,0,"[?25lh 7[?25h",,terminal_output +478,14399108,"TERMINAL",0,0,"[?25ll 8[?25h",,terminal_output +479,14399834,"TERMINAL",0,0,"[?25ll 9[?25h[?25ll 50[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h",,terminal_output +480,14400314,"TERMINAL",0,0,"[?25ll 8[?25h[?25ll 9[?25h[?25ll 60[?25h[?25ll 1[?25h[?25ll 2[?25h[?25ll 3[?25h[?25ll 4[?25h[?25ll 5[?25h[?25ll 6[?25h[?25ll 7[?25h[?25ll 8[?25h[?25ll 9[?25h[?25ll 70[?25h",,terminal_output +481,14400329,"TERMINAL",0,0,"[?25ll 1[?25h",,terminal_output +482,14400453,"TERMINAL",0,0,"[?25ll 2[?25h",,terminal_output +483,14400611,"TERMINAL",0,0,"[?25ll 3[?25h",,terminal_output +484,14400798,"TERMINAL",0,0,"[?25ll 4[?25h",,terminal_output +485,14400941,"TERMINAL",0,0,"[?25ll 5[?25h",,terminal_output +486,14401049,"TERMINAL",0,0,"[?25ll 6[?25h",,terminal_output +487,14401346,"TERMINAL",0,0,"[?25li -- INSERT --27,76Top[?25h",,terminal_output +488,14404749,"TERMINAL",0,0,"[?25l_/""7[?25h",,terminal_output +489,14405162,"TERMINAL",0,0,"[?25lj/""8[?25h",,terminal_output +490,14405316,"TERMINAL",0,0,"[?25lo/""9[?25h",,terminal_output +491,14405531,"TERMINAL",0,0,"[?25lb/""80[?25h",,terminal_output +492,14405599,"TERMINAL",0,0,"[?25ls/""1[?25h",,terminal_output +493,14406023,"TERMINAL",0,0,"[?25l^[",,terminal_output +494,14406132,"TERMINAL",0,0," 27,80Top[?25h",,terminal_output +495,14407393,"TERMINAL",0,0,"[?25l::[?25h",,terminal_output +496,14407643,"TERMINAL",0,0,"w",,terminal_output +497,14407710,"TERMINAL",0,0,"q",,terminal_output +498,14410463,"TERMINAL",0,0,"\r[?25l[?2004l[>4;m""~/.bashrc"" 48L, 2579B written\r\r\r\n[?1004l[?2004l[?1l>[?25h[>4;m[?1049l]0;tum_cte0515@hkn1993:~/Projects]633;D;0",,terminal_output +499,14414478,"TERMINAL",0,0,"source ~/.bashrc",,terminal_command +500,14417268,"TERMINAL",0,0,"dev",,terminal_command +501,14417277,"TERMINAL",0,0,"]633;E;2025-09-03 17:41:16 dev;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C]0;tum_cte0515@hkn1993:~/Projects/jasmine]633;D;0",,terminal_output +502,14418906,"TERMINAL",0,0,"runner",,terminal_command +503,14423535,"TERMINAL",0,0,"runner-jafar",,terminal_command +504,14423585,"TERMINAL",0,0,"]633;E;2025-09-03 17:41:22 runner-jafar;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;Cbash: runner-jafar: command not found...\r\n",,terminal_output +505,14424639,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine_jobs]633;D;130",,terminal_output +506,14427421,"TERMINAL",0,0,"runner_jafar",,terminal_command +507,14427455,"TERMINAL",0,0,"]633;E;2025-09-03 17:41:26 runner_jafar;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;Cbash: runner_jafar: command not found...\r\n",,terminal_output +508,14427534,"TERMINAL",0,0,"]0;tum_cte0515@hkn1993:~/Projects/jasmine_jobs]633;D;127",,terminal_output +509,14432135,"TERMINAL",0,0,"vim ~/.bashrc",,terminal_command +510,14432431,"TERMINAL",0,0,"]633;E;2025-09-03 17:41:31 vim ~/.bashrc;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C[?1049h[>4;2m[?1h=[?2004h[?1004h[?12h[?12l[?25l""~/.bashrc"" 48L, 2579B▽ Pzz\[0%m [>c]10;?]11;?# .bashrc# Source global definitionsif [ -f /etc/bashrc ]; then\r\n . /etc/bashrc\r\nfi\r\n\r\n# User specific environment\r\nif ! [[ ""$PATH"" =~ ""$HOME/.local/bin:$HOME/bin:"" ]]\r\nthen\r\n PATH=""$HOME/.local/bin:$HOME/bin:$PATH""\r\nfi\r\nexport PATH\r\n\r\nalias idle='sinfo_t_idle'\r\nexport ws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\nalias salloc_node='salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8'\r\nalias salloc_cpu='salloc --time=01:00:00 --partition=dev_cpuonly --nodes=1 --cpus-per-task=128'\r\nalias sync-runner=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs/""\r\nalias sync-jafar=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jafar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""\r\nalias sync-runner-2=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_2/""\r\nalias sync-runner-3=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""\r\nalias runner=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs/""\r\nalias runner-2=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_2/""\r\nalias runner-3=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""\r\nalias jafar=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar/""\r\nalias jafar-runner=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""\r\nalias dev=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine/""\r\nalias smi=""watch -n1 nvidia-smi""\r\nalias idling=""watch -n1 sinfo_t_idle""\r\nalias queue='watch -n1 squeue --me'\r\nalias fqueue='watch -n 1 ""squeue -o \""%.10i %.16P %.30j %.8u %.8T %.10M %.9l %.6D %R\""""'\r\nalias fsacct_week='sacct --format=""JobID%15,JobName%30,Partition%16,AllocCPUS%3,State%12,Elapsed%10,Timelimit%10"" --starttime $(date -d ""last week"" +%Y-%m-%d) | grep -vE ""*.batch|*.extern|*.inter|bash|python|CANCELLED|echo""'\r\nalias logs=""cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir""27,80Top[?25hP+q436f\P+q6b75\P+q6b64\P+q6b72\P+q6b6c\P+q2332\P+q2334\P+q2569\P+q2a37\P+q6b31\[?12$p[?25l/3333/3333 [?25h[?25l/f6f6/e3e3 [?25h",,terminal_output +511,14434131,"TERMINAL",0,0,"[?25l^[",,terminal_output +512,14434231,"TERMINAL",0,0," ^[ [?25h",,terminal_output +513,14435468,"TERMINAL",0,0,"[?25lh 79[?25h",,terminal_output +514,14437466,"TERMINAL",0,0,"[?25lh 8[?25h[?25lh 7[?25h[?25lh 6[?25h[?25lh 5[?25h[?25lh 4[?25h[?25lh 3[?25h[?25lh 2[?25h[?25lh 1[?25h[?25lh 0[?25h[?25lh 69[?25h[?25lh 8[?25h[?25lh 7[?25h[?25lh 6[?25h[?25lh 5[?25h[?25lh 4[?25h[?25lh 3[?25h[?25lh 2[?25h[?25lh 1[?25h[?25lh 0[?25h[?25lh 59[?25h[?25lh 8[?25h[?25lh 7[?25h[?25lh 6[?25h[?25lh 5[?25h[?25lh 4[?25h[?25lh 3[?25h[?25lh 2[?25h[?25lh 1[?25h[?25lh 0[?25h[?25lh 49[?25h[?25lh 8[?25h[?25lh 7[?25h[?25lh 6[?25h[?25lh 5[?25h[?25lh 4[?25h[?25lh 3[?25h[?25lh 2[?25h[?25lh 1[?25h[?25lh 0[?25h[?25lh 39[?25h[?25lh 8[?25h[?25lh 7[?25h[?25lh 6[?25h[?25lh 5[?25h[?25lh 4[?25h[?25lh 3[?25h[?25lh 2[?25h[?25lh 1[?25h[?25lh 0[?25h[?25lh 29[?25h[?25lh 8[?25h",,terminal_output +515,14437750,"TERMINAL",0,0,"[?25lh 7[?25h[?25lh 6[?25h[?25lh 5[?25h[?25lh 4[?25h[?25lh 3[?25h[?25lh 2[?25h[?25lh 1[?25h[?25lh 0[?25h[?25lh 19[?25h[?25lh 8[?25h",,terminal_output +516,14437861,"TERMINAL",0,0,"[?25lh 7[?25h[?25lh 6[?25h[?25lh 5[?25h[?25lh 4[?25h",,terminal_output +517,14438181,"TERMINAL",0,0,"[?25lh 3[?25h",,terminal_output +518,14438407,"TERMINAL",0,0,"[?25lh 2[?25h",,terminal_output +519,14440737,"TERMINAL",0,0,"[?25l^[",,terminal_output +520,14440803,"TERMINAL",0,0," ^[ [?25h",,terminal_output +521,14441075,"TERMINAL",0,0,"[?25l::[?25h",,terminal_output +522,14441141,"TERMINAL",0,0,"w",,terminal_output +523,14441343,"TERMINAL",0,0,"q",,terminal_output +524,14442739,"TERMINAL",0,0,"\r[?25l[?2004l[>4;m""~/.bashrc"" 48L, 2579B written\r\r\r\n[?1004l[?2004l[?1l>[?25h[>4;m[?1049l]0;tum_cte0515@hkn1993:~/Projects/jasmine_jobs]633;D;0",,terminal_output +525,14446598,"TERMINAL",0,0,"jafar-runner",,terminal_command +526,14557860,"TERMINAL",0,0,"dev",,terminal_command +527,14558629,"TERMINAL",0,0,"cd ..",,terminal_command +528,14587698,"TERMINAL",0,0,"git clone git@github.com:FLAIROx/jafar.git",,terminal_command +529,14587755,"TERMINAL",0,0,"]633;E;2025-09-03 17:44:06 git clone git@github.com:FLAIROx/jafar.git;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;CCloning into 'jafar'...\r\n",,terminal_output +530,14589372,"TERMINAL",0,0,"remote: Enumerating objects: 71, done.\r\nremote: Counting objects: 2% (1/43)\rremote: Counting objects: 4% (2/43)\rremote: Counting objects: 6% (3/43)\rremote: Counting objects: 9% (4/43)\rremote: Counting objects: 11% (5/43)\rremote: Counting objects: 13% (6/43)\rremote: Counting objects: 16% (7/43)\rremote: Counting objects: 18% (8/43)\rremote: Counting objects: 20% (9/43)\rremote: Counting objects: 23% (10/43)\rremote: Counting objects: 25% (11/43)\rremote: Counting objects: 27% (12/43)\rremote: Counting objects: 30% (13/43)\rremote: Counting objects: 32% (14/43)\rremote: Counting objects: 34% (15/43)\rremote: Counting objects: 37% (16/43)\rremote: Counting objects: 39% (17/43)\rremote: Counting objects: 41% (18/43)\rremote: Counting objects: 44% (19/43)\rremote: Counting objects: 46% (20/43)\rremote: Counting objects: 48% (21/43)\rremote: Counting objects: 51% (22/43)\rremote: Counting objects: 53% (23/43)\rremote: Counting objects: 55% (24/43)\rremote: Counting objects: 58% (25/43)\rremote: Counting objects: 60% (26/43)\rremote: Counting objects: 62% (27/43)\rremote: Counting objects: 65% (28/43)\rremote: Counting objects: 67% (29/43)\rremote: Counting objects: 69% (30/43)\rremote: Counting objects: 72% (31/43)\rremote: Counting objects: 74% (32/43)\rremote: Counting objects: 76% (33/43)\rremote: Counting objects: 79% (34/43)\rremote: Counting objects: 81% (35/43)\rremote: Counting objects: 83% (36/43)\rremote: Counting objects: 86% (37/43)\rremote: Counting objects: 88% (38/43)\rremote: Counting objects: 90% (39/43)\rremote: Counting objects: 93% (40/43)\rremote: Counting objects: 95% (41/43)\rremote: Counting objects: 97% (42/43)\rremote: Counting objects: 100% (43/43)\rremote: Counting objects: 100% (43/43), done.\r\nremote: Compressing objects: 4% (1/22)\rremote: Compressing objects: 9% (2/22)\rremote: Compressing objects: 13% (3/22)\rremote: Compressing objects: 18% (4/22)\rremote: Compressing objects: 22% (5/22)\rremote: Compressing objects: 27% (6/22)\rremote: Compressing objects: 31% (7/22)\r",,terminal_output +531,14589487,"TERMINAL",0,0,"remote: Compressing objects: 36% (8/22)\rremote: Compressing objects: 40% (9/22)\rremote: Compressing objects: 45% (10/22)\rremote: Compressing objects: 50% (11/22)\rremote: Compressing objects: 54% (12/22)\rremote: Compressing objects: 59% (13/22)\rremote: Compressing objects: 63% (14/22)\rremote: Compressing objects: 68% (15/22)\rremote: Compressing objects: 72% (16/22)\rremote: Compressing objects: 77% (17/22)\rremote: Compressing objects: 81% (18/22)\rremote: Compressing objects: 86% (19/22)\rremote: Compressing objects: 90% (20/22)\rremote: Compressing objects: 95% (21/22)\rremote: Compressing objects: 100% (22/22)\rremote: Compressing objects: 100% (22/22), done.\r\nReceiving objects: 1% (1/71)\rReceiving objects: 2% (2/71)\rReceiving objects: 4% (3/71)\rReceiving objects: 5% (4/71)\rReceiving objects: 7% (5/71)\rReceiving objects: 8% (6/71)\rReceiving objects: 9% (7/71)\rReceiving objects: 11% (8/71)\rReceiving objects: 12% (9/71)\rReceiving objects: 14% (10/71)\rReceiving objects: 15% (11/71)\rReceiving objects: 16% (12/71)\rReceiving objects: 18% (13/71)\rReceiving objects: 19% (14/71)\rReceiving objects: 21% (15/71)\r",,terminal_output +532,14589571,"TERMINAL",0,0,"Receiving objects: 22% (16/71)\rReceiving objects: 23% (17/71)\rReceiving objects: 25% (18/71)\rReceiving objects: 26% (19/71)\rReceiving objects: 28% (20/71)\rReceiving objects: 29% (21/71)\rReceiving objects: 30% (22/71)\rReceiving objects: 32% (23/71)\rReceiving objects: 33% (24/71)\rReceiving objects: 35% (25/71)\rReceiving objects: 36% (26/71)\rReceiving objects: 38% (27/71)\rReceiving objects: 39% (28/71)\rReceiving objects: 40% (29/71)\rReceiving objects: 42% (30/71)\rReceiving objects: 43% (31/71)\rReceiving objects: 45% (32/71)\rReceiving objects: 46% (33/71)\rReceiving objects: 47% (34/71)\rReceiving objects: 49% (35/71)\rReceiving objects: 50% (36/71)\rReceiving objects: 52% (37/71)\rReceiving objects: 53% (38/71)\rReceiving objects: 54% (39/71)\rReceiving objects: 56% (40/71)\rReceiving objects: 57% (41/71)\rReceiving objects: 59% (42/71)\rReceiving objects: 60% (43/71)\rReceiving objects: 61% (44/71)\rReceiving objects: 63% (45/71)\rReceiving objects: 64% (46/71)\rReceiving objects: 66% (47/71)\rReceiving objects: 67% (48/71)\rReceiving objects: 69% (49/71)\rReceiving objects: 70% (50/71)\rReceiving objects: 71% (51/71)\rReceiving objects: 73% (52/71)\rReceiving objects: 74% (53/71)\rReceiving objects: 76% (54/71)\rReceiving objects: 77% (55/71)\rReceiving objects: 78% (56/71)\rReceiving objects: 80% (57/71)\rReceiving objects: 81% (58/71)\rReceiving objects: 83% (59/71)\rReceiving objects: 84% (60/71)\rReceiving objects: 85% (61/71)\rReceiving objects: 87% (62/71)\rReceiving objects: 88% (63/71)\rReceiving objects: 90% (64/71)\rReceiving objects: 91% (65/71)\rReceiving objects: 92% (66/71)\rReceiving objects: 94% (67/71)\rReceiving objects: 95% (68/71)\rReceiving objects: 97% (69/71)\rReceiving objects: 98% (70/71)\rReceiving objects: 100% (71/71)\rReceiving objects: 100% (71/71), 81.99 KiB | 792.00 KiB/s, done.\r\nResolving deltas: 0% (0/34)\rResolving deltas: 2% (1/34)\rResolving deltas: 5% (2/34)\rResolving deltas: 11% (4/34)\rResolving deltas: 14% (5/34)\rResolving deltas: 20% (7/34)\rResolving deltas: 23% (8/34)\rResolving deltas: 26% (9/34)\rResolving deltas: 32% (11/34)\rResolving deltas: 35% (12/34)\rResolving deltas: 41% (14/34)\rResolving deltas: 44% (15/34)\rResolving deltas: 47% (16/34)\rResolving deltas: 52% (18/34)\rResolving deltas: 55% (19/34)\rResolving deltas: 61% (21/34)\rResolving deltas: 67% (23/34)\rResolving deltas: 70% (24/34)\rResolving deltas: 73% (25/34)\rResolving deltas: 76% (26/34)\rResolving deltas: 79% (27/34)\rResolving deltas: 82% (28/34)\rResolving deltas: 85% (29/34)\rResolving deltas: 88% (30/34)\rResolving deltas: 91% (31/34)\rResolving deltas: 94% (32/34)\rResolving deltas: 97% (33/34)\rResolving deltas: 100% (34/34)\rResolving deltas: 100% (34/34), done.\r\n",,terminal_output +533,14589666,"TERMINAL",0,0,"remote: Total 71 (delta 33), reused 21 (delta 21), pack-reused 28 (from 1)\r\n",,terminal_output +534,14589730,"TERMINAL",0,0,"]0;tum_cte0515@hkn1993:~/Projects]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects",,terminal_output +535,14591447,"",0,0,"Switched from branch 'feat/darkness-filter' to 'main'",,git_branch_checkout +536,14638061,"TERMINAL",0,0,"jafar",,terminal_command +537,14639618,"TERMINAL",0,0,"ls",,terminal_command +538,14674257,"TERMINAL",0,0,"dev",,terminal_command +539,14675452,"TERMINAL",0,0,"cd slurm/",,terminal_command +540,14675794,"TERMINAL",0,0,"ls",,terminal_command +541,14675846,"TERMINAL",0,0,"]633;E;2025-09-03 17:45:34 ls;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C",,terminal_output +542,14675928,"TERMINAL",0,0,"common dev jobs README.md templates utils\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine/slurm]633;D;0",,terminal_output +543,14677718,"TERMINAL",0,0,"git pull",,terminal_command +544,14677788,"TERMINAL",0,0,"]633;E;2025-09-03 17:45:36 git pull;135ca8e6-3085-4cb9-9ca6-3533ce46ed81]633;C",,terminal_output +545,14679568,"TERMINAL",0,0,"remote: Enumerating objects: 381, done.\r\nremote: Counting objects: 0% (1/217)\rremote: Counting objects: 1% (3/217)\rremote: Counting objects: 2% (5/217)\rremote: Counting objects: 3% (7/217)\rremote: Counting objects: 4% (9/217)\rremote: Counting objects: 5% (11/217)\rremote: Counting objects: 6% (14/217)\rremote: Counting objects: 7% (16/217)\rremote: Counting objects: 8% (18/217)\rremote: Counting objects: 9% (20/217)\rremote: Counting objects: 10% (22/217)\rremote: Counting objects: 11% (24/217)\rremote: Counting objects: 12% (27/217)\rremote: Counting objects: 13% (29/217)\rremote: Counting objects: 14% (31/217)\rremote: Counting objects: 15% (33/217)\rremote: Counting objects: 16% (35/217)\rremote: Counting objects: 17% (37/217)\rremote: Counting objects: 18% (40/217)\rremote: Counting objects: 19% (42/217)\rremote: Counting objects: 20% (44/217)\rremote: Counting objects: 21% (46/217)\rremote: Counting objects: 22% (48/217)\rremote: Counting objects: 23% (50/217)\rremote: Counting objects: 24% (53/217)\rremote: Counting objects: 25% (55/217)\rremote: Counting objects: 26% (57/217)\rremote: Counting objects: 27% (59/217)\rremote: Counting objects: 28% (61/217)\rremote: Counting objects: 29% (63/217)\rremote: Counting objects: 30% (66/217)\rremote: Counting objects: 31% (68/217)\rremote: Counting objects: 32% (70/217)\rremote: Counting objects: 33% (72/217)\rremote: Counting objects: 34% (74/217)\rremote: Counting objects: 35% (76/217)\rremote: Counting objects: 36% (79/217)\rremote: Counting objects: 37% (81/217)\rremote: Counting objects: 38% (83/217)\r",,terminal_output +546,14680234,"TERMINAL",0,0,"remote: Counting objects: 39% (85/217)\rremote: Counting objects: 40% (87/217)\rremote: Counting objects: 41% (89/217)\rremote: Counting objects: 42% (92/217)\rremote: Counting objects: 43% (94/217)\rremote: Counting objects: 44% (96/217)\rremote: Counting objects: 45% (98/217)\rremote: Counting objects: 46% (100/217)\rremote: Counting objects: 47% (102/217)\rremote: Counting objects: 48% (105/217)\rremote: Counting objects: 49% (107/217)\rremote: Counting objects: 50% (109/217)\rremote: Counting objects: 51% (111/217)\rremote: Counting objects: 52% (113/217)\rremote: Counting objects: 53% (116/217)\rremote: Counting objects: 54% (118/217)\rremote: Counting objects: 55% (120/217)\rremote: Counting objects: 56% (122/217)\rremote: Counting objects: 57% (124/217)\rremote: Counting objects: 58% (126/217)\rremote: Counting objects: 59% (129/217)\rremote: Counting objects: 60% (131/217)\rremote: Counting objects: 61% (133/217)\rremote: Counting objects: 62% (135/217)\rremote: Counting objects: 63% (137/217)\rremote: Counting objects: 64% (139/217)\rremote: Counting objects: 65% (142/217)\rremote: Counting objects: 66% (144/217)\rremote: Counting objects: 67% (146/217)\rremote: Counting objects: 68% (148/217)\rremote: Counting objects: 69% (150/217)\rremote: Counting objects: 70% (152/217)\rremote: Counting objects: 71% (155/217)\rremote: Counting objects: 72% (157/217)\rremote: Counting objects: 73% (159/217)\rremote: Counting objects: 74% (161/217)\rremote: Counting objects: 75% (163/217)\rremote: Counting objects: 76% (165/217)\rremote: Counting objects: 77% (168/217)\rremote: Counting objects: 78% (170/217)\rremote: Counting objects: 79% (172/217)\rremote: Counting objects: 80% (174/217)\rremote: Counting objects: 81% (176/217)\rremote: Counting objects: 82% (178/217)\rremote: Counting objects: 83% (181/217)\rremote: Counting objects: 84% (183/217)\rremote: Counting objects: 85% (185/217)\rremote: Counting objects: 86% (187/217)\rremote: Counting objects: 87% (189/217)\rremote: Counting objects: 88% (191/217)\rremote: Counting objects: 89% (194/217)\rremote: Counting objects: 90% (196/217)\rremote: Counting objects: 91% (198/217)\rremote: Counting objects: 92% (200/217)\rremote: Counting objects: 93% (202/217)\rremote: Counting objects: 94% (204/217)\rremote: Counting objects: 95% (207/217)\rremote: Counting objects: 96% (209/217)\rremote: Counting objects: 97% (211/217)\rremote: Counting objects: 98% (213/217)\rremote: Counting objects: 99% (215/217)\rremote: Counting objects: 100% (217/217)\rremote: Counting objects: 100% (217/217), done.\r\nremote: Compressing objects: 0% (1/108)\rremote: Compressing objects: 1% (2/108)\rremote: Compressing objects: 2% (3/108)\rremote: Compressing objects: 3% (4/108)\rremote: Compressing objects: 4% (5/108)\rremote: Compressing objects: 5% (6/108)\rremote: Compressing objects: 6% (7/108)\rremote: Compressing objects: 7% (8/108)\rremote: Compressing objects: 8% (9/108)\rremote: Compressing objects: 9% (10/108)\rremote: Compressing objects: 10% (11/108)\rremote: Compressing objects: 11% (12/108)\rremote: Compressing objects: 12% (13/108)\rremote: Compressing objects: 13% (15/108)\rremote: Compressing objects: 14% (16/108)\rremote: Compressing objects: 15% (17/108)\rremote: Compressing objects: 16% (18/108)\rremote: Compressing objects: 17% (19/108)\rremote: Compressing objects: 18% (20/108)\rremote: Compressing objects: 19% (21/108)\rremote: Compressing objects: 20% (22/108)\rremote: Compressing objects: 21% (23/108)\rremote: Compressing objects: 22% (24/108)\rremote: Compressing objects: 23% (25/108)\rremote: Compressing objects: 24% (26/108)\rremote: Compressing objects: 25% (27/108)\rremote: Compressing objects: 26% (29/108)\rremote: Compressing objects: 27% (30/108)\rremote: Compressing objects: 28% (31/108)\rremote: Compressing objects: 29% (32/108)\rremote: Compressing objects: 30% (33/108)\rremote: Compressing objects: 31% (34/108)\rremote: Compressing objects: 32% (35/108)\rremote: Compressing objects: 33% (36/108)\rremote: Compressing objects: 34% (37/108)\rremote: Compressing objects: 35% (38/108)\rremote: Compressing objects: 36% (39/108)\rremote: Compressing objects: 37% (40/108)\rremote: Compressing objects: 38% (42/108)\rremote: Compressing objects: 39% (43/108)\rremote: Compressing objects: 40% (44/108)\rremote: Compressing objects: 41% (45/108)\rremote: Compressing objects: 42% (46/108)\rremote: Compressing objects: 43% (47/108)\rremote: Compressing objects: 44% (48/108)\rremote: Compressing objects: 45% (49/108)\rremote: Compressing objects: 46% (50/108)\rremote: Compressing objects: 47% (51/108)\rremote: Compressing objects: 48% (52/108)\rremote: Compressing objects: 49% (53/108)\rremote: Compressing objects: 50% (54/108)\rremote: Compressing objects: 51% (56/108)\rremote: Compressing objects: 52% (57/108)\rremote: Compressing objects: 53% (58/108)\rremote: Compressing objects: 54% (59/108)\rremote: Compressing objects: 55% (60/108)\rremote: Compressing objects: 56% (61/108)\rremote: Compressing objects: 57% (62/108)\rremote: Compressing objects: 58% (63/108)\rremote: Compressing objects: 59% (64/108)\rremote: Compressing objects: 60% (65/108)\rremote: Compressing objects: 61% (66/108)\rremote: Compressing objects: 62% (67/108)\rremote: Compressing objects: 63% (69/108)\rremote: Compressing objects: 64% (70/108)\rremote: Compressing objects: 65% (71/108)\rremote: Compressing objects: 66% (72/108)\rremote: Compressing objects: 67% (73/108)\rremote: Compressing objects: 68% (74/108)\rremote: Compressing objects: 69% (75/108)\rremote: Compressing objects: 70% (76/108)\rremote: Compressing objects: 71% (77/108)\rremote: Compressing objects: 72% (78/108)\rremote: Compressing objects: 73% (79/108)\rremote: Compressing objects: 74% (80/108)\rremote: Compressing objects: 75% (81/108)\rremote: Compressing objects: 76% (83/108)\rremote: Compressing objects: 77% (84/108)\rremote: Compressing objects: 78% (85/108)\rremote: Compressing objects: 79% (86/108)\rremote: Compressing objects: 80% (87/108)\rremote: Compressing objects: 81% (88/108)\rremote: Compressing objects: 82% (89/108)\rremote: Compressing objects: 83% (90/108)\rremote: Compressing objects: 84% (91/108)\rremote: Compressing objects: 85% (92/108)\rremote: Compressing objects: 86% (93/108)\rremote: Compressing objects: 87% (94/108)\rremote: Compressing objects: 88% (96/108)\rremote: Compressing objects: 89% (97/108)\rremote: Compressing objects: 90% (98/108)\rremote: Compressing objects: 91% (99/108)\rremote: Compressing objects: 92% (100/108)\rremote: Compressing objects: 93% (101/108)\rremote: Compressing objects: 94% (102/108)\rremote: Compressing objects: 95% (103/108)\rremote: Compressing objects: 96% (104/108)\rremote: Compressing objects: 97% (105/108)\rremote: Compressing objects: 98% (106/108)\rremote: Compressing objects: 99% (107/108)\rremote: Compressing objects: 100% (108/108)\rremote: Compressing objects: 100% (108/108), done.\r\nReceiving objects: 0% (1/381)\rReceiving objects: 1% (4/381)\rReceiving objects: 2% (8/381)\rReceiving objects: 3% (12/381)\rReceiving objects: 4% (16/381)\rReceiving objects: 5% (20/381)\rReceiving objects: 6% (23/381)\rReceiving objects: 7% (27/381)\rReceiving objects: 8% (31/381)\rReceiving objects: 9% (35/381)\rReceiving objects: 10% (39/381)\rReceiving objects: 11% (42/381)\rReceiving objects: 12% (46/381)\rReceiving objects: 13% (50/381)\rReceiving objects: 14% (54/381)\rReceiving objects: 15% (58/381)\rReceiving objects: 16% (61/381)\rReceiving objects: 17% (65/381)\rReceiving objects: 18% (69/381)\rReceiving objects: 19% (73/381)\rremote: Total 381 (delta 127), reused 188 (delta 102), pack-reused 164 (from 1)\r\nReceiving objects: 20% (77/381)\rReceiving objects: 21% (81/381)\rReceiving objects: 22% (84/381)\rReceiving objects: 23% (88/381)\rReceiving objects: 24% (92/381)\rReceiving objects: 25% (96/381)\rReceiving objects: 26% (100/381)\rReceiving objects: 27% (103/381)\rReceiving objects: 28% (107/381)\rReceiving objects: 29% (111/381)\rReceiving objects: 30% (115/381)\rReceiving objects: 31% (119/381)\rReceiving objects: 32% (122/381)\rReceiving objects: 33% (126/381)\rReceiving objects: 34% (130/381)\rReceiving objects: 35% (134/381)\rReceiving objects: 36% (138/381)\rReceiving objects: 37% (141/381)\rReceiving objects: 38% (145/381)\rReceiving objects: 39% (149/381)\rReceiving objects: 40% (153/381)\rReceiving objects: 41% (157/381)\rReceiving objects: 42% (161/381)\rReceiving objects: 43% (164/381)\rReceiving objects: 44% (168/381)\rReceiving objects: 45% (172/381)\rReceiving objects: 46% (176/381)\rReceiving objects: 47% (180/381)\rReceiving objects: 48% (183/381)\rReceiving objects: 49% (187/381)\rReceiving objects: 50% (191/381)\rReceiving objects: 51% (195/381)\rReceiving objects: 52% (199/381)\rReceiving objects: 53% (202/381)\rReceiving objects: 54% (206/381)\rReceiving objects: 55% (210/381)\rReceiving objects: 56% (214/381)\rReceiving objects: 57% (218/381)\rReceiving objects: 58% (221/381)\rReceiving objects: 59% (225/381)\rReceiving objects: 60% (229/381)\rReceiving objects: 61% (233/381)\rReceiving objects: 62% (237/381)\rReceiving objects: 63% (241/381)\rReceiving objects: 64% (244/381)\rReceiving objects: 65% (248/381)\rReceiving objects: 66% (252/381)\rReceiving objects: 67% (256/381)\rReceiving objects: 68% (260/381)\rReceiving objects: 69% (263/381)\rReceiving objects: 70% (267/381)\rReceiving objects: 71% (271/381)\rReceiving objects: 72% (275/381)\rReceiving objects: 73% (279/381)\rReceiving objects: 74% (282/381)\rReceiving objects: 75% (286/381)\rReceiving objects: 76% (290/381)\rReceiving objects: 77% (294/381)\rReceiving objects: 78% (298/381)\rReceiving objects: 79% (301/381)\rReceiving objects: 80% (305/381)\rReceiving objects: 81% (309/381)\rReceiving objects: 82% (313/381)\rReceiving objects: 83% (317/381)\rReceiving objects: 84% (321/381)\rReceiving objects: 85% (324/381)\rReceiving objects: 86% (328/381)\rReceiving objects: 87% (332/381)\rReceiving objects: 88% (336/381)\rReceiving objects: 89% (340/381)\rReceiving objects: 90% (343/381)\rReceiving objects: 91% (347/381)\rReceiving objects: 92% (351/381)\rReceiving objects: 93% (355/381)\rReceiving objects: 94% (359/381)\rReceiving objects: 95% (362/381)\rReceiving objects: 96% (366/381)\rReceiving objects: 97% (370/381)\rReceiving objects: 98% (374/381)\rReceiving objects: 99% (378/381)\rReceiving objects: 100% (381/381)\rReceiving objects: 100% (381/381), 71.62 KiB | 261.00 KiB/s, done.\r\nResolving deltas: 0% (0/211)\rResolving deltas: 1% (3/211)\rResolving deltas: 2% (5/211)\rResolving deltas: 3% (7/211)\rResolving deltas: 4% (9/211)\rResolving deltas: 5% (11/211)\rResolving deltas: 6% (13/211)\rResolving deltas: 7% (15/211)\rResolving deltas: 8% (17/211)\rResolving deltas: 9% (19/211)\rResolving deltas: 10% (22/211)\rResolving deltas: 11% (24/211)\rResolving deltas: 12% (26/211)\rResolving deltas: 13% (28/211)\rResolving deltas: 14% (30/211)\rResolving deltas: 15% (32/211)\rResolving deltas: 16% (34/211)\rResolving deltas: 17% (36/211)\rResolving deltas: 18% (38/211)\rResolving deltas: 19% (41/211)\rResolving deltas: 20% (43/211)\rResolving deltas: 21% (45/211)\rResolving deltas: 22% (47/211)\rResolving deltas: 23% (49/211)\rResolving deltas: 24% (51/211)\rResolving deltas: 25% (53/211)\rResolving deltas: 26% (55/211)\rResolving deltas: 27% (57/211)\rResolving deltas: 28% (60/211)\rResolving deltas: 29% (62/211)\rResolving deltas: 30% (64/211)\rResolving deltas: 31% (66/211)\rResolving deltas: 32% (68/211)\rResolving deltas: 33% (70/211)\rResolving deltas: 34% (72/211)\rResolving deltas: 35% (74/211)\rResolving deltas: 36% (76/211)\rResolving deltas: 37% (79/211)\rResolving deltas: 38% (81/211)\rResolving deltas: 39% (83/211)\rResolving deltas: 40% (85/211)\rResolving deltas: 41% (87/211)\rResolving deltas: 42% (89/211)\rResolving deltas: 43% (91/211)\rResolving deltas: 44% (93/211)\rResolving deltas: 45% (95/211)\rResolving deltas: 46% (98/211)\rResolving deltas: 47% (100/211)\rResolving deltas: 48% (102/211)\rResolving deltas: 49% (104/211)\rResolving deltas: 50% (106/211)\rResolving deltas: 51% (108/211)\rResolving deltas: 52% (110/211)\rResolving deltas: 53% (112/211)\rResolving deltas: 54% (114/211)\rResolving deltas: 55% (117/211)\rResolving deltas: 56% (119/211)\rResolving deltas: 57% (121/211)\rResolving deltas: 58% (123/211)\rResolving deltas: 59% (125/211)\rResolving deltas: 60% (127/211)\rResolving deltas: 61% (129/211)\rResolving deltas: 62% (131/211)\rResolving deltas: 63% (133/211)\rResolving deltas: 64% (136/211)\rResolving deltas: 65% (138/211)\rResolving deltas: 66% (140/211)\rResolving deltas: 67% (142/211)\rResolving deltas: 68% (144/211)\rResolving deltas: 69% (146/211)\rResolving deltas: 70% (148/211)\rResolving deltas: 71% (150/211)\rResolving deltas: 72% (152/211)\rResolving deltas: 73% (155/211)\rResolving deltas: 74% (157/211)\rResolving deltas: 75% (159/211)\rResolving deltas: 76% (161/211)\rResolving deltas: 77% (163/211)\rResolving deltas: 78% (165/211)\rResolving deltas: 79% (167/211)\rResolving deltas: 80% (169/211)\rResolving deltas: 81% (171/211)\rResolving deltas: 82% (174/211)\rResolving deltas: 83% (176/211)\rResolving deltas: 84% (178/211)\rResolving deltas: 85% (180/211)\rResolving deltas: 86% (182/211)\rResolving deltas: 87% (184/211)\rResolving deltas: 88% (186/211)\rResolving deltas: 89% (188/211)\rResolving deltas: 90% (190/211)\rResolving deltas: 91% (193/211)\rResolving deltas: 92% (195/211)\rResolving deltas: 93% (197/211)\rResolving deltas: 94% (199/211)\rResolving deltas: 95% (201/211)\rResolving deltas: 96% (203/211)\rResolving deltas: 97% (205/211)\rResolving deltas: 98% (207/211)\rResolving deltas: 99% (209/211)\rResolving deltas: 100% (211/211)\rResolving deltas: 100% (211/211), completed with 13 local objects.\r\n",,terminal_output +547,14680608,"TERMINAL",0,0,"From github.com:p-doom/slurm\r\n 94549c5..63f0e09 main -> origin/main\r\n",,terminal_output +548,14680751,"TERMINAL",0,0,"Updating 94549c5..63f0e09\r\n",,terminal_output +549,14682075,"TERMINAL",0,0,"Fast-forward\r\n",,terminal_output +550,14682247,"TERMINAL",0,0," dev/alfred/berlin/gt_actions/sample_causal_32gpus.sbatch | 29 +++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_actions/sample_darkness_filter.sbatch | 33 +++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_actions/train_dynacmis_on_3nodes_2gpu.sbatch | 52 ++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_actions/train_dynacmis_on_8gpu.sbatch | 52 ++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_actions/train_dynacmis_on_more_than4.sbatch | 52 ++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_actions/train_dynacmis_overfit.sbatch | 53 +++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_actions/train_dynacmis_overfit_2gpus.sbatch | 51 +++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_actions/train_dynacmis_overfit_4gpu.sbatch | 53 +++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_actions/train_dynacmis_overfit_4gpu_to_8gpu.sbatch | 52 ++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_actions/train_tok_topology_one_gpu copy.sbatch | 45 +++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_actions/train_tok_topology_one_gpu.sbatch | 45 +++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_actions/train_tok_topology_restore_to_on_gpu.sbatch | 48 ++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_actions/train_tok_topology_two_gpus.sbatch | 45 +++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/gt_sampling/sampling_mock.sbatch | 29 +++++++++++++++++++++++++++++\r\n dev/alfred/berlin/job_requeueing/notes.md | 47 -----------------------------------------------\r\n dev/alfred/berlin/test_franz_pr/train_dynacmis_overfit.sbatch | 52 ++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/test_franz_pr/train_lam_overfit.sbatch | 40 ++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/test_franz_pr/train_tokenizer_overfit.sbatch | 40 ++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology/sample_overfit_single_gpu.sbatch | 30 ++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology/train_dynacmis_overfit_1.sbatch | 50 ++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology/train_dynacmis_overfit_1_gt_actions.sbatch | 51 +++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology/train_dynacmis_overfit_1_gt_actions_noise.sbatch | 51 +++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology/train_dynacmis_overfit_1_noise.sbatch | 50 ++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology/train_dynacmis_overfit_2_nodes_2_gpu.sbatch | 51 +++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology/train_dynacmis_overfit_2_nodes_4_gpu.sbatch | 51 +++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology/train_dynacmis_overfit_4gpu.sbatch | 52 ++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology/train_lam_overfit.sbatch | 40 ++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology/train_tokenizer_overfit_4_gpu.sbatch | 41 +++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology/train_tokenizer_overfit_single_record.sbatch | 42 ++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology/train_tokenizer_overfit_single_record_requeue.sbatch | 40 ++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology_restore_fix/train_tokenizer_overfit_4_gpu.sbatch | 42 ++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/topology_restore_fix/train_tokenizer_overfit_4_to_1_gpu.sbatch | 48 ++++++++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/berlin/train_dyn_dev/{train_dynacmis.sbatch => train_dynacmis_overfit.sbatch} | 26 ++++++++++++--------------\r\n dev/alfred/berlin/train_lam_dev/train_lam.sbatch | 19 ++++++++++++++++---\r\n dev/alfred/berlin/train_tok_dev/train_tok.sbatch | 7 ++++---\r\n dev/alfred/berlin/train_tok_dev/train_tok_overfit.sbatch | 43 +++++++++++++++++++++++++++++++++++++++++++\r\n dev/alfred/helmholtz_cluster/train_tok_dev/train_tok.sbatch | 30 ++++++++++++++++++++++++++++++\r\n dev/alfred/helmholtz_cluster/train_tok_dev/train_tok_overfit.sbatch | 43 +++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_dynamics/coinrun_dynamics.sbatch | 2 +-\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_dynamics_reproduction.sbatch | 3 ++-\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_full_prec.sbatch | 69 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_full_prec_cosine.sbatch | 70 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_full_prec_w_restore.sbatch | 70 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_mix_prec_cosine.sbatch | 70 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_mix_prec_cosine_min_init_lt_3e-6.sbatch | 72 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction/sample.sbatch | 30 ++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction/sample_cotrain.sbatch | 28 ++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_dynamics_reproduction.sbatch | 81 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_dynamics_reproduction_cotrain.sbatch | 76 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction.sbatch | 76 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_dc_0.sbatch | 76 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_ffn_512.sbatch | 77 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_ffn_512_num_blocks_8.sbatch | 78 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_lower_lr.sbatch | 76 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_lower_lr_0.5x.sbatch | 77 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_lower_lr_3e6_1e5_0.sbatch | 76 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_lower_lr_3e6_3e5_0.sbatch | 76 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_tokenizer_repoduction.sbatch | 75 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_tokenizer_repoduction_ffn_512_n_blocks_8.sbatch | 77 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/generate_data.sbatch | 15 +++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/sample.sbatch | 30 ++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/sample_cotrain.sbatch | 28 ++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch | 42 ++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_dynamics_reproduction_requeue.sbatch | 47 +++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_dynamics_reproduction_requeue_2.0.sbatch | 47 +++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_tokenizer_repoduction_requeue.sbatch | 47 +++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_tokenizer_repoduction_requeue_2.0.sbatch | 48 ++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/berlin/jafar_og_reproduction/sample.sbatch | 25 +++++++++++++++++++++++++\r\n jobs/alfred/berlin/jafar_og_reproduction/sample_og_150k.sbatch | 27 +++++++++++++++++++++++++++\r\n jobs/alfred/berlin/jafar_og_reproduction/sample_og_175k.sbatch | 27 +++++++++++++++++++++++++++\r\n jobs/alfred/berlin/jafar_og_reproduction/sample_og_200k.sbatch | 26 ++++++++++++++++++++++++++\r\n jobs/alfred/berlin/jafar_og_reproduction/sample_og_spawner.sh | 10 ++++++++++\r\n jobs/alfred/berlin/jafar_og_reproduction/spawner_sample.sh | 11 +++++++++++\r\n jobs/alfred/helmholtz_cluster/jafar_og_reproduction/generate_dataset.sbatch | 18 ++++++++++++++++++\r\n jobs/alfred/helmholtz_cluster/jafar_og_reproduction/generate_dataset_10m.sbatch | 21 +++++++++++++++++++++\r\n jobs/alfred/helmholtz_cluster/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch | 44 ++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/helmholtz_cluster/jafar_og_reproduction/og_coinrun_lam_reproduction.sbatch | 40 ++++++++++++++++++++++++++++++++++++++++\r\n jobs/alfred/helmholtz_cluster/jafar_og_reproduction/og_coinrun_tokenizer_repoduction.sbatch | 41 +++++++++++++++++++++++++++++++++++++++++\r\n jobs/franz/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_fp32_adam_moments.sh | 84 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/franz/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_fp32_layernorm.sh | 84 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/franz/berlin/coinrun/coinrun_lam/lam_coinrun_nan_investigation_100k_to_107k.sbatch | 49 +++++++++++++++++++++++++++++++++++++++++++++++++\r\n jobs/mihir/horeka/causal_big_runs/train_dynamics_8_nodes_filter_dark_req.sbatch | 1 +\r\n utils/alfred/scp_scripts/copy_to_local.sh | 7 +++++++\r\n utils/exclude.txt | 4 ++++\r\n 84 files changed, 3714 insertions(+), 69 deletions(-)\r\n create mode 100644 dev/alfred/berlin/gt_actions/sample_causal_32gpus.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_actions/sample_darkness_filter.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_actions/train_dynacmis_on_3nodes_2gpu.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_actions/train_dynacmis_on_8gpu.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_actions/train_dynacmis_on_more_than4.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_actions/train_dynacmis_overfit.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_actions/train_dynacmis_overfit_2gpus.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_actions/train_dynacmis_overfit_4gpu.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_actions/train_dynacmis_overfit_4gpu_to_8gpu.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_actions/train_tok_topology_one_gpu copy.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_actions/train_tok_topology_one_gpu.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_actions/train_tok_topology_restore_to_on_gpu.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_actions/train_tok_topology_two_gpus.sbatch\r\n create mode 100644 dev/alfred/berlin/gt_sampling/sampling_mock.sbatch\r\n delete mode 100644 dev/alfred/berlin/job_requeueing/notes.md\r\n create mode 100644 dev/alfred/berlin/test_franz_pr/train_dynacmis_overfit.sbatch\r\n create mode 100644 dev/alfred/berlin/test_franz_pr/train_lam_overfit.sbatch\r\n create mode 100644 dev/alfred/berlin/test_franz_pr/train_tokenizer_overfit.sbatch\r\n create mode 100644 dev/alfred/berlin/topology/sample_overfit_single_gpu.sbatch\r\n create mode 100644 dev/alfred/berlin/topology/train_dynacmis_overfit_1.sbatch\r\n create mode 100644 dev/alfred/berlin/topology/train_dynacmis_overfit_1_gt_actions.sbatch\r\n create mode 100644 dev/alfred/berlin/topology/train_dynacmis_overfit_1_gt_actions_noise.sbatch\r\n create mode 100644 dev/alfred/berlin/topology/train_dynacmis_overfit_1_noise.sbatch\r\n create mode 100644 dev/alfred/berlin/topology/train_dynacmis_overfit_2_nodes_2_gpu.sbatch\r\n create mode 100644 dev/alfred/berlin/topology/train_dynacmis_overfit_2_nodes_4_gpu.sbatch\r\n create mode 100644 dev/alfred/berlin/topology/train_dynacmis_overfit_4gpu.sbatch\r\n create mode 100644 dev/alfred/berlin/topology/train_lam_overfit.sbatch\r\n create mode 100644 dev/alfred/berlin/topology/train_tokenizer_overfit_4_gpu.sbatch\r\n create mode 100644 dev/alfred/berlin/topology/train_tokenizer_overfit_single_record.sbatch\r\n create mode 100644 dev/alfred/berlin/topology/train_tokenizer_overfit_single_record_requeue.sbatch\r\n create mode 100644 dev/alfred/berlin/topology_restore_fix/train_tokenizer_overfit_4_gpu.sbatch\r\n create mode 100644 dev/alfred/berlin/topology_restore_fix/train_tokenizer_overfit_4_to_1_gpu.sbatch\r\n rename dev/alfred/berlin/train_dyn_dev/{train_dynacmis.sbatch => train_dynacmis_overfit.sbatch} (71%)\r\n mode change 100644 => 100755 dev/alfred/berlin/train_lam_dev/train_lam.sbatch\r\n create mode 100644 dev/alfred/berlin/train_tok_dev/train_tok_overfit.sbatch\r\n create mode 100644 dev/alfred/helmholtz_cluster/train_tok_dev/train_tok.sbatch\r\n create mode 100644 dev/alfred/helmholtz_cluster/train_tok_dev/train_tok_overfit.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_full_prec.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_full_prec_cosine.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_full_prec_w_restore.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_mix_prec_cosine.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction/coinrun_lam_reproduction_mix_prec_cosine_min_init_lt_3e-6.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction/sample.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction/sample_cotrain.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_dynamics_reproduction.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_dynamics_reproduction_cotrain.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_dc_0.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_ffn_512.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_ffn_512_num_blocks_8.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_lower_lr.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_lower_lr_0.5x.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_lower_lr_3e6_1e5_0.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_lam_reproduction_lower_lr_3e6_3e5_0.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_tokenizer_repoduction.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/coinrun_tokenizer_repoduction_ffn_512_n_blocks_8.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/generate_data.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/sample.sbatch\r\n create mode 100644 jobs/alfred/berlin/coinrun/coinrun_reproduction_10k/sample_cotrain.sbatch\r\n create mode 100644 jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch\r\n create mode 100644 jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_dynamics_reproduction_requeue.sbatch\r\n create mode 100644 jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_dynamics_reproduction_requeue_2.0.sbatch\r\n create mode 100644 jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_tokenizer_repoduction_requeue.sbatch\r\n create mode 100644 jobs/alfred/berlin/jafar_og_reproduction/og_coinrun_tokenizer_repoduction_requeue_2.0.sbatch\r\n create mode 100644 jobs/alfred/berlin/jafar_og_reproduction/sample.sbatch\r\n create mode 100644 jobs/alfred/berlin/jafar_og_reproduction/sample_og_150k.sbatch\r\n create mode 100644 jobs/alfred/berlin/jafar_og_reproduction/sample_og_175k.sbatch\r\n create mode 100644 jobs/alfred/berlin/jafar_og_reproduction/sample_og_200k.sbatch\r\n create mode 100644 jobs/alfred/berlin/jafar_og_reproduction/sample_og_spawner.sh\r\n create mode 100644 jobs/alfred/berlin/jafar_og_reproduction/spawner_sample.sh\r\n create mode 100644 jobs/alfred/helmholtz_cluster/jafar_og_reproduction/generate_dataset.sbatch\r\n create mode 100644 jobs/alfred/helmholtz_cluster/jafar_og_reproduction/generate_dataset_10m.sbatch\r\n create mode 100644 jobs/alfred/helmholtz_cluster/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch\r\n create mode 100644 jobs/alfred/helmholtz_cluster/jafar_og_reproduction/og_coinrun_lam_reproduction.sbatch\r\n create mode 100644 jobs/alfred/helmholtz_cluster/jafar_og_reproduction/og_coinrun_tokenizer_repoduction.sbatch\r\n create mode 100644 jobs/franz/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_fp32_adam_moments.sh\r\n create mode 100644 jobs/franz/berlin/coinrun/coinrun_dynamics/coinrun_dynamics_fp32_layernorm.sh\r\n create mode 100644 jobs/franz/berlin/coinrun/coinrun_lam/lam_coinrun_nan_investigation_100k_to_107k.sbatch\r\n create mode 100644 utils/alfred/scp_scripts/copy_to_local.sh\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine/slurm",,terminal_output diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-33d627dc-2555-472b-a3d4-76d3b06878631756723263490-2025_09_01-12.41.38.380/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-33d627dc-2555-472b-a3d4-76d3b06878631756723263490-2025_09_01-12.41.38.380/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..97197aaea46aa8e34487c7a794363867d0a8d5c9 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-33d627dc-2555-472b-a3d4-76d3b06878631756723263490-2025_09_01-12.41.38.380/source.csv @@ -0,0 +1,3463 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +1,4,"train_lam.py",0,0,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom models.lam import LatentActionModel\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n vq_beta: float = 0.25\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n vq_reset_thresh: int = 50\n darkness_threshold: float = 0.0\n # LAM\n model_dim: int = 512\n ffn_dim: int = 2048\n latent_dim: int = 32\n num_latents: int = 6\n patch_size: int = 16\n num_blocks: int = 4\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.0\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_lam""\n tags: list[str] = field(default_factory=lambda: [""lam""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\n\ndef lam_loss_fn(\n model: LatentActionModel, inputs: dict\n) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n # --- Compute loss ---\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n model.train()\n outputs = model(inputs, training=True)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=(index_counts != 0).mean(),\n )\n return loss, (outputs[""recon""], index_counts, metrics)\n\n\n@nnx.jit\ndef train_step(\n lam: LatentActionModel,\n optimizer: nnx.Optimizer,\n inputs: dict,\n action_last_active: jax.Array,\n rng: jax.Array,\n) -> tuple[jax.Array, jax.Array, jax.Array, dict]:\n def loss_fn(\n model: LatentActionModel,\n ) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n return lam_loss_fn(model, inputs)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)\n\n # --- Reset inactive latent actions ---\n codebook = lam.vq.codebook\n num_codes = len(codebook)\n active_codes = idx_counts != 0.0\n action_last_active = jnp.where(active_codes, 0, action_last_active + 1)\n p_code = active_codes / active_codes.sum()\n reset_idxs = jax.random.choice(rng, num_codes, shape=(num_codes,), p=p_code)\n do_reset = action_last_active >= args.vq_reset_thresh\n new_codebook = jnp.where(\n jnp.expand_dims(do_reset, -1), codebook[reset_idxs], codebook.value\n )\n lam.vq.codebook.value = new_codebook\n action_last_active = jnp.where(do_reset, 0, action_last_active)\n return loss, recon, action_last_active, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n ffn_dim=args.ffn_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n\n # Count parameters\n _, params, _ = nnx.split(lam, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n optimizer = nnx.Optimizer(lam, tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n darkness_threshold=args.darkness_threshold,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in grain_iterator\n )\n print(f""Starting training from step {step}..."")\n action_last_active = jnp.zeros(args.num_latents, dtype=jnp.int32)\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng = jax.random.split(rng)\n\n inputs = dict(videos=videos, rng=_rng)\n rng, _rng = jax.random.split(rng)\n loss, recon, action_last_active, metrics = train_step(\n lam, optimizer, inputs, action_last_active, _rng\n )\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0, 1:].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n grain_iterator # type: ignore\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +2,512,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:41:38 PM [info] Activating crowd-code\n12:41:38 PM [info] Recording started\n12:41:38 PM [info] Initializing git provider using file system watchers...\n12:41:38 PM [info] Git repository found\n12:41:38 PM [info] Git provider initialized successfully\n12:41:38 PM [info] Initial git state: [object Object]\n",Log,tab +3,37789,"train_lam.py",0,0,"",python,tab +4,38854,"train_lam.py",9464,0,"",python,selection_mouse +5,40228,"train_lam.py",9405,0,"",python,selection_mouse +6,40759,"train_lam.py",9466,0,"",python,selection_command +7,40917,"train_lam.py",9488,0,"",python,selection_command +8,41316,"train_lam.py",9517,0,"",python,selection_command +9,41537,"train_lam.py",9563,0,"",python,selection_command +10,41833,"train_lam.py",9517,0,"",python,selection_command +11,41985,"train_lam.py",9563,0,"",python,selection_command +12,42137,"train_lam.py",9517,0,"",python,selection_command +13,47931,"TERMINAL",0,0,"",,terminal_focus +14,50060,"TERMINAL",0,0,"source /home/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/bin/activate",,terminal_command +15,50088,"TERMINAL",0,0,"]633;E;2025-09-01 12:42:28 source /home/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/bin/activate;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +16,52744,"TERMINAL",0,0,"clear",,terminal_command +17,52778,"TERMINAL",0,0,"]633;E;2025-09-01 12:42:31 clear;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +18,54926,"TERMINAL",0,0,"idling",,terminal_command +19,54984,"TERMINAL",0,0,"]633;E;2025-09-01 12:42:33 idling;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Mon Sep 1 12:42:33 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 131 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated: 33 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 0 nodes idle\rPartition accelerated-h200:\t 0 nodes idle",,terminal_output +20,56067,"TERMINAL",0,0,"4\t ",,terminal_output +21,57070,"TERMINAL",0,0,"5\t ",,terminal_output +22,58129,"TERMINAL",0,0,"6\t ",,terminal_output +23,59188,"TERMINAL",0,0,"7\t ",,terminal_output +24,64570,"train_dynamics.py",0,0,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n darkness_threshold: float = 0.0\n # Tokenizer\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_type: str = ""maskgit"" # supported options: maskgit, causal\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n wandb_id: str = """"\n\n\nargs = tyro.cli(Args)\n\n\ndef dynamics_loss_fn(\n model: Genie, inputs: dict\n) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n """"""Compute masked dynamics loss""""""\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n model.train()\n outputs = model(inputs, training=True)\n mask = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n gt = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]), size=args.num_latent_actions, fill_value=0\n )\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]), size=args.num_patch_latents, fill_value=0\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n psnr=psnr,\n ssim=ssim,\n codebook_usage_lam=codebook_usage_lam,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n\n@nnx.jit\ndef train_step(\n model: Genie, optimizer: nnx.Optimizer, inputs: dict\n) -> tuple[jax.Array, jax.Array, dict]:\n """"""Update state and compute metrics""""""\n\n def loss_fn(model: Genie) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n return dynamics_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(model)\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n decode=False,\n rngs=rngs,\n )\n\n _, params, _ = nnx.split(genie, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n optimizer = nnx.Optimizer(genie, tx)\n del genie\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n darkness_threshold=args.darkness_threshold,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n optimizer = restore_genie_components(optimizer, replicated_sharding, rng, args)\n # NOTE: We have to remove the (unused) tokenizer vq dropout due flax.nnx lazily initializing modules.\n # Specifically, the first dynamics model checkpoint will contain the vq dropout module,\n # but the first full restore will fail due to nnx not initializing the module when\n # dropout is set to 0.0.\n del optimizer.model.tokenizer.vq.drop\n\n # --- TRAIN LOOP ---\n dataloader = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in grain_iterator\n )\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng_mask = jax.random.split(rng, 2)\n inputs = dict(videos=videos, mask_rng=_rng_mask)\n loss, recon, metrics = train_step(optimizer.model, optimizer, inputs)\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n grain_iterator # type: ignore\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +25,406654,"TERMINAL",0,0,"idle",,terminal_command +26,406694,"TERMINAL",0,0,"]633;E;2025-09-01 12:48:24 idle;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;CPartition dev_cpuonly : 2 nodes idle\r\nPartition cpuonly : 132 nodes idle\r\nPartition dev_accelerated : 1 nodes idle\r\nPartition accelerated : 33 nodes idle\r\nPartition dev_accelerated-h100 : 0 nodes idle\r\nPartition accelerated-h100 : 0 nodes idle\r\nPartition large : 0 nodes idle\r\nPartition accelerated-h200 : 0 nodes idle\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +27,413321,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --job-name=train_dynamics_maskgit_8_node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/holiday/maskgit/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --darkness_threshold=50 \\n --dyna_type=maskgit \\n --num_latent_actions=100 \\n --max_lr=8e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-maskgit-8-node-darkness-filter-$slurm_job_id \\n --tags dynamics maskgit 8-node post-launch-main darkness-filter \\n --entity instant-uv \\n --project jafar \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16 \\n --dyna_ffn_dim=4096 \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +28,416914,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +29,420545,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=00:10:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/train_lam_minecraft_8node_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/train_lam_minecraft_8node_%j.log\n#SBATCH --job-name=train_lam_minecraft_1node_dev\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=20 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-1-node-dev-$slurm_job_id \\n --tags lam minecraft 1-node \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +30,423680,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1688,0,"",shellscript,selection_mouse +31,423686,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1687,0,"",shellscript,selection_command +32,424256,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1650,0,"",shellscript,selection_mouse +33,424290,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1649,0,"",shellscript,selection_command +34,424894,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1688,0,"",shellscript,selection_mouse +35,424895,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1687,0,"",shellscript,selection_command +36,425478,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1688,0,"",shellscript,selection_mouse +37,425490,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1687,0,"",shellscript,selection_command +38,426207,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1668,0,"",shellscript,selection_mouse +39,426213,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1667,0,"",shellscript,selection_command +40,442135,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",38,0,"",shellscript,selection_mouse +41,442145,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",37,0,"",shellscript,selection_command +42,479533,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1732,0,"",shellscript,selection_mouse +43,479538,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1731,0,"",shellscript,selection_command +44,480001,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1720,0,"",shellscript,selection_mouse +45,480012,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1719,0,"",shellscript,selection_command +46,481122,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +47,483951,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2074,0,"",shellscript,selection_mouse +48,483986,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",2073,0,"",shellscript,selection_command +49,484763,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",1973,0,"",shellscript,selection_mouse +50,485330,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",1943,0,"",shellscript,selection_mouse +51,487847,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +52,490060,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1720,0,"\n --darkness_threshold=50 \",shellscript,content +53,490118,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1725,0,"",shellscript,selection_command +54,496206,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +55,504449,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark_req.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit/dynamics-cotraining/%x_%j.log\n#SBATCH --job-name=train_dynamics_maskgit_8_node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=3423250\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/holiday/maskgit/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_1e-4/3412401\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --restore_ckpt \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --darkness_threshold=50 \\n --dyna_type=maskgit \\n --num_latent_actions=100 \\n --max_lr=8e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-maskgit-8-node-darkness-filter-$slurm_job_id \\n --tags dynamics maskgit 8-node post-launch-main darkness-filter \\n --entity instant-uv \\n --project jafar \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16 \\n --dyna_ffn_dim=4096 \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +56,509893,"slurm/jobs/mihir/horeka/mask_prob_fix/train_dynamics_8_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit-maskprob-fix/dynamics-cotraining/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/maskgit-maskprob-fix/dynamics-cotraining/%x_%j.log\n#SBATCH --job-name=train_dynamics_maskprob_fix_8_node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\n# CHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/maskgit-maskprob-fix/$job_name/$slurm_job_id\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/maskgit-maskprob-fix/train_dynamics_maskprob_fix_8_node/3371237\nmkdir -p $CHECKPOINT_DIR\n\n# tokenizer with the new structure supporting larger ffn_dim\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/train_tokenizer_lr_sweep_1e-4_larger_ffn/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --restore_ckpt \\n --wandb_id=3371237 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --init_lr=0 \\n --max_lr=8e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-maskprob-fix-8-node-3371237 \\n --tags dynamics maskprob-fix 8-node \\n --entity instant-uv \\n --project jafar \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16 \\n --dyna_ffn_dim=4096 \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +57,518730,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/atari/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/atari/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_3e-5\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n # optional: trigger checkpoint saving here\n # e.g., touch $checkpoint_dir/requeue_trigger\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_atari/array_records\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n --image_height=84 \\n --image_width=64 \\n $restore_ckpt_flag \\n --wandb_id $SLURM_JOB_ID \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=48 \\n --init_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=atari-tokenizer-3e-5-$slurm_job_id \\n --tags tokenizer atari 3e-5 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n\n",shellscript,tab +58,520317,"slurm/jobs/mihir/horeka/atari/train_tokenizer_lr_3e-5.sbatch",456,0,"",shellscript,selection_mouse +59,522754,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +60,524822,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +61,525775,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",483,0,"",shellscript,selection_mouse +62,525966,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",483,19,"ueue\n#SBATCH --sign",shellscript,selection_mouse +63,525967,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",483,56,"ueue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n",shellscript,selection_mouse +64,526729,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",436,0,"",shellscript,selection_mouse +65,527758,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",469,0,"\n#SBATCH --reservation=llmtum",shellscript,content +66,527777,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",470,0,"",shellscript,selection_command +67,541847,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +68,550189,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",,terminal_command +69,550221,"TERMINAL",0,0,"]633;E;2025-09-01 12:50:48 sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C",,terminal_output +70,550956,"TERMINAL",0,0,"queue",,terminal_command +71,551044,"TERMINAL",0,0,"]633;E;2025-09-01 12:50:49 queue;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 12:50:49 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454890 accelerat train_la tum_cte0 R\t0:00\t 1 hkn0410",,terminal_output +72,552085,"TERMINAL",0,0,"501",,terminal_output +73,553128,"TERMINAL",0,0,"12",,terminal_output +74,554194,"TERMINAL",0,0,"23",,terminal_output +75,554790,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +76,555964,"TERMINAL",0,0,"idling",,terminal_command +77,556042,"TERMINAL",0,0,"]633;E;2025-09-01 12:50:54 idling;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Mon Sep 1 12:50:54 2025Partition dev_cpuonly:\t 6 nodes idle\rPartition cpuonly: 131 nodes idle\rPartition dev_accelerated:\t 1 nodes idle\rPartition accelerated: 32 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 0 nodes idle\rPartition accelerated-h200:\t 0 nodes idle",,terminal_output +78,557084,"TERMINAL",0,0,"5\t ",,terminal_output +79,557885,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +80,560084,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +81,595042,"TERMINAL",0,0,"queue",,terminal_command +82,595103,"TERMINAL",0,0,"]633;E;2025-09-01 12:51:33 queue;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C",,terminal_output +83,595167,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 12:51:33 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454890 accelerat train_la tum_cte0 R\t0:44\t 1 hkn0410",,terminal_output +84,596173,"TERMINAL",0,0,"45",,terminal_output +85,597213,"TERMINAL",0,0,"56",,terminal_output +86,598095,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +87,606660,"TERMINAL",0,0,"",,terminal_focus +88,608790,"TERMINAL",0,0,"source /home/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/bin/activate",,terminal_command +89,608814,"TERMINAL",0,0,"]633;E;2025-09-01 12:51:47 source /home/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/bin/activate;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +90,609423,"TERMINAL",0,0,"logs",,terminal_command +91,609460,"TERMINAL",0,0,"]633;E;2025-09-01 12:51:47 logs;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +92,610920,"TERMINAL",0,0,"ls",,terminal_command +93,610983,"TERMINAL",0,0,"]633;E;2025-09-01 12:51:49 ls;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C",,terminal_output +94,611300,"TERMINAL",0,0,"atari train_lam_action_space_scaling_50_3329804.log\r\nbig_run train_lam_action_space_scaling_50_3331286.log\r\nbig-runs train_lam_action_space_scaling_6_3318549.log\r\ncausal train_lam_action_space_scaling_6_3320178.log\r\ncoinrun train_lam_action_space_scaling_6_3321528.log\r\nlam train_lam_action_space_scaling_6_3329790.log\r\nmaskgit train_lam_action_space_scaling_6_3329805.log\r\nmaskgit-maskprob-fix train_lam_action_space_scaling_6_3331287.log\r\npreprocess train_lam_action_space_scaling_8_3318550.log\r\ntrain_dyn_causal_180M_3372931.log train_lam_action_space_scaling_8_3329791.log\r\ntrain_dyn_causal_180M_3372963.log train_lam_action_space_scaling_8_3329806.log\r\ntrain_dyn_causal_180M_3372969.log train_lam_action_space_scaling_8_3331288.log\r\ntrain_dyn_causal_180M_3373107.log train_lam_minecraft_overfit_sample_3309655.log\r\ntrain_dyn_causal_255M_3372932.log train_lam_model_size_scaling_38M_3317098.log\r\ntrain_dyn_causal_255M_3372970.log train_lam_model_size_scaling_38M_3317115.log\r\ntrain_dyn_causal_255M_3373108.log train_lam_model_size_scaling_38M_3317231.log\r\ntrain_dyn_causal_356M_3372934.log train_tokenizer_batch_size_scaling_16_node_3321526.log\r\ntrain_dyn_causal_356M_3372971.log train_tokenizer_batch_size_scaling_1_node_3318551.log\r\ntrain_dyn_causal_356M_3373109.log train_tokenizer_batch_size_scaling_2_node_3318552.log\r\ntrain_dyn_causal_500M_3372936.log train_tokenizer_batch_size_scaling_2_node_3330806.log\r\ntrain_dyn_causal_500M_3372972.log train_tokenizer_batch_size_scaling_2_node_3330848.log\r\ntrain_dyn_causal_500M_3373110.log train_tokenizer_batch_size_scaling_2_node_3331282.log\r\ntrain_dyn_new_arch-bugfixed-spatial-shift_3359343.log train_tokenizer_batch_size_scaling_4_node_3318553.log\r\ntrain_dyn_new_arch-bugfixed-temporal-shift_3359349.log train_tokenizer_batch_size_scaling_4_node_3320175.log\r\ntrain_dyn_yolorun_3333026.log train_tokenizer_batch_size_scaling_4_node_3321524.log\r\ntrain_dyn_yolorun_3333448.log train_tokenizer_batch_size_scaling_8_node_3320176.log\r\ntrain_dyn_yolorun_3335345.log train_tokenizer_batch_size_scaling_8_node_3321525.log\r\ntrain_dyn_yolorun_3335362.log train_tokenizer_minecraft_overfit_sample_3309656.log\r\ntrain_dyn_yolorun_3348592.log train_tokenizer_model_size_scaling_127M_3317233.log\r\ntrain_dyn_yolorun_new_arch_3351743.log train_tokenizer_model_size_scaling_127M_3318554.log\r\ntrain_dyn_yolorun_new_arch_3352103.log train_tokenizer_model_size_scaling_140M_3313562.log\r\ntrain_dyn_yolorun_new_arch_3352115.log train_tokenizer_model_size_scaling_140M_3316019.log\r\ntrain_dyn_yolorun_new_arch_3358457.log train_tokenizer_model_size_scaling_200M_3313563.log\r\ntrain_lam_action_space_scaling_10_3320179.log train_tokenizer_model_size_scaling_200M_3316020.log\r\ntrain_lam_action_space_scaling_10_3321529.log train_tokenizer_model_size_scaling_227M_3317234.log\r\ntrain_lam_action_space_scaling_10_3329786.log train_tokenizer_model_size_scaling_227M_3318555.log\r\ntrain_lam_action_space_scaling_10_3329801.log train_tokenizer_model_size_scaling_227M_3320173.log\r\ntrain_lam_action_space_scaling_10_3331283.log train_tokenizer_model_size_scaling_227M_3321523.log\r\ntrain_lam_action_space_scaling_12_3318546.log train_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_lam_action_space_scaling_12_3320177.log train_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_lam_action_space_scaling_12_3321527.log train_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_lam_action_space_scaling_12_3329787.log train_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_lam_action_space_scaling_12_3329802.log train_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_lam_action_space_scaling_12_3331284.log train_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_lam_action_space_scaling_20_3318547.log train_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_lam_action_space_scaling_20_3329788.log train_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_lam_action_space_scaling_20_3329803.log train_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_lam_action_space_scaling_20_3331285.log train_tokenizer_model_size_scaling_80M_3316026.log\r\ntrain_lam_action_space_scaling_50_3320180.log yoloruns\r\ntrain_lam_action_space_scaling_50_3329789.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +95,620547,"TERMINAL",0,0,"cd lam/",,terminal_command +96,620817,"TERMINAL",0,0,"ls",,terminal_command +97,620889,"TERMINAL",0,0,"]633;E;2025-09-01 12:51:59 ls;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C",,terminal_output +98,621063,"TERMINAL",0,0,"train_lam_minecraft_8node_3431870.log train_lam_minecraft_8node_3431876.log train_lam_minecraft_8node_3431895.log\r\ntrain_lam_minecraft_8node_3431875.log train_lam_minecraft_8node_3431885.log train_lam_minecraft_8node_3454890.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +99,626411,"TERMINAL",0,0,"ls",,terminal_command +100,631974,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +101,650494,"TERMINAL",0,0,"queue",,terminal_command +102,650598,"TERMINAL",0,0,"]633;E;2025-09-01 12:52:28 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 12:52:28 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454890 accelerat train_la tum_cte0 R\t1:39\t 1 hkn0410",,terminal_output +103,651626,"TERMINAL",0,0,"940",,terminal_output +104,652282,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +105,656326,"TERMINAL",0,0,"tail -f train_lam_minecraft_8node_3454890.log",,terminal_command +106,656368,"TERMINAL",0,0,"]633;E;2025-09-01 12:52:34 tail -f train_lam_minecraft_8node_3454890.log;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;CSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0410\r\nGpuFreq=control_disabled\r\nwandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\nwandb: Tracking run with wandb version 0.19.11\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250901_125152-3454890\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run lam-minecraft-1-node-dev-3454890\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/3454890\r\n",,terminal_output +107,710344,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 4).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 4).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 4).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 4).\r\nRunning on 4 devices.\r\nCounting all components: ['action_in', 'action_up', 'decoder', 'encoder', 'patch_up', 'vq']\r\nParameter counts:\r\n{'action_in': 768, 'action_up': 16896, 'decoder': 17474816, 'encoder': 17228832, 'patch_up': 393728, 'vq': 3200, 'total': 35118240}\r\nStarting training from step 0...\r\nRunning on 4 devices.\r\nCounting all components: ['action_in', 'action_up', 'decoder', 'encoder', 'patch_up', 'vq']\r\nParameter counts:\r\n{'action_in': 768, 'action_up': 16896, 'decoder': 17474816, 'encoder': 17228832, 'patch_up': 393728, 'vq': 3200, 'total': 35118240}\r\nStarting training from step 0...\r\nRunning on 4 devices.\r\nCounting all components: ['action_in', 'action_up', 'decoder', 'encoder', 'patch_up', 'vq']\r\nParameter counts:\r\n{'action_in': 768, 'action_up': 16896, 'decoder': 17474816, 'encoder': 17228832, 'patch_up': 393728, 'vq': 3200, 'total': 35118240}\r\nStarting training from step 0...\r\nRunning on 4 devices.\r\nCounting all components: ['action_in', 'action_up', 'decoder', 'encoder', 'patch_up', 'vq']\r\nParameter counts:\r\n{'action_in': 768, 'action_up': 16896, 'decoder': 17474816, 'encoder': 17228832, 'patch_up': 393728, 'vq': 3200, 'total': 35118240}\r\nStarting training from step 0...\r\n",,terminal_output +108,730357,"TERMINAL",0,0,"2025-09-01 12:53:47.644825: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +109,731365,"TERMINAL",0,0,"2025-09-01 12:53:49.181617: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-01 12:53:49.182147: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +110,737379,"TERMINAL",0,0,"2025-09-01 12:53:54.741874: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-01 12:53:54.741904: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +111,772400,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +112,785818,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1697,0,"",shellscript,selection_mouse +113,785851,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1696,0,"",shellscript,selection_command +114,786354,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1791,0,"",shellscript,selection_mouse +115,786362,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1790,0,"",shellscript,selection_command +116,786980,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1962,0,"",shellscript,selection_mouse +117,786997,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1961,0,"",shellscript,selection_command +118,811809,"slurm/utils/mihir/model_sizes.md",0,0,"# Genie 1 - Model Sizes and their configs\n\n## Tokenizer model: sizes\n\ndefault: \n| Model | model_dim | num_blocks | num_heads | latent_dim | num_latents | Est. Params |\n|-------|-----------|------------|-----------|------------|-------------|-------------|\n| default | 512 | 8 | 8 | 32 | 1024 | ~38M |\n\n### scaling up \n#### (not tested yet - TODO @mihir)\n\n| Model | model_dim | num_blocks | num_heads | latent_dim | num_latents | Est. Params |\n|-------|-----------|------------|-----------|------------|-------------|-------------|\n| L1 | 768 | 12 | 12 | 64 | 2048 | ~80M |\n| L2 | 1024 | 12 | 16 | 128 | 2048 | ~140M |\n| L3 | 1152 | 16 | 16 | 128 | 4096 | ~200M |\n| L4 | 896 | 16 | 14 | 96 | 4096 | ~120M |\n| L5 | 1536 | 12 | 24 | 256 | 2048 | ~190M |\n\n\n### tiny models\n| Model | model_dim | num_blocks | num_heads | latent_dim | num_latents | Est. Params |\n|-------|-----------|------------|-----------|------------|-------------|-------------|\n| S1 | 128 | 2 | 2 | 8 | 128 | ~0.6M |\n| S2 | 192 | 2 | 3 | 16 | 128 | ~1.3M |\n| S3 | 256 | 3 | 4 | 16 | 256 | ~3.6M |\n| S4 | 320 | 4 | 5 | 24 | 256 | ~7.4M |\n| S5 | 384 | 4 | 6 | 32 | 512 | ~10M |\n\n\n## Latent Action model: sizes\ndefault: \n| Model | model_dim | num_blocks | num_heads | latent_dim | num_latents | Est. Params |\n|-------|-----------|------------|-----------|------------|-------------|-------------|\n| default | 512 | 8 | 8 | 32 | 6 | ~39M |\n\n### scaling up \n#### (not tested yet - TODO @mihir)\n\n| Name | model_dim | num_blocks | num_heads | latent_dim | num_latents | Est. Params |\n|--------------|-----------|------------|-----------|------------|-------------|-------------|\n| XL | 1024 | 12 | 16 | 64 | 12 | ~200M |\n| L | 896 | 12 | 14 | 48 | 8 | ~150M |\n| M+ | 768 | 10 | 12 | 48 | 8 | ~100M |\n| M | 640 | 10 | 10 | 32 | 8 | ~70M |\n| Base+ | 512 | 12 | 8 | 32 | 8 | ~55M |\n\n\n### tiny models\n| Name | model_dim | num_blocks | num_heads | latent_dim | num_latents | Est. Params |\n|--------------|-----------|------------|-----------|------------|-------------|-------------|\n| XS | 128 | 2 | 2 | 8 | 4 | ~0.9M |\n| S | 160 | 2 | 2 | 8 | 4 | ~1.3M |\n| S+ | 192 | 3 | 3 | 8 | 4 | ~2.4M |\n| M- | 256 | 4 | 4 | 16 | 6 | ~5.4M |\n| M | 320 | 6 | 4 | 16 | 6 | ~12M |\n\n\n## Dynamics model: sizes \n\n| Config | dyna_dim | dyna_num_blocks | dyna_num_heads | dyna_ffn_dim | Approx. Params |\n|--------|----------|-----------------|---------------|----------------|----------------|\n| 1 | 512 | 12 | 8 || ~36M (77M total) |\n| 2 | 768 | 16 | 12 || ~110M (180M total) |\n| 3 | 1024 | 16 | 16 | 4096 | ~180M (255M total) |\n| 4 | 1024 | 24 | 16 | 4096 | ~270M (356M total) |\n| 5 | 1536 | 24 | 24 || ~500M |\n\n\n### tiny models\n| Config | dyna_dim | dyna_num_blocks | dyna_num_heads | Approx. Params |\n|--------|----------|-----------------|---------------|----------------|\n| A | 128 | 2 | 4 | ~1.5M |\n| B | 256 | 2 | 4 | ~3.5M |\n| C | 256 | 4 | 4 | ~6M |\n| D | 384 | 4 | 6 | ~12M |\n| E | 512 | 4 | 8 | ~18M |\n",markdown,tab +119,823405,"TERMINAL",0,0,"Filtering out sequence with average brightness 24.848534413194443, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.483316305121527, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.379828300347222, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.673830663194455, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.52555222526041, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.932063086371528, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.47289565581598, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.331307913628475, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.90293144574653, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.358255647135415, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.67461656380208, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.858823306857644, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.6497311453993, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.868902018663185, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 22.599402040364577, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 6.572883183159724, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.96038383506946, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 8.203265067708331, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.007096327256946, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.5221109392361, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.40983376866319, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.34761852473958, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.87565206553818, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.588127897569443, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.84445749262152, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.924731419270834, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.91030475651041, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.05798519618057, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.91061917274305, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.42965113845485, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.0773638298611, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.14640914453125, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.54003396831597, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.25781172048611, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 34.07691591536458, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.227902725260414, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.417728900173614, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.773970571614585, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.487862317708334, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.97858213107639, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.753541560329865, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.766726998697905, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.476055635850699, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.055581654947915, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 5.627149081163195, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.31705893880208, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.00060356206597, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.88665690364583, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 45.50409288498264, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 31.61641055902778, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.5580957096354, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.720037359809027, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 31.478678345052078, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.79012843446179, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.436763125434034, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.983562127604166, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.378709736979154, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.00539932725694, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.565465722656253, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.30490064756945, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.1476609592014, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.61714330729165, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 37.23104319097222, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 31.846979019965282, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.43747165147569, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.2896229765625, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.63308103428819, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.74481558463542, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.217998243055554, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 47.57220169444444, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.944998164062508, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.762338558159723, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 23.760039276041663, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.60974422829861, which is below the darkness threshold 50.0.\r\n",,terminal_output +120,833416,"TERMINAL",0,0,"Filtering out sequence with average brightness 24.139636642795132, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.84985505468751, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 8.405905884982642, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.82792859288194, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.425166213975697, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.74526238064237, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.82678295529514, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.942426434461805, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.580835393229165, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.77418353168403, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.001715821180554, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.403384335069443, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.892703191406248, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.41814324392361, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.46294229730904, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.23780273263889, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 16.99111554817708, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.177946543402776, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.17766344184027, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.860548177517362, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.05504833940973, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 29.932533608072916, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.527453131944444, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 7.137750802083334, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.58781443142361, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.79947780121527, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.68925760416666, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 30.60300879687501, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 33.95528398654513, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 15.448870614583333, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.355783140190965, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.60541607118055, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.53008347786458, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.785689658854174, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 49.518580784722225, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.513503248263895, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 46.004052375434014, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 26.30585684027778, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 7.372377047743056, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.12718165060765, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.04163297352432, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 17.51940406640625, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 27.589010426215275, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 48.519365423611106, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 35.50765792187501, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 39.19852556336807, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 24.427056894965272, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 44.802222088107634, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.85009925086805, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.442516197048604, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 5.21316979861111, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 14.449708920138889, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 32.485645563802095, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.516910422309028, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.553891733940965, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.65755801692708, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 13.071391086371525, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.324552201822904, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 19.327773061197917, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 18.491498563368054, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 42.80723423220486, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 40.19733232638888, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 43.33489234418402, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 7.683369060329861, which is below the darkness threshold 50.0.\r\nFiltering out episode with length 14, which is shorter than the requested sequence length 16.\r\nFiltering out sequence with average brightness 20.030340970920133, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 36.607144654947916, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 9.051733965277778, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.172398512586808, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 25.251918238715277, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 38.64909593532985, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 34.981573673611095, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 11.190590286892363, which is below the darkness threshold 50.0.\r\nFiltering out sequence with average brightness 41.483688669704854, which is below the darkness threshold 50.0.\r\n[Mon Sep 1 12:55:31 PM CEST 2025] caught sigusr1 (timeout warning), requeueing slurm job 3454890...\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3454890.0 ON hkn0410 CANCELLED AT 2025-09-01T12:55:31 DUE TO JOB REQUEUE ***\r\nslurmstepd: error: *** JOB 3454890 ON hkn0410 CANCELLED AT 2025-09-01T12:55:31 DUE TO JOB REQUEUE ***\r\n",,terminal_output +121,845130,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;130",,terminal_output +122,846244,"TERMINAL",0,0,"queue",,terminal_command +123,846297,"TERMINAL",0,0,"]633;E;2025-09-01 12:55:44 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C",,terminal_output +124,846361,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 12:55:44 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454890 accelerat train_la tum_cte0 CG\t0:00\t 1 hkn0410",,terminal_output +125,847394,"TERMINAL",0,0,"5\t ",,terminal_output +126,848446,"TERMINAL",0,0,"6\t ",,terminal_output +127,849170,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +128,851432,"TERMINAL",0,0,"scancel 3454890",,terminal_command +129,851500,"TERMINAL",0,0,"]633;E;2025-09-01 12:55:49 scancel 3454890;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +130,854904,"TERMINAL",0,0,"queue",,terminal_command +131,854966,"TERMINAL",0,0,"]633;E;2025-09-01 12:55:53 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 12:55:53 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454890 accelerat train_la tum_cte0 CG\t4:42\t 1 hkn0410",,terminal_output +132,856016,"TERMINAL",0,0,"4\t ",,terminal_output +133,857060,"TERMINAL",0,0,"5\t ",,terminal_output +134,858106,"TERMINAL",0,0,"6\t ",,terminal_output +135,858412,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +136,863399,"TERMINAL",0,0,"bash",,terminal_focus +137,867312,"TERMINAL",0,0,"bash",,terminal_focus +138,869867,"TERMINAL",0,0,"queu",,terminal_command +139,869946,"TERMINAL",0,0,"]633;E;2025-09-01 12:56:08 queu;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;Cbash: queu: command not found...\r\n",,terminal_output +140,871014,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;130",,terminal_output +141,871712,"TERMINAL",0,0,"queue",,terminal_command +142,871788,"TERMINAL",0,0,"]633;E;2025-09-01 12:56:09 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C",,terminal_output +143,871821,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 12:56:09 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454890 accelerat train_la tum_cte0 CG\t4:42\t 1 hkn0410",,terminal_output +144,872787,"TERMINAL",0,0,"11\t ",,terminal_output +145,873869,"TERMINAL",0,0,"2\t ",,terminal_output +146,875093,"slurm/utils/mihir/model_sizes.md",0,0,"",markdown,tab +147,875153,"TERMINAL",0,0,"3\t ",,terminal_output +148,875970,"TERMINAL",0,0,"4\t ",,terminal_output +149,876969,"TERMINAL",0,0,"5\t ",,terminal_output +150,877198,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +151,878050,"slurm/utils/mihir/model_sizes.md",0,0,"",markdown,tab +152,878175,"TERMINAL",0,0,"6\t ",,terminal_output +153,879089,"TERMINAL",0,0,"\r7\t ",,terminal_output +154,880118,"TERMINAL",0,0,"8\t ",,terminal_output +155,881205,"TERMINAL",0,0,"9\t ",,terminal_output +156,882211,"TERMINAL",0,0,"20\t ",,terminal_output +157,883255,"TERMINAL",0,0,"1\t ",,terminal_output +158,884341,"TERMINAL",0,0,"2\t ",,terminal_output +159,885336,"TERMINAL",0,0,"3\t ",,terminal_output +160,886386,"TERMINAL",0,0,"4\t ",,terminal_output +161,887436,"TERMINAL",0,0,"5\t ",,terminal_output +162,888477,"TERMINAL",0,0,"6\t ",,terminal_output +163,889532,"TERMINAL",0,0,"7\t ",,terminal_output +164,890608,"TERMINAL",0,0,"8\t ",,terminal_output +165,891619,"TERMINAL",0,0,"9\t ",,terminal_output +166,892676,"TERMINAL",0,0,"30\t ",,terminal_output +167,893757,"TERMINAL",0,0,"1\t ",,terminal_output +168,894765,"TERMINAL",0,0,"3\t ",,terminal_output +169,895823,"TERMINAL",0,0,"4\t ",,terminal_output +170,896861,"TERMINAL",0,0,"5\t ",,terminal_output +171,897906,"TERMINAL",0,0,"6\t ",,terminal_output +172,898955,"TERMINAL",0,0,"7\t ",,terminal_output +173,899992,"TERMINAL",0,0,"8\t ",,terminal_output +174,901037,"TERMINAL",0,0,"9\t ",,terminal_output +175,902103,"TERMINAL",0,0,"40\t ",,terminal_output +176,903115,"TERMINAL",0,0,"1\t ",,terminal_output +177,904183,"TERMINAL",0,0,"2\t ",,terminal_output +178,905223,"TERMINAL",0,0,"3\t ",,terminal_output +179,906252,"TERMINAL",0,0,"4\t ",,terminal_output +180,907288,"TERMINAL",0,0,"5\t ",,terminal_output +181,908343,"TERMINAL",0,0,"6\t ",,terminal_output +182,909387,"TERMINAL",0,0,"7\t ",,terminal_output +183,910436,"TERMINAL",0,0,"8\t ",,terminal_output +184,911477,"TERMINAL",0,0,"9\t ",,terminal_output +185,912526,"TERMINAL",0,0,"50\t ",,terminal_output +186,913573,"TERMINAL",0,0,"1\t ",,terminal_output +187,914620,"TERMINAL",0,0,"2\t ",,terminal_output +188,915656,"TERMINAL",0,0,"3\t ",,terminal_output +189,916701,"TERMINAL",0,0,"4\t ",,terminal_output +190,917745,"TERMINAL",0,0,"5\t ",,terminal_output +191,918794,"TERMINAL",0,0,"7\t ",,terminal_output +192,919850,"TERMINAL",0,0,"8\t ",,terminal_output +193,920882,"TERMINAL",0,0,"9\t ",,terminal_output +194,921927,"TERMINAL",0,0,"7:00\t ",,terminal_output +195,922960,"TERMINAL",0,0,"1\t ",,terminal_output +196,923996,"TERMINAL",0,0,"2\t ",,terminal_output +197,925044,"TERMINAL",0,0,"3\t ",,terminal_output +198,926094,"TERMINAL",0,0,"4\t ",,terminal_output +199,927148,"TERMINAL",0,0,"5\t ",,terminal_output +200,928196,"TERMINAL",0,0,"6\t ",,terminal_output +201,929273,"TERMINAL",0,0,"7\t ",,terminal_output +202,930296,"TERMINAL",0,0,"8\t ",,terminal_output +203,931347,"TERMINAL",0,0,"9\t ",,terminal_output +204,932390,"TERMINAL",0,0,"10\t ",,terminal_output +205,933440,"TERMINAL",0,0,"1\t ",,terminal_output +206,937297,"TERMINAL",0,0,"2\t ",,terminal_output +207,938340,"TERMINAL",0,0,"6\t ",,terminal_output +208,939387,"TERMINAL",0,0,"7\t ",,terminal_output +209,940422,"TERMINAL",0,0,"8\t ",,terminal_output +210,941470,"TERMINAL",0,0,"9\t ",,terminal_output +211,942507,"TERMINAL",0,0,"20\t ",,terminal_output +212,943596,"TERMINAL",0,0,"1\t ",,terminal_output +213,944621,"TERMINAL",0,0,"2\t ",,terminal_output +214,945649,"TERMINAL",0,0,"3\t ",,terminal_output +215,946687,"TERMINAL",0,0,"4\t ",,terminal_output +216,947745,"TERMINAL",0,0,"5\t ",,terminal_output +217,948820,"TERMINAL",0,0,"7\t ",,terminal_output +218,949836,"TERMINAL",0,0,"8\t ",,terminal_output +219,950875,"TERMINAL",0,0,"9\t ",,terminal_output +220,951941,"TERMINAL",0,0,"30\t ",,terminal_output +221,952977,"TERMINAL",0,0,"1\t ",,terminal_output +222,954024,"TERMINAL",0,0,"2\t ",,terminal_output +223,955068,"TERMINAL",0,0,"3\t ",,terminal_output +224,956118,"TERMINAL",0,0,"4\t ",,terminal_output +225,957158,"TERMINAL",0,0,"5\t ",,terminal_output +226,958190,"TERMINAL",0,0,"6\t ",,terminal_output +227,959263,"TERMINAL",0,0,"7\t ",,terminal_output +228,960286,"TERMINAL",0,0,"8\t ",,terminal_output +229,961336,"TERMINAL",0,0,"9\t ",,terminal_output +230,962378,"TERMINAL",0,0,"40\t ",,terminal_output +231,963425,"TERMINAL",0,0,"1\t ",,terminal_output +232,964470,"TERMINAL",0,0,"2\t ",,terminal_output +233,965523,"TERMINAL",0,0,"3\t ",,terminal_output +234,966590,"TERMINAL",0,0,"4\t ",,terminal_output +235,967608,"TERMINAL",0,0,"5\t ",,terminal_output +236,968661,"TERMINAL",0,0,"6\t ",,terminal_output +237,969702,"TERMINAL",0,0,"7\t ",,terminal_output +238,970756,"TERMINAL",0,0,"8\t ",,terminal_output +239,971803,"TERMINAL",0,0,"50\t ",,terminal_output +240,972839,"TERMINAL",0,0,"1\t ",,terminal_output +241,973881,"TERMINAL",0,0,"2\t ",,terminal_output +242,974926,"TERMINAL",0,0,"3\t ",,terminal_output +243,975966,"TERMINAL",0,0,"4\t ",,terminal_output +244,977014,"TERMINAL",0,0,"5\t ",,terminal_output +245,978066,"TERMINAL",0,0,"6\t ",,terminal_output +246,979115,"TERMINAL",0,0,"7\t ",,terminal_output +247,980155,"TERMINAL",0,0,"8\t ",,terminal_output +248,981200,"TERMINAL",0,0,"9\t ",,terminal_output +249,982248,"TERMINAL",0,0,"8:00\t ",,terminal_output +250,983279,"TERMINAL",0,0,"1\t ",,terminal_output +251,984332,"TERMINAL",0,0,"2\t ",,terminal_output +252,985363,"TERMINAL",0,0,"3\t ",,terminal_output +253,986420,"TERMINAL",0,0,"4\t ",,terminal_output +254,987473,"TERMINAL",0,0,"5\t ",,terminal_output +255,988528,"TERMINAL",0,0,"6\t ",,terminal_output +256,989569,"TERMINAL",0,0,"7\t ",,terminal_output +257,990680,"TERMINAL",0,0,"8\t ",,terminal_output +258,991661,"TERMINAL",0,0,"9\t ",,terminal_output +259,992710,"TERMINAL",0,0,"10\t ",,terminal_output +260,993755,"TERMINAL",0,0,"1\t ",,terminal_output +261,994797,"TERMINAL",0,0,"3\t ",,terminal_output +262,995867,"TERMINAL",0,0,"4\t ",,terminal_output +263,996887,"TERMINAL",0,0,"5\t ",,terminal_output +264,997025,"slurm/utils/mihir/model_sizes.md",1797,0,"",markdown,selection_mouse +265,997051,"slurm/utils/mihir/model_sizes.md",1796,0,"",markdown,selection_command +266,997521,"slurm/utils/mihir/model_sizes.md",1759,0,"",markdown,selection_mouse +267,997525,"slurm/utils/mihir/model_sizes.md",1758,0,"",markdown,selection_command +268,997693,"slurm/utils/mihir/model_sizes.md",1759,0,"",markdown,selection_mouse +269,997698,"slurm/utils/mihir/model_sizes.md",1758,0,"",markdown,selection_command +270,997817,"slurm/utils/mihir/model_sizes.md",1720,40,"| M+ | 768 | 10 | 12 | 48 | 8 | ~100M |\n",markdown,selection_mouse +271,997826,"slurm/utils/mihir/model_sizes.md",1721,39," M+ | 768 | 10 | 12 | 48 | 8 | ~100M |\n",markdown,selection_command +272,997973,"TERMINAL",0,0,"6\t ",,terminal_output +273,999012,"TERMINAL",0,0,"7\t ",,terminal_output +274,999090,"slurm/utils/mihir/model_sizes.md",1753,0,"",markdown,selection_mouse +275,999160,"slurm/utils/mihir/model_sizes.md",1753,4,"100M",markdown,selection_mouse +276,999325,"slurm/utils/mihir/model_sizes.md",1720,40,"| M+ | 768 | 10 | 12 | 48 | 8 | ~100M |\n",markdown,selection_mouse +277,1000047,"TERMINAL",0,0,"8\t ",,terminal_output +278,1000174,"slurm/utils/mihir/model_sizes.md",1727,0,"",markdown,selection_mouse +279,1001002,"slurm/utils/mihir/model_sizes.md",1754,0,"",markdown,selection_mouse +280,1001082,"TERMINAL",0,0,"9\t ",,terminal_output +281,1001764,"slurm/utils/mihir/model_sizes.md",1580,0,"",markdown,selection_mouse +282,1001933,"slurm/utils/mihir/model_sizes.md",1544,94,"|--------------|-----------|------------|-----------|------------|-------------|-------------|",markdown,selection_mouse +283,1002118,"TERMINAL",0,0,"20\t ",,terminal_output +284,1003163,"TERMINAL",0,0,"1\t ",,terminal_output +285,1004230,"TERMINAL",0,0,"2\t ",,terminal_output +286,1005254,"TERMINAL",0,0,"3\t ",,terminal_output +287,1006317,"TERMINAL",0,0,"4\t ",,terminal_output +288,1007349,"TERMINAL",0,0,"5\t ",,terminal_output +289,1007852,"slurm/utils/mihir/model_sizes.md",1666,0,"",markdown,selection_mouse +290,1007995,"slurm/utils/mihir/model_sizes.md",1665,1," ",markdown,selection_mouse +291,1008383,"TERMINAL",0,0,"6\t ",,terminal_output +292,1008775,"slurm/utils/mihir/model_sizes.md",1659,0,"",markdown,selection_mouse +293,1008908,"slurm/utils/mihir/model_sizes.md",1658,2,"16",markdown,selection_mouse +294,1009427,"TERMINAL",0,0,"7\t ",,terminal_output +295,1009545,"slurm/utils/mihir/model_sizes.md",1653,0,"",markdown,selection_mouse +296,1009661,"slurm/utils/mihir/model_sizes.md",1653,2,"12",markdown,selection_mouse +297,1010460,"TERMINAL",0,0,"8\t ",,terminal_output +298,1011510,"TERMINAL",0,0,"9\t ",,terminal_output +299,1012568,"TERMINAL",0,0,"30\t ",,terminal_output +300,1013630,"TERMINAL",0,0,"1\t ",,terminal_output +301,1014646,"TERMINAL",0,0,"2\t ",,terminal_output +302,1015293,"slurm/utils/mihir/model_sizes.md",1502,0,"",markdown,selection_mouse +303,1015679,"TERMINAL",0,0,"3\t ",,terminal_output +304,1015898,"slurm/utils/mihir/model_sizes.md",1663,0,"",markdown,selection_mouse +305,1016488,"slurm/utils/mihir/model_sizes.md",1647,0,"",markdown,selection_mouse +306,1016621,"slurm/utils/mihir/model_sizes.md",1646,4,"1024",markdown,selection_mouse +307,1016752,"TERMINAL",0,0,"4\t ",,terminal_output +308,1017650,"slurm/utils/mihir/model_sizes.md",1653,0,"",markdown,selection_mouse +309,1017745,"slurm/utils/mihir/model_sizes.md",1653,2,"12",markdown,selection_mouse +310,1017809,"TERMINAL",0,0,"6\t ",,terminal_output +311,1018264,"slurm/utils/mihir/model_sizes.md",1660,0,"",markdown,selection_mouse +312,1018418,"slurm/utils/mihir/model_sizes.md",1658,2,"16",markdown,selection_mouse +313,1018840,"TERMINAL",0,0,"7\t ",,terminal_output +314,1019183,"slurm/utils/mihir/model_sizes.md",1659,0,"",markdown,selection_mouse +315,1019864,"TERMINAL",0,0,"8\t ",,terminal_output +316,1019948,"slurm/utils/mihir/model_sizes.md",1648,0,"",markdown,selection_mouse +317,1020121,"slurm/utils/mihir/model_sizes.md",1646,4,"1024",markdown,selection_mouse +318,1020897,"TERMINAL",0,0,"9\t ",,terminal_output +319,1021934,"TERMINAL",0,0,"40\t ",,terminal_output +320,1022980,"TERMINAL",0,0,"1\t ",,terminal_output +321,1024032,"TERMINAL",0,0,"2\t ",,terminal_output +322,1024300,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +323,1025072,"TERMINAL",0,0,"3\t ",,terminal_output +324,1026135,"TERMINAL",0,0,"4\t ",,terminal_output +325,1027172,"TERMINAL",0,0,"5\t ",,terminal_output +326,1028282,"slurm/utils/mihir/model_sizes.md",0,0,"",markdown,tab +327,1028430,"TERMINAL",0,0,"6\t ",,terminal_output +328,1029268,"TERMINAL",0,0,"7\t ",,terminal_output +329,1030301,"TERMINAL",0,0,"8\t ",,terminal_output +330,1031344,"TERMINAL",0,0,"9\t ",,terminal_output +331,1031438,"slurm/utils/mihir/model_sizes.md",1647,0,"",markdown,selection_mouse +332,1032489,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +333,1032613,"TERMINAL",0,0,"50\t ",,terminal_output +334,1033362,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1849,0,"",shellscript,selection_mouse +335,1033436,"TERMINAL",0,0,"1\t ",,terminal_output +336,1033886,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1773,0,"",shellscript,selection_mouse +337,1034487,"TERMINAL",0,0,"2\t ",,terminal_output +338,1034712,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1986,0,"",shellscript,selection_mouse +339,1034716,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1985,0,"",shellscript,selection_command +340,1035535,"TERMINAL",0,0,"3\t ",,terminal_output +341,1035806,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1986,0,"\n ",shellscript,content +342,1036636,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1991,0,"-",shellscript,content +343,1036638,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1992,0,"",shellscript,selection_keyboard +344,1036672,"TERMINAL",0,0,"4\t ",,terminal_output +345,1036765,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1992,0,"-",shellscript,content +346,1036766,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1993,0,"",shellscript,selection_keyboard +347,1036997,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1993,0,"m",shellscript,content +348,1036999,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1994,0,"",shellscript,selection_keyboard +349,1037137,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1994,0,"o",shellscript,content +350,1037138,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1995,0,"",shellscript,selection_keyboard +351,1037304,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1995,0,"d",shellscript,content +352,1037306,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1996,0,"",shellscript,selection_keyboard +353,1037435,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1996,0,"e",shellscript,content +354,1037436,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1997,0,"",shellscript,selection_keyboard +355,1037670,"TERMINAL",0,0,"5\t ",,terminal_output +356,1038179,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1997,0,"k",shellscript,content +357,1038180,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1998,0,"",shellscript,selection_keyboard +358,1038680,"TERMINAL",0,0,"6\t ",,terminal_output +359,1038873,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1997,1,"",shellscript,content +360,1039727,"TERMINAL",0,0,"7\t ",,terminal_output +361,1040369,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1997,0,"l",shellscript,content +362,1040371,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1998,0,"",shellscript,selection_keyboard +363,1040737,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1998,0,"_",shellscript,content +364,1040737,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1999,0,"",shellscript,selection_keyboard +365,1040844,"TERMINAL",0,0,"9\t ",,terminal_output +366,1041817,"TERMINAL",0,0,"9:00\t ",,terminal_output +367,1042410,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1999,0,"d",shellscript,content +368,1042411,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2000,0,"",shellscript,selection_keyboard +369,1042486,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2000,0,"i",shellscript,content +370,1042487,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2001,0,"",shellscript,selection_keyboard +371,1042522,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2001,0,"m",shellscript,content +372,1042523,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2002,0,"",shellscript,selection_keyboard +373,1042933,"TERMINAL",0,0,"1\t ",,terminal_output +374,1043946,"TERMINAL",0,0,"2\t ",,terminal_output +375,1044198,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2002,0,"=1024 \",shellscript,content +376,1044956,"TERMINAL",0,0,"3\t ",,terminal_output +377,1045995,"TERMINAL",0,0,"4\t ",,terminal_output +378,1046292,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2016,0,"model_ffn_dim=4096 \\n --model_num_blocks=16 \\n --model_num_heads=16 \\n --",shellscript,content +379,1047031,"TERMINAL",0,0,"5\t ",,terminal_output +380,1048099,"TERMINAL",0,0,"6\t ",,terminal_output +381,1049135,"TERMINAL",0,0,"7\t ",,terminal_output +382,1049557,"slurm/utils/mihir/model_sizes.md",0,0,"",markdown,tab +383,1050166,"TERMINAL",0,0,"8\t ",,terminal_output +384,1051245,"TERMINAL",0,0,"9\t ",,terminal_output +385,1052060,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +386,1052281,"TERMINAL",0,0,"10\t ",,terminal_output +387,1053142,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2016,0,"",shellscript,selection_mouse +388,1053324,"TERMINAL",0,0,"1\t ",,terminal_output +389,1054205,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2016,13,"",shellscript,content +390,1054371,"TERMINAL",0,0,"2\t ",,terminal_output +391,1054577,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2016,0,"n",shellscript,content +392,1054579,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2017,0,"",shellscript,selection_keyboard +393,1054772,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2017,0,"u",shellscript,content +394,1054773,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2018,0,"",shellscript,selection_keyboard +395,1054926,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2018,0,"m",shellscript,content +396,1054927,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2019,0,"",shellscript,selection_keyboard +397,1055308,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2019,0,"_",shellscript,content +398,1055309,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2020,0,"",shellscript,selection_keyboard +399,1055427,"TERMINAL",0,0,"3\t ",,terminal_output +400,1055604,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2020,0,"b",shellscript,content +401,1055605,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2021,0,"",shellscript,selection_keyboard +402,1055832,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2021,0,"l",shellscript,content +403,1055833,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2022,0,"",shellscript,selection_keyboard +404,1055937,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2022,0,"o",shellscript,content +405,1055938,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2023,0,"",shellscript,selection_keyboard +406,1056117,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2023,0,"c",shellscript,content +407,1056118,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2024,0,"",shellscript,selection_keyboard +408,1056144,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2024,0,"k",shellscript,content +409,1056145,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2025,0,"",shellscript,selection_keyboard +410,1056486,"TERMINAL",0,0,"4\t ",,terminal_output +411,1056776,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2025,0,"s",shellscript,content +412,1056777,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2026,0,"",shellscript,selection_keyboard +413,1058930,"slurm/utils/mihir/model_sizes.md",0,0,"",markdown,tab +414,1059089,"TERMINAL",0,0,"5\t ",,terminal_output +415,1059874,"TERMINAL",0,0,"8\t ",,terminal_output +416,1060917,"TERMINAL",0,0,"9\t ",,terminal_output +417,1061964,"TERMINAL",0,0,"20\t ",,terminal_output +418,1063009,"TERMINAL",0,0,"1\t ",,terminal_output +419,1064061,"TERMINAL",0,0,"2\t ",,terminal_output +420,1065111,"TERMINAL",0,0,"3\t ",,terminal_output +421,1066166,"TERMINAL",0,0,"4\t ",,terminal_output +422,1067231,"TERMINAL",0,0,"5\t ",,terminal_output +423,1068284,"TERMINAL",0,0,"6\t ",,terminal_output +424,1069366,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +425,1069537,"TERMINAL",0,0,"7\t ",,terminal_output +426,1070365,"TERMINAL",0,0,"8\t ",,terminal_output +427,1071414,"TERMINAL",0,0,"9\t ",,terminal_output +428,1072463,"TERMINAL",0,0,"30\t ",,terminal_output +429,1073514,"TERMINAL",0,0,"1\t ",,terminal_output +430,1074150,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2030,0,"",shellscript,selection_mouse +431,1074283,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2027,4,"4096",shellscript,selection_mouse +432,1074579,"TERMINAL",0,0,"2\t ",,terminal_output +433,1075624,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2027,4,"1",shellscript,content +434,1075626,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2028,0,"",shellscript,selection_keyboard +435,1075648,"TERMINAL",0,0,"3\t ",,terminal_output +436,1075724,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2028,0,"2",shellscript,content +437,1075724,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2029,0,"",shellscript,selection_keyboard +438,1076677,"TERMINAL",0,0,"4\t ",,terminal_output +439,1077705,"TERMINAL",0,0,"5\t ",,terminal_output +440,1078748,"TERMINAL",0,0,"6\t ",,terminal_output +441,1078924,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2083,0,"409",shellscript,content +442,1078925,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2082,1,"",shellscript,content +443,1078925,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2081,0,"im",shellscript,content +444,1078925,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2080,1,"",shellscript,content +445,1078925,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2076,3,"",shellscript,content +446,1078925,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2073,2,"",shellscript,content +447,1078925,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2072,0,"ff",shellscript,content +448,1078925,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2066,6,"",shellscript,content +449,1078925,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2057,0,"2",shellscript,content +450,1078925,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2056,1,"",shellscript,content +451,1078926,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2053,0,"head",shellscript,content +452,1078926,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2048,5,"",shellscript,content +453,1078926,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2038,6,"",shellscript,content +454,1079801,"TERMINAL",0,0,"8\t ",,terminal_output +455,1079913,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2049,0,"",shellscript,selection_mouse +456,1080733,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2049,1,"",shellscript,content +457,1080853,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2049,0,"6",shellscript,content +458,1080854,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2050,0,"",shellscript,selection_keyboard +459,1080887,"TERMINAL",0,0,"9\t ",,terminal_output +460,1081899,"TERMINAL",0,0,"40\t ",,terminal_output +461,1082952,"TERMINAL",0,0,"1\t ",,terminal_output +462,1084011,"TERMINAL",0,0,"2\t ",,terminal_output +463,1084826,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2066,0,"",shellscript,selection_mouse +464,1085049,"TERMINAL",0,0,"3\t ",,terminal_output +465,1085565,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2052,0,"",shellscript,selection_mouse +466,1086080,"TERMINAL",0,0,"4\t ",,terminal_output +467,1086219,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2052,0,"\n ",shellscript,content +468,1087082,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2057,0,"-",shellscript,content +469,1087083,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2058,0,"",shellscript,selection_keyboard +470,1087167,"TERMINAL",0,0,"5\t ",,terminal_output +471,1087261,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2058,0,"-",shellscript,content +472,1087262,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2059,0,"",shellscript,selection_keyboard +473,1087467,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2059,0,"l",shellscript,content +474,1087468,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2060,0,"",shellscript,selection_keyboard +475,1087649,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2060,0,"a",shellscript,content +476,1087650,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2061,0,"",shellscript,selection_keyboard +477,1088182,"TERMINAL",0,0,"6\t ",,terminal_output +478,1089215,"TERMINAL",0,0,"7\t ",,terminal_output +479,1089623,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2061,0,"t",shellscript,content +480,1089624,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2062,0,"",shellscript,selection_keyboard +481,1089937,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2062,0,"e",shellscript,content +482,1089938,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2063,0,"",shellscript,selection_keyboard +483,1090054,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2063,0,"n",shellscript,content +484,1090055,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2064,0,"",shellscript,selection_keyboard +485,1090294,"TERMINAL",0,0,"8\t ",,terminal_output +486,1090597,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2064,0,"t",shellscript,content +487,1090598,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2065,0,"",shellscript,selection_keyboard +488,1090780,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2065,0,"_",shellscript,content +489,1090781,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2066,0,"",shellscript,selection_keyboard +490,1090942,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2066,0,"d",shellscript,content +491,1090943,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2067,0,"",shellscript,selection_keyboard +492,1091077,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2067,0,"i",shellscript,content +493,1091079,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2068,0,"",shellscript,selection_keyboard +494,1091172,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2068,0,"m",shellscript,content +495,1091173,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2069,0,"",shellscript,selection_keyboard +496,1091309,"TERMINAL",0,0,"9\t ",,terminal_output +497,1091759,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2069,0,"=",shellscript,content +498,1091760,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2070,0,"",shellscript,selection_keyboard +499,1092366,"TERMINAL",0,0,"50\t ",,terminal_output +500,1093343,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2070,0,"6",shellscript,content +501,1093344,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2071,0,"",shellscript,selection_keyboard +502,1093382,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2071,0,"4",shellscript,content +503,1093383,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2072,0,"",shellscript,selection_keyboard +504,1093418,"TERMINAL",0,0,"1\t ",,terminal_output +505,1094457,"TERMINAL",0,0,"2\t ",,terminal_output +506,1095518,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2072,0," \",shellscript,content +507,1095556,"TERMINAL",0,0,"3\t ",,terminal_output +508,1096551,"TERMINAL",0,0,"4\t ",,terminal_output +509,1097591,"TERMINAL",0,0,"5\t ",,terminal_output +510,1098312,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2095,0,"",shellscript,selection_command +511,1098653,"TERMINAL",0,0,"6\t ",,terminal_output +512,1098821,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2094,0,"",shellscript,selection_command +513,1098925,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2093,0,"",shellscript,selection_command +514,1099676,"TERMINAL",0,0,"7\t ",,terminal_output +515,1100671,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2093,0,"48",shellscript,content +516,1100671,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2091,2,"",shellscript,content +517,1100671,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,0,"2",shellscript,content +518,1100671,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2089,1,"",shellscript,content +519,1100734,"TERMINAL",0,0,"8\t ",,terminal_output +520,1101783,"TERMINAL",0,0,"3:00:00\t ",,terminal_output +521,1102818,"TERMINAL",0,0,"1\t ",,terminal_output +522,1103487,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,0,"",shellscript,selection_mouse +523,1103630,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2089,4,"2048",shellscript,selection_mouse +524,1103862,"TERMINAL",0,0,"2\t ",,terminal_output +525,1104407,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2113,0,"",shellscript,selection_mouse +526,1104920,"TERMINAL",0,0,"3\t ",,terminal_output +527,1105043,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,0,"",shellscript,selection_mouse +528,1105200,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2089,4,"2048",shellscript,selection_mouse +529,1105946,"TERMINAL",0,0,"4\t ",,terminal_output +530,1105987,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2064,0,"",shellscript,selection_mouse +531,1106120,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2059,10,"latent_dim",shellscript,selection_mouse +532,1107006,"TERMINAL",0,0,"5\t ",,terminal_output +533,1107049,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1978,0,"",shellscript,selection_mouse +534,1107196,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1969,11,"num_latents",shellscript,selection_mouse +535,1107733,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2000,0,"",shellscript,selection_mouse +536,1107932,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1993,9,"model_dim",shellscript,selection_mouse +537,1108060,"TERMINAL",0,0,"6\t ",,terminal_output +538,1108932,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,0,"",shellscript,selection_mouse +539,1109141,"TERMINAL",0,0,"7\t ",,terminal_output +540,1109665,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2095,0,"",shellscript,selection_mouse +541,1110137,"TERMINAL",0,0,"8\t ",,terminal_output +542,1111187,"TERMINAL",0,0,"9\t ",,terminal_output +543,1112227,"TERMINAL",0,0,"10\t ",,terminal_output +544,1113280,"TERMINAL",0,0,"1\t ",,terminal_output +545,1114334,"TERMINAL",0,0,"2\t ",,terminal_output +546,1115397,"TERMINAL",0,0,"3\t ",,terminal_output +547,1116419,"TERMINAL",0,0,"4\t ",,terminal_output +548,1117468,"TERMINAL",0,0,"5\t ",,terminal_output +549,1118505,"TERMINAL",0,0,"6\t ",,terminal_output +550,1119219,"train_lam.py",0,0,"",python,tab +551,1119614,"TERMINAL",0,0,"7\t ",,terminal_output +552,1120608,"TERMINAL",0,0,"8\t ",,terminal_output +553,1121678,"TERMINAL",0,0,"9\t ",,terminal_output +554,1122699,"TERMINAL",0,0,"20\t ",,terminal_output +555,1123746,"TERMINAL",0,0,"1\t ",,terminal_output +556,1124790,"TERMINAL",0,0,"3\t ",,terminal_output +557,1125851,"TERMINAL",0,0,"4\t ",,terminal_output +558,1126868,"TERMINAL",0,0,"5\t ",,terminal_output +559,1127928,"TERMINAL",0,0,"6\t ",,terminal_output +560,1128976,"TERMINAL",0,0,"7\t ",,terminal_output +561,1130045,"TERMINAL",0,0,"8\t ",,terminal_output +562,1131065,"TERMINAL",0,0,"9\t ",,terminal_output +563,1132114,"TERMINAL",0,0,"30\t ",,terminal_output +564,1133150,"TERMINAL",0,0,"1\t ",,terminal_output +565,1134201,"TERMINAL",0,0,"2\t ",,terminal_output +566,1135244,"TERMINAL",0,0,"3\t ",,terminal_output +567,1136285,"TERMINAL",0,0,"4\t ",,terminal_output +568,1137323,"TERMINAL",0,0,"5\t ",,terminal_output +569,1138367,"TERMINAL",0,0,"6\t ",,terminal_output +570,1139418,"TERMINAL",0,0,"7\t ",,terminal_output +571,1140456,"TERMINAL",0,0,"8\t ",,terminal_output +572,1141518,"TERMINAL",0,0,"9\t ",,terminal_output +573,1142549,"TERMINAL",0,0,"40\t ",,terminal_output +574,1143588,"TERMINAL",0,0,"1\t ",,terminal_output +575,1144653,"TERMINAL",0,0,"2\t ",,terminal_output +576,1145700,"TERMINAL",0,0,"3\t ",,terminal_output +577,1146734,"TERMINAL",0,0,"4\t ",,terminal_output +578,1147787,"TERMINAL",0,0,"6\t ",,terminal_output +579,1148830,"TERMINAL",0,0,"7\t ",,terminal_output +580,1149866,"TERMINAL",0,0,"8\t ",,terminal_output +581,1150906,"TERMINAL",0,0,"9\t ",,terminal_output +582,1151945,"TERMINAL",0,0,"50\t ",,terminal_output +583,1152988,"TERMINAL",0,0,"1\t ",,terminal_output +584,1154039,"TERMINAL",0,0,"2\t ",,terminal_output +585,1155079,"TERMINAL",0,0,"3\t ",,terminal_output +586,1156127,"TERMINAL",0,0,"4\t ",,terminal_output +587,1157179,"TERMINAL",0,0,"5\t ",,terminal_output +588,1158237,"TERMINAL",0,0,"6\t ",,terminal_output +589,1159273,"TERMINAL",0,0,"7\t ",,terminal_output +590,1159721,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +591,1160354,"TERMINAL",0,0,"8\t ",,terminal_output +592,1161351,"TERMINAL",0,0,"9\t ",,terminal_output +593,1162399,"TERMINAL",0,0,"1:00\t ",,terminal_output +594,1162643,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2089,0,"",shellscript,selection_mouse +595,1162893,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2089,4,"2048",shellscript,selection_mouse +596,1163440,"TERMINAL",0,0,"1\t ",,terminal_output +597,1163833,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2089,4,"4",shellscript,content +598,1163835,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,0,"",shellscript,selection_keyboard +599,1164490,"TERMINAL",0,0,"2\t ",,terminal_output +600,1165098,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,0,"0",shellscript,content +601,1165100,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2091,0,"",shellscript,selection_keyboard +602,1165544,"TERMINAL",0,0,"3\t ",,terminal_output +603,1166440,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,1,"",shellscript,content +604,1166575,"TERMINAL",0,0,"4\t ",,terminal_output +605,1167638,"TERMINAL",0,0,"5\t ",,terminal_output +606,1168656,"TERMINAL",0,0,"6\t ",,terminal_output +607,1169698,"TERMINAL",0,0,"7\t ",,terminal_output +608,1170741,"TERMINAL",0,0,"8\t ",,terminal_output +609,1171782,"TERMINAL",0,0,"10\t ",,terminal_output +610,1171992,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,0,"0",shellscript,content +611,1171994,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2091,0,"",shellscript,selection_keyboard +612,1172517,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2091,0,"9",shellscript,content +613,1172517,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2092,0,"",shellscript,selection_keyboard +614,1172837,"TERMINAL",0,0,"1\t ",,terminal_output +615,1172950,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2092,0,"6",shellscript,content +616,1172951,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2093,0,"",shellscript,selection_keyboard +617,1173870,"TERMINAL",0,0,"2\t ",,terminal_output +618,1174922,"TERMINAL",0,0,"3\t ",,terminal_output +619,1175800,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2095,0,"",shellscript,selection_mouse +620,1176010,"TERMINAL",0,0,"4\t ",,terminal_output +621,1177012,"TERMINAL",0,0,"5\t ",,terminal_output +622,1178089,"TERMINAL",0,0,"6\t ",,terminal_output +623,1180035,"train_lam.py",0,0,"",python,tab +624,1180231,"TERMINAL",0,0,"7\t ",,terminal_output +625,1180920,"train_lam.py",0,0,"",python,tab +626,1181179,"TERMINAL",0,0,"9\t ",,terminal_output +627,1182037,"TERMINAL",0,0,"20\t ",,terminal_output +628,1183071,"TERMINAL",0,0,"1\t ",,terminal_output +629,1183248,"train_lam.py",1297,0,"",python,selection_mouse +630,1183368,"train_lam.py",1291,9,"model_dim",python,selection_mouse +631,1184167,"TERMINAL",0,0,"2\t ",,terminal_output +632,1185165,"TERMINAL",0,0,"3\t ",,terminal_output +633,1186212,"TERMINAL",0,0,"4\t ",,terminal_output +634,1186849,"train_lam.py",1318,0,"",python,selection_mouse +635,1186995,"train_lam.py",1316,7,"ffn_dim",python,selection_mouse +636,1187265,"TERMINAL",0,0,"5\t ",,terminal_output +637,1188305,"TERMINAL",0,0,"6\t ",,terminal_output +638,1188853,"train_lam.py",1344,0,"",python,selection_mouse +639,1189007,"train_lam.py",1340,10,"latent_dim",python,selection_mouse +640,1189369,"TERMINAL",0,0,"7\t ",,terminal_output +641,1190414,"TERMINAL",0,0,"8\t ",,terminal_output +642,1191466,"TERMINAL",0,0,"9\t ",,terminal_output +643,1192516,"TERMINAL",0,0,"30\t ",,terminal_output +644,1193571,"TERMINAL",0,0,"1\t ",,terminal_output +645,1194616,"TERMINAL",0,0,"2\t ",,terminal_output +646,1195132,"train_lam.py",1421,0,"",python,selection_mouse +647,1195670,"TERMINAL",0,0,"3\t ",,terminal_output +648,1196483,"train_lam.py",1422,0,"",python,selection_mouse +649,1196614,"train_lam.py",1415,10,"num_blocks",python,selection_mouse +650,1196753,"TERMINAL",0,0,"4\t ",,terminal_output +651,1197792,"TERMINAL",0,0,"5\t ",,terminal_output +652,1198811,"TERMINAL",0,0,"7\t ",,terminal_output +653,1199864,"TERMINAL",0,0,"8\t ",,terminal_output +654,1200662,"train_lam.py",1446,0,"",python,selection_mouse +655,1200804,"train_lam.py",1439,9,"num_heads",python,selection_mouse +656,1200937,"TERMINAL",0,0,"9\t ",,terminal_output +657,1201939,"TERMINAL",0,0,"40\t ",,terminal_output +658,1202986,"TERMINAL",0,0,"1\t ",,terminal_output +659,1204032,"TERMINAL",0,0,"2\t ",,terminal_output +660,1205073,"TERMINAL",0,0,"3\t ",,terminal_output +661,1205460,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +662,1205461,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1697,0,"",shellscript,selection_mouse +663,1206011,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1986,0,"",shellscript,selection_mouse +664,1206125,"TERMINAL",0,0,"4\t ",,terminal_output +665,1207164,"TERMINAL",0,0,"5\t ",,terminal_output +666,1208002,"TERMINAL",0,0,"bash",,terminal_focus +667,1208216,"TERMINAL",0,0,"6\t ",,terminal_output +668,1208613,"TERMINAL",0,0,"watch",,terminal_focus +669,1208803,"TERMINAL",0,0,"Every 1.0s: squeu... hkn1993.localdomain: Mon Sep 1 13:01:47 2025JOBID PARTITION NAME USER ST\tTIME NODESNODELIST(REASON)",,terminal_output +670,1209798,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +671,1210963,"TERMINAL",0,0,"bash",,terminal_focus +672,1216049,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",,terminal_command +673,1216083,"TERMINAL",0,0,"]633;E;2025-09-01 13:01:54 sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;CSubmitted batch job 3454917\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +674,1217695,"TERMINAL",0,0,"bash",,terminal_focus +675,1219797,"TERMINAL",0,0,"queue",,terminal_command +676,1219864,"TERMINAL",0,0,"]633;E;2025-09-01 13:01:58 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C",,terminal_output +677,1219928,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeu... hkn1993.localdomain: Mon Sep 1 13:01:58 2025JOBID PARTITION NAME USER ST\tTIME NODESNODELIST(REASON)3454917 accelerat train_la tum_cte0 R\t0:03\t 1hkn0410",,terminal_output +678,1220921,"TERMINAL",0,0,"94",,terminal_output +679,1221969,"TERMINAL",0,0,"2:005",,terminal_output +680,1222947,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 13:02:01 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454917 accelerat train_la tum_cte0 R\t0:06\t 1 hkn0410",,terminal_output +681,1223964,"TERMINAL",0,0,"27",,terminal_output +682,1225013,"TERMINAL",0,0,"38",,terminal_output +683,1226058,"TERMINAL",0,0,"49",,terminal_output +684,1227103,"TERMINAL",0,0,"510",,terminal_output +685,1228173,"TERMINAL",0,0,"61",,terminal_output +686,1228299,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +687,1230937,"TERMINAL",0,0,"ls",,terminal_command +688,1231004,"TERMINAL",0,0,"]633;E;2025-09-01 13:02:09 ls;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C",,terminal_output +689,1231125,"TERMINAL",0,0,"train_lam_minecraft_8node_3431870.log train_lam_minecraft_8node_3431876.log train_lam_minecraft_8node_3431895.log\r\ntrain_lam_minecraft_8node_3431875.log train_lam_minecraft_8node_3431885.log train_lam_minecraft_8node_3454890.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0]633;P;Cwd=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam",,terminal_output +690,1232852,"TERMINAL",0,0,"ls",,terminal_command +691,1234023,"TERMINAL",0,0,"queue",,terminal_command +692,1234097,"TERMINAL",0,0,"]633;E;2025-09-01 13:02:12 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 13:02:12 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454917 accelerat train_la tum_cte0 R\t0:17\t 1 hkn0410",,terminal_output +693,1235165,"TERMINAL",0,0,"38",,terminal_output +694,1236204,"TERMINAL",0,0,"49",,terminal_output +695,1237252,"TERMINAL",0,0,"520",,terminal_output +696,1238301,"TERMINAL",0,0,"61",,terminal_output +697,1239353,"TERMINAL",0,0,"72",,terminal_output +698,1240394,"TERMINAL",0,0,"83",,terminal_output +699,1241445,"TERMINAL",0,0,"94",,terminal_output +700,1242492,"TERMINAL",0,0,"205",,terminal_output +701,1243539,"TERMINAL",0,0,"16",,terminal_output +702,1244593,"TERMINAL",0,0,"27",,terminal_output +703,1245644,"TERMINAL",0,0,"38",,terminal_output +704,1246695,"TERMINAL",0,0,"49",,terminal_output +705,1247768,"TERMINAL",0,0,"531",,terminal_output +706,1248785,"TERMINAL",0,0,"72",,terminal_output +707,1249828,"TERMINAL",0,0,"83",,terminal_output +708,1250925,"TERMINAL",0,0,"94",,terminal_output +709,1251929,"TERMINAL",0,0,"305",,terminal_output +710,1252969,"TERMINAL",0,0,"16",,terminal_output +711,1254013,"TERMINAL",0,0,"27",,terminal_output +712,1255068,"TERMINAL",0,0,"38",,terminal_output +713,1256119,"TERMINAL",0,0,"49",,terminal_output +714,1257168,"TERMINAL",0,0,"540",,terminal_output +715,1257501,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +716,1258222,"TERMINAL",0,0,"queue",,terminal_command +717,1258304,"TERMINAL",0,0,"]633;E;2025-09-01 13:02:36 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 13:02:36 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454917 accelerat train_la tum_cte0 R\t0:41\t 1 hkn0410",,terminal_output +718,1259270,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +719,1260087,"TERMINAL",0,0,"idling",,terminal_command +720,1260175,"TERMINAL",0,0,"]633;E;2025-09-01 13:02:38 idling;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Mon Sep 1 13:02:38 2025Partition dev_cpuonly:\t 8 nodes idle\rPartition cpuonly: 123 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated: 32 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 0 nodes idle\rPartition accelerated-h200:\t 0 nodes idle",,terminal_output +721,1261207,"TERMINAL",0,0,"9\t ",,terminal_output +722,1261671,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +723,1262775,"TERMINAL",0,0,"ls",,terminal_command +724,1262868,"TERMINAL",0,0,"]633;E;2025-09-01 13:02:41 ls;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;Ctrain_lam_minecraft_8node_3431870.log train_lam_minecraft_8node_3431885.log train_lam_minecraft_8node_3454917.log\r\ntrain_lam_minecraft_8node_3431875.log train_lam_minecraft_8node_3431895.log\r\ntrain_lam_minecraft_8node_3431876.log train_lam_minecraft_8node_3454890.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +725,1268280,"TERMINAL",0,0,"tail -f train_lam_minecraft_8node_3454917.log",,terminal_command +726,1268352,"TERMINAL",0,0,"]633;E;2025-09-01 13:02:46 tail -f train_lam_minecraft_8node_3454917.log;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;CSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3454917\r\nSLURM_NODEID=0\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=train_lam_minecraft_1node_dev\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0410\r\nGpuFreq=control_disabled\r\nwandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +727,1269313,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.19.11\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250901_130246-3454917\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run lam-minecraft-1-node-dev-3454917\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/3454917\r\n",,terminal_output +728,1346905,"slurm/utils/mihir/model_sizes.md",0,0,"",markdown,tab +729,1384360,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 4).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 4).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 4).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 4).\r\nRunning on 4 devices.\r\nCounting all components: ['action_in', 'action_up', 'decoder', 'encoder', 'patch_up', 'vq']\r\nParameter counts:\r\n{'action_in': 768, 'action_up': 66560, 'decoder': 203400960, 'encoder': 202416704, 'patch_up': 787456, 'vq': 6400, 'total': 406678848}\r\nStarting training from step 0...\r\nRunning on 4 devices.\r\nCounting all components: ['action_in', 'action_up', 'decoder', 'encoder', 'patch_up', 'vq']\r\nParameter counts:\r\n{'action_in': 768, 'action_up': 66560, 'decoder': 203400960, 'encoder': 202416704, 'patch_up': 787456, 'vq': 6400, 'total': 406678848}\r\nStarting training from step 0...\r\nRunning on 4 devices.\r\nCounting all components: ['action_in', 'action_up', 'decoder', 'encoder', 'patch_up', 'vq']\r\nParameter counts:\r\n{'action_in': 768, 'action_up': 66560, 'decoder': 203400960, 'encoder': 202416704, 'patch_up': 787456, 'vq': 6400, 'total': 406678848}\r\nStarting training from step 0...\r\nRunning on 4 devices.\r\nCounting all components: ['action_in', 'action_up', 'decoder', 'encoder', 'patch_up', 'vq']\r\nParameter counts:\r\n{'action_in': 768, 'action_up': 66560, 'decoder': 203400960, 'encoder': 202416704, 'patch_up': 787456, 'vq': 6400, 'total': 406678848}\r\nStarting training from step 0...\r\n",,terminal_output +730,1411368,"TERMINAL",0,0,"2025-09-01 13:05:09.574071: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +731,1413376,"TERMINAL",0,0,"2025-09-01 13:05:10.947126: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +732,1417388,"TERMINAL",0,0,"2025-09-01 13:05:14.805231: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-01 13:05:14.805270: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +733,1493412,"TERMINAL",0,0,"[Mon Sep 1 01:06:31 PM CEST 2025] caught sigusr1 (timeout warning), requeueing slurm job 3454917...\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** JOB 3454917 ON hkn0410 CANCELLED AT 2025-09-01T13:06:31 DUE TO JOB REQUEUE ***\r\nslurmstepd: error: *** STEP 3454917.0 ON hkn0410 CANCELLED AT 2025-09-01T13:06:31 DUE TO JOB REQUEUE ***\r\nsrun: got SIGCONT\r\n",,terminal_output +734,1508166,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;130",,terminal_output +735,1509752,"TERMINAL",0,0,"queue",,terminal_command +736,1509799,"TERMINAL",0,0,"]633;E;2025-09-01 13:06:48 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C",,terminal_output +737,1509864,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 13:06:48 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454917 accelerat train_la tum_cte0 CG\t0:00\t 1 hkn0410",,terminal_output +738,1510874,"TERMINAL",0,0,"9\t ",,terminal_output +739,1511946,"TERMINAL",0,0,"50\t ",,terminal_output +740,1512963,"TERMINAL",0,0,"1\t ",,terminal_output +741,1513752,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +742,1516753,"TERMINAL",0,0,"scancel 3454917",,terminal_command +743,1516801,"TERMINAL",0,0,"]633;E;2025-09-01 13:06:55 scancel 3454917;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0]633;P;Cwd=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam",,terminal_output +744,1519368,"TERMINAL",0,0,"queue",,terminal_command +745,1519449,"TERMINAL",0,0,"]633;E;2025-09-01 13:06:57 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 13:06:57 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454917 accelerat train_la tum_cte0 CG\t4:36\t 1 hkn0410",,terminal_output +746,1520485,"TERMINAL",0,0,"8\t ",,terminal_output +747,1521538,"TERMINAL",0,0,"9\t ",,terminal_output +748,1522526,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +749,1525052,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +750,1526660,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",86,0,"",shellscript,selection_mouse +751,1527685,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",87,0,"",shellscript,selection_command +752,1528351,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",86,0,"",shellscript,selection_command +753,1529038,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",85,1,"",shellscript,content +754,1529126,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",85,0,"3",shellscript,content +755,1529127,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",86,0,"",shellscript,selection_keyboard +756,1536900,"TERMINAL",0,0,"bash",,terminal_focus +757,1565199,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/train_lam_minecraft_8node_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/train_lam_minecraft_8node_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-$slurm_job_id \\n --tags lam minecraft 8-node \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +758,1569686,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +759,1575877,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node.sbatch",0,0,"",shellscript,tab +760,1577181,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node copy.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/train_lam_minecraft_8node_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/train_lam_minecraft_8node_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-$slurm_job_id \\n --tags lam minecraft 8-node \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +761,1593683,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/train_lam_minecraft_8node_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/train_lam_minecraft_8node_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-$slurm_job_id \\n --tags lam minecraft 8-node \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +762,1606125,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +763,1606972,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2031,0,"",shellscript,selection_mouse +764,1607476,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2074,0,"",shellscript,selection_mouse +765,1608761,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1791,0,"",shellscript,selection_mouse +766,1609230,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1774,0,"",shellscript,selection_mouse +767,1609513,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1773,0,"",shellscript,selection_command +768,1611922,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",0,0,"",shellscript,tab +769,1612784,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1683,0,"",shellscript,selection_mouse +770,1613355,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1685,0,"\n --darkness_threshold=50 \",shellscript,content +771,1613401,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1690,0,"",shellscript,selection_command +772,1616311,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1829,0,"darkness-filter-base-",shellscript,content +773,1616312,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1850,0,"",shellscript,selection_command +774,1617887,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1897,0,"",shellscript,selection_mouse +775,1618968,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1898,0,"darkness-filter-base ",shellscript,content +776,1618969,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1919,0,"",shellscript,selection_command +777,1619781,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1968,0,"",shellscript,selection_mouse +778,1619787,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1967,0,"",shellscript,selection_command +779,1621145,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1913,0,"",shellscript,selection_mouse +780,1622191,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1913,1,"",shellscript,content +781,1622672,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1913,0," ",shellscript,content +782,1622673,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1914,0,"",shellscript,selection_keyboard +783,1623221,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1913,0,"",shellscript,selection_command +784,1624121,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",2028,0,"",shellscript,selection_mouse +785,1624122,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",2027,0,"",shellscript,selection_command +786,1625052,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1845,0,"",shellscript,selection_mouse +787,1626755,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1845,4,"",shellscript,content +788,1628671,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1845,0,"3",shellscript,content +789,1628672,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1846,0,"",shellscript,selection_keyboard +790,1628765,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1846,0,"7",shellscript,content +791,1628765,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1847,0,"",shellscript,selection_keyboard +792,1629529,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1847,0,"M",shellscript,content +793,1629530,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1848,0,"",shellscript,selection_keyboard +794,1630889,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1917,0,"37M",shellscript,content +795,1630889,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1913,4,"",shellscript,content +796,1631703,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",1966,0,"",shellscript,selection_mouse +797,1634558,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",465,0,"",shellscript,selection_mouse +798,1637450,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",478,0,"servation=llmtum\n#SBATCH --re",shellscript,content +799,1638850,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",465,0,"_",shellscript,content +800,1638851,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",466,0,"",shellscript,selection_keyboard +801,1641563,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",466,0,"darkness_filter_37M",shellscript,content +802,1644075,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,0,"",shellscript,selection_mouse +803,1644320,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,1,"t",shellscript,selection_mouse +804,1644321,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,3,"tra",shellscript,selection_mouse +805,1644321,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,6,"train_",shellscript,selection_mouse +806,1644321,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,7,"train_l",shellscript,selection_mouse +807,1644321,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,8,"train_la",shellscript,selection_mouse +808,1644322,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,9,"train_lam",shellscript,selection_mouse +809,1644357,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,10,"train_lam_",shellscript,selection_mouse +810,1644395,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,11,"train_lam_m",shellscript,selection_mouse +811,1644395,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,12,"train_lam_mi",shellscript,selection_mouse +812,1644430,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,13,"train_lam_min",shellscript,selection_mouse +813,1644471,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,14,"train_lam_mine",shellscript,selection_mouse +814,1644506,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,15,"train_lam_minec",shellscript,selection_mouse +815,1644591,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,16,"train_lam_minecr",shellscript,selection_mouse +816,1644591,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,17,"train_lam_minecra",shellscript,selection_mouse +817,1644689,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,18,"train_lam_minecraf",shellscript,selection_mouse +818,1644723,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,19,"train_lam_minecraft",shellscript,selection_mouse +819,1644942,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,20,"train_lam_minecraft_",shellscript,selection_mouse +820,1644982,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,21,"train_lam_minecraft_8",shellscript,selection_mouse +821,1645027,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,22,"train_lam_minecraft_8n",shellscript,selection_mouse +822,1645101,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,23,"train_lam_minecraft_8no",shellscript,selection_mouse +823,1645111,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,24,"train_lam_minecraft_8nod",shellscript,selection_mouse +824,1645157,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,25,"train_lam_minecraft_8node",shellscript,selection_mouse +825,1645350,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,26,"train_lam_minecraft_8node_",shellscript,selection_mouse +826,1680410,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,26,"",shellscript,content +827,1682966,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",263,0,"%",shellscript,content +828,1682967,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",264,0,"",shellscript,selection_keyboard +829,1683238,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",264,0,"x",shellscript,content +830,1683239,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",265,0,"",shellscript,selection_keyboard +831,1683507,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",265,0,"_",shellscript,content +832,1683508,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",266,0,"",shellscript,selection_keyboard +833,1684982,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",390,0,"%x",shellscript,content +834,1684983,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",365,25,"",shellscript,content +835,1686327,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",537,0,"",shellscript,selection_mouse +836,1690937,"slurm/jobs/mihir/horeka/maskgit_big_runs/train_dynamics_8_nodes_filter_dark.sbatch",0,0,"",shellscript,tab +837,1695443,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +838,1699498,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",0,0,"",shellscript,tab +839,1763109,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",2043,0,"",shellscript,selection_mouse +840,1763679,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",2044,0,"",shellscript,selection_mouse +841,1764266,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",2043,0,"",shellscript,selection_mouse +842,1764878,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",2044,0,"",shellscript,selection_mouse +843,1765440,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",2060,0,"",shellscript,selection_mouse +844,1766067,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",2044,0,"",shellscript,selection_mouse +845,1766556,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",2043,0,"",shellscript,selection_mouse +846,1767217,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",2060,0,"",shellscript,selection_mouse +847,1767872,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",2044,0,"",shellscript,selection_mouse +848,1771371,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base copy.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node_darkness_filter_37M\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --darkness_threshold=50 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-darkness-filter-37M-$slurm_job_id \\n --tags lam minecraft 8-node darkness-filter 37M \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +849,1783059,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base-400M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node_darkness_filter_37M\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --darkness_threshold=50 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-darkness-filter-37M-$slurm_job_id \\n --tags lam minecraft 8-node darkness-filter 37M \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +850,1794461,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node_darkness_filter_37M\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --darkness_threshold=50 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-darkness-filter-37M-$slurm_job_id \\n --tags lam minecraft 8-node darkness-filter 37M \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +851,1798068,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1848,0,"",shellscript,selection_mouse +852,1800291,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1848,1,"",shellscript,content +853,1800437,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1848,1,"",shellscript,content +854,1801688,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1848,0,"37",shellscript,content +855,1802241,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1847,0,"",shellscript,selection_command +856,1803791,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1918,0,"400",shellscript,content +857,1803792,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1916,2,"",shellscript,content +858,1803792,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1850,0,"400",shellscript,content +859,1803792,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1848,2,"",shellscript,content +860,1807122,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",438,0,"400",shellscript,content +861,1807123,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",436,2,"",shellscript,content +862,1817836,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +863,1820023,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2095,0,"",shellscript,selection_mouse +864,1820039,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2094,0,"",shellscript,selection_command +865,1820168,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2094,1,"\",shellscript,selection_mouse +866,1820173,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2095,0,"",shellscript,selection_command +867,1820277,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2094,1,"\",shellscript,selection_mouse +868,1820278,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,5,"096 \",shellscript,selection_mouse +869,1820279,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2087,8,"m=4096 \",shellscript,selection_mouse +870,1820279,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2086,9,"im=4096 \",shellscript,selection_mouse +871,1820279,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2085,10,"dim=4096 \",shellscript,selection_mouse +872,1820315,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2084,11,"_dim=4096 \",shellscript,selection_mouse +873,1820316,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2083,12,"n_dim=4096 \",shellscript,selection_mouse +874,1820351,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2060,35,"atent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +875,1820351,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2059,36,"latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +876,1820393,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2058,37,"-latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +877,1820423,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2057,38,"--latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +878,1820424,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2056,39," --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +879,1820470,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2033,62," --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +880,1820511,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2032,63," --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +881,1820588,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2010,85," --num_blocks=12 \\n --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +882,1820663,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1987,108," --model_dim=1024 \\n --num_blocks=12 \\n --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +883,1821065,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1963,132," --num_latents=100 \\n --model_dim=1024 \\n --num_blocks=12 \\n --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +884,1822000,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1987,108," --model_dim=1024 \\n --num_blocks=12 \\n --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +885,1824522,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",0,0,"",shellscript,tab +886,1826803,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1991,0,"",shellscript,selection_mouse +887,1827550,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1992,0,"base",shellscript,content +888,1827552,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1995,0,"",shellscript,selection_command +889,1828738,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1992,4,"",shellscript,content +890,1828760,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1918,3,"37",shellscript,content +891,1828763,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1849,3,"37",shellscript,content +892,1828776,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",436,3,"37",shellscript,content +893,1828794,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1848,0,"",shellscript,selection_command +894,1829733,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1916,0,"",shellscript,selection_command +895,1829907,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1946,0,"",shellscript,selection_command +896,1830113,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1968,0,"",shellscript,selection_command +897,1830212,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1992,0,"",shellscript,selection_command +898,1830352,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2028,0,"",shellscript,selection_command +899,1830837,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1992,0,"",shellscript,selection_command +900,1831393,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1993,0,"\n ",shellscript,content +901,1831736,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1998,0," --model_dim=1024 \\n --num_blocks=12 \\n --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,content +902,1832609,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2105,0,"",shellscript,selection_command +903,1832710,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2083,0,"",shellscript,selection_command +904,1832940,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2062,0,"",shellscript,selection_command +905,1833029,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2040,0,"",shellscript,selection_command +906,1833141,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2013,0,"",shellscript,selection_command +907,1833268,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1989,0,"",shellscript,selection_command +908,1833575,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2013,0,"",shellscript,selection_command +909,1833818,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2002,0,"",shellscript,selection_command +910,1834099,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1998,4,"",shellscript,content +911,1834275,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1997,0,"",shellscript,selection_command +912,1835990,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2079,0,"",shellscript,selection_mouse +913,1836556,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2102,0,"",shellscript,selection_mouse +914,1836566,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2101,0,"",shellscript,selection_command +915,1881618,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",0,0,"",shellscript,tab +916,1882697,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",0,0,"",shellscript,tab +917,1885565,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",606,0,"",shellscript,selection_mouse +918,1885570,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",605,0,"",shellscript,selection_command +919,1887327,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +920,1890438,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",0,0,"",shellscript,tab +921,1890928,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",858,0,"",shellscript,selection_mouse +922,1893041,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1998,0," ",shellscript,content +923,1894046,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1998,108,"",shellscript,content +924,1895006,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1993,5,"",shellscript,content +925,1895342,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",436,2,"400",shellscript,content +926,1895698,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1849,2,"400",shellscript,content +927,1896591,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1918,2,"400",shellscript,content +928,1899323,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1992,0,"base",shellscript,content +929,1900300,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1992,4,"",shellscript,content +930,1901489,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1991,0,"",shellscript,selection_command +931,1902213,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",1996,0,"\n ",shellscript,content +932,1902790,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2001,0," --model_dim=1024 \\n --num_blocks=12 \\n --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,content +933,1903754,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2087,0,"",shellscript,selection_command +934,1903896,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2066,0,"",shellscript,selection_command +935,1904032,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2044,0,"",shellscript,selection_command +936,1904422,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2017,0,"",shellscript,selection_command +937,1905044,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2016,0,"",shellscript,selection_command +938,1905538,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2005,0,"",shellscript,selection_command +939,1905738,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2001,4,"",shellscript,content +940,1905972,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",2000,0,"",shellscript,selection_command +941,1914365,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M copy.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node_darkness_filter_400M\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --darkness_threshold=50 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-darkness-filter-400M-$slurm_job_id \\n --tags lam minecraft 8-node darkness-filter 400M \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --model_dim=1024 \\n --num_blocks=12 \\n --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +942,1919891,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node_darkness_filter_400M\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --darkness_threshold=50 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-darkness-filter-400M-$slurm_job_id \\n --tags lam minecraft 8-node darkness-filter 400M \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --model_dim=1024 \\n --num_blocks=12 \\n --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +943,1923696,"slurm/utils/mihir/model_sizes.md",0,0,"",markdown,tab +944,1932775,"slurm/utils/mihir/model_sizes.md",1780,0,"",markdown,selection_mouse +945,1933510,"slurm/utils/mihir/model_sizes.md",1726,0,"",markdown,selection_mouse +946,1934062,"slurm/utils/mihir/model_sizes.md",1762,0,"",markdown,selection_mouse +947,1935035,"TERMINAL",0,0,"bash",,terminal_focus +948,1936446,"TERMINAL",0,0,"queue",,terminal_command +949,1936526,"TERMINAL",0,0,"]633;E;2025-09-01 13:13:54 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 13:13:54 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)",,terminal_output +950,1937565,"TERMINAL",0,0,"5",,terminal_output +951,1937734,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +952,1939158,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +953,1940441,"slurm/utils/mihir/model_sizes.md",0,0,"",markdown,tab +954,1948727,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +955,1949539,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2031,0,"",shellscript,selection_mouse +956,1949571,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2030,0,"",shellscript,selection_command +957,1950256,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2007,0,"",shellscript,selection_mouse +958,1960741,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2003,4,"",shellscript,content +959,1961991,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2003,0,"6",shellscript,content +960,1961991,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2004,0,"",shellscript,selection_keyboard +961,1962314,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2004,0,"4",shellscript,content +962,1962315,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2005,0,"",shellscript,selection_keyboard +963,1962345,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2005,0,"0",shellscript,content +964,1962346,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2006,0,"",shellscript,selection_keyboard +965,1963061,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2005,0,"",shellscript,selection_command +966,1963206,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2027,0,"",shellscript,selection_command +967,1963673,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2028,0,"",shellscript,selection_command +968,1964202,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2027,1,"",shellscript,content +969,1964596,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2026,1,"",shellscript,content +970,1967095,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2026,0,"1",shellscript,content +971,1967096,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2027,0,"",shellscript,selection_keyboard +972,1967109,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2027,0,"0",shellscript,content +973,1967110,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2028,0,"",shellscript,selection_keyboard +974,1967607,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2027,0,"",shellscript,selection_command +975,1967776,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2049,0,"",shellscript,selection_command +976,1968389,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2048,1,"",shellscript,content +977,1968833,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2048,0,"0",shellscript,content +978,1968834,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2049,0,"",shellscript,selection_keyboard +979,1969067,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2048,0,"",shellscript,selection_command +980,1969239,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2069,0,"",shellscript,selection_command +981,1970738,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2070,0,"",shellscript,selection_command +982,1970982,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2071,0,"",shellscript,selection_command +983,1971528,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2070,1,"",shellscript,content +984,1971626,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2069,1,"",shellscript,content +985,1972304,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2069,0,"3",shellscript,content +986,1972305,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2070,0,"",shellscript,selection_keyboard +987,1972438,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2070,0,"2",shellscript,content +988,1972439,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2071,0,"",shellscript,selection_keyboard +989,1972650,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2070,0,"",shellscript,selection_command +990,1972796,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2092,0,"",shellscript,selection_command +991,1977306,"TERMINAL",0,0,"bash",,terminal_focus +992,1979036,"TERMINAL",0,0,"python",,terminal_command +993,1979069,"TERMINAL",0,0,"]633;E;2025-09-01 13:14:37 python;b7f1e619-a2b5-49a5-a127-74d220979cb5]633;C",,terminal_output +994,1979393,"TERMINAL",0,0,"Python 3.10.18 (main, Jun 4 2025, 17:36:27) [Clang 20.1.4 ] on linux\r\nType ""help"", ""copyright"", ""credits"" or ""license"" for more information.\r\n",,terminal_output +995,1979799,"TERMINAL",0,0,">>> ",,terminal_output +996,1981877,"TERMINAL",0,0,"6",,terminal_output +997,1982117,"TERMINAL",0,0,"[?25l4[?25h",,terminal_output +998,1982449,"TERMINAL",0,0,"[?25l0[?25h",,terminal_output +999,1983228,"TERMINAL",0,0,"[?25l*[?25h",,terminal_output +1000,1983792,"TERMINAL",0,0,"[?25l4[?25h",,terminal_output +1001,1983885,"TERMINAL",0,0,"\r\n2560\r\n>>> ",,terminal_output +1002,1987719,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2091,1,"",shellscript,content +1003,1987844,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,1,"",shellscript,content +1004,1987992,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2089,1,"",shellscript,content +1005,1989140,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2088,1,"",shellscript,content +1006,1992109,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2088,0,"2560",shellscript,content +1007,2032264,"TERMINAL",0,0,"bash",,terminal_focus +1008,2032833,"TERMINAL",0,0,"bash",,terminal_focus +1009,2035804,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",,terminal_command +1010,2035836,"TERMINAL",0,0,"]633;E;2025-09-01 13:15:34 sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;CSubmitted batch job 3454941\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +1011,2037484,"TERMINAL",0,0,"bash",,terminal_focus +1012,2038980,"TERMINAL",0,0,"queue",,terminal_command +1013,2039053,"TERMINAL",0,0,"]633;E;2025-09-01 13:15:37 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: s... hkn1993.localdomain: Mon Sep 1 13:15:37 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454941 accelerat train_la tum_cte0 R\t0:021 hkn0813",,terminal_output +1014,2040089,"TERMINAL",0,0,"803 ",,terminal_output +1015,2041147,"TERMINAL",0,0,"904 ",,terminal_output +1016,2041627,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1017,2051948,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +1018,2053318,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2094,0,"",shellscript,selection_mouse +1019,2053495,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2091,3,"0 \",shellscript,selection_mouse +1020,2053562,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2089,5,"560 \",shellscript,selection_mouse +1021,2053563,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2064,30,"_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1022,2053564,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2063,31,"t_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1023,2053564,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2062,32,"nt_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1024,2053594,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2061,33,"ent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1025,2053595,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2060,34,"tent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1026,2053628,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2038,56,"um_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1027,2053662,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2037,57,"num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1028,2053663,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2036,58,"-num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1029,2053702,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2035,59,"--num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1030,2053737,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2034,60," --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1031,2053769,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2033,61," --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1032,2053806,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2011,83," --num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1033,2053896,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2010,84," --num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1034,2053914,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1988,106," --model_dim=640 \\n --num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1035,2054314,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1987,107," --model_dim=640 \\n --num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1036,2058037,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",0,0,"",shellscript,tab +1037,2060686,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2105,0,"",shellscript,selection_mouse +1038,2060687,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2104,0,"",shellscript,selection_command +1039,2060804,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2104,1,"\",shellscript,selection_mouse +1040,2060810,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2105,0,"",shellscript,selection_command +1041,2060873,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2103,2," \",shellscript,selection_mouse +1042,2060873,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2076,29,"dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +1043,2060908,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2073,32,"nt_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +1044,2060985,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2071,34,"tent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +1045,2060985,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2048,57,"num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +1046,2060986,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2047,58,"-num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +1047,2060986,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2046,59,"--num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +1048,2060986,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2045,60," --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +1049,2060986,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2044,61," --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +1050,2061011,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2043,62," --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +1051,2061039,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",2020,85," --num_blocks=12 \\n --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +1052,2061161,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",1997,108," --model_dim=1024 \\n --num_blocks=12 \\n --num_heads=16 \\n --latent_dim=64 \\n --ffn_dim=4096 \",shellscript,selection_mouse +1053,2062057,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",1997,108,"",shellscript,content +1054,2062694,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-200M.sbatch",1997,0," --model_dim=640 \\n --num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,content +1055,2071868,"TERMINAL",0,0,"ls",,terminal_command +1056,2071914,"TERMINAL",0,0,"]633;E;2025-09-01 13:16:10 ls;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C",,terminal_output +1057,2072189,"TERMINAL",0,0,"train_lam_minecraft_8node_3431870.log\r\ntrain_lam_minecraft_8node_3431875.log\r\ntrain_lam_minecraft_8node_3431876.log\r\ntrain_lam_minecraft_8node_3431885.log\r\ntrain_lam_minecraft_8node_3431895.log\r\ntrain_lam_minecraft_8node_3454890.log\r\ntrain_lam_minecraft_8node_3454917.log\r\ntrain_lam_minecraft_8node_3454941.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1058,2074531,"TERMINAL",0,0,"queue",,terminal_command +1059,2074581,"TERMINAL",0,0,"]633;E;2025-09-01 13:16:12 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C",,terminal_output +1060,2074643,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: s... hkn1993.localdomain: Mon Sep 1 13:16:12 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454941 accelerat train_la tum_cte0 R\t0:371 hkn0813",,terminal_output +1061,2075686,"TERMINAL",0,0,"338 ",,terminal_output +1062,2076752,"TERMINAL",0,0,"440 ",,terminal_output +1063,2077758,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1064,2081572,"TERMINAL",0,0,"tail -f train_lam_minecraft_8node_3454941.log",,terminal_command +1065,2081604,"TERMINAL",0,0,"]633;E;2025-09-01 13:16:19 tail -f train_lam_minecraft_8node_3454941.log;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;CSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3454941\r\nSLURM_NODEID=0\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=train_lam_minecraft_1node_dev\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0813\r\nGpuFreq=control_disabled\r\n",,terminal_output +1066,2101567,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +1067,2102568,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.19.11\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250901_131639-3454941\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run lam-minecraft-1-node-dev-3454941\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/3454941\r\n",,terminal_output +1068,2171005,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node_darkness_filter_400M\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --darkness_threshold=50 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-darkness-filter-400M-$slurm_job_id \\n --tags lam minecraft 8-node darkness-filter 400M \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --model_dim=640 \\n --num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +1069,2172236,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1921,0,"",shellscript,selection_mouse +1070,2173153,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1920,1,"",shellscript,content +1071,2173271,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1919,1,"",shellscript,content +1072,2173610,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1918,1,"",shellscript,content +1073,2173974,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1918,0,"1",shellscript,content +1074,2173975,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1919,0,"",shellscript,selection_keyboard +1075,2174196,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1919,0,"3",shellscript,content +1076,2174197,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1920,0,"",shellscript,selection_keyboard +1077,2174299,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1920,0,"3",shellscript,content +1078,2174300,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1921,0,"",shellscript,selection_keyboard +1079,2174641,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1920,0,"",shellscript,selection_command +1080,2174840,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1851,0,"",shellscript,selection_command +1081,2175238,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1852,0,"",shellscript,selection_command +1082,2175644,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1851,1,"",shellscript,content +1083,2175754,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1850,1,"",shellscript,content +1084,2176049,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1849,1,"",shellscript,content +1085,2176222,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1849,0,"1",shellscript,content +1086,2176223,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1850,0,"",shellscript,selection_keyboard +1087,2176639,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1850,0,"3",shellscript,content +1088,2176640,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1851,0,"",shellscript,selection_keyboard +1089,2176782,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1851,0,"3",shellscript,content +1090,2176783,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1852,0,"",shellscript,selection_keyboard +1091,2177208,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1851,0,"",shellscript,selection_command +1092,2177318,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1799,0,"",shellscript,selection_command +1093,2177923,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1762,0,"",shellscript,selection_command +1094,2178050,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1750,0,"",shellscript,selection_command +1095,2178051,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1718,0,"",shellscript,selection_command +1096,2178051,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1688,0,"",shellscript,selection_command +1097,2178087,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1668,0,"",shellscript,selection_command +1098,2178088,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1650,0,"",shellscript,selection_command +1099,2178088,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1627,0,"",shellscript,selection_command +1100,2178123,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1594,0,"",shellscript,selection_command +1101,2178123,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1563,0,"",shellscript,selection_command +1102,2178151,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1538,0,"",shellscript,selection_command +1103,2178152,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1520,0,"",shellscript,selection_command +1104,2178194,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1494,0,"",shellscript,selection_command +1105,2178219,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1492,0,"",shellscript,selection_command +1106,2178251,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1476,0,"",shellscript,selection_command +1107,2178306,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1474,0,"",shellscript,selection_command +1108,2178377,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1384,0,"",shellscript,selection_command +1109,2178377,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1333,0,"",shellscript,selection_command +1110,2178414,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1331,0,"",shellscript,selection_command +1111,2178415,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1304,0,"",shellscript,selection_command +1112,2178475,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1280,0,"",shellscript,selection_command +1113,2178476,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1208,0,"",shellscript,selection_command +1114,2178541,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1157,0,"",shellscript,selection_command +1115,2178542,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1155,0,"",shellscript,selection_command +1116,2178542,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1129,0,"",shellscript,selection_command +1117,2178578,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1099,0,"",shellscript,selection_command +1118,2178657,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1070,0,"",shellscript,selection_command +1119,2178657,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1068,0,"",shellscript,selection_command +1120,2178689,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1061,0,"",shellscript,selection_command +1121,2178724,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1038,0,"",shellscript,selection_command +1122,2178757,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1036,0,"",shellscript,selection_command +1123,2178841,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",1033,0,"",shellscript,selection_command +1124,2178874,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",994,0,"",shellscript,selection_command +1125,2178875,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",989,0,"",shellscript,selection_command +1126,2178965,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",994,0,"",shellscript,selection_command +1127,2179186,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",989,0,"",shellscript,selection_command +1128,2179335,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",947,0,"",shellscript,selection_command +1129,2179491,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",914,0,"",shellscript,selection_command +1130,2179636,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",871,0,"",shellscript,selection_command +1131,2179794,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",819,0,"",shellscript,selection_command +1132,2179917,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",775,0,"",shellscript,selection_command +1133,2180065,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",773,0,"",shellscript,selection_command +1134,2180223,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",749,0,"",shellscript,selection_command +1135,2180390,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",747,0,"",shellscript,selection_command +1136,2180520,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",745,0,"",shellscript,selection_command +1137,2180647,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",734,0,"",shellscript,selection_command +1138,2180846,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",658,0,"",shellscript,selection_command +1139,2180983,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",606,0,"",shellscript,selection_command +1140,2181077,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",590,0,"",shellscript,selection_command +1141,2181210,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",539,0,"",shellscript,selection_command +1142,2181353,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",537,0,"",shellscript,selection_command +1143,2181667,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",486,0,"",shellscript,selection_command +1144,2181883,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",468,0,"",shellscript,selection_command +1145,2182052,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",425,0,"",shellscript,selection_command +1146,2183506,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",439,0,"133",shellscript,content +1147,2183506,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",436,3,"",shellscript,content +1148,2209627,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 4).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 4).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 4).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 4).\r\nRunning on 4 devices.\r\nCounting all components: ['action_in', 'action_up', 'decoder', 'encoder', 'patch_up', 'vq']\r\nParameter counts:\r\n{'action_in': 768, 'action_up': 21120, 'decoder': 66562688, 'encoder': 66173088, 'patch_up': 492160, 'vq': 3200, 'total': 133253024}\r\nStarting training from step 0...\r\nRunning on 4 devices.\r\nCounting all components: ['action_in', 'action_up', 'decoder', 'encoder', 'patch_up', 'vq']\r\nParameter counts:\r\n{'action_in': 768, 'action_up': 21120, 'decoder': 66562688, 'encoder': 66173088, 'patch_up': 492160, 'vq': 3200, 'total': 133253024}\r\nStarting training from step 0...\r\nRunning on 4 devices.\r\nCounting all components: ['action_in', 'action_up', 'decoder', 'encoder', 'patch_up', 'vq']\r\nParameter counts:\r\n{'action_in': 768, 'action_up': 21120, 'decoder': 66562688, 'encoder': 66173088, 'patch_up': 492160, 'vq': 3200, 'total': 133253024}\r\nStarting training from step 0...\r\nRunning on 4 devices.\r\nCounting all components: ['action_in', 'action_up', 'decoder', 'encoder', 'patch_up', 'vq']\r\nParameter counts:\r\n{'action_in': 768, 'action_up': 21120, 'decoder': 66562688, 'encoder': 66173088, 'patch_up': 492160, 'vq': 3200, 'total': 133253024}\r\nStarting training from step 0...\r\n",,terminal_output +1149,2229737,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node_darkness_filter_133M\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --darkness_threshold=50 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-darkness-filter-133M-$slurm_job_id \\n --tags lam minecraft 8-node darkness-filter 133M \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --model_dim=640 \\n --num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +1150,2231905,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",995,0,"",shellscript,selection_mouse +1151,2231905,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",994,0,"",shellscript,selection_command +1152,2235917,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +1153,2237400,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2072,0,"",shellscript,selection_mouse +1154,2237762,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2091,0,"",shellscript,selection_mouse +1155,2240214,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2004,0,"",shellscript,selection_mouse +1156,2240330,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2003,3,"640",shellscript,selection_mouse +1157,2242891,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2004,0,"",shellscript,selection_mouse +1158,2242892,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2003,3,"640",shellscript,selection_mouse +1159,2244624,"TERMINAL",0,0,"2025-09-01 13:19:02.407350: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1160,2245629,"TERMINAL",0,0,"2025-09-01 13:19:03.604053: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-01 13:19:03.604689: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1161,2248342,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2003,3,"7",shellscript,content +1162,2248344,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2004,0,"",shellscript,selection_keyboard +1163,2248644,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2004,0,"6",shellscript,content +1164,2248645,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2005,0,"",shellscript,selection_keyboard +1165,2248920,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2005,0,"8",shellscript,content +1166,2248921,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2006,0,"",shellscript,selection_keyboard +1167,2251636,"TERMINAL",0,0,"2025-09-01 13:19:09.018693: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-01 13:19:09.018731: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1168,2254503,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2049,0,"",shellscript,selection_mouse +1169,2255179,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2048,1,"",shellscript,content +1170,2255184,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2048,0,"2",shellscript,content +1171,2255184,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2049,0,"",shellscript,selection_keyboard +1172,2260607,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2070,0,"",shellscript,selection_mouse +1173,2260741,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2069,2,"32",shellscript,selection_mouse +1174,2262143,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2069,2,"4",shellscript,content +1175,2262144,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2070,0,"",shellscript,selection_keyboard +1176,2262185,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2070,0,"8",shellscript,content +1177,2262186,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2071,0,"",shellscript,selection_keyboard +1178,2263371,"TERMINAL",0,0,"python",,terminal_focus +1179,2266659,"TERMINAL",0,0,"7",,terminal_output +1180,2268083,"TERMINAL",0,0,"[?25l6[?25h",,terminal_output +1181,2268353,"TERMINAL",0,0,"[?25l8[?25h",,terminal_output +1182,2269050,"TERMINAL",0,0,"[?25l*[?25h",,terminal_output +1183,2269340,"TERMINAL",0,0,"[?25l4[?25h",,terminal_output +1184,2269428,"TERMINAL",0,0,"\r\n3072\r\n>>> ",,terminal_output +1185,2272902,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2113,0,"",shellscript,selection_mouse +1186,2273709,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2091,0,"",shellscript,selection_mouse +1187,2274582,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2092,0,"",shellscript,selection_command +1188,2274863,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2091,1,"",shellscript,content +1189,2275065,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,1,"",shellscript,content +1190,2275190,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2089,1,"",shellscript,content +1191,2275330,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2088,1,"",shellscript,content +1192,2275690,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2088,0,"3072",shellscript,content +1193,2289186,"TERMINAL",0,0,"tail",,terminal_focus +1194,2376877,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;130",,terminal_output +1195,2378934,"TERMINAL",0,0,"queue",,terminal_command +1196,2379023,"TERMINAL",0,0,"]633;E;2025-09-01 13:21:17 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: s... hkn1993.localdomain: Mon Sep 1 13:21:17 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454941 accelerat train_la tum_cte0 R\t5:421 hkn0813",,terminal_output +1197,2380081,"TERMINAL",0,0,"843 ",,terminal_output +1198,2381117,"TERMINAL",0,0,"944 ",,terminal_output +1199,2381656,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1200,2385920,"TERMINAL",0,0,"scancel 3454941",,terminal_command +1201,2386005,"TERMINAL",0,0,"]633;E;2025-09-01 13:21:24 scancel 3454941;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1202,2393834,"TERMINAL",0,0,"queue",,terminal_command +1203,2393920,"TERMINAL",0,0,"]633;E;2025-09-01 13:21:32 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: s... hkn1993.localdomain: Mon Sep 1 13:21:32 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454941 accelerat train_la tum_cte0 CG\t5:491 hkn0813",,terminal_output +1204,2394969,"TERMINAL",0,0,"3",,terminal_output +1205,2395658,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1206,2397287,"slurm/utils/mihir/model_sizes.md",0,0,"",markdown,tab +1207,2405382,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +1208,2409303,"TERMINAL",0,0,"bash",,terminal_focus +1209,2410952,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",,terminal_command +1210,2411012,"TERMINAL",0,0,"]633;E;2025-09-01 13:21:49 sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;CSubmitted batch job 3454944\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +1211,2412385,"TERMINAL",0,0,"queue",,terminal_command +1212,2412474,"TERMINAL",0,0,"]633;E;2025-09-01 13:21:50 queue;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: s... hkn1993.localdomain: Mon Sep 1 13:21:50 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454941 accelerat train_la tum_cte0 CG\t5:491 hkn08133454944 accelerat train_la tum_cte0 R\t0:001 hkn0604",,terminal_output +1213,2413509,"TERMINAL",0,0,"101 ",,terminal_output +1214,2414577,"TERMINAL",0,0,"202 ",,terminal_output +1215,2415601,"TERMINAL",0,0,"303 ",,terminal_output +1216,2416649,"TERMINAL",0,0,"404 ",,terminal_output +1217,2417702,"TERMINAL",0,0,"505 ",,terminal_output +1218,2418749,"TERMINAL",0,0,"607 ",,terminal_output +1219,2419803,"TERMINAL",0,0,"808 ",,terminal_output +1220,2420833,"TERMINAL",0,0,"909 ",,terminal_output +1221,2421881,"TERMINAL",0,0,"2:0010 ",,terminal_output +1222,2422934,"TERMINAL",0,0,"111 ",,terminal_output +1223,2423980,"TERMINAL",0,0,"212 ",,terminal_output +1224,2425031,"TERMINAL",0,0,"313 ",,terminal_output +1225,2426083,"TERMINAL",0,0,"414 ",,terminal_output +1226,2427124,"TERMINAL",0,0,"515 ",,terminal_output +1227,2428173,"TERMINAL",0,0,"616 ",,terminal_output +1228,2429222,"TERMINAL",0,0,"717 ",,terminal_output +1229,2430310,"TERMINAL",0,0,"818 ",,terminal_output +1230,2431376,"TERMINAL",0,0,"919 ",,terminal_output +1231,2432382,"TERMINAL",0,0,"\r1020 ",,terminal_output +1232,2433426,"TERMINAL",0,0,"121 ",,terminal_output +1233,2434508,"TERMINAL",0,0,"222 ",,terminal_output +1234,2435516,"TERMINAL",0,0,"323 ",,terminal_output +1235,2436567,"TERMINAL",0,0,"424 ",,terminal_output +1236,2437625,"TERMINAL",0,0,"525 ",,terminal_output +1237,2438673,"TERMINAL",0,0,"626 ",,terminal_output +1238,2439724,"TERMINAL",0,0,"728 ",,terminal_output +1239,2440778,"TERMINAL",0,0,"929 ",,terminal_output +1240,2441830,"TERMINAL",0,0,"2030 ",,terminal_output +1241,2442876,"TERMINAL",0,0,"131 ",,terminal_output +1242,2443944,"TERMINAL",0,0,"232 ",,terminal_output +1243,2445015,"TERMINAL",0,0,"333 ",,terminal_output +1244,2446052,"TERMINAL",0,0,"434 ",,terminal_output +1245,2447090,"TERMINAL",0,0,"535 ",,terminal_output +1246,2448137,"TERMINAL",0,0,"636 ",,terminal_output +1247,2448318,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",0,0,"",shellscript,tab +1248,2449190,"TERMINAL",0,0,"737 ",,terminal_output +1249,2450271,"TERMINAL",0,0,"838 ",,terminal_output +1250,2451294,"TERMINAL",0,0,"939 ",,terminal_output +1251,2452032,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +1252,2452344,"TERMINAL",0,0,"3040 ",,terminal_output +1253,2453096,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2094,0,"",shellscript,selection_mouse +1254,2453426,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2093,1,"\",shellscript,selection_mouse +1255,2453427,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2092,2," \",shellscript,selection_mouse +1256,2453427,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2091,3,"2 \",shellscript,selection_mouse +1257,2453427,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,4,"72 \",shellscript,selection_mouse +1258,2453428,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2089,5,"072 \",shellscript,selection_mouse +1259,2453428,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2088,6,"3072 \",shellscript,selection_mouse +1260,2453428,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2087,7,"=3072 \",shellscript,selection_mouse +1261,2453429,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2086,8,"m=3072 \",shellscript,selection_mouse +1262,2453429,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2085,9,"im=3072 \",shellscript,selection_mouse +1263,2453465,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2084,10,"dim=3072 \",shellscript,selection_mouse +1264,2453466,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2083,11,"_dim=3072 \",shellscript,selection_mouse +1265,2453481,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2059,35,"atent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1266,2453506,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2058,36,"latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1267,2453517,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2057,37,"-latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1268,2453546,"TERMINAL",0,0,"141 ",,terminal_output +1269,2453554,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2056,38,"--latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1270,2453586,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2055,39," --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1271,2453601,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2033,61," --num_heads=12 \\n --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1272,2453674,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2032,62," --num_heads=12 \\n --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1273,2453685,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2010,84," --num_blocks=10 \\n --num_heads=12 \\n --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1274,2453746,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2009,85," --num_blocks=10 \\n --num_heads=12 \\n --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1275,2453894,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1987,107," --model_dim=768 \\n --num_blocks=10 \\n --num_heads=12 \\n --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1276,2454439,"TERMINAL",0,0,"242 ",,terminal_output +1277,2455471,"TERMINAL",0,0,"343 ",,terminal_output +1278,2456532,"TERMINAL",0,0,"444 ",,terminal_output +1279,2457625,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",0,0,"",shellscript,tab +1280,2457739,"TERMINAL",0,0,"545 ",,terminal_output +1281,2458632,"TERMINAL",0,0,"646 ",,terminal_output +1282,2458687,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2104,0,"",shellscript,selection_mouse +1283,2458724,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2103,0,"",shellscript,selection_command +1284,2458799,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2103,1,"\",shellscript,selection_mouse +1285,2458800,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2104,0,"",shellscript,selection_command +1286,2458917,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2101,3,"0 \",shellscript,selection_mouse +1287,2458918,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2075,29,"dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1288,2458918,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2072,32,"nt_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1289,2458919,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2071,33,"ent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1290,2458919,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2049,55,"m_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1291,2458956,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2048,56,"um_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1292,2459077,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2047,57,"num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1293,2459078,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2046,58,"-num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1294,2459078,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2024,80,"-num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1295,2459078,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2023,81,"--num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1296,2459122,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2022,82," --num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1297,2459166,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",2000,104," --model_dim=640 \\n --num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1298,2459281,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",1999,105," --model_dim=640 \\n --num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1299,2459287,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",1998,106," --model_dim=640 \\n --num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1300,2459407,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",1997,107," --model_dim=640 \\n --num_blocks=10 \\n --num_heads=10 \\n --latent_dim=32 \\n --ffn_dim=2560 \",shellscript,selection_mouse +1301,2459731,"TERMINAL",0,0,"747 ",,terminal_output +1302,2460740,"TERMINAL",0,0,"849 ",,terminal_output +1303,2461707,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",1997,107,"",shellscript,content +1304,2461840,"TERMINAL",0,0,"4050 ",,terminal_output +1305,2462396,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M copy.sbatch",1997,0," --model_dim=768 \\n --num_blocks=10 \\n --num_heads=12 \\n --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,content +1306,2462834,"TERMINAL",0,0,"151 ",,terminal_output +1307,2464065,"TERMINAL",0,0,"252 ",,terminal_output +1308,2464975,"TERMINAL",0,0,"353 ",,terminal_output +1309,2465982,"TERMINAL",0,0,"454 ",,terminal_output +1310,2467032,"TERMINAL",0,0,"555 ",,terminal_output +1311,2468079,"TERMINAL",0,0,"656 ",,terminal_output +1312,2469189,"TERMINAL",0,0,"757 ",,terminal_output +1313,2470165,"TERMINAL",0,0,"858 ",,terminal_output +1314,2471296,"TERMINAL",0,0,"959 ",,terminal_output +1315,2471382,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node_darkness_filter_133M\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --darkness_threshold=50 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-darkness-filter-133M-$slurm_job_id \\n --tags lam minecraft 8-node darkness-filter 133M \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --model_dim=768 \\n --num_blocks=10 \\n --num_heads=12 \\n --latent_dim=48 \\n --ffn_dim=3072 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +1316,2472277,"TERMINAL",0,0,"501:00 ",,terminal_output +1317,2473359,"TERMINAL",0,0,"101 ",,terminal_output +1318,2474366,"TERMINAL",0,0,"202 ",,terminal_output +1319,2475420,"TERMINAL",0,0,"303 ",,terminal_output +1320,2476475,"TERMINAL",0,0,"404 ",,terminal_output +1321,2477578,"TERMINAL",0,0,"505 ",,terminal_output +1322,2478573,"TERMINAL",0,0,"606 ",,terminal_output +1323,2479623,"TERMINAL",0,0,"707 ",,terminal_output +1324,2480678,"TERMINAL",0,0,"808 ",,terminal_output +1325,2481734,"TERMINAL",0,0,"910 ",,terminal_output +1326,2481957,"TERMINAL",0,0,"bash",,terminal_focus +1327,2482768,"TERMINAL",0,0,"3:0111 ",,terminal_output +1328,2483743,"TERMINAL",0,0,"watch",,terminal_focus +1329,2483832,"TERMINAL",0,0,"212 ",,terminal_output +1330,2484871,"TERMINAL",0,0,"313 ",,terminal_output +1331,2485327,"TERMINAL",0,0,"bash",,terminal_focus +1332,2485915,"TERMINAL",0,0,"414 ",,terminal_output +1333,2486360,"TERMINAL",0,0,"ls",,terminal_command +1334,2486393,"TERMINAL",0,0,"]633;E;2025-09-01 13:23:04 ls;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C",,terminal_output +1335,2486556,"TERMINAL",0,0,"train_lam_minecraft_8node_3431870.log\r\ntrain_lam_minecraft_8node_3431875.log\r\ntrain_lam_minecraft_8node_3431876.log\r\ntrain_lam_minecraft_8node_3431885.log\r\ntrain_lam_minecraft_8node_3431895.log\r\ntrain_lam_minecraft_8node_3454890.log\r\ntrain_lam_minecraft_8node_3454917.log\r\ntrain_lam_minecraft_8node_3454941.log\r\ntrain_lam_minecraft_8node_3454944.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1336,2487009,"TERMINAL",0,0,"515 ",,terminal_output +1337,2488012,"TERMINAL",0,0,"616 ",,terminal_output +1338,2489056,"TERMINAL",0,0,"717 ",,terminal_output +1339,2489873,"TERMINAL",0,0,"tail -f train_lam_minecraft_8node_3454944.log",,terminal_command +1340,2489919,"TERMINAL",0,0,"]633;E;2025-09-01 13:23:08 tail -f train_lam_minecraft_8node_3454944.log;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;CSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0604\r\nGpuFreq=control_disabled\r\nwandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\nwandb: Tracking run with wandb version 0.19.11\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250901_132247-3454944\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run lam-minecraft-1-node-dev-3454944\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/3454944\r\n",,terminal_output +1341,2490140,"TERMINAL",0,0,"818 ",,terminal_output +1342,2491152,"TERMINAL",0,0,"919 ",,terminal_output +1343,2492212,"TERMINAL",0,0,"1020 ",,terminal_output +1344,2493265,"TERMINAL",0,0,"121 ",,terminal_output +1345,2494305,"TERMINAL",0,0,"222 ",,terminal_output +1346,2495362,"TERMINAL",0,0,"323 ",,terminal_output +1347,2496428,"TERMINAL",0,0,"424 ",,terminal_output +1348,2497468,"TERMINAL",0,0,"525 ",,terminal_output +1349,2498514,"TERMINAL",0,0,"626 ",,terminal_output +1350,2499566,"TERMINAL",0,0,"727 ",,terminal_output +1351,2500617,"TERMINAL",0,0,"828 ",,terminal_output +1352,2501668,"TERMINAL",0,0,"929 ",,terminal_output +1353,2502712,"TERMINAL",0,0,"2030 ",,terminal_output +1354,2503761,"TERMINAL",0,0,"232 ",,terminal_output +1355,2504849,"TERMINAL",0,0,"333 ",,terminal_output +1356,2505851,"TERMINAL",0,0,"434 ",,terminal_output +1357,2506901,"TERMINAL",0,0,"535 ",,terminal_output +1358,2507952,"TERMINAL",0,0,"636 ",,terminal_output +1359,2509010,"TERMINAL",0,0,"737 ",,terminal_output +1360,2510057,"TERMINAL",0,0,"838 ",,terminal_output +1361,2511111,"TERMINAL",0,0,"939 ",,terminal_output +1362,2512172,"TERMINAL",0,0,"3040 ",,terminal_output +1363,2513973,"TERMINAL",0,0,"142 ",,terminal_output +1364,2515014,"TERMINAL",0,0,"343 ",,terminal_output +1365,2516094,"TERMINAL",0,0,"444 ",,terminal_output +1366,2517113,"TERMINAL",0,0,"545 ",,terminal_output +1367,2518175,"TERMINAL",0,0,"646 ",,terminal_output +1368,2519224,"TERMINAL",0,0,"747 ",,terminal_output +1369,2519290,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;130",,terminal_output +1370,2519477,"TERMINAL",0,0,"^C",,terminal_command +1371,2519549,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]633;E;;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D",,terminal_output +1372,2520310,"TERMINAL",0,0,"848 ",,terminal_output +1373,2521328,"TERMINAL",0,0,"949 ",,terminal_output +1374,2522394,"TERMINAL",0,0,"4050 ",,terminal_output +1375,2522991,"TERMINAL",0,0,"queue",,terminal_command +1376,2523084,"TERMINAL",0,0,"]633;E;2025-09-01 13:23:41 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: s... hkn1993.localdomain: Mon Sep 1 13:23:41 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454944 accelerat train_la tum_cte0 R\t1:511 hkn0604",,terminal_output +1377,2523446,"TERMINAL",0,0,"151 ",,terminal_output +1378,2524137,"TERMINAL",0,0,"252 ",,terminal_output +1379,2524470,"TERMINAL",0,0,"252 ",,terminal_output +1380,2525205,"TERMINAL",0,0,"353 [?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1381,2525517,"TERMINAL",0,0,"353 ",,terminal_output +1382,2526567,"TERMINAL",0,0,"454 ",,terminal_output +1383,2527388,"TERMINAL",0,0,"scancel 3454944",,terminal_command +1384,2527419,"TERMINAL",0,0,"]633;E;2025-09-01 13:23:45 scancel 3454944;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1385,2527608,"TERMINAL",0,0,"5CG\t1:55 ",,terminal_output +1386,2528661,"TERMINAL",0,0,"6",,terminal_output +1387,2529714,"TERMINAL",0,0,"7",,terminal_output +1388,2530741,"TERMINAL",0,0,"8",,terminal_output +1389,2531082,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +1390,2531783,"TERMINAL",0,0,"50",,terminal_output +1391,2532311,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2008,0,"",shellscript,selection_mouse +1392,2532850,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2005,0,"",shellscript,selection_mouse +1393,2532882,"TERMINAL",0,0,"1",,terminal_output +1394,2533604,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2004,1,"",shellscript,content +1395,2533908,"TERMINAL",0,0,"2",,terminal_output +1396,2534032,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2003,1,"",shellscript,content +1397,2534298,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2003,1,"",shellscript,content +1398,2534933,"TERMINAL",0,0,"3",,terminal_output +1399,2535405,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2003,0,"8",shellscript,content +1400,2535406,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2004,0,"",shellscript,selection_keyboard +1401,2535976,"TERMINAL",0,0,"4",,terminal_output +1402,2537036,"TERMINAL",0,0,"5",,terminal_output +1403,2538039,"TERMINAL",0,0,"6",,terminal_output +1404,2538769,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2047,0,"4",shellscript,content +1405,2538769,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2046,1,"",shellscript,content +1406,2538769,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2004,0,"96",shellscript,content +1407,2539113,"TERMINAL",0,0,"7",,terminal_output +1408,2540177,"TERMINAL",0,0,"8",,terminal_output +1409,2541187,"TERMINAL",0,0,"9",,terminal_output +1410,2542251,"TERMINAL",0,0,"4:00",,terminal_output +1411,2543026,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2028,0,"",shellscript,selection_mouse +1412,2543293,"TERMINAL",0,0,"1",,terminal_output +1413,2543853,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2027,0,"",shellscript,selection_command +1414,2544166,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2028,0,"",shellscript,selection_command +1415,2544316,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2027,1,"",shellscript,content +1416,2544387,"TERMINAL",0,0,"2",,terminal_output +1417,2544440,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2027,0,"2",shellscript,content +1418,2544441,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2028,0,"",shellscript,selection_keyboard +1419,2545388,"TERMINAL",0,0,"3",,terminal_output +1420,2546468,"TERMINAL",0,0,"4",,terminal_output +1421,2547522,"TERMINAL",0,0,"5",,terminal_output +1422,2548552,"TERMINAL",0,0,"6",,terminal_output +1423,2549598,"TERMINAL",0,0,"7",,terminal_output +1424,2550649,"TERMINAL",0,0,"8",,terminal_output +1425,2551699,"TERMINAL",0,0,"9",,terminal_output +1426,2552546,"TERMINAL",0,0,"python",,terminal_focus +1427,2552740,"TERMINAL",0,0,"10",,terminal_output +1428,2553784,"TERMINAL",0,0,"2",,terminal_output +1429,2554434,"TERMINAL",0,0,"8",,terminal_output +1430,2554830,"TERMINAL",0,0,"3",,terminal_output +1431,2555385,"TERMINAL",0,0,"[?25l9[?25h",,terminal_output +1432,2555711,"TERMINAL",0,0,"[?25l6[?25h",,terminal_output +1433,2555894,"TERMINAL",0,0,"4",,terminal_output +1434,2556385,"TERMINAL",0,0,"[?25l*[?25h",,terminal_output +1435,2556779,"TERMINAL",0,0,"[?25l4[?25h",,terminal_output +1436,2556830,"TERMINAL",0,0,"\r\n3584\r\n>>> ",,terminal_output +1437,2556955,"TERMINAL",0,0,"5",,terminal_output +1438,2558026,"TERMINAL",0,0,"6",,terminal_output +1439,2559028,"TERMINAL",0,0,"7",,terminal_output +1440,2559497,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2090,0,"",shellscript,selection_mouse +1441,2559609,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2088,4,"3072",shellscript,selection_mouse +1442,2560013,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2088,4,"x",shellscript,content +1443,2560014,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2089,0,"",shellscript,selection_keyboard +1444,2560097,"TERMINAL",0,0,"8",,terminal_output +1445,2560565,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2088,1,"",shellscript,content +1446,2560970,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2088,0,"3584",shellscript,content +1447,2561112,"TERMINAL",0,0,"9",,terminal_output +1448,2562169,"TERMINAL",0,0,"20",,terminal_output +1449,2562823,"TERMINAL",0,0,"watch",,terminal_focus +1450,2563191,"TERMINAL",0,0,"1",,terminal_output +1451,2564257,"TERMINAL",0,0,"2",,terminal_output +1452,2564408,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +1453,2566495,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",,terminal_command +1454,2566558,"TERMINAL",0,0,"]633;E;2025-09-01 13:24:24 sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;CSubmitted batch job 3454948\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +1455,2567471,"TERMINAL",0,0,"bash",,terminal_focus +1456,2570896,"TERMINAL",0,0,"queue",,terminal_command +1457,2570980,"TERMINAL",0,0,"]633;E;2025-09-01 13:24:29 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: s... hkn1993.localdomain: Mon Sep 1 13:24:29 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454944 accelerat train_la tum_cte0 CG\t1:551 hkn06043454948 accelerat train_la tum_cte0 R\t0:021 hkn0813",,terminal_output +1458,2572021,"TERMINAL",0,0,"3003 ",,terminal_output +1459,2573058,"TERMINAL",0,0,"104 ",,terminal_output +1460,2573968,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1461,2574851,"TERMINAL",0,0,"ls",,terminal_command +1462,2574885,"TERMINAL",0,0,"]633;E;2025-09-01 13:24:33 ls;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;Ctrain_lam_minecraft_8node_3431870.log\r\ntrain_lam_minecraft_8node_3431875.log\r\ntrain_lam_minecraft_8node_3431876.log\r\ntrain_lam_minecraft_8node_3431885.log\r\ntrain_lam_minecraft_8node_3431895.log\r\ntrain_lam_minecraft_8node_3454890.log\r\ntrain_lam_minecraft_8node_3454917.log\r\ntrain_lam_minecraft_8node_3454941.log\r\ntrain_lam_minecraft_8node_3454944.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1463,2576103,"TERMINAL",0,0,"ls",,terminal_command +1464,2576139,"TERMINAL",0,0,"]633;E;2025-09-01 13:24:34 ls;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;Ctrain_lam_minecraft_8node_3431870.log\r\ntrain_lam_minecraft_8node_3431875.log\r\ntrain_lam_minecraft_8node_3431876.log\r\ntrain_lam_minecraft_8node_3431885.log\r\ntrain_lam_minecraft_8node_3431895.log\r\ntrain_lam_minecraft_8node_3454890.log\r\ntrain_lam_minecraft_8node_3454917.log\r\ntrain_lam_minecraft_8node_3454941.log\r\ntrain_lam_minecraft_8node_3454944.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1465,2576908,"TERMINAL",0,0,"ls",,terminal_command +1466,2577783,"TERMINAL",0,0,"ls",,terminal_command +1467,2577822,"TERMINAL",0,0,"]633;E;2025-09-01 13:24:36 ls;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;Ctrain_lam_minecraft_8node_3431870.log\r\ntrain_lam_minecraft_8node_3431875.log\r\ntrain_lam_minecraft_8node_3431876.log\r\ntrain_lam_minecraft_8node_3431885.log\r\ntrain_lam_minecraft_8node_3431895.log\r\ntrain_lam_minecraft_8node_3454890.log\r\ntrain_lam_minecraft_8node_3454917.log\r\ntrain_lam_minecraft_8node_3454941.log\r\ntrain_lam_minecraft_8node_3454944.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1468,2578509,"TERMINAL",0,0,"ls",,terminal_command +1469,2580129,"TERMINAL",0,0,"clear",,terminal_command +1470,2580169,"TERMINAL",0,0,"]633;E;2025-09-01 13:24:38 clear;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1471,2581009,"TERMINAL",0,0,"ls",,terminal_command +1472,2581047,"TERMINAL",0,0,"]633;E;2025-09-01 13:24:39 ls;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;Ctrain_lam_minecraft_8node_3431870.log\r\ntrain_lam_minecraft_8node_3431875.log\r\ntrain_lam_minecraft_8node_3431876.log\r\ntrain_lam_minecraft_8node_3431885.log\r\ntrain_lam_minecraft_8node_3431895.log\r\ntrain_lam_minecraft_8node_3454890.log\r\ntrain_lam_minecraft_8node_3454917.log\r\ntrain_lam_minecraft_8node_3454941.log\r\ntrain_lam_minecraft_8node_3454944.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1473,2589729,"TERMINAL",0,0,"tail -f train_lam_minecraft_8node_3454948.log",,terminal_command +1474,2589765,"TERMINAL",0,0,"]633;E;2025-09-01 13:24:48 tail -f train_lam_minecraft_8node_3454948.log;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;Ctail: cannot open 'train_lam_minecraft_8node_3454948.log' for reading: No such file or directory\r\ntail: no files remaining\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam",,terminal_output +1475,2591476,"TERMINAL",0,0,"tail -f train_lam_minecraft_8node_3454948.log",,terminal_command +1476,2592350,"TERMINAL",0,0,"tail -f train_lam_minecraft_8node_3454948.log",,terminal_command +1477,2592386,"TERMINAL",0,0,"]633;E;2025-09-01 13:24:50 tail -f train_lam_minecraft_8node_3454948.log;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;Ctail: cannot open 'train_lam_minecraft_8node_3454948.log' for reading: No such file or directory\r\ntail: no files remaining\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;1",,terminal_output +1478,2593064,"TERMINAL",0,0,"ls",,terminal_command +1479,2593121,"TERMINAL",0,0,"]633;E;2025-09-01 13:24:51 ls;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;Ctrain_lam_minecraft_8node_3431870.log\r\ntrain_lam_minecraft_8node_3431875.log\r\ntrain_lam_minecraft_8node_3431876.log\r\ntrain_lam_minecraft_8node_3431885.log\r\ntrain_lam_minecraft_8node_3431895.log\r\ntrain_lam_minecraft_8node_3454890.log\r\ntrain_lam_minecraft_8node_3454917.log\r\ntrain_lam_minecraft_8node_3454941.log\r\ntrain_lam_minecraft_8node_3454944.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1480,2594471,"TERMINAL",0,0,"queue",,terminal_command +1481,2594508,"TERMINAL",0,0,"]633;E;2025-09-01 13:24:52 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C",,terminal_output +1482,2594606,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: s... hkn1993.localdomain: Mon Sep 1 13:24:52 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454948 accelerat train_la tum_cte0 R\t0:251 hkn0813",,terminal_output +1483,2595625,"TERMINAL",0,0,"326 ",,terminal_output +1484,2596241,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1485,2601540,"TERMINAL",0,0,"tail -f train_lam_minecraft_8node_3454948.log",,terminal_command +1486,2601606,"TERMINAL",0,0,"]633;E;2025-09-01 13:24:59 tail -f train_lam_minecraft_8node_3454948.log;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;CSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3454948\r\nSLURM_NODEID=0\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=train_lam_minecraft_1node_dev\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0813\r\nGpuFreq=control_disabled\r\n",,terminal_output +1487,2603757,"slurm/utils/mihir/model_sizes.md",0,0,"",markdown,tab +1488,2606591,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",0,0,"",shellscript,tab +1489,2607577,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2094,0,"",shellscript,selection_mouse +1490,2607851,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2093,1,"\",shellscript,selection_mouse +1491,2607852,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2089,5,"584 \",shellscript,selection_mouse +1492,2607852,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2064,30,"_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1493,2607853,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2062,32,"nt_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1494,2607853,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2061,33,"ent_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1495,2607853,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2060,34,"tent_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1496,2607854,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2058,36,"latent_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1497,2607881,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2057,37,"-latent_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1498,2607882,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2056,38,"--latent_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1499,2607899,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2055,39," --latent_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1500,2607916,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2054,40," --latent_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1501,2607938,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2032,62," --num_heads=14 \\n --latent_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1502,2607949,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2031,63," --num_heads=14 \\n --latent_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1503,2608058,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",2009,85," --num_blocks=12 \\n --num_heads=14 \\n --latent_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1504,2608341,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch",1987,107," --model_dim=896 \\n --num_blocks=12 \\n --num_heads=14 \\n --latent_dim=48 \\n --ffn_dim=3584 \",shellscript,selection_mouse +1505,2617085,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",0,0,"",shellscript,tab +1506,2618386,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2102,2," \",shellscript,selection_mouse +1507,2618387,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2101,3,"2 \",shellscript,selection_mouse +1508,2618387,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2099,5,"072 \",shellscript,selection_mouse +1509,2618388,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2098,6,"3072 \",shellscript,selection_mouse +1510,2618388,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2097,7,"=3072 \",shellscript,selection_mouse +1511,2618454,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2096,8,"m=3072 \",shellscript,selection_mouse +1512,2618455,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2095,9,"im=3072 \",shellscript,selection_mouse +1513,2618455,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2094,10,"dim=3072 \",shellscript,selection_mouse +1514,2618516,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2093,11,"_dim=3072 \",shellscript,selection_mouse +1515,2618581,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +1516,2618632,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2092,12,"n_dim=3072 \",shellscript,selection_mouse +1517,2618702,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2091,13,"fn_dim=3072 \",shellscript,selection_mouse +1518,2618703,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2090,14,"ffn_dim=3072 \",shellscript,selection_mouse +1519,2618738,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2089,15,"-ffn_dim=3072 \",shellscript,selection_mouse +1520,2618739,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2066,38,"--latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1521,2618776,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2065,39," --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1522,2618811,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2064,40," --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1523,2618812,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2043,61," --num_heads=12 \\n --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1524,2618846,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2042,62," --num_heads=12 \\n --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1525,2618881,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2041,63," --num_heads=12 \\n --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1526,2618882,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",2019,85," --num_blocks=10 \\n --num_heads=12 \\n --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1527,2619074,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",1997,107," --model_dim=768 \\n --num_blocks=10 \\n --num_heads=12 \\n --latent_dim=48 \\n --ffn_dim=3072 \",shellscript,selection_mouse +1528,2619586,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.19.11\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250901_132516-3454948\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run lam-minecraft-1-node-dev-3454948\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/3454948\r\n",,terminal_output +1529,2619725,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",1997,107,"",shellscript,content +1530,2620227,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-132M.sbatch",1997,0," --model_dim=896 \\n --num_blocks=12 \\n --num_heads=14 \\n --latent_dim=48 \\n --ffn_dim=3584 \",shellscript,content +1531,2677387,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;130",,terminal_output +1532,2679594,"TERMINAL",0,0,"queue",,terminal_command +1533,2679665,"TERMINAL",0,0,"]633;E;2025-09-01 13:26:17 queue;494c8224-3c77-4dff-9ba2-7c9ac811f0d6]633;C[?1049h(B[?7hEvery 1.0s: s... hkn1993.localdomain: Mon Sep 1 13:26:17 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454948 accelerat train_la tum_cte0 R\t1:501 hkn0813",,terminal_output +1534,2680740,"TERMINAL",0,0,"851 ",,terminal_output +1535,2681758,"TERMINAL",0,0,"2053 ",,terminal_output +1536,2682356,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam]633;D;0",,terminal_output +1537,2684568,"TERMINAL",0,0,"scancel 3454948",,terminal_command +1538,2687410,"TERMINAL",0,0,"bash",,terminal_focus +1539,2691122,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/lam/%x_%j.log\n#SBATCH --job-name=train_lam_minecraft_8node_darkness_filter_133M\n#SBATCH --reservation=llmtum\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# --- signal trap to requeue job before timeout ---\nrequeue_job() {\n echo ""[$(date)] caught sigusr1 (timeout warning), requeueing slurm job $SLURM_JOB_ID...""\n scontrol requeue $SLURM_JOB_ID\n exit 0\n}\n\ntrap requeue_job sigusr1\n\n# set checkpoint flag based on restart count\nrestart_count=$(scontrol show job $SLURM_JOB_ID | grep -o 'Restarts=[0-9]*' | cut -d'=' -f2)\n\nif [ $restart_count -eq 0 ]; then\n restore_ckpt_flag=""--no-restore-ckpt""\nelse\n restore_ckpt_flag=""--restore-ckpt""\nfi\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/lam/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --save_ckpt \\n $restore_ckpt_flag \\n --wandb_id $slurm_job_id \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=160 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --darkness_threshold=50 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=lam-minecraft-8-node-darkness-filter-133M-$slurm_job_id \\n --tags lam minecraft 8-node darkness-filter 133M \\n --entity instant-uv \\n --project jafar \\n --num_latents=100 \\n --model_dim=896 \\n --num_blocks=12 \\n --num_heads=14 \\n --latent_dim=48 \\n --ffn_dim=3584 \\n --data_dir $array_records_dir &\n\nchild_pid=$!\n\nwait $child_pid\n",shellscript,tab +1540,2693567,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1921,0,"",shellscript,selection_mouse +1541,2695300,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1921,1,"M",shellscript,selection_command +1542,2696296,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1852,1,"M",shellscript,selection_command +1543,2700159,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1852,0,"",shellscript,selection_command +1544,2700893,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1920,1,"",shellscript,content +1545,2700893,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1851,1,"",shellscript,content +1546,2701020,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1918,1,"",shellscript,content +1547,2701021,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1850,1,"",shellscript,content +1548,2701345,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1916,1,"",shellscript,content +1549,2701346,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1849,1,"",shellscript,content +1550,2702152,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1915,0,"3",shellscript,content +1551,2702153,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1849,0,"3",shellscript,content +1552,2702153,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1850,0,"",shellscript,selection_keyboard +1553,2702307,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1917,0,"1",shellscript,content +1554,2702307,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1850,0,"1",shellscript,content +1555,2702308,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1851,0,"",shellscript,selection_keyboard +1556,2702827,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1919,0,"1",shellscript,content +1557,2702827,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1851,0,"1",shellscript,content +1558,2702828,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1852,0,"",shellscript,selection_keyboard +1559,2704512,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",1851,0,"",shellscript,selection_command +1560,2705987,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",2172,0,"",shellscript,selection_command +1561,2707429,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",0,0,"",shellscript,selection_command +1562,2707835,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",20,0,"",shellscript,selection_command +1563,2708324,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",21,0,"",shellscript,selection_command +1564,2708404,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",39,0,"",shellscript,selection_command +1565,2708404,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",67,0,"",shellscript,selection_command +1566,2708426,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",91,0,"",shellscript,selection_command +1567,2708462,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",123,0,"",shellscript,selection_command +1568,2708539,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",149,0,"",shellscript,selection_command +1569,2708550,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",170,0,"",shellscript,selection_command +1570,2708629,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",273,0,"",shellscript,selection_command +1571,2708800,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",375,0,"",shellscript,selection_command +1572,2709144,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",440,0,"",shellscript,selection_command +1573,2710139,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",439,0,"11",shellscript,content +1574,2710139,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",436,2,"",shellscript,content +1575,2720174,"TERMINAL",0,0,"queue",,terminal_command +1576,2720233,"TERMINAL",0,0,"]633;E;2025-09-01 13:26:58 queue;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: s... hkn1993.localdomain: Mon Sep 1 13:26:58 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454948 accelerat train_la tum_cte0 CG\t1:551 hkn0813",,terminal_output +1577,2720836,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +1578,2722514,"TERMINAL",0,0,"idöing",,terminal_command +1579,2722567,"TERMINAL",0,0,"]633;E;2025-09-01 13:27:00 idöing;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;Cbash: idöing: command not found...\r\n",,terminal_output +1580,2723401,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;130",,terminal_output +1581,2724473,"TERMINAL",0,0,"idling",,terminal_command +1582,2724548,"TERMINAL",0,0,"]633;E;2025-09-01 13:27:02 idling;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: s... hkn1993.localdomain: Mon Sep 1 13:27:02 2025Partition dev_cpuonly:\t 7 nodes idle\rPartition cpuonly: 118 nodes idle\rPartition dev_accelerated:\t 1 nodes idle\rPartition accelerated: 32 nodes idle\rPartition dev_accelerated-h100 :\t 1 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 0 nodes idle\rPartition accelerated-h200:\t 0 nodes idle",,terminal_output +1583,2725565,"TERMINAL",0,0,"3",,terminal_output +1584,2726076,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +1585,2727289,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",0,0,"",shellscript,tab +1586,2744300,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch^C",,terminal_command +1587,2744338,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]633;E;;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D",,terminal_output +1588,2751985,"TERMINAL",0,0,"sync-runner",,terminal_command +1589,2752059,"TERMINAL",0,0,"]633;E;2025-09-01 13:27:30 sync-runner;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;Csending incremental file list\r\n",,terminal_output +1590,2752769,"TERMINAL",0,0,"train_lam.py\r\n",,terminal_output +1591,2753735,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/lam/\r\nslurm/jobs/mihir/horeka/lam/train_lam_minecraft_1node_dev.sbatch\r\nslurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch\r\nslurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch\r\nslurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch\r\nslurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch\r\n",,terminal_output +1592,2753799,"TERMINAL",0,0,"\r\nsent 50,462 bytes received 306 bytes 20,307.20 bytes/sec\r\ntotal size is 128,230,617 speedup is 2,525.82\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +1593,2755162,"TERMINAL",0,0,"runer",,terminal_command +1594,2755199,"TERMINAL",0,0,"]633;E;2025-09-01 13:27:33 runer;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;Cbash: runer: command not found...\r\n",,terminal_output +1595,2755343,"TERMINAL",0,0,"]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;127]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +1596,2758242,"TERMINAL",0,0,"runner",,terminal_command +1597,2758279,"TERMINAL",0,0,"]633;E;2025-09-01 13:27:36 runner;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1598,2763120,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",,terminal_command +1599,2763169,"TERMINAL",0,0,"]633;E;2025-09-01 13:27:41 sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C",,terminal_output +1600,2763261,"TERMINAL",0,0,"Submitted batch job 3454953\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1601,2767106,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",,terminal_command +1602,2767170,"TERMINAL",0,0,"]633;E;2025-09-01 13:27:45 sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;CSubmitted batch job 3454954\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1603,2771344,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",,terminal_command +1604,2771481,"TERMINAL",0,0,"]633;E;2025-09-01 13:27:49 sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;CSubmitted batch job 3454955\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1605,2774983,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",,terminal_command +1606,2775018,"TERMINAL",0,0,"]633;E;2025-09-01 13:27:53 sbatch slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;CSubmitted batch job 3454956\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1607,2776181,"TERMINAL",0,0,"queueu",,terminal_command +1608,2776245,"TERMINAL",0,0,"]633;E;2025-09-01 13:27:54 queueu;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;Cbash: queueu: command not found...\r\n",,terminal_output +1609,2776428,"TERMINAL",0,0,"]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;127]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs",,terminal_output +1610,2777826,"TERMINAL",0,0,"queue",,terminal_command +1611,2777905,"TERMINAL",0,0,"]633;E;2025-09-01 13:27:56 queue;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 13:27:56 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454956 accelerat train_la tum_cte0 PD\t0:00\t 8 (Resources)3454955 accelerat train_la tum_cte0 R\t0:06\t 8 hkn[0504,0520,0720,0722-0724,0728,0731]3454954 accelerat train_la tum_cte0 R\t0:10\t 8 hkn[0521-0528]3454953 accelerat train_la tum_cte0 R\t0:14\t 8 hkn[0703,0706-0707,0711-0715]",,terminal_output +1612,2778980,"TERMINAL",0,0,"7715",,terminal_output +1613,2780002,"TERMINAL",0,0,"8826",,terminal_output +1614,2781056,"TERMINAL",0,0,"9937",,terminal_output +1615,2782089,"TERMINAL",0,0,"8:001048",,terminal_output +1616,2783139,"TERMINAL",0,0,"1159",,terminal_output +1617,2784222,"TERMINAL",0,0,"22620",,terminal_output +1618,2785240,"TERMINAL",0,0,"3371",,terminal_output +1619,2786283,"TERMINAL",0,0,"4482",,terminal_output +1620,2787401,"TERMINAL",0,0,"5593",,terminal_output +1621,2788385,"TERMINAL",0,0,"66204",,terminal_output +1622,2788550,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1623,2789882,"TERMINAL",0,0,"idling",,terminal_command +1624,2789954,"TERMINAL",0,0,"]633;E;2025-09-01 13:28:08 idling;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Mon Sep 1 13:28:08 2025Partition dev_cpuonly:\t 7 nodes idle\rPartition cpuonly: 117 nodes idle\rPartition dev_accelerated:\t 1 nodes idle\rPartition accelerated:\t 8 nodes idle\rPartition dev_accelerated-h100 :\t 1 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 0 nodes idle\rPartition accelerated-h200:\t 0 nodes idle",,terminal_output +1625,2790986,"TERMINAL",0,0,"9",,terminal_output +1626,2791597,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1627,2792710,"TERMINAL",0,0,"queue",,terminal_command +1628,2792793,"TERMINAL",0,0,"]633;E;2025-09-01 13:28:11 queue;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 13:28:11 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454956 accelerat train_la tum_cte0 PD\t0:00\t 8 (Resources)3454955 accelerat train_la tum_cte0 R\t0:21\t 8 hkn[0504,0520,0720,0722-0724,0728,0731]3454954 accelerat train_la tum_cte0 R\t0:25\t 8 hkn[0521-0528]3454953 accelerat train_la tum_cte0 R\t0:29\t 8 hkn[0703,0706-0707,0711-0715]",,terminal_output +1629,2793852,"TERMINAL",0,0,"22630",,terminal_output +1630,2794897,"TERMINAL",0,0,"3371",,terminal_output +1631,2795948,"TERMINAL",0,0,"4482",,terminal_output +1632,2796997,"TERMINAL",0,0,"5593",,terminal_output +1633,2798041,"TERMINAL",0,0,"66304",,terminal_output +1634,2798684,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1635,2799838,"TERMINAL",0,0,"fqueue",,terminal_command +1636,2799928,"TERMINAL",0,0,"]633;E;2025-09-01 13:28:18 fqueue;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: squeue -o ""%.10i %.16P %.30j %.8u %.8T %.10M %.9l %.6D %R""hkn1993.localdomain: Mon Sep 1 13:28:18 2025JOBIDPARTITIONNAME USER STATE\t TIME TIME_LIMI NODES NODELIST(REASON)3454956\taccelerated train_lam_minecraft_8node_dark tum_cte0 PENDING\t 0:00 2-00:00:00\t8 (Resources)3454955\taccelerated train_lam_minecraft_8node_dark tum_cte0 RUNNING\t 0:28 2-00:00:00\t8 hkn[0504,0520,0720,0722-0724,0728,0731]3454954\taccelerated train_lam_minecraft_8node_dark tum_cte0 RUNNING\t 0:32 2-00:00:00\t8 hkn[0521-0528]3454953\taccelerated train_lam_minecraft_8node_dark tum_cte0 RUNNING\t 0:36 2-00:00:00\t8 hkn[0703,0706-0707,0711-0715]",,terminal_output +1637,2800941,"TERMINAL",0,0,"9937",,terminal_output +1638,2801982,"TERMINAL",0,0,"203048",,terminal_output +1639,2803018,"TERMINAL",0,0,"1159",,terminal_output +1640,2804047,"TERMINAL",0,0,"22640",,terminal_output +1641,2805078,"TERMINAL",0,0,"3371",,terminal_output +1642,2806124,"TERMINAL",0,0,"4482",,terminal_output +1643,2807143,"TERMINAL",0,0,"5593",,terminal_output +1644,2808207,"TERMINAL",0,0,"66404",,terminal_output +1645,2809120,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",0,0,"",shellscript,tab +1646,2809242,"TERMINAL",0,0,"7715",,terminal_output +1647,2810259,"TERMINAL",0,0,"8826",,terminal_output +1648,2811305,"TERMINAL",0,0,"9937",,terminal_output +1649,2812340,"TERMINAL",0,0,"304048",,terminal_output +1650,2813404,"TERMINAL",0,0,"1159",,terminal_output +1651,2814423,"TERMINAL",0,0,"22650",,terminal_output +1652,2815448,"TERMINAL",0,0,"3371",,terminal_output +1653,2815696,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1654,2824023,"TERMINAL",0,0,"queue",,terminal_command +1655,2824111,"TERMINAL",0,0,"]633;E;2025-09-01 13:28:42 queue;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 13:28:42 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454956 accelerat train_la tum_cte0 PD\t0:00\t 8 (Resources)3454955 accelerat train_la tum_cte0 R\t0:52\t 8 hkn[0504,0520,0720,0722-0724,0728,0731]3454954 accelerat train_la tum_cte0 R\t0:56\t 8 hkn[0521-0528]3454953 accelerat train_la tum_cte0 R\t1:00\t 8 hkn[0703,0706-0707,0711-0715]",,terminal_output +1656,2825030,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +1657,2989524,"TERMINAL",0,0,"queue",,terminal_command +1658,2989581,"TERMINAL",0,0,"]633;E;2025-09-01 13:31:27 queue;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 13:31:27 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454956 accelerat train_la tum_cte0 PD\t0:00\t 8 (Resources)3454955 accelerat train_la tum_cte0 R\t3:37\t 8 hkn[0504,0520,0720,0722-0724,0728,0731]3454954 accelerat train_la tum_cte0 R\t3:41\t 8 hkn[0521-0528]3454953 accelerat train_la tum_cte0 R\t3:45\t 8 hkn[0703,0706-0707,0711-0715]",,terminal_output +1659,2990626,"TERMINAL",0,0,"8826",,terminal_output +1660,2991678,"TERMINAL",0,0,"9937",,terminal_output +1661,2992736,"TERMINAL",0,0,"304159",,terminal_output +1662,2993758,"TERMINAL",0,0,"22650",,terminal_output +1663,2994813,"TERMINAL",0,0,"3371",,terminal_output +1664,2995867,"TERMINAL",0,0,"4482",,terminal_output +1665,2996908,"TERMINAL",0,0,"5593",,terminal_output +1666,2999340,"TERMINAL",0,0,"67515",,terminal_output +1667,3000352,"TERMINAL",0,0,"8826",,terminal_output +1668,3001401,"TERMINAL",0,0,"9937",,terminal_output +1669,3002493,"TERMINAL",0,0,"405048",,terminal_output +1670,3003494,"TERMINAL",0,0,"1159",,terminal_output +1671,3004549,"TERMINAL",0,0,"2264:00",,terminal_output +1672,3005601,"TERMINAL",0,0,"3371",,terminal_output +1673,3006682,"TERMINAL",0,0,"4482",,terminal_output +1674,3007690,"TERMINAL",0,0,"5593",,terminal_output +1675,3008737,"TERMINAL",0,0,"674:015",,terminal_output +1676,3009790,"TERMINAL",0,0,"8826",,terminal_output +1677,3010845,"TERMINAL",0,0,"9937",,terminal_output +1678,3011891,"TERMINAL",0,0,"504:0048",,terminal_output +1679,3012945,"TERMINAL",0,0,"1159",,terminal_output +1680,3013982,"TERMINAL",0,0,"22610",,terminal_output +1681,3015032,"TERMINAL",0,0,"3371",,terminal_output +1682,3016080,"TERMINAL",0,0,"4482",,terminal_output +1683,3017130,"TERMINAL",0,0,"5593",,terminal_output +1684,3018178,"TERMINAL",0,0,"66104",,terminal_output +1685,3019215,"TERMINAL",0,0,"7715",,terminal_output +1686,3020273,"TERMINAL",0,0,"8826",,terminal_output +1687,3021331,"TERMINAL",0,0,"9937",,terminal_output +1688,3022380,"TERMINAL",0,0,"2:001048",,terminal_output +1689,3023425,"TERMINAL",0,0,"1159",,terminal_output +1690,3024481,"TERMINAL",0,0,"22620",,terminal_output +1691,3025529,"TERMINAL",0,0,"3371",,terminal_output +1692,3026565,"TERMINAL",0,0,"4482",,terminal_output +1693,3027611,"TERMINAL",0,0,"5593",,terminal_output +1694,3028656,"TERMINAL",0,0,"66204",,terminal_output +1695,3029710,"TERMINAL",0,0,"7715",,terminal_output +1696,3030762,"TERMINAL",0,0,"8937",,terminal_output +1697,3031811,"TERMINAL",0,0,"102048",,terminal_output +1698,3032866,"TERMINAL",0,0,"1159",,terminal_output +1699,3033918,"TERMINAL",0,0,"22630",,terminal_output +1700,3034967,"TERMINAL",0,0,"3371",,terminal_output +1701,3036020,"TERMINAL",0,0,"4482",,terminal_output +1702,3037076,"TERMINAL",0,0,"5593",,terminal_output +1703,3038124,"TERMINAL",0,0,"66304",,terminal_output +1704,3039169,"TERMINAL",0,0,"7715",,terminal_output +1705,3040221,"TERMINAL",0,0,"8826",,terminal_output +1706,3041271,"TERMINAL",0,0,"9937",,terminal_output +1707,3042323,"TERMINAL",0,0,"203048",,terminal_output +1708,3043369,"TERMINAL",0,0,"1159",,terminal_output +1709,3044409,"TERMINAL",0,0,"22640",,terminal_output +1710,3045499,"TERMINAL",0,0,"3371",,terminal_output +1711,3046509,"TERMINAL",0,0,"4482",,terminal_output +1712,3047555,"TERMINAL",0,0,"5593",,terminal_output +1713,3048610,"TERMINAL",0,0,"66404",,terminal_output +1714,3049631,"TERMINAL",0,0,"7715",,terminal_output +1715,3050672,"TERMINAL",0,0,"8826",,terminal_output +1716,3051725,"TERMINAL",0,0,"94048",,terminal_output +1717,3052775,"TERMINAL",0,0,"31159",,terminal_output +1718,3053821,"TERMINAL",0,0,"22650",,terminal_output +1719,3054871,"TERMINAL",0,0,"3371",,terminal_output +1720,3055916,"TERMINAL",0,0,"4482",,terminal_output +1721,3056964,"TERMINAL",0,0,"5593",,terminal_output +1722,3058021,"TERMINAL",0,0,"66504",,terminal_output +1723,3059066,"TERMINAL",0,0,"7715",,terminal_output +1724,3060113,"TERMINAL",0,0,"8826",,terminal_output +1725,3061165,"TERMINAL",0,0,"9937",,terminal_output +1726,3062241,"TERMINAL",0,0,"405048",,terminal_output +1727,3063254,"TERMINAL",0,0,"1159",,terminal_output +1728,3064343,"TERMINAL",0,0,"2265:00",,terminal_output +1729,3065348,"TERMINAL",0,0,"3371",,terminal_output +1730,3066430,"TERMINAL",0,0,"4482",,terminal_output +1731,3067448,"TERMINAL",0,0,"5593",,terminal_output +1732,3068526,"TERMINAL",0,0,"665:004",,terminal_output +1733,3069570,"TERMINAL",0,0,"7715",,terminal_output +1734,3070606,"TERMINAL",0,0,"8826",,terminal_output +1735,3071658,"TERMINAL",0,0,"9937",,terminal_output +1736,3072709,"TERMINAL",0,0,"505:0048",,terminal_output +1737,3073762,"TERMINAL",0,0,"12610",,terminal_output +1738,3074804,"TERMINAL",0,0,"3371",,terminal_output +1739,3075855,"TERMINAL",0,0,"4482",,terminal_output +1740,3076906,"TERMINAL",0,0,"5593",,terminal_output +1741,3077954,"TERMINAL",0,0,"66104",,terminal_output +1742,3079009,"TERMINAL",0,0,"7715",,terminal_output +1743,3080058,"TERMINAL",0,0,"8826",,terminal_output +1744,3081105,"TERMINAL",0,0,"9937",,terminal_output +1745,3082145,"TERMINAL",0,0,"3:001048",,terminal_output +1746,3083196,"TERMINAL",0,0,"1159",,terminal_output +1747,3084244,"TERMINAL",0,0,"22620",,terminal_output +1748,3085299,"TERMINAL",0,0,"3371",,terminal_output +1749,3086343,"TERMINAL",0,0,"4482",,terminal_output +1750,3087389,"TERMINAL",0,0,"5593",,terminal_output +1751,3088438,"TERMINAL",0,0,"66204",,terminal_output +1752,3089486,"TERMINAL",0,0,"7715",,terminal_output +1753,3090540,"TERMINAL",0,0,"8826",,terminal_output +1754,3091593,"TERMINAL",0,0,"9937",,terminal_output +1755,3092633,"TERMINAL",0,0,"102048",,terminal_output +1756,3093689,"TERMINAL",0,0,"1159",,terminal_output +1757,3094740,"TERMINAL",0,0,"23731",,terminal_output +1758,3095772,"TERMINAL",0,0,"4482",,terminal_output +1759,3096821,"TERMINAL",0,0,"5593",,terminal_output +1760,3097874,"TERMINAL",0,0,"66304",,terminal_output +1761,3098914,"TERMINAL",0,0,"7715",,terminal_output +1762,3099958,"TERMINAL",0,0,"8826",,terminal_output +1763,3101014,"TERMINAL",0,0,"9937",,terminal_output +1764,3102061,"TERMINAL",0,0,"203048",,terminal_output +1765,3103112,"TERMINAL",0,0,"1159",,terminal_output +1766,3104205,"TERMINAL",0,0,"22640",,terminal_output +1767,3105213,"TERMINAL",0,0,"3371",,terminal_output +1768,3106263,"TERMINAL",0,0,"4482",,terminal_output +1769,3107300,"TERMINAL",0,0,"5593",,terminal_output +1770,3108346,"TERMINAL",0,0,"66404",,terminal_output +1771,3109401,"TERMINAL",0,0,"7715",,terminal_output +1772,3110451,"TERMINAL",0,0,"8826",,terminal_output +1773,3111524,"TERMINAL",0,0,"9937",,terminal_output +1774,3112552,"TERMINAL",0,0,"304048",,terminal_output +1775,3113598,"TERMINAL",0,0,"1159",,terminal_output +1776,3114642,"TERMINAL",0,0,"22650",,terminal_output +1777,3115697,"TERMINAL",0,0,"3371",,terminal_output +1778,3116751,"TERMINAL",0,0,"4593",,terminal_output +1779,3117800,"TERMINAL",0,0,"66504",,terminal_output +1780,3118841,"TERMINAL",0,0,"7715",,terminal_output +1781,3120455,"TERMINAL",0,0,"8826",,terminal_output +1782,3121509,"TERMINAL",0,0,"9937",,terminal_output +1783,3122555,"TERMINAL",0,0,"405048",,terminal_output +1784,3123610,"TERMINAL",0,0,"1159",,terminal_output +1785,3124655,"TERMINAL",0,0,"2266:00",,terminal_output +1786,3125709,"TERMINAL",0,0,"3371",,terminal_output +1787,3126758,"TERMINAL",0,0,"4593",,terminal_output +1788,3127800,"TERMINAL",0,0,"666:004",,terminal_output +1789,3128852,"TERMINAL",0,0,"7715",,terminal_output +1790,3129900,"TERMINAL",0,0,"8826",,terminal_output +1791,3130947,"TERMINAL",0,0,"9937",,terminal_output +1792,3132000,"TERMINAL",0,0,"506:0048",,terminal_output +1793,3133055,"TERMINAL",0,0,"1159",,terminal_output +1794,3134138,"TERMINAL",0,0,"22610",,terminal_output +1795,3135149,"TERMINAL",0,0,"3371",,terminal_output +1796,3136201,"TERMINAL",0,0,"4482",,terminal_output +1797,3137250,"TERMINAL",0,0,"5593",,terminal_output +1798,3138282,"TERMINAL",0,0,"66104",,terminal_output +1799,3139339,"TERMINAL",0,0,"7715",,terminal_output +1800,3140354,"TERMINAL",0,0,"8826",,terminal_output +1801,3141442,"TERMINAL",0,0,"9937",,terminal_output +1802,3142452,"TERMINAL",0,0,"4:001048",,terminal_output +1803,3143505,"TERMINAL",0,0,"1159",,terminal_output +1804,3144550,"TERMINAL",0,0,"22620",,terminal_output +1805,3145597,"TERMINAL",0,0,"3371",,terminal_output +1806,3146644,"TERMINAL",0,0,"4482",,terminal_output +1807,3147681,"TERMINAL",0,0,"5593",,terminal_output +1808,3148730,"TERMINAL",0,0,"67215",,terminal_output +1809,3149778,"TERMINAL",0,0,"8826",,terminal_output +1810,3150826,"TERMINAL",0,0,"9937",,terminal_output +1811,3151881,"TERMINAL",0,0,"102048",,terminal_output +1812,3152936,"TERMINAL",0,0,"1159",,terminal_output +1813,3153986,"TERMINAL",0,0,"22630",,terminal_output +1814,3155043,"TERMINAL",0,0,"3371",,terminal_output +1815,3156088,"TERMINAL",0,0,"4482",,terminal_output +1816,3157142,"TERMINAL",0,0,"5593",,terminal_output +1817,3158184,"TERMINAL",0,0,"66304",,terminal_output +1818,3159235,"TERMINAL",0,0,"7715",,terminal_output +1819,3160263,"TERMINAL",0,0,"8826",,terminal_output +1820,3161304,"TERMINAL",0,0,"9937",,terminal_output +1821,3162355,"TERMINAL",0,0,"203048",,terminal_output +1822,3163394,"TERMINAL",0,0,"1159",,terminal_output +1823,3164455,"TERMINAL",0,0,"22640",,terminal_output +1824,3165534,"TERMINAL",0,0,"3371",,terminal_output +1825,3166558,"TERMINAL",0,0,"4482",,terminal_output +1826,3167617,"TERMINAL",0,0,"5593",,terminal_output +1827,3168644,"TERMINAL",0,0,"66404",,terminal_output +1828,3169696,"TERMINAL",0,0,"7715",,terminal_output +1829,3170745,"TERMINAL",0,0,"8937",,terminal_output +1830,3171795,"TERMINAL",0,0,"304048",,terminal_output +1831,3172844,"TERMINAL",0,0,"1159",,terminal_output +1832,3173895,"TERMINAL",0,0,"22650",,terminal_output +1833,3174937,"TERMINAL",0,0,"3371",,terminal_output +1834,3175984,"TERMINAL",0,0,"4482",,terminal_output +1835,3177035,"TERMINAL",0,0,"5593",,terminal_output +1836,3178088,"TERMINAL",0,0,"66504",,terminal_output +1837,3179140,"TERMINAL",0,0,"7715",,terminal_output +1838,3180192,"TERMINAL",0,0,"8826",,terminal_output +1839,3181240,"TERMINAL",0,0,"9937",,terminal_output +1840,3182291,"TERMINAL",0,0,"405048",,terminal_output +1841,3183344,"TERMINAL",0,0,"1159",,terminal_output +1842,3184382,"TERMINAL",0,0,"2267:00",,terminal_output +1843,3185422,"TERMINAL",0,0,"3371",,terminal_output +1844,3186471,"TERMINAL",0,0,"4482",,terminal_output +1845,3187544,"TERMINAL",0,0,"5593",,terminal_output +1846,3188565,"TERMINAL",0,0,"667:004",,terminal_output +1847,3189597,"TERMINAL",0,0,"7715",,terminal_output +1848,3190650,"TERMINAL",0,0,"8826",,terminal_output +1849,3191681,"TERMINAL",0,0,"9937",,terminal_output +1850,3192737,"TERMINAL",0,0,"507:0159",,terminal_output +1851,3193789,"TERMINAL",0,0,"22610",,terminal_output +1852,3194832,"TERMINAL",0,0,"3371",,terminal_output +1853,3195884,"TERMINAL",0,0,"4482",,terminal_output +1854,3196924,"TERMINAL",0,0,"5593",,terminal_output +1855,3197976,"TERMINAL",0,0,"66104",,terminal_output +1856,3199032,"TERMINAL",0,0,"7715",,terminal_output +1857,3200083,"TERMINAL",0,0,"8826",,terminal_output +1858,3201134,"TERMINAL",0,0,"9937",,terminal_output +1859,3202185,"TERMINAL",0,0,"5:001048",,terminal_output +1860,3203243,"TERMINAL",0,0,"1159",,terminal_output +1861,3204288,"TERMINAL",0,0,"22620",,terminal_output +1862,3205336,"TERMINAL",0,0,"3371",,terminal_output +1863,3206386,"TERMINAL",0,0,"4482",,terminal_output +1864,3207442,"TERMINAL",0,0,"5593",,terminal_output +1865,3208481,"TERMINAL",0,0,"66204",,terminal_output +1866,3209550,"TERMINAL",0,0,"7715",,terminal_output +1867,3210577,"TERMINAL",0,0,"8826",,terminal_output +1868,3211622,"TERMINAL",0,0,"9937",,terminal_output +1869,3212665,"TERMINAL",0,0,"102048",,terminal_output +1870,3213716,"TERMINAL",0,0,"12630",,terminal_output +1871,3214775,"TERMINAL",0,0,"3371",,terminal_output +1872,3215814,"TERMINAL",0,0,"4482",,terminal_output +1873,3216866,"TERMINAL",0,0,"5593",,terminal_output +1874,3217913,"TERMINAL",0,0,"66304",,terminal_output +1875,3218957,"TERMINAL",0,0,"7715",,terminal_output +1876,3220031,"TERMINAL",0,0,"8826",,terminal_output +1877,3221072,"TERMINAL",0,0,"9937",,terminal_output +1878,3222134,"TERMINAL",0,0,"203048",,terminal_output +1879,3223185,"TERMINAL",0,0,"1159",,terminal_output +1880,3224218,"TERMINAL",0,0,"22640",,terminal_output +1881,3225262,"TERMINAL",0,0,"3371",,terminal_output +1882,3226557,"TERMINAL",0,0,"4482",,terminal_output +1883,3227606,"TERMINAL",0,0,"5593",,terminal_output +1884,3228654,"TERMINAL",0,0,"66404",,terminal_output +1885,3229702,"TERMINAL",0,0,"7715",,terminal_output +1886,3230786,"TERMINAL",0,0,"8937",,terminal_output +1887,3231848,"TERMINAL",0,0,"304048",,terminal_output +1888,3232895,"TERMINAL",0,0,"1159",,terminal_output +1889,3233946,"TERMINAL",0,0,"22650",,terminal_output +1890,3234986,"TERMINAL",0,0,"3371",,terminal_output +1891,3236031,"TERMINAL",0,0,"4482",,terminal_output +1892,3237085,"TERMINAL",0,0,"5593",,terminal_output +1893,3238185,"TERMINAL",0,0,"66504",,terminal_output +1894,3239236,"TERMINAL",0,0,"7715",,terminal_output +1895,3240280,"TERMINAL",0,0,"8826",,terminal_output +1896,3241610,"TERMINAL",0,0,"9937",,terminal_output +1897,3242665,"TERMINAL",0,0,"405048",,terminal_output +1898,3243698,"TERMINAL",0,0,"1159",,terminal_output +1899,3244749,"TERMINAL",0,0,"2378:01",,terminal_output +1900,3245796,"TERMINAL",0,0,"4482",,terminal_output +1901,3246849,"TERMINAL",0,0,"5593",,terminal_output +1902,3247909,"TERMINAL",0,0,"668:004",,terminal_output +1903,3248958,"TERMINAL",0,0,"7715",,terminal_output +1904,3250002,"TERMINAL",0,0,"8826",,terminal_output +1905,3251069,"TERMINAL",0,0,"9937",,terminal_output +1906,3252101,"TERMINAL",0,0,"508:0048",,terminal_output +1907,3253150,"TERMINAL",0,0,"1159",,terminal_output +1908,3254202,"TERMINAL",0,0,"22610",,terminal_output +1909,3255253,"TERMINAL",0,0,"3371",,terminal_output +1910,3256301,"TERMINAL",0,0,"4482",,terminal_output +1911,3257354,"TERMINAL",0,0,"5593",,terminal_output +1912,3258411,"TERMINAL",0,0,"66104",,terminal_output +1913,3259450,"TERMINAL",0,0,"7715",,terminal_output +1914,3260496,"TERMINAL",0,0,"8826",,terminal_output +1915,3261551,"TERMINAL",0,0,"9937",,terminal_output +1916,3262595,"TERMINAL",0,0,"6:001048",,terminal_output +1917,3263636,"TERMINAL",0,0,"1159",,terminal_output +1918,3264691,"TERMINAL",0,0,"22620",,terminal_output +1919,3265732,"TERMINAL",0,0,"3482",,terminal_output +1920,3266778,"TERMINAL",0,0,"5593",,terminal_output +1921,3267870,"TERMINAL",0,0,"66204",,terminal_output +1922,3268882,"TERMINAL",0,0,"7715",,terminal_output +1923,3269925,"TERMINAL",0,0,"8826",,terminal_output +1924,3270972,"TERMINAL",0,0,"9937",,terminal_output +1925,3272018,"TERMINAL",0,0,"102048",,terminal_output +1926,3273060,"TERMINAL",0,0,"1159",,terminal_output +1927,3274103,"TERMINAL",0,0,"22630",,terminal_output +1928,3275153,"TERMINAL",0,0,"3371",,terminal_output +1929,3276225,"TERMINAL",0,0,"4482",,terminal_output +1930,3277255,"TERMINAL",0,0,"5593",,terminal_output +1931,3278307,"TERMINAL",0,0,"66304",,terminal_output +1932,3279361,"TERMINAL",0,0,"7715",,terminal_output +1933,3280410,"TERMINAL",0,0,"8826",,terminal_output +1934,3281454,"TERMINAL",0,0,"9937",,terminal_output +1935,3282502,"TERMINAL",0,0,"203048",,terminal_output +1936,3283639,"TERMINAL",0,0,"1159",,terminal_output +1937,3284587,"TERMINAL",0,0,"22640",,terminal_output +1938,3285633,"TERMINAL",0,0,"3371",,terminal_output +1939,3286680,"TERMINAL",0,0,"4482",,terminal_output +1940,3287724,"TERMINAL",0,0,"56404",,terminal_output +1941,3288773,"TERMINAL",0,0,"7715",,terminal_output +1942,3289822,"TERMINAL",0,0,"8826",,terminal_output +1943,3290875,"TERMINAL",0,0,"9937",,terminal_output +1944,3291923,"TERMINAL",0,0,"304048",,terminal_output +1945,3292972,"TERMINAL",0,0,"1159",,terminal_output +1946,3294013,"TERMINAL",0,0,"22650",,terminal_output +1947,3295064,"TERMINAL",0,0,"3371",,terminal_output +1948,3296101,"TERMINAL",0,0,"4482",,terminal_output +1949,3297159,"TERMINAL",0,0,"5593",,terminal_output +1950,3298201,"TERMINAL",0,0,"66504",,terminal_output +1951,3299250,"TERMINAL",0,0,"7715",,terminal_output +1952,3300304,"TERMINAL",0,0,"8826",,terminal_output +1953,3301347,"TERMINAL",0,0,"9937",,terminal_output +1954,3302400,"TERMINAL",0,0,"405048",,terminal_output +1955,3303452,"TERMINAL",0,0,"1159",,terminal_output +1956,3304502,"TERMINAL",0,0,"2269:00",,terminal_output +1957,3305586,"TERMINAL",0,0,"3371",,terminal_output +1958,3306598,"TERMINAL",0,0,"4482",,terminal_output +1959,3307641,"TERMINAL",0,0,"5593",,terminal_output +1960,3308684,"TERMINAL",0,0,"669:004",,terminal_output +1961,3309729,"TERMINAL",0,0,"7826",,terminal_output +1962,3310784,"TERMINAL",0,0,"9937",,terminal_output +1963,3311838,"TERMINAL",0,0,"509:0048",,terminal_output +1964,3312878,"TERMINAL",0,0,"1159",,terminal_output +1965,3313932,"TERMINAL",0,0,"22610",,terminal_output +1966,3314979,"TERMINAL",0,0,"3371",,terminal_output +1967,3316023,"TERMINAL",0,0,"4482",,terminal_output +1968,3317070,"TERMINAL",0,0,"5593",,terminal_output +1969,3318124,"TERMINAL",0,0,"66104",,terminal_output +1970,3319175,"TERMINAL",0,0,"7715",,terminal_output +1971,3320228,"TERMINAL",0,0,"8826",,terminal_output +1972,3321290,"TERMINAL",0,0,"9937",,terminal_output +1973,3322330,"TERMINAL",0,0,"7:001048",,terminal_output +1974,3323383,"TERMINAL",0,0,"1159",,terminal_output +1975,3324430,"TERMINAL",0,0,"22620",,terminal_output +1976,3325476,"TERMINAL",0,0,"3371",,terminal_output +1977,3326528,"TERMINAL",0,0,"4482",,terminal_output +1978,3327617,"TERMINAL",0,0,"5593",,terminal_output +1979,3328635,"TERMINAL",0,0,"66204",,terminal_output +1980,3329729,"TERMINAL",0,0,"7715",,terminal_output +1981,3330729,"TERMINAL",0,0,"8937",,terminal_output +1982,3331775,"TERMINAL",0,0,"102048",,terminal_output +1983,3332827,"TERMINAL",0,0,"1159",,terminal_output +1984,3333901,"TERMINAL",0,0,"22630",,terminal_output +1985,3334951,"TERMINAL",0,0,"3371",,terminal_output +1986,3335998,"TERMINAL",0,0,"4482",,terminal_output +1987,3337075,"TERMINAL",0,0,"5593",,terminal_output +1988,3338105,"TERMINAL",0,0,"66304",,terminal_output +1989,3339140,"TERMINAL",0,0,"7715",,terminal_output +1990,3340190,"TERMINAL",0,0,"8826",,terminal_output +1991,3341250,"TERMINAL",0,0,"9937",,terminal_output +1992,3342295,"TERMINAL",0,0,"203048",,terminal_output +1993,3343343,"TERMINAL",0,0,"1159",,terminal_output +1994,3344396,"TERMINAL",0,0,"22640",,terminal_output +1995,3345444,"TERMINAL",0,0,"3371",,terminal_output +1996,3346496,"TERMINAL",0,0,"4482",,terminal_output +1997,3347536,"TERMINAL",0,0,"5593",,terminal_output +1998,3348606,"TERMINAL",0,0,"66404",,terminal_output +1999,3349646,"TERMINAL",0,0,"7715",,terminal_output +2000,3350676,"TERMINAL",0,0,"8826",,terminal_output +2001,3351731,"TERMINAL",0,0,"94048",,terminal_output +2002,3352778,"TERMINAL",0,0,"31159",,terminal_output +2003,3353829,"TERMINAL",0,0,"22650",,terminal_output +2004,3354926,"TERMINAL",0,0,"3371",,terminal_output +2005,3355973,"TERMINAL",0,0,"4482",,terminal_output +2006,3357015,"TERMINAL",0,0,"5593",,terminal_output +2007,3358058,"TERMINAL",0,0,"66504",,terminal_output +2008,3359118,"TERMINAL",0,0,"7715",,terminal_output +2009,3360163,"TERMINAL",0,0,"8826",,terminal_output +2010,3361195,"TERMINAL",0,0,"9937",,terminal_output +2011,3362722,"TERMINAL",0,0,"405159",,terminal_output +2012,3363777,"TERMINAL",0,0,"22610:00",,terminal_output +2013,3364843,"TERMINAL",0,0,"3371",,terminal_output +2014,3365898,"TERMINAL",0,0,"4482",,terminal_output +2015,3366916,"TERMINAL",0,0,"5593",,terminal_output +2016,3367973,"TERMINAL",0,0,"6610:004",,terminal_output +2017,3368234,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",0,0,"",shellscript,tab +2018,3369022,"TERMINAL",0,0,"7715",,terminal_output +2019,3370068,"TERMINAL",0,0,"8826",,terminal_output +2020,3371120,"TERMINAL",0,0,"9937",,terminal_output +2021,3372208,"TERMINAL",0,0,"5010:0048",,terminal_output +2022,3373225,"TERMINAL",0,0,"1159",,terminal_output +2023,3374260,"TERMINAL",0,0,"22610",,terminal_output +2024,3375307,"TERMINAL",0,0,"3371",,terminal_output +2025,3375678,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",0,0,"",shellscript,tab +2026,3376560,"TERMINAL",0,0,"4482",,terminal_output +2027,3377398,"TERMINAL",0,0,"5593",,terminal_output +2028,3378588,"TERMINAL",0,0,"66104",,terminal_output +2029,3379382,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",0,0,"",shellscript,tab +2030,3379499,"TERMINAL",0,0,"7715",,terminal_output +2031,3380525,"TERMINAL",0,0,"8826",,terminal_output +2032,3381602,"TERMINAL",0,0,"9937",,terminal_output +2033,3381899,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",0,0,"",shellscript,tab +2034,3382622,"TERMINAL",0,0,"8:001048",,terminal_output +2035,3383675,"TERMINAL",0,0,"1159",,terminal_output +2036,3384728,"TERMINAL",0,0,"23721",,terminal_output +2037,3385773,"TERMINAL",0,0,"4482",,terminal_output +2038,3386822,"TERMINAL",0,0,"5593",,terminal_output +2039,3387895,"TERMINAL",0,0,"66204",,terminal_output +2040,3388912,"TERMINAL",0,0,"7715",,terminal_output +2041,3389959,"TERMINAL",0,0,"8826",,terminal_output +2042,3391028,"TERMINAL",0,0,"9937",,terminal_output +2043,3392069,"TERMINAL",0,0,"102048",,terminal_output +2044,3393114,"TERMINAL",0,0,"1159",,terminal_output +2045,3394192,"TERMINAL",0,0,"22630",,terminal_output +2046,3395212,"TERMINAL",0,0,"3371",,terminal_output +2047,3396265,"TERMINAL",0,0,"4482",,terminal_output +2048,3397305,"TERMINAL",0,0,"5593",,terminal_output +2049,3398356,"TERMINAL",0,0,"66304",,terminal_output +2050,3399397,"TERMINAL",0,0,"7715",,terminal_output +2051,3400440,"TERMINAL",0,0,"8826",,terminal_output +2052,3401490,"TERMINAL",0,0,"9937",,terminal_output +2053,3402528,"TERMINAL",0,0,"203048",,terminal_output +2054,3403610,"TERMINAL",0,0,"1159",,terminal_output +2055,3404635,"TERMINAL",0,0,"22640",,terminal_output +2056,3405691,"TERMINAL",0,0,"3371",,terminal_output +2057,3406731,"TERMINAL",0,0,"4593",,terminal_output +2058,3407782,"TERMINAL",0,0,"66404",,terminal_output +2059,3408823,"TERMINAL",0,0,"7715",,terminal_output +2060,3409878,"TERMINAL",0,0,"8826",,terminal_output +2061,3410917,"TERMINAL",0,0,"9937",,terminal_output +2062,3411977,"TERMINAL",0,0,"304048",,terminal_output +2063,3413019,"TERMINAL",0,0,"1159",,terminal_output +2064,3414072,"TERMINAL",0,0,"22650",,terminal_output +2065,3415122,"TERMINAL",0,0,"3371",,terminal_output +2066,3416169,"TERMINAL",0,0,"4482",,terminal_output +2067,3417200,"TERMINAL",0,0,"5593",,terminal_output +2068,3418296,"TERMINAL",0,0,"66504",,terminal_output +2069,3419295,"TERMINAL",0,0,"7715",,terminal_output +2070,3420348,"TERMINAL",0,0,"8826",,terminal_output +2071,3421397,"TERMINAL",0,0,"9937",,terminal_output +2072,3422459,"TERMINAL",0,0,"405048",,terminal_output +2073,3423492,"TERMINAL",0,0,"1159",,terminal_output +2074,3424543,"TERMINAL",0,0,"2261:00",,terminal_output +2075,3425634,"TERMINAL",0,0,"3371",,terminal_output +2076,3426632,"TERMINAL",0,0,"4482",,terminal_output +2077,3427691,"TERMINAL",0,0,"5593",,terminal_output +2078,3428737,"TERMINAL",0,0,"671:015",,terminal_output +2079,3429782,"TERMINAL",0,0,"8826",,terminal_output +2080,3430829,"TERMINAL",0,0,"9937",,terminal_output +2081,3431882,"TERMINAL",0,0,"501:0048",,terminal_output +2082,3432932,"TERMINAL",0,0,"1159",,terminal_output +2083,3433989,"TERMINAL",0,0,"22610",,terminal_output +2084,3435036,"TERMINAL",0,0,"3371",,terminal_output +2085,3436083,"TERMINAL",0,0,"4482",,terminal_output +2086,3437133,"TERMINAL",0,0,"5593",,terminal_output +2087,3438209,"TERMINAL",0,0,"66104",,terminal_output +2088,3439241,"TERMINAL",0,0,"7715",,terminal_output +2089,3440292,"TERMINAL",0,0,"8826",,terminal_output +2090,3441340,"TERMINAL",0,0,"9937",,terminal_output +2091,3442397,"TERMINAL",0,0,"9:001048",,terminal_output +2092,3443439,"TERMINAL",0,0,"1159",,terminal_output +2093,3444490,"TERMINAL",0,0,"22620",,terminal_output +2094,3445521,"TERMINAL",0,0,"3371",,terminal_output +2095,3446581,"TERMINAL",0,0,"4482",,terminal_output +2096,3447662,"TERMINAL",0,0,"5593",,terminal_output +2097,3448675,"TERMINAL",0,0,"66204",,terminal_output +2098,3449729,"TERMINAL",0,0,"7826",,terminal_output +2099,3450777,"TERMINAL",0,0,"9937",,terminal_output +2100,3451819,"TERMINAL",0,0,"102048",,terminal_output +2101,3452873,"TERMINAL",0,0,"1159",,terminal_output +2102,3453968,"TERMINAL",0,0,"22630",,terminal_output +2103,3454972,"TERMINAL",0,0,"3371",,terminal_output +2104,3456035,"TERMINAL",0,0,"4482",,terminal_output +2105,3456698,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-base.sbatch",0,0,"",shellscript,tab +2106,3457177,"TERMINAL",0,0,"5593",,terminal_output +2107,3458121,"TERMINAL",0,0,"66304",,terminal_output +2108,3459180,"TERMINAL",0,0,"7715",,terminal_output +2109,3460241,"TERMINAL",0,0,"8826",,terminal_output +2110,3461181,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",0,0,"",shellscript,tab +2111,3461304,"TERMINAL",0,0,"9937",,terminal_output +2112,3462331,"TERMINAL",0,0,"203048",,terminal_output +2113,3463454,"TERMINAL",0,0,"1159",,terminal_output +2114,3464288,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-400M.sbatch",0,0,"",shellscript,tab +2115,3464639,"TERMINAL",0,0,"22640",,terminal_output +2116,3465489,"TERMINAL",0,0,"3371",,terminal_output +2117,3466526,"TERMINAL",0,0,"4482",,terminal_output +2118,3467580,"TERMINAL",0,0,"5593",,terminal_output +2119,3468657,"TERMINAL",0,0,"66404",,terminal_output +2120,3469683,"TERMINAL",0,0,"7715",,terminal_output +2121,3470493,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-311M.sbatch",0,0,"",shellscript,tab +2122,3470746,"TERMINAL",0,0,"8937",,terminal_output +2123,3471790,"TERMINAL",0,0,"304048",,terminal_output +2124,3472869,"TERMINAL",0,0,"1159",,terminal_output +2125,3473890,"TERMINAL",0,0,"22650",,terminal_output +2126,3474940,"TERMINAL",0,0,"3371",,terminal_output +2127,3475994,"TERMINAL",0,0,"4482",,terminal_output +2128,3477062,"TERMINAL",0,0,"5593",,terminal_output +2129,3478123,"TERMINAL",0,0,"66504",,terminal_output +2130,3479149,"TERMINAL",0,0,"7715",,terminal_output +2131,3480194,"TERMINAL",0,0,"8826",,terminal_output +2132,3481252,"TERMINAL",0,0,"9937",,terminal_output +2133,3482299,"TERMINAL",0,0,"405048",,terminal_output +2134,3483788,"TERMINAL",0,0,"1262:00",,terminal_output +2135,3484822,"TERMINAL",0,0,"3371",,terminal_output +2136,3485870,"TERMINAL",0,0,"4482",,terminal_output +2137,3486926,"TERMINAL",0,0,"5593",,terminal_output +2138,3487970,"TERMINAL",0,0,"662:004",,terminal_output +2139,3489024,"TERMINAL",0,0,"7715",,terminal_output +2140,3490070,"TERMINAL",0,0,"8826",,terminal_output +2141,3491112,"TERMINAL",0,0,"9937",,terminal_output +2142,3492152,"TERMINAL",0,0,"502:0048",,terminal_output +2143,3493200,"TERMINAL",0,0,"1159",,terminal_output +2144,3494259,"TERMINAL",0,0,"22610",,terminal_output +2145,3495292,"TERMINAL",0,0,"3371",,terminal_output +2146,3496331,"TERMINAL",0,0,"4482",,terminal_output +2147,3497381,"TERMINAL",0,0,"5593",,terminal_output +2148,3498438,"TERMINAL",0,0,"66104",,terminal_output +2149,3499491,"TERMINAL",0,0,"7715",,terminal_output +2150,3500547,"TERMINAL",0,0,"8826",,terminal_output +2151,3501579,"TERMINAL",0,0,"9937",,terminal_output +2152,3502660,"TERMINAL",0,0,"40:001048",,terminal_output +2153,3503686,"TERMINAL",0,0,"1159",,terminal_output +2154,3504738,"TERMINAL",0,0,"23721",,terminal_output +2155,3505792,"TERMINAL",0,0,"4482",,terminal_output +2156,3506842,"TERMINAL",0,0,"5593",,terminal_output +2157,3507885,"TERMINAL",0,0,"66204",,terminal_output +2158,3508943,"TERMINAL",0,0,"7715",,terminal_output +2159,3509983,"TERMINAL",0,0,"8826",,terminal_output +2160,3511036,"TERMINAL",0,0,"9937",,terminal_output +2161,3512086,"TERMINAL",0,0,"102048",,terminal_output +2162,3513130,"TERMINAL",0,0,"1159",,terminal_output +2163,3514184,"TERMINAL",0,0,"22630",,terminal_output +2164,3515246,"TERMINAL",0,0,"3371",,terminal_output +2165,3516298,"TERMINAL",0,0,"4482",,terminal_output +2166,3517342,"TERMINAL",0,0,"5593",,terminal_output +2167,3518393,"TERMINAL",0,0,"66304",,terminal_output +2168,3519441,"TERMINAL",0,0,"7715",,terminal_output +2169,3520491,"TERMINAL",0,0,"8826",,terminal_output +2170,3521538,"TERMINAL",0,0,"9937",,terminal_output +2171,3522586,"TERMINAL",0,0,"203048",,terminal_output +2172,3523666,"TERMINAL",0,0,"1159",,terminal_output +2173,3524685,"TERMINAL",0,0,"22640",,terminal_output +2174,3525720,"TERMINAL",0,0,"3482",,terminal_output +2175,3526768,"TERMINAL",0,0,"5593",,terminal_output +2176,3527817,"TERMINAL",0,0,"66404",,terminal_output +2177,3528872,"TERMINAL",0,0,"7715",,terminal_output +2178,3529919,"TERMINAL",0,0,"8826",,terminal_output +2179,3530958,"TERMINAL",0,0,"9937",,terminal_output +2180,3532008,"TERMINAL",0,0,"304048",,terminal_output +2181,3533061,"TERMINAL",0,0,"1159",,terminal_output +2182,3534134,"TERMINAL",0,0,"22650",,terminal_output +2183,3535145,"TERMINAL",0,0,"3371",,terminal_output +2184,3536191,"TERMINAL",0,0,"4482",,terminal_output +2185,3537244,"TERMINAL",0,0,"5593",,terminal_output +2186,3538298,"TERMINAL",0,0,"66504",,terminal_output +2187,3539366,"TERMINAL",0,0,"7715",,terminal_output +2188,3540389,"TERMINAL",0,0,"8826",,terminal_output +2189,3541461,"TERMINAL",0,0,"9937",,terminal_output +2190,3542496,"TERMINAL",0,0,"405048",,terminal_output +2191,3543542,"TERMINAL",0,0,"1159",,terminal_output +2192,3544586,"TERMINAL",0,0,"2263:00",,terminal_output +2193,3545654,"TERMINAL",0,0,"3371",,terminal_output +2194,3546690,"TERMINAL",0,0,"4482",,terminal_output +2195,3547734,"TERMINAL",0,0,"563:004",,terminal_output +2196,3548790,"TERMINAL",0,0,"7715",,terminal_output +2197,3549840,"TERMINAL",0,0,"8826",,terminal_output +2198,3550899,"TERMINAL",0,0,"9937",,terminal_output +2199,3551947,"TERMINAL",0,0,"503:0048",,terminal_output +2200,3553025,"TERMINAL",0,0,"1159",,terminal_output +2201,3554070,"TERMINAL",0,0,"22610",,terminal_output +2202,3555120,"TERMINAL",0,0,"3371",,terminal_output +2203,3556170,"TERMINAL",0,0,"4482",,terminal_output +2204,3557219,"TERMINAL",0,0,"5593",,terminal_output +2205,3558269,"TERMINAL",0,0,"66104",,terminal_output +2206,3559322,"TERMINAL",0,0,"7715",,terminal_output +2207,3560362,"TERMINAL",0,0,"8826",,terminal_output +2208,3561468,"TERMINAL",0,0,"9937",,terminal_output +2209,3562464,"TERMINAL",0,0,"1:001048",,terminal_output +2210,3563522,"TERMINAL",0,0,"1159",,terminal_output +2211,3564561,"TERMINAL",0,0,"22620",,terminal_output +2212,3565625,"TERMINAL",0,0,"3371",,terminal_output +2213,3566705,"TERMINAL",0,0,"4482",,terminal_output +2214,3567721,"TERMINAL",0,0,"56204",,terminal_output +2215,3568767,"TERMINAL",0,0,"7715",,terminal_output +2216,3569814,"TERMINAL",0,0,"8826",,terminal_output +2217,3570852,"TERMINAL",0,0,"9937",,terminal_output +2218,3571904,"TERMINAL",0,0,"102048",,terminal_output +2219,3572954,"TERMINAL",0,0,"1159",,terminal_output +2220,3574016,"TERMINAL",0,0,"22630",,terminal_output +2221,3575060,"TERMINAL",0,0,"3371",,terminal_output +2222,3576110,"TERMINAL",0,0,"4482",,terminal_output +2223,3577156,"TERMINAL",0,0,"5593",,terminal_output +2224,3578210,"TERMINAL",0,0,"66304",,terminal_output +2225,3579260,"TERMINAL",0,0,"7715",,terminal_output +2226,3580313,"TERMINAL",0,0,"8826",,terminal_output +2227,3581366,"TERMINAL",0,0,"9937",,terminal_output +2228,3582400,"TERMINAL",0,0,"203048",,terminal_output +2229,3583455,"TERMINAL",0,0,"1159",,terminal_output +2230,3584497,"TERMINAL",0,0,"22640",,terminal_output +2231,3585555,"TERMINAL",0,0,"3371",,terminal_output +2232,3586611,"TERMINAL",0,0,"4482",,terminal_output +2233,3587654,"TERMINAL",0,0,"5593",,terminal_output +2234,3588752,"TERMINAL",0,0,"66404",,terminal_output +2235,3589757,"TERMINAL",0,0,"7826",,terminal_output +2236,3590797,"TERMINAL",0,0,"9937",,terminal_output +2237,3591850,"TERMINAL",0,0,"304048",,terminal_output +2238,3592895,"TERMINAL",0,0,"1159",,terminal_output +2239,3593980,"TERMINAL",0,0,"22650",,terminal_output +2240,3594985,"TERMINAL",0,0,"3371",,terminal_output +2241,3596025,"TERMINAL",0,0,"4482",,terminal_output +2242,3597092,"TERMINAL",0,0,"5593",,terminal_output +2243,3598133,"TERMINAL",0,0,"66504",,terminal_output +2244,3599177,"TERMINAL",0,0,"7715",,terminal_output +2245,3600226,"TERMINAL",0,0,"8826",,terminal_output +2246,3601291,"TERMINAL",0,0,"9937",,terminal_output +2247,3602338,"TERMINAL",0,0,"405048",,terminal_output +2248,3603392,"TERMINAL",0,0,"1159",,terminal_output +2249,3604724,"TERMINAL",0,0,"2374:01",,terminal_output +2250,3605780,"TERMINAL",0,0,"4482",,terminal_output +2251,3606819,"TERMINAL",0,0,"5593",,terminal_output +2252,3607867,"TERMINAL",0,0,"664:004",,terminal_output +2253,3608905,"TERMINAL",0,0,"7715",,terminal_output +2254,3609946,"TERMINAL",0,0,"8826",,terminal_output +2255,3610999,"TERMINAL",0,0,"9937",,terminal_output +2256,3612048,"TERMINAL",0,0,"504:0048",,terminal_output +2257,3613096,"TERMINAL",0,0,"1159",,terminal_output +2258,3614147,"TERMINAL",0,0,"22610",,terminal_output +2259,3615173,"TERMINAL",0,0,"3371",,terminal_output +2260,3616234,"TERMINAL",0,0,"4482",,terminal_output +2261,3617268,"TERMINAL",0,0,"5593",,terminal_output +2262,3618326,"TERMINAL",0,0,"66104",,terminal_output +2263,3619364,"TERMINAL",0,0,"7715",,terminal_output +2264,3620412,"TERMINAL",0,0,"8826",,terminal_output +2265,3621464,"TERMINAL",0,0,"9937",,terminal_output +2266,3622499,"TERMINAL",0,0,"2:001048",,terminal_output +2267,3623546,"TERMINAL",0,0,"1159",,terminal_output +2268,3624594,"TERMINAL",0,0,"22620",,terminal_output +2269,3625640,"TERMINAL",0,0,"3371",,terminal_output +2270,3626696,"TERMINAL",0,0,"4482",,terminal_output +2271,3627742,"TERMINAL",0,0,"56204",,terminal_output +2272,3628797,"TERMINAL",0,0,"7715",,terminal_output +2273,3629862,"TERMINAL",0,0,"8826",,terminal_output +2274,3630907,"TERMINAL",0,0,"9937",,terminal_output +2275,3631950,"TERMINAL",0,0,"102048",,terminal_output +2276,3633013,"TERMINAL",0,0,"1159",,terminal_output +2277,3634050,"TERMINAL",0,0,"22630",,terminal_output +2278,3635109,"TERMINAL",0,0,"3371",,terminal_output +2279,3636152,"TERMINAL",0,0,"4482",,terminal_output +2280,3637201,"TERMINAL",0,0,"5593",,terminal_output +2281,3638278,"TERMINAL",0,0,"66304",,terminal_output +2282,3639299,"TERMINAL",0,0,"7715",,terminal_output +2283,3640351,"TERMINAL",0,0,"8826",,terminal_output +2284,3641388,"TERMINAL",0,0,"9937",,terminal_output +2285,3642445,"TERMINAL",0,0,"203048",,terminal_output +2286,3643481,"TERMINAL",0,0,"1159",,terminal_output +2287,3644512,"TERMINAL",0,0,"22640",,terminal_output +2288,3645572,"TERMINAL",0,0,"3371",,terminal_output +2289,3646623,"TERMINAL",0,0,"4482",,terminal_output +2290,3647664,"TERMINAL",0,0,"5593",,terminal_output +2291,3648728,"TERMINAL",0,0,"66404",,terminal_output +2292,3649755,"TERMINAL",0,0,"7826",,terminal_output +2293,3650806,"TERMINAL",0,0,"9937",,terminal_output +2294,3651858,"TERMINAL",0,0,"304048",,terminal_output +2295,3652905,"TERMINAL",0,0,"1159",,terminal_output +2296,3653959,"TERMINAL",0,0,"22650",,terminal_output +2297,3654996,"TERMINAL",0,0,"3371",,terminal_output +2298,3656038,"TERMINAL",0,0,"4482",,terminal_output +2299,3657087,"TERMINAL",0,0,"5593",,terminal_output +2300,3658124,"TERMINAL",0,0,"66504",,terminal_output +2301,3659169,"TERMINAL",0,0,"7715",,terminal_output +2302,3660214,"TERMINAL",0,0,"8826",,terminal_output +2303,3661264,"TERMINAL",0,0,"9937",,terminal_output +2304,3662315,"TERMINAL",0,0,"405048",,terminal_output +2305,3663347,"TERMINAL",0,0,"1159",,terminal_output +2306,3664404,"TERMINAL",0,0,"2265:00",,terminal_output +2307,3665450,"TERMINAL",0,0,"3371",,terminal_output +2308,3666508,"TERMINAL",0,0,"4482",,terminal_output +2309,3667914,"TERMINAL",0,0,"565:004",,terminal_output +2310,3668848,"TERMINAL",0,0,"7715",,terminal_output +2311,3669884,"TERMINAL",0,0,"8826",,terminal_output +2312,3670934,"TERMINAL",0,0,"9937",,terminal_output +2313,3671984,"TERMINAL",0,0,"505:0048",,terminal_output +2314,3673026,"TERMINAL",0,0,"1159",,terminal_output +2315,3674077,"TERMINAL",0,0,"22610",,terminal_output +2316,3675134,"TERMINAL",0,0,"3371",,terminal_output +2317,3676188,"TERMINAL",0,0,"4482",,terminal_output +2318,3677229,"TERMINAL",0,0,"5593",,terminal_output +2319,3678281,"TERMINAL",0,0,"66104",,terminal_output +2320,3679367,"TERMINAL",0,0,"7715",,terminal_output +2321,3680392,"TERMINAL",0,0,"8826",,terminal_output +2322,3681464,"TERMINAL",0,0,"9937",,terminal_output +2323,3682492,"TERMINAL",0,0,"3:001048",,terminal_output +2324,3683593,"TERMINAL",0,0,"1159",,terminal_output +2325,3684592,"TERMINAL",0,0,"22620",,terminal_output +2326,3685636,"TERMINAL",0,0,"3371",,terminal_output +2327,3686727,"TERMINAL",0,0,"4482",,terminal_output +2328,3687734,"TERMINAL",0,0,"56204",,terminal_output +2329,3688783,"TERMINAL",0,0,"7715",,terminal_output +2330,3689829,"TERMINAL",0,0,"8826",,terminal_output +2331,3690884,"TERMINAL",0,0,"9937",,terminal_output +2332,3691931,"TERMINAL",0,0,"102048",,terminal_output +2333,3692976,"TERMINAL",0,0,"1159",,terminal_output +2334,3694028,"TERMINAL",0,0,"22630",,terminal_output +2335,3695073,"TERMINAL",0,0,"3371",,terminal_output +2336,3696127,"TERMINAL",0,0,"4482",,terminal_output +2337,3697161,"TERMINAL",0,0,"5593",,terminal_output +2338,3698211,"TERMINAL",0,0,"66304",,terminal_output +2339,3699256,"TERMINAL",0,0,"7715",,terminal_output +2340,3700290,"TERMINAL",0,0,"8826",,terminal_output +2341,3701340,"TERMINAL",0,0,"9937",,terminal_output +2342,3702396,"TERMINAL",0,0,"203048",,terminal_output +2343,3703440,"TERMINAL",0,0,"1159",,terminal_output +2344,3704468,"TERMINAL",0,0,"22640",,terminal_output +2345,3705523,"TERMINAL",0,0,"3371",,terminal_output +2346,3706579,"TERMINAL",0,0,"4482",,terminal_output +2347,3707622,"TERMINAL",0,0,"5593",,terminal_output +2348,3708696,"TERMINAL",0,0,"66404",,terminal_output +2349,3709721,"TERMINAL",0,0,"7826",,terminal_output +2350,3710770,"TERMINAL",0,0,"9937",,terminal_output +2351,3711810,"TERMINAL",0,0,"304048",,terminal_output +2352,3712866,"TERMINAL",0,0,"1159",,terminal_output +2353,3713929,"TERMINAL",0,0,"22650",,terminal_output +2354,3714985,"TERMINAL",0,0,"3371",,terminal_output +2355,3716044,"TERMINAL",0,0,"4482",,terminal_output +2356,3717081,"TERMINAL",0,0,"5593",,terminal_output +2357,3718113,"TERMINAL",0,0,"66504",,terminal_output +2358,3719180,"TERMINAL",0,0,"7715",,terminal_output +2359,3720201,"TERMINAL",0,0,"8826",,terminal_output +2360,3721250,"TERMINAL",0,0,"9937",,terminal_output +2361,3722288,"TERMINAL",0,0,"405048",,terminal_output +2362,3723340,"TERMINAL",0,0,"1159",,terminal_output +2363,3724384,"TERMINAL",0,0,"2266:00",,terminal_output +2364,3725707,"TERMINAL",0,0,"3371",,terminal_output +2365,3726737,"TERMINAL",0,0,"4593",,terminal_output +2366,3727787,"TERMINAL",0,0,"666:004",,terminal_output +2367,3728841,"TERMINAL",0,0,"7715",,terminal_output +2368,3729881,"TERMINAL",0,0,"8826",,terminal_output +2369,3730928,"TERMINAL",0,0,"9937",,terminal_output +2370,3731980,"TERMINAL",0,0,"506:0048",,terminal_output +2371,3733028,"TERMINAL",0,0,"1159",,terminal_output +2372,3734074,"TERMINAL",0,0,"22610",,terminal_output +2373,3735106,"TERMINAL",0,0,"3371",,terminal_output +2374,3736159,"TERMINAL",0,0,"4482",,terminal_output +2375,3737216,"TERMINAL",0,0,"5593",,terminal_output +2376,3738261,"TERMINAL",0,0,"66104",,terminal_output +2377,3739319,"TERMINAL",0,0,"7715",,terminal_output +2378,3740352,"TERMINAL",0,0,"8826",,terminal_output +2379,3741410,"TERMINAL",0,0,"9937",,terminal_output +2380,3742452,"TERMINAL",0,0,"4:001048",,terminal_output +2381,3743502,"TERMINAL",0,0,"1159",,terminal_output +2382,3744545,"TERMINAL",0,0,"22620",,terminal_output +2383,3745592,"TERMINAL",0,0,"3371",,terminal_output +2384,3746637,"TERMINAL",0,0,"4482",,terminal_output +2385,3747693,"TERMINAL",0,0,"5593",,terminal_output +2386,3748773,"TERMINAL",0,0,"67215",,terminal_output +2387,3749797,"TERMINAL",0,0,"8826",,terminal_output +2388,3750843,"TERMINAL",0,0,"9937",,terminal_output +2389,3751890,"TERMINAL",0,0,"102048",,terminal_output +2390,3752949,"TERMINAL",0,0,"1159",,terminal_output +2391,3754010,"TERMINAL",0,0,"22630",,terminal_output +2392,3755046,"TERMINAL",0,0,"3371",,terminal_output +2393,3756097,"TERMINAL",0,0,"4482",,terminal_output +2394,3757143,"TERMINAL",0,0,"5593",,terminal_output +2395,3758199,"TERMINAL",0,0,"66304",,terminal_output +2396,3759234,"TERMINAL",0,0,"7715",,terminal_output +2397,3760280,"TERMINAL",0,0,"8826",,terminal_output +2398,3761321,"TERMINAL",0,0,"9937",,terminal_output +2399,3762373,"TERMINAL",0,0,"203048",,terminal_output +2400,3763415,"TERMINAL",0,0,"1159",,terminal_output +2401,3764470,"TERMINAL",0,0,"22640",,terminal_output +2402,3765522,"TERMINAL",0,0,"3371",,terminal_output +2403,3766565,"TERMINAL",0,0,"4482",,terminal_output +2404,3767612,"TERMINAL",0,0,"5593",,terminal_output +2405,3768658,"TERMINAL",0,0,"66404",,terminal_output +2406,3769754,"TERMINAL",0,0,"7715",,terminal_output +2407,3770753,"TERMINAL",0,0,"8937",,terminal_output +2408,3771797,"TERMINAL",0,0,"304048",,terminal_output +2409,3772850,"TERMINAL",0,0,"1159",,terminal_output +2410,3773902,"TERMINAL",0,0,"22650",,terminal_output +2411,3774955,"TERMINAL",0,0,"3371",,terminal_output +2412,3776002,"TERMINAL",0,0,"4482",,terminal_output +2413,3777059,"TERMINAL",0,0,"5593",,terminal_output +2414,3778133,"TERMINAL",0,0,"66504",,terminal_output +2415,3779185,"TERMINAL",0,0,"7715",,terminal_output +2416,3780230,"TERMINAL",0,0,"8826",,terminal_output +2417,3781287,"TERMINAL",0,0,"9937",,terminal_output +2418,3782327,"TERMINAL",0,0,"405048",,terminal_output +2419,3783374,"TERMINAL",0,0,"1159",,terminal_output +2420,3784427,"TERMINAL",0,0,"2267:00",,terminal_output +2421,3785465,"TERMINAL",0,0,"3371",,terminal_output +2422,3786518,"TERMINAL",0,0,"4482",,terminal_output +2423,3787566,"TERMINAL",0,0,"5593",,terminal_output +2424,3788620,"TERMINAL",0,0,"667:004",,terminal_output +2425,3789669,"TERMINAL",0,0,"7715",,terminal_output +2426,3790739,"TERMINAL",0,0,"8937",,terminal_output +2427,3791773,"TERMINAL",0,0,"507:0048",,terminal_output +2428,3792822,"TERMINAL",0,0,"1159",,terminal_output +2429,3793873,"TERMINAL",0,0,"22610",,terminal_output +2430,3794930,"TERMINAL",0,0,"3371",,terminal_output +2431,3795989,"TERMINAL",0,0,"4482",,terminal_output +2432,3797029,"TERMINAL",0,0,"5593",,terminal_output +2433,3798082,"TERMINAL",0,0,"66104",,terminal_output +2434,3799162,"TERMINAL",0,0,"7715",,terminal_output +2435,3800179,"TERMINAL",0,0,"8826",,terminal_output +2436,3801228,"TERMINAL",0,0,"9937",,terminal_output +2437,3802280,"TERMINAL",0,0,"5:001048",,terminal_output +2438,3803330,"TERMINAL",0,0,"1159",,terminal_output +2439,3804383,"TERMINAL",0,0,"22620",,terminal_output +2440,3805426,"TERMINAL",0,0,"3371",,terminal_output +2441,3806500,"TERMINAL",0,0,"4482",,terminal_output +2442,3807535,"TERMINAL",0,0,"5593",,terminal_output +2443,3808581,"TERMINAL",0,0,"66204",,terminal_output +2444,3809626,"TERMINAL",0,0,"7715",,terminal_output +2445,3810684,"TERMINAL",0,0,"8826",,terminal_output +2446,3811723,"TERMINAL",0,0,"92048",,terminal_output +2447,3812805,"TERMINAL",0,0,"11159",,terminal_output +2448,3813830,"TERMINAL",0,0,"22630",,terminal_output +2449,3814875,"TERMINAL",0,0,"3371",,terminal_output +2450,3815916,"TERMINAL",0,0,"4482",,terminal_output +2451,3816970,"TERMINAL",0,0,"5593",,terminal_output +2452,3818021,"TERMINAL",0,0,"66304",,terminal_output +2453,3819076,"TERMINAL",0,0,"7715",,terminal_output +2454,3820123,"TERMINAL",0,0,"8826",,terminal_output +2455,3821172,"TERMINAL",0,0,"9937",,terminal_output +2456,3822217,"TERMINAL",0,0,"203048",,terminal_output +2457,3823269,"TERMINAL",0,0,"1159",,terminal_output +2458,3824316,"TERMINAL",0,0,"22640",,terminal_output +2459,3825367,"TERMINAL",0,0,"3371",,terminal_output +2460,3826430,"TERMINAL",0,0,"4482",,terminal_output +2461,3827446,"TERMINAL",0,0,"5593",,terminal_output +2462,3828498,"TERMINAL",0,0,"66404",,terminal_output +2463,3829543,"TERMINAL",0,0,"7715",,terminal_output +2464,3830599,"TERMINAL",0,0,"8826",,terminal_output +2465,3831650,"TERMINAL",0,0,"9937",,terminal_output +2466,3832710,"TERMINAL",0,0,"304048",,terminal_output +2467,3833771,"TERMINAL",0,0,"12650",,terminal_output +2468,3834811,"TERMINAL",0,0,"3371",,terminal_output +2469,3835860,"TERMINAL",0,0,"4482",,terminal_output +2470,3836927,"TERMINAL",0,0,"5593",,terminal_output +2471,3837958,"TERMINAL",0,0,"66504",,terminal_output +2472,3839007,"TERMINAL",0,0,"7715",,terminal_output +2473,3840063,"TERMINAL",0,0,"8826",,terminal_output +2474,3841110,"TERMINAL",0,0,"9937",,terminal_output +2475,3842204,"TERMINAL",0,0,"405048",,terminal_output +2476,3843252,"TERMINAL",0,0,"1159",,terminal_output +2477,3844376,"TERMINAL",0,0,"2268:00",,terminal_output +2478,3845394,"TERMINAL",0,0,"3371",,terminal_output +2479,3846622,"TERMINAL",0,0,"4482",,terminal_output +2480,3847681,"TERMINAL",0,0,"5593",,terminal_output +2481,3848722,"TERMINAL",0,0,"678:015",,terminal_output +2482,3849810,"TERMINAL",0,0,"8826",,terminal_output +2483,3850828,"TERMINAL",0,0,"9937",,terminal_output +2484,3851880,"TERMINAL",0,0,"508:0048",,terminal_output +2485,3852938,"TERMINAL",0,0,"1159",,terminal_output +2486,3853984,"TERMINAL",0,0,"22610",,terminal_output +2487,3855042,"TERMINAL",0,0,"3371",,terminal_output +2488,3856078,"TERMINAL",0,0,"4482",,terminal_output +2489,3857128,"TERMINAL",0,0,"5593",,terminal_output +2490,3858174,"TERMINAL",0,0,"66104",,terminal_output +2491,3859224,"TERMINAL",0,0,"7715",,terminal_output +2492,3860282,"TERMINAL",0,0,"8826",,terminal_output +2493,3861327,"TERMINAL",0,0,"9937",,terminal_output +2494,3862374,"TERMINAL",0,0,"6:001048",,terminal_output +2495,3863412,"TERMINAL",0,0,"1159",,terminal_output +2496,3864455,"TERMINAL",0,0,"22620",,terminal_output +2497,3865506,"TERMINAL",0,0,"3371",,terminal_output +2498,3866560,"TERMINAL",0,0,"4482",,terminal_output +2499,3867610,"TERMINAL",0,0,"5593",,terminal_output +2500,3868653,"TERMINAL",0,0,"66204",,terminal_output +2501,3869698,"TERMINAL",0,0,"7715",,terminal_output +2502,3870750,"TERMINAL",0,0,"8937",,terminal_output +2503,3871783,"TERMINAL",0,0,"102048",,terminal_output +2504,3872815,"TERMINAL",0,0,"1159",,terminal_output +2505,3873860,"TERMINAL",0,0,"22630",,terminal_output +2506,3874903,"TERMINAL",0,0,"3371",,terminal_output +2507,3875955,"TERMINAL",0,0,"4482",,terminal_output +2508,3877010,"TERMINAL",0,0,"5593",,terminal_output +2509,3878080,"TERMINAL",0,0,"66304",,terminal_output +2510,3879133,"TERMINAL",0,0,"7715",,terminal_output +2511,3880198,"TERMINAL",0,0,"8826",,terminal_output +2512,3881253,"TERMINAL",0,0,"9937",,terminal_output +2513,3882302,"TERMINAL",0,0,"203048",,terminal_output +2514,3883360,"TERMINAL",0,0,"1159",,terminal_output +2515,3884391,"TERMINAL",0,0,"22640",,terminal_output +2516,3885438,"TERMINAL",0,0,"3371",,terminal_output +2517,3886489,"TERMINAL",0,0,"4482",,terminal_output +2518,3887541,"TERMINAL",0,0,"5593",,terminal_output +2519,3888593,"TERMINAL",0,0,"66404",,terminal_output +2520,3889654,"TERMINAL",0,0,"7715",,terminal_output +2521,3890699,"TERMINAL",0,0,"8826",,terminal_output +2522,3891742,"TERMINAL",0,0,"94048",,terminal_output +2523,3892840,"TERMINAL",0,0,"31159",,terminal_output +2524,3893851,"TERMINAL",0,0,"22650",,terminal_output +2525,3894896,"TERMINAL",0,0,"3371",,terminal_output +2526,3895943,"TERMINAL",0,0,"4482",,terminal_output +2527,3896999,"TERMINAL",0,0,"5593",,terminal_output +2528,3898050,"TERMINAL",0,0,"66504",,terminal_output +2529,3899134,"TERMINAL",0,0,"7715",,terminal_output +2530,3900145,"TERMINAL",0,0,"8826",,terminal_output +2531,3901205,"TERMINAL",0,0,"9937",,terminal_output +2532,3902248,"TERMINAL",0,0,"405048",,terminal_output +2533,3903322,"TERMINAL",0,0,"1159",,terminal_output +2534,3904373,"TERMINAL",0,0,"2269:00",,terminal_output +2535,3905386,"TERMINAL",0,0,"3371",,terminal_output +2536,3906472,"TERMINAL",0,0,"4482",,terminal_output +2537,3907482,"TERMINAL",0,0,"5593",,terminal_output +2538,3908535,"TERMINAL",0,0,"669:004",,terminal_output +2539,3909585,"TERMINAL",0,0,"7715",,terminal_output +2540,3910625,"TERMINAL",0,0,"8826",,terminal_output +2541,3911658,"TERMINAL",0,0,"9937",,terminal_output +2542,3912718,"TERMINAL",0,0,"509:0048",,terminal_output +2543,3913768,"TERMINAL",0,0,"22610",,terminal_output +2544,3914814,"TERMINAL",0,0,"3371",,terminal_output +2545,3915867,"TERMINAL",0,0,"4482",,terminal_output +2546,3916913,"TERMINAL",0,0,"5593",,terminal_output +2547,3917955,"TERMINAL",0,0,"66104",,terminal_output +2548,3919009,"TERMINAL",0,0,"7715",,terminal_output +2549,3920042,"TERMINAL",0,0,"8826",,terminal_output +2550,3921085,"TERMINAL",0,0,"9937",,terminal_output +2551,3922129,"TERMINAL",0,0,"7:001048",,terminal_output +2552,3923180,"TERMINAL",0,0,"1159",,terminal_output +2553,3924226,"TERMINAL",0,0,"22620",,terminal_output +2554,3925278,"TERMINAL",0,0,"3371",,terminal_output +2555,3926345,"TERMINAL",0,0,"4482",,terminal_output +2556,3927375,"TERMINAL",0,0,"5593",,terminal_output +2557,3928402,"TERMINAL",0,0,"66204",,terminal_output +2558,3929447,"TERMINAL",0,0,"7715",,terminal_output +2559,3930507,"TERMINAL",0,0,"8826",,terminal_output +2560,3931533,"TERMINAL",0,0,"9937",,terminal_output +2561,3932587,"TERMINAL",0,0,"102048",,terminal_output +2562,3933622,"TERMINAL",0,0,"1159",,terminal_output +2563,3934668,"TERMINAL",0,0,"22630",,terminal_output +2564,3935722,"TERMINAL",0,0,"3482",,terminal_output +2565,3936756,"TERMINAL",0,0,"5593",,terminal_output +2566,3937827,"TERMINAL",0,0,"66304",,terminal_output +2567,3938866,"TERMINAL",0,0,"7715",,terminal_output +2568,3939916,"TERMINAL",0,0,"8826",,terminal_output +2569,3940937,"TERMINAL",0,0,"9937",,terminal_output +2570,3941991,"TERMINAL",0,0,"203048",,terminal_output +2571,3943031,"TERMINAL",0,0,"1159",,terminal_output +2572,3944137,"TERMINAL",0,0,"22640",,terminal_output +2573,3945126,"TERMINAL",0,0,"3371",,terminal_output +2574,3946183,"TERMINAL",0,0,"4482",,terminal_output +2575,3947217,"TERMINAL",0,0,"5593",,terminal_output +2576,3948274,"TERMINAL",0,0,"66404",,terminal_output +2577,3949316,"TERMINAL",0,0,"7715",,terminal_output +2578,3950372,"TERMINAL",0,0,"8826",,terminal_output +2579,3951424,"TERMINAL",0,0,"9937",,terminal_output +2580,3952482,"TERMINAL",0,0,"304048",,terminal_output +2581,3953531,"TERMINAL",0,0,"1159",,terminal_output +2582,3954587,"TERMINAL",0,0,"22650",,terminal_output +2583,3955642,"TERMINAL",0,0,"3371",,terminal_output +2584,3956694,"TERMINAL",0,0,"4482",,terminal_output +2585,3957743,"TERMINAL",0,0,"56504",,terminal_output +2586,3958810,"TERMINAL",0,0,"7715",,terminal_output +2587,3959852,"TERMINAL",0,0,"8826",,terminal_output +2588,3960897,"TERMINAL",0,0,"9937",,terminal_output +2589,3961945,"TERMINAL",0,0,"405048",,terminal_output +2590,3962985,"TERMINAL",0,0,"1159",,terminal_output +2591,3964036,"TERMINAL",0,0,"22620:00",,terminal_output +2592,3965097,"TERMINAL",0,0,"3371",,terminal_output +2593,3966147,"TERMINAL",0,0,"4482",,terminal_output +2594,3967650,"TERMINAL",0,0,"5593",,terminal_output +2595,3968691,"TERMINAL",0,0,"6620:004",,terminal_output +2596,3969778,"TERMINAL",0,0,"7826",,terminal_output +2597,3970776,"TERMINAL",0,0,"9937",,terminal_output +2598,3971881,"TERMINAL",0,0,"5020:0048",,terminal_output +2599,3972869,"TERMINAL",0,0,"1159",,terminal_output +2600,3973921,"TERMINAL",0,0,"22610",,terminal_output +2601,3974978,"TERMINAL",0,0,"3371",,terminal_output +2602,3976025,"TERMINAL",0,0,"4482",,terminal_output +2603,3977072,"TERMINAL",0,0,"5593",,terminal_output +2604,3978129,"TERMINAL",0,0,"66104",,terminal_output +2605,3979182,"TERMINAL",0,0,"7715",,terminal_output +2606,3980228,"TERMINAL",0,0,"8826",,terminal_output +2607,3981270,"TERMINAL",0,0,"9937",,terminal_output +2608,3982333,"TERMINAL",0,0,"8:001048",,terminal_output +2609,3983378,"TERMINAL",0,0,"1159",,terminal_output +2610,3984424,"TERMINAL",0,0,"22620",,terminal_output +2611,3985459,"TERMINAL",0,0,"3371",,terminal_output +2612,3986508,"TERMINAL",0,0,"4482",,terminal_output +2613,3987565,"TERMINAL",0,0,"5593",,terminal_output +2614,3988610,"TERMINAL",0,0,"66204",,terminal_output +2615,3989664,"TERMINAL",0,0,"7715",,terminal_output +2616,3990715,"TERMINAL",0,0,"8937",,terminal_output +2617,3991768,"TERMINAL",0,0,"102048",,terminal_output +2618,3992822,"TERMINAL",0,0,"1159",,terminal_output +2619,3993850,"TERMINAL",0,0,"22630",,terminal_output +2620,3994900,"TERMINAL",0,0,"3371",,terminal_output +2621,3995956,"TERMINAL",0,0,"4482",,terminal_output +2622,3997000,"TERMINAL",0,0,"5593",,terminal_output +2623,3998047,"TERMINAL",0,0,"66304",,terminal_output +2624,3999139,"TERMINAL",0,0,"7715",,terminal_output +2625,4000148,"TERMINAL",0,0,"8826",,terminal_output +2626,4001195,"TERMINAL",0,0,"9937",,terminal_output +2627,4002244,"TERMINAL",0,0,"203048",,terminal_output +2628,4003842,"TERMINAL",0,0,"12640",,terminal_output +2629,4004857,"TERMINAL",0,0,"3371",,terminal_output +2630,4005929,"TERMINAL",0,0,"4482",,terminal_output +2631,4006929,"TERMINAL",0,0,"5593",,terminal_output +2632,4007979,"TERMINAL",0,0,"66404",,terminal_output +2633,4009022,"TERMINAL",0,0,"7715",,terminal_output +2634,4010074,"TERMINAL",0,0,"8826",,terminal_output +2635,4011115,"TERMINAL",0,0,"9937",,terminal_output +2636,4012162,"TERMINAL",0,0,"304048",,terminal_output +2637,4013203,"TERMINAL",0,0,"1159",,terminal_output +2638,4014249,"TERMINAL",0,0,"22650",,terminal_output +2639,4015299,"TERMINAL",0,0,"3371",,terminal_output +2640,4016345,"TERMINAL",0,0,"4482",,terminal_output +2641,4017390,"TERMINAL",0,0,"5593",,terminal_output +2642,4018442,"TERMINAL",0,0,"66504",,terminal_output +2643,4019494,"TERMINAL",0,0,"7715",,terminal_output +2644,4020541,"TERMINAL",0,0,"8826",,terminal_output +2645,4021586,"TERMINAL",0,0,"9937",,terminal_output +2646,4021797,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",0,0,"",shellscript,tab +2647,4022637,"TERMINAL",0,0,"405048",,terminal_output +2648,4023683,"TERMINAL",0,0,"1159",,terminal_output +2649,4024160,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",465,0,"",shellscript,selection_mouse +2650,4024719,"TERMINAL",0,0,"2371:01",,terminal_output +2651,4025046,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",469,0,"",shellscript,selection_mouse +2652,4025060,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",468,0,"",shellscript,selection_command +2653,4025607,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",487,0,"",shellscript,selection_mouse +2654,4025607,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",486,0,"",shellscript,selection_command +2655,4025779,"TERMINAL",0,0,"4482",,terminal_output +2656,4026572,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",487,0,"",shellscript,selection_mouse +2657,4026573,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",486,0,"",shellscript,selection_command +2658,4026848,"TERMINAL",0,0,"5593",,terminal_output +2659,4027543,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",487,0,"",shellscript,selection_mouse +2660,4027544,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",486,0,"",shellscript,selection_command +2661,4027896,"TERMINAL",0,0,"661:004",,terminal_output +2662,4028148,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",538,0,"",shellscript,selection_mouse +2663,4028186,"slurm/jobs/mihir/horeka/lam/train_lam_minecraft_8node-darkness-filter-133M.sbatch",537,0,"",shellscript,selection_command +2664,4028969,"TERMINAL",0,0,"7715",,terminal_output +2665,4030045,"TERMINAL",0,0,"8826",,terminal_output +2666,4031034,"TERMINAL",0,0,"9937",,terminal_output +2667,4032246,"TERMINAL",0,0,"501:0048",,terminal_output +2668,4033139,"TERMINAL",0,0,"1159",,terminal_output +2669,4034192,"TERMINAL",0,0,"22610",,terminal_output +2670,4035288,"TERMINAL",0,0,"3371",,terminal_output +2671,4036304,"TERMINAL",0,0,"4482",,terminal_output +2672,4037357,"TERMINAL",0,0,"5593",,terminal_output +2673,4038405,"TERMINAL",0,0,"66104",,terminal_output +2674,4039446,"TERMINAL",0,0,"7715",,terminal_output +2675,4040494,"TERMINAL",0,0,"8826",,terminal_output +2676,4041539,"TERMINAL",0,0,"9937",,terminal_output +2677,4042595,"TERMINAL",0,0,"9:001048",,terminal_output +2678,4043645,"TERMINAL",0,0,"1159",,terminal_output +2679,4044701,"TERMINAL",0,0,"22620",,terminal_output +2680,4045750,"TERMINAL",0,0,"3482",,terminal_output +2681,4046802,"TERMINAL",0,0,"5593",,terminal_output +2682,4047884,"TERMINAL",0,0,"66204",,terminal_output +2683,4048902,"TERMINAL",0,0,"7715",,terminal_output +2684,4049950,"TERMINAL",0,0,"8826",,terminal_output +2685,4051000,"TERMINAL",0,0,"9937",,terminal_output +2686,4052050,"TERMINAL",0,0,"102048",,terminal_output +2687,4053102,"TERMINAL",0,0,"1159",,terminal_output +2688,4054150,"TERMINAL",0,0,"22630",,terminal_output +2689,4055191,"TERMINAL",0,0,"3371",,terminal_output +2690,4056242,"TERMINAL",0,0,"4482",,terminal_output +2691,4057297,"TERMINAL",0,0,"5593",,terminal_output +2692,4058335,"TERMINAL",0,0,"66304",,terminal_output +2693,4059392,"TERMINAL",0,0,"7715",,terminal_output +2694,4060438,"TERMINAL",0,0,"8826",,terminal_output +2695,4061489,"TERMINAL",0,0,"9937",,terminal_output +2696,4062527,"TERMINAL",0,0,"203048",,terminal_output +2697,4063581,"TERMINAL",0,0,"1159",,terminal_output +2698,4064626,"TERMINAL",0,0,"22640",,terminal_output +2699,4065665,"TERMINAL",0,0,"3371",,terminal_output +2700,4066712,"TERMINAL",0,0,"4482",,terminal_output +2701,4067773,"TERMINAL",0,0,"56404",,terminal_output +2702,4068850,"TERMINAL",0,0,"7715",,terminal_output +2703,4069874,"TERMINAL",0,0,"8826",,terminal_output +2704,4070925,"TERMINAL",0,0,"9937",,terminal_output +2705,4071972,"TERMINAL",0,0,"304048",,terminal_output +2706,4073025,"TERMINAL",0,0,"1159",,terminal_output +2707,4074072,"TERMINAL",0,0,"22650",,terminal_output +2708,4075119,"TERMINAL",0,0,"3371",,terminal_output +2709,4076170,"TERMINAL",0,0,"4482",,terminal_output +2710,4077222,"TERMINAL",0,0,"5593",,terminal_output +2711,4078279,"TERMINAL",0,0,"66504",,terminal_output +2712,4079321,"TERMINAL",0,0,"7715",,terminal_output +2713,4080376,"TERMINAL",0,0,"8826",,terminal_output +2714,4081425,"TERMINAL",0,0,"9937",,terminal_output +2715,4082469,"TERMINAL",0,0,"405048",,terminal_output +2716,4083523,"TERMINAL",0,0,"1159",,terminal_output +2717,4084583,"TERMINAL",0,0,"2262:00",,terminal_output +2718,4085622,"TERMINAL",0,0,"3371",,terminal_output +2719,4086671,"TERMINAL",0,0,"4482",,terminal_output +2720,4088579,"TERMINAL",0,0,"562:004",,terminal_output +2721,4089625,"TERMINAL",0,0,"7715",,terminal_output +2722,4090666,"TERMINAL",0,0,"8826",,terminal_output +2723,4091701,"TERMINAL",0,0,"9937",,terminal_output +2724,4092740,"TERMINAL",0,0,"502:0159",,terminal_output +2725,4093796,"TERMINAL",0,0,"22610",,terminal_output +2726,4094873,"TERMINAL",0,0,"3371",,terminal_output +2727,4095882,"TERMINAL",0,0,"4482",,terminal_output +2728,4096939,"TERMINAL",0,0,"5593",,terminal_output +2729,4097999,"TERMINAL",0,0,"66104",,terminal_output +2730,4099046,"TERMINAL",0,0,"7715",,terminal_output +2731,4100096,"TERMINAL",0,0,"8826",,terminal_output +2732,4101149,"TERMINAL",0,0,"9937",,terminal_output +2733,4102202,"TERMINAL",0,0,"50:001048",,terminal_output +2734,4103244,"TERMINAL",0,0,"1159",,terminal_output +2735,4104288,"TERMINAL",0,0,"22620",,terminal_output +2736,4105336,"TERMINAL",0,0,"3371",,terminal_output +2737,4106385,"TERMINAL",0,0,"4482",,terminal_output +2738,4107430,"TERMINAL",0,0,"5593",,terminal_output +2739,4108479,"TERMINAL",0,0,"66204",,terminal_output +2740,4109532,"TERMINAL",0,0,"7715",,terminal_output +2741,4110581,"TERMINAL",0,0,"8826",,terminal_output +2742,4111634,"TERMINAL",0,0,"9937",,terminal_output +2743,4112682,"TERMINAL",0,0,"102048",,terminal_output +2744,4113736,"TERMINAL",0,0,"12630",,terminal_output +2745,4114784,"TERMINAL",0,0,"3371",,terminal_output +2746,4115864,"TERMINAL",0,0,"4482",,terminal_output +2747,4116893,"TERMINAL",0,0,"5593",,terminal_output +2748,4117940,"TERMINAL",0,0,"66304",,terminal_output +2749,4118992,"TERMINAL",0,0,"7715",,terminal_output +2750,4120040,"TERMINAL",0,0,"8826",,terminal_output +2751,4121092,"TERMINAL",0,0,"9937",,terminal_output +2752,4122146,"TERMINAL",0,0,"203048",,terminal_output +2753,4123197,"TERMINAL",0,0,"1159",,terminal_output +2754,4124253,"TERMINAL",0,0,"22640",,terminal_output +2755,4125309,"TERMINAL",0,0,"3371",,terminal_output +2756,4126346,"TERMINAL",0,0,"4482",,terminal_output +2757,4127397,"TERMINAL",0,0,"5593",,terminal_output +2758,4128448,"TERMINAL",0,0,"66404",,terminal_output +2759,4129499,"TERMINAL",0,0,"7715",,terminal_output +2760,4130560,"TERMINAL",0,0,"8826",,terminal_output +2761,4131604,"TERMINAL",0,0,"9937",,terminal_output +2762,4132676,"TERMINAL",0,0,"304048",,terminal_output +2763,4133734,"TERMINAL",0,0,"12650",,terminal_output +2764,4134782,"TERMINAL",0,0,"3371",,terminal_output +2765,4135811,"TERMINAL",0,0,"4482",,terminal_output +2766,4136891,"TERMINAL",0,0,"5593",,terminal_output +2767,4137910,"TERMINAL",0,0,"66504",,terminal_output +2768,4138960,"TERMINAL",0,0,"7715",,terminal_output +2769,4140016,"TERMINAL",0,0,"8826",,terminal_output +2770,4141063,"TERMINAL",0,0,"9937",,terminal_output +2771,4142115,"TERMINAL",0,0,"405048",,terminal_output +2772,4143166,"TERMINAL",0,0,"1159",,terminal_output +2773,4144225,"TERMINAL",0,0,"2263:00",,terminal_output +2774,4145282,"TERMINAL",0,0,"3371",,terminal_output +2775,4146330,"TERMINAL",0,0,"4482",,terminal_output +2776,4147421,"TERMINAL",0,0,"5593",,terminal_output +2777,4148421,"TERMINAL",0,0,"663:004",,terminal_output +2778,4149470,"TERMINAL",0,0,"7715",,terminal_output +2779,4150514,"TERMINAL",0,0,"8826",,terminal_output +2780,4151558,"TERMINAL",0,0,"9937",,terminal_output +2781,4152616,"TERMINAL",0,0,"503:0048",,terminal_output +2782,4153667,"TERMINAL",0,0,"1159",,terminal_output +2783,4154712,"TERMINAL",0,0,"22610",,terminal_output +2784,4155764,"TERMINAL",0,0,"4482",,terminal_output +2785,4156817,"TERMINAL",0,0,"5593",,terminal_output +2786,4157881,"TERMINAL",0,0,"66104",,terminal_output +2787,4158936,"TERMINAL",0,0,"7715",,terminal_output +2788,4159958,"TERMINAL",0,0,"8826",,terminal_output +2789,4161007,"TERMINAL",0,0,"9937",,terminal_output +2790,4162057,"TERMINAL",0,0,"1:001048",,terminal_output +2791,4163111,"TERMINAL",0,0,"1159",,terminal_output +2792,4164191,"TERMINAL",0,0,"22620",,terminal_output +2793,4165210,"TERMINAL",0,0,"3371",,terminal_output +2794,4166260,"TERMINAL",0,0,"4482",,terminal_output +2795,4167325,"TERMINAL",0,0,"5593",,terminal_output +2796,4168365,"TERMINAL",0,0,"66204",,terminal_output +2797,4169411,"TERMINAL",0,0,"7715",,terminal_output +2798,4170454,"TERMINAL",0,0,"8826",,terminal_output +2799,4171507,"TERMINAL",0,0,"9937",,terminal_output +2800,4172556,"TERMINAL",0,0,"102048",,terminal_output +2801,4173606,"TERMINAL",0,0,"1159",,terminal_output +2802,4174666,"TERMINAL",0,0,"22630",,terminal_output +2803,4175708,"TERMINAL",0,0,"3371",,terminal_output +2804,4176758,"TERMINAL",0,0,"4593",,terminal_output +2805,4177810,"TERMINAL",0,0,"66304",,terminal_output +2806,4178863,"TERMINAL",0,0,"7715",,terminal_output +2807,4179923,"TERMINAL",0,0,"8826",,terminal_output +2808,4180980,"TERMINAL",0,0,"9937",,terminal_output +2809,4182029,"TERMINAL",0,0,"203048",,terminal_output +2810,4183075,"TERMINAL",0,0,"1159",,terminal_output +2811,4184162,"TERMINAL",0,0,"22640",,terminal_output +2812,4185176,"TERMINAL",0,0,"3371",,terminal_output +2813,4186226,"TERMINAL",0,0,"4482",,terminal_output +2814,4187284,"TERMINAL",0,0,"5593",,terminal_output +2815,4188336,"TERMINAL",0,0,"66404",,terminal_output +2816,4189386,"TERMINAL",0,0,"7715",,terminal_output +2817,4190434,"TERMINAL",0,0,"8826",,terminal_output +2818,4191499,"TERMINAL",0,0,"9937",,terminal_output +2819,4192559,"TERMINAL",0,0,"304048",,terminal_output +2820,4193600,"TERMINAL",0,0,"1159",,terminal_output +2821,4194665,"TERMINAL",0,0,"22650",,terminal_output +2822,4195690,"TERMINAL",0,0,"3371",,terminal_output +2823,4196744,"TERMINAL",0,0,"4593",,terminal_output +2824,4197794,"TERMINAL",0,0,"66504",,terminal_output +2825,4198844,"TERMINAL",0,0,"7715",,terminal_output +2826,4199940,"TERMINAL",0,0,"8826",,terminal_output +2827,4200957,"TERMINAL",0,0,"9937",,terminal_output +2828,4202006,"TERMINAL",0,0,"405048",,terminal_output +2829,4203052,"TERMINAL",0,0,"1159",,terminal_output +2830,4204139,"TERMINAL",0,0,"2264:00",,terminal_output +2831,4205140,"TERMINAL",0,0,"3371",,terminal_output +2832,4206197,"TERMINAL",0,0,"4482",,terminal_output +2833,4207247,"TERMINAL",0,0,"5593",,terminal_output +2834,4208294,"TERMINAL",0,0,"664:004",,terminal_output +2835,4209517,"TERMINAL",0,0,"7715",,terminal_output +2836,4210561,"TERMINAL",0,0,"8826",,terminal_output +2837,4211598,"TERMINAL",0,0,"9937",,terminal_output +2838,4212653,"TERMINAL",0,0,"504:0048",,terminal_output +2839,4213699,"TERMINAL",0,0,"1159",,terminal_output +2840,4214755,"TERMINAL",0,0,"23711",,terminal_output +2841,4215799,"TERMINAL",0,0,"4482",,terminal_output +2842,4216857,"TERMINAL",0,0,"5593",,terminal_output +2843,4217938,"TERMINAL",0,0,"66104",,terminal_output +2844,4218947,"TERMINAL",0,0,"7715",,terminal_output +2845,4219993,"TERMINAL",0,0,"8826",,terminal_output +2846,4221047,"TERMINAL",0,0,"9937",,terminal_output +2847,4222092,"TERMINAL",0,0,"2:001048",,terminal_output +2848,4223147,"TERMINAL",0,0,"1159",,terminal_output +2849,4224204,"TERMINAL",0,0,"22620",,terminal_output +2850,4225257,"TERMINAL",0,0,"3371",,terminal_output +2851,4226313,"TERMINAL",0,0,"4482",,terminal_output +2852,4227361,"TERMINAL",0,0,"5593",,terminal_output +2853,4228423,"TERMINAL",0,0,"66204",,terminal_output +2854,4229472,"TERMINAL",0,0,"7715",,terminal_output +2855,4230533,"TERMINAL",0,0,"8826",,terminal_output +2856,4231576,"TERMINAL",0,0,"9937",,terminal_output +2857,4232634,"TERMINAL",0,0,"102048",,terminal_output +2858,4233684,"TERMINAL",0,0,"1159",,terminal_output +2859,4234738,"TERMINAL",0,0,"23731",,terminal_output +2860,4235791,"TERMINAL",0,0,"4482",,terminal_output +2861,4236841,"TERMINAL",0,0,"5593",,terminal_output +2862,4237900,"TERMINAL",0,0,"66304",,terminal_output +2863,4238978,"TERMINAL",0,0,"7715",,terminal_output +2864,4240007,"TERMINAL",0,0,"8826",,terminal_output +2865,4241050,"TERMINAL",0,0,"9937",,terminal_output +2866,4242103,"TERMINAL",0,0,"203048",,terminal_output +2867,4243161,"TERMINAL",0,0,"1159",,terminal_output +2868,4244210,"TERMINAL",0,0,"22640",,terminal_output +2869,4245284,"TERMINAL",0,0,"3371",,terminal_output +2870,4246321,"TERMINAL",0,0,"4482",,terminal_output +2871,4247385,"TERMINAL",0,0,"5593",,terminal_output +2872,4248454,"TERMINAL",0,0,"66404",,terminal_output +2873,4249508,"TERMINAL",0,0,"7715",,terminal_output +2874,4250552,"TERMINAL",0,0,"8826",,terminal_output +2875,4251593,"TERMINAL",0,0,"9937",,terminal_output +2876,4252644,"TERMINAL",0,0,"304048",,terminal_output +2877,4253700,"TERMINAL",0,0,"1159",,terminal_output +2878,4254747,"TERMINAL",0,0,"23751",,terminal_output +2879,4255802,"TERMINAL",0,0,"4482",,terminal_output +2880,4256851,"TERMINAL",0,0,"5593",,terminal_output +2881,4257895,"TERMINAL",0,0,"66504",,terminal_output +2882,4258985,"TERMINAL",0,0,"7715",,terminal_output +2883,4260019,"TERMINAL",0,0,"8826",,terminal_output +2884,4261073,"TERMINAL",0,0,"9937",,terminal_output +2885,4262114,"TERMINAL",0,0,"405048",,terminal_output +2886,4263169,"TERMINAL",0,0,"1159",,terminal_output +2887,4264220,"TERMINAL",0,0,"2265:00",,terminal_output +2888,4265265,"TERMINAL",0,0,"3371",,terminal_output +2889,4266310,"TERMINAL",0,0,"4482",,terminal_output +2890,4267360,"TERMINAL",0,0,"5593",,terminal_output +2891,4268412,"TERMINAL",0,0,"665:004",,terminal_output +2892,4269477,"TERMINAL",0,0,"7715",,terminal_output +2893,4270510,"TERMINAL",0,0,"8826",,terminal_output +2894,4271565,"TERMINAL",0,0,"9937",,terminal_output +2895,4272615,"TERMINAL",0,0,"505:0048",,terminal_output +2896,4273677,"TERMINAL",0,0,"1159",,terminal_output +2897,4274708,"TERMINAL",0,0,"22610",,terminal_output +2898,4275756,"TERMINAL",0,0,"3482",,terminal_output +2899,4276828,"TERMINAL",0,0,"5593",,terminal_output +2900,4277876,"TERMINAL",0,0,"66104",,terminal_output +2901,4278938,"TERMINAL",0,0,"7715",,terminal_output +2902,4279970,"TERMINAL",0,0,"8826",,terminal_output +2903,4281020,"TERMINAL",0,0,"9937",,terminal_output +2904,4282074,"TERMINAL",0,0,"3:001048",,terminal_output +2905,4283122,"TERMINAL",0,0,"1159",,terminal_output +2906,4284173,"TERMINAL",0,0,"22620",,terminal_output +2907,4285216,"TERMINAL",0,0,"3371",,terminal_output +2908,4286256,"TERMINAL",0,0,"4482",,terminal_output +2909,4287316,"TERMINAL",0,0,"5593",,terminal_output +2910,4288347,"TERMINAL",0,0,"66204",,terminal_output +2911,4289404,"TERMINAL",0,0,"7715",,terminal_output +2912,4290462,"TERMINAL",0,0,"8826",,terminal_output +2913,4291500,"TERMINAL",0,0,"9937",,terminal_output +2914,4292549,"TERMINAL",0,0,"102048",,terminal_output +2915,4293605,"TERMINAL",0,0,"1159",,terminal_output +2916,4294656,"TERMINAL",0,0,"22630",,terminal_output +2917,4295704,"TERMINAL",0,0,"3371",,terminal_output +2918,4296760,"TERMINAL",0,0,"4593",,terminal_output +2919,4297830,"TERMINAL",0,0,"66304",,terminal_output +2920,4298880,"TERMINAL",0,0,"7715",,terminal_output +2921,4299962,"TERMINAL",0,0,"8826",,terminal_output +2922,4300981,"TERMINAL",0,0,"9937",,terminal_output +2923,4302026,"TERMINAL",0,0,"203048",,terminal_output +2924,4303075,"TERMINAL",0,0,"1159",,terminal_output +2925,4304163,"TERMINAL",0,0,"22640",,terminal_output +2926,4305175,"TERMINAL",0,0,"3371",,terminal_output +2927,4306228,"TERMINAL",0,0,"4482",,terminal_output +2928,4307279,"TERMINAL",0,0,"5593",,terminal_output +2929,4308338,"TERMINAL",0,0,"66404",,terminal_output +2930,4309394,"TERMINAL",0,0,"7715",,terminal_output +2931,4310428,"TERMINAL",0,0,"8826",,terminal_output +2932,4311505,"TERMINAL",0,0,"9937",,terminal_output +2933,4312560,"TERMINAL",0,0,"304048",,terminal_output +2934,4313576,"TERMINAL",0,0,"1159",,terminal_output +2935,4314624,"TERMINAL",0,0,"22650",,terminal_output +2936,4315677,"TERMINAL",0,0,"3371",,terminal_output +2937,4316735,"TERMINAL",0,0,"4593",,terminal_output +2938,4317770,"TERMINAL",0,0,"66504",,terminal_output +2939,4318825,"TERMINAL",0,0,"7715",,terminal_output +2940,4319870,"TERMINAL",0,0,"8826",,terminal_output +2941,4322042,"TERMINAL",0,0,"95048",,terminal_output +2942,4323096,"TERMINAL",0,0,"41159",,terminal_output +2943,4324136,"TERMINAL",0,0,"2266:00",,terminal_output +2944,4325182,"TERMINAL",0,0,"3371",,terminal_output +2945,4326219,"TERMINAL",0,0,"4482",,terminal_output +2946,4327271,"TERMINAL",0,0,"5593",,terminal_output +2947,4328328,"TERMINAL",0,0,"666:004",,terminal_output +2948,4329396,"TERMINAL",0,0,"7715",,terminal_output +2949,4330541,"TERMINAL",0,0,"8826",,terminal_output +2950,4331581,"TERMINAL",0,0,"9937",,terminal_output +2951,4332638,"TERMINAL",0,0,"506:0048",,terminal_output +2952,4333682,"TERMINAL",0,0,"1159",,terminal_output +2953,4334744,"TERMINAL",0,0,"23711",,terminal_output +2954,4335810,"TERMINAL",0,0,"4482",,terminal_output +2955,4336854,"TERMINAL",0,0,"5593",,terminal_output +2956,4337908,"TERMINAL",0,0,"66104",,terminal_output +2957,4338986,"TERMINAL",0,0,"7715",,terminal_output +2958,4340001,"TERMINAL",0,0,"8826",,terminal_output +2959,4341059,"TERMINAL",0,0,"9937",,terminal_output +2960,4342105,"TERMINAL",0,0,"4:001048",,terminal_output +2961,4343156,"TERMINAL",0,0,"1159",,terminal_output +2962,4344210,"TERMINAL",0,0,"22620",,terminal_output +2963,4345263,"TERMINAL",0,0,"3371",,terminal_output +2964,4346354,"TERMINAL",0,0,"4482",,terminal_output +2965,4347376,"TERMINAL",0,0,"5593",,terminal_output +2966,4348440,"TERMINAL",0,0,"66204",,terminal_output +2967,4349502,"TERMINAL",0,0,"7715",,terminal_output +2968,4350558,"TERMINAL",0,0,"8826",,terminal_output +2969,4351621,"TERMINAL",0,0,"9937",,terminal_output +2970,4352677,"TERMINAL",0,0,"102048",,terminal_output +2971,4353733,"TERMINAL",0,0,"12630",,terminal_output +2972,4354794,"TERMINAL",0,0,"3371",,terminal_output +2973,4355845,"TERMINAL",0,0,"4482",,terminal_output +2974,4356891,"TERMINAL",0,0,"5593",,terminal_output +2975,4357940,"TERMINAL",0,0,"66304",,terminal_output +2976,4359027,"TERMINAL",0,0,"7715",,terminal_output +2977,4360034,"TERMINAL",0,0,"8826",,terminal_output +2978,4361084,"TERMINAL",0,0,"9937",,terminal_output +2979,4362138,"TERMINAL",0,0,"203048",,terminal_output +2980,4363195,"TERMINAL",0,0,"1159",,terminal_output +2981,4364254,"TERMINAL",0,0,"22640",,terminal_output +2982,4365314,"TERMINAL",0,0,"3371",,terminal_output +2983,4366343,"TERMINAL",0,0,"4482",,terminal_output +2984,4367392,"TERMINAL",0,0,"5593",,terminal_output +2985,4368443,"TERMINAL",0,0,"66404",,terminal_output +2986,4369496,"TERMINAL",0,0,"7715",,terminal_output +2987,4370551,"TERMINAL",0,0,"8826",,terminal_output +2988,4371592,"TERMINAL",0,0,"9937",,terminal_output +2989,4372652,"TERMINAL",0,0,"304048",,terminal_output +2990,4373733,"TERMINAL",0,0,"1159",,terminal_output +2991,4374770,"TERMINAL",0,0,"33751",,terminal_output +2992,4375829,"TERMINAL",0,0,"4482",,terminal_output +2993,4376893,"TERMINAL",0,0,"5593",,terminal_output +2994,4377948,"TERMINAL",0,0,"66504",,terminal_output +2995,4379013,"TERMINAL",0,0,"7715",,terminal_output +2996,4380054,"TERMINAL",0,0,"8826",,terminal_output +2997,4381118,"TERMINAL",0,0,"9937",,terminal_output +2998,4382156,"TERMINAL",0,0,"405048",,terminal_output +2999,4383196,"TERMINAL",0,0,"1159",,terminal_output +3000,4384251,"TERMINAL",0,0,"2267:00",,terminal_output +3001,4385308,"TERMINAL",0,0,"3371",,terminal_output +3002,4386345,"TERMINAL",0,0,"4482",,terminal_output +3003,4387405,"TERMINAL",0,0,"5593",,terminal_output +3004,4388452,"TERMINAL",0,0,"667:004",,terminal_output +3005,4389598,"TERMINAL",0,0,"7715",,terminal_output +3006,4390645,"TERMINAL",0,0,"8826",,terminal_output +3007,4391695,"TERMINAL",0,0,"9937",,terminal_output +3008,4392759,"TERMINAL",0,0,"507:0159",,terminal_output +3009,4393917,"TERMINAL",0,0,"22610",,terminal_output +3010,4394997,"TERMINAL",0,0,"3371",,terminal_output +3011,4396027,"TERMINAL",0,0,"4482",,terminal_output +3012,4397091,"TERMINAL",0,0,"5593",,terminal_output +3013,4398150,"TERMINAL",0,0,"66104",,terminal_output +3014,4399204,"TERMINAL",0,0,"7715",,terminal_output +3015,4400255,"TERMINAL",0,0,"8826",,terminal_output +3016,4401310,"TERMINAL",0,0,"9937",,terminal_output +3017,4402358,"TERMINAL",0,0,"5:001048",,terminal_output +3018,4403416,"TERMINAL",0,0,"1159",,terminal_output +3019,4404478,"TERMINAL",0,0,"22620",,terminal_output +3020,4405536,"TERMINAL",0,0,"3371",,terminal_output +3021,4406583,"TERMINAL",0,0,"4482",,terminal_output +3022,4407635,"TERMINAL",0,0,"5593",,terminal_output +3023,4408690,"TERMINAL",0,0,"66204",,terminal_output +3024,4409737,"TERMINAL",0,0,"7826",,terminal_output +3025,4410780,"TERMINAL",0,0,"9937",,terminal_output +3026,4411834,"TERMINAL",0,0,"102048",,terminal_output +3027,4412882,"TERMINAL",0,0,"1159",,terminal_output +3028,4413927,"TERMINAL",0,0,"22630",,terminal_output +3029,4414998,"TERMINAL",0,0,"3371",,terminal_output +3030,4416054,"TERMINAL",0,0,"4482",,terminal_output +3031,4417072,"TERMINAL",0,0,"5593",,terminal_output +3032,4418128,"TERMINAL",0,0,"66304",,terminal_output +3033,4419193,"TERMINAL",0,0,"7715",,terminal_output +3034,4420236,"TERMINAL",0,0,"8826",,terminal_output +3035,4421277,"TERMINAL",0,0,"9937",,terminal_output +3036,4422328,"TERMINAL",0,0,"203048",,terminal_output +3037,4423373,"TERMINAL",0,0,"1159",,terminal_output +3038,4424430,"TERMINAL",0,0,"22640",,terminal_output +3039,4425481,"TERMINAL",0,0,"3371",,terminal_output +3040,4426533,"TERMINAL",0,0,"4482",,terminal_output +3041,4427574,"TERMINAL",0,0,"5593",,terminal_output +3042,4428625,"TERMINAL",0,0,"66404",,terminal_output +3043,4429664,"TERMINAL",0,0,"7715",,terminal_output +3044,4430711,"TERMINAL",0,0,"8826",,terminal_output +3045,4431765,"TERMINAL",0,0,"94048",,terminal_output +3046,4432813,"TERMINAL",0,0,"31159",,terminal_output +3047,4433867,"TERMINAL",0,0,"22650",,terminal_output +3048,4434913,"TERMINAL",0,0,"3371",,terminal_output +3049,4435962,"TERMINAL",0,0,"4482",,terminal_output +3050,4437036,"TERMINAL",0,0,"5593",,terminal_output +3051,4438056,"TERMINAL",0,0,"66504",,terminal_output +3052,4439138,"TERMINAL",0,0,"7715",,terminal_output +3053,4440146,"TERMINAL",0,0,"8826",,terminal_output +3054,4441201,"TERMINAL",0,0,"9937",,terminal_output +3055,4442233,"TERMINAL",0,0,"405048",,terminal_output +3056,4443286,"TERMINAL",0,0,"1159",,terminal_output +3057,4444340,"TERMINAL",0,0,"2268:00",,terminal_output +3058,4445405,"TERMINAL",0,0,"3371",,terminal_output +3059,4446450,"TERMINAL",0,0,"4482",,terminal_output +3060,4447489,"TERMINAL",0,0,"5593",,terminal_output +3061,4448538,"TERMINAL",0,0,"668:004",,terminal_output +3062,4449584,"TERMINAL",0,0,"7715",,terminal_output +3063,4451534,"TERMINAL",0,0,"8937",,terminal_output +3064,4452582,"TERMINAL",0,0,"508:0048",,terminal_output +3065,4453636,"TERMINAL",0,0,"1159",,terminal_output +3066,4454686,"TERMINAL",0,0,"22610",,terminal_output +3067,4455741,"TERMINAL",0,0,"3482",,terminal_output +3068,4456787,"TERMINAL",0,0,"5593",,terminal_output +3069,4457832,"TERMINAL",0,0,"66104",,terminal_output +3070,4458885,"TERMINAL",0,0,"7715",,terminal_output +3071,4459930,"TERMINAL",0,0,"8826",,terminal_output +3072,4460979,"TERMINAL",0,0,"9937",,terminal_output +3073,4462066,"TERMINAL",0,0,"6:001048",,terminal_output +3074,4463077,"TERMINAL",0,0,"1159",,terminal_output +3075,4464145,"TERMINAL",0,0,"22620",,terminal_output +3076,4465170,"TERMINAL",0,0,"3371",,terminal_output +3077,4466218,"TERMINAL",0,0,"4482",,terminal_output +3078,4467268,"TERMINAL",0,0,"5593",,terminal_output +3079,4468323,"TERMINAL",0,0,"66204",,terminal_output +3080,4469406,"TERMINAL",0,0,"7715",,terminal_output +3081,4470414,"TERMINAL",0,0,"8826",,terminal_output +3082,4471462,"TERMINAL",0,0,"9937",,terminal_output +3083,4472513,"TERMINAL",0,0,"102048",,terminal_output +3084,4473578,"TERMINAL",0,0,"1159",,terminal_output +3085,4474615,"TERMINAL",0,0,"22630",,terminal_output +3086,4475662,"TERMINAL",0,0,"3371",,terminal_output +3087,4477048,"TERMINAL",0,0,"4482",,terminal_output +3088,4477748,"TERMINAL",0,0,"56304",,terminal_output +3089,4478797,"TERMINAL",0,0,"7715",,terminal_output +3090,4479836,"TERMINAL",0,0,"8826",,terminal_output +3091,4480890,"TERMINAL",0,0,"9937",,terminal_output +3092,4481934,"TERMINAL",0,0,"203048",,terminal_output +3093,4482987,"TERMINAL",0,0,"1159",,terminal_output +3094,4484065,"TERMINAL",0,0,"22640",,terminal_output +3095,4485099,"TERMINAL",0,0,"3371",,terminal_output +3096,4486148,"TERMINAL",0,0,"4482",,terminal_output +3097,4487336,"TERMINAL",0,0,"5593",,terminal_output +3098,4488394,"TERMINAL",0,0,"66404",,terminal_output +3099,4489446,"TERMINAL",0,0,"7715",,terminal_output +3100,4490507,"TERMINAL",0,0,"8826",,terminal_output +3101,4491555,"TERMINAL",0,0,"9937",,terminal_output +3102,4492613,"TERMINAL",0,0,"304048",,terminal_output +3103,4493667,"TERMINAL",0,0,"1159",,terminal_output +3104,4494733,"TERMINAL",0,0,"23751",,terminal_output +3105,4495798,"TERMINAL",0,0,"4482",,terminal_output +3106,4496830,"TERMINAL",0,0,"5593",,terminal_output +3107,4497883,"TERMINAL",0,0,"66504",,terminal_output +3108,4498935,"TERMINAL",0,0,"7715",,terminal_output +3109,4499973,"TERMINAL",0,0,"8826",,terminal_output +3110,4501015,"TERMINAL",0,0,"9937",,terminal_output +3111,4502060,"TERMINAL",0,0,"405048",,terminal_output +3112,4503132,"TERMINAL",0,0,"1159",,terminal_output +3113,4504200,"TERMINAL",0,0,"2269:00",,terminal_output +3114,4505212,"TERMINAL",0,0,"3371",,terminal_output +3115,4506253,"TERMINAL",0,0,"4482",,terminal_output +3116,4507298,"TERMINAL",0,0,"5593",,terminal_output +3117,4508349,"TERMINAL",0,0,"669:004",,terminal_output +3118,4509410,"TERMINAL",0,0,"7715",,terminal_output +3119,4510444,"TERMINAL",0,0,"8826",,terminal_output +3120,4511518,"TERMINAL",0,0,"9937",,terminal_output +3121,4512555,"TERMINAL",0,0,"509:0048",,terminal_output +3122,4513604,"TERMINAL",0,0,"1159",,terminal_output +3123,4514653,"TERMINAL",0,0,"22610",,terminal_output +3124,4515693,"TERMINAL",0,0,"3371",,terminal_output +3125,4516740,"TERMINAL",0,0,"4593",,terminal_output +3126,4517796,"TERMINAL",0,0,"66104",,terminal_output +3127,4518854,"TERMINAL",0,0,"7715",,terminal_output +3128,4519892,"TERMINAL",0,0,"8826",,terminal_output +3129,4520945,"TERMINAL",0,0,"9937",,terminal_output +3130,4522000,"TERMINAL",0,0,"7:001048",,terminal_output +3131,4523087,"TERMINAL",0,0,"1159",,terminal_output +3132,4524095,"TERMINAL",0,0,"22620",,terminal_output +3133,4525153,"TERMINAL",0,0,"3371",,terminal_output +3134,4526199,"TERMINAL",0,0,"4482",,terminal_output +3135,4527256,"TERMINAL",0,0,"5593",,terminal_output +3136,4528302,"TERMINAL",0,0,"66204",,terminal_output +3137,4529348,"TERMINAL",0,0,"7715",,terminal_output +3138,4530395,"TERMINAL",0,0,"8826",,terminal_output +3139,4531444,"TERMINAL",0,0,"9937",,terminal_output +3140,4532499,"TERMINAL",0,0,"102048",,terminal_output +3141,4533553,"TERMINAL",0,0,"1159",,terminal_output +3142,4534607,"TERMINAL",0,0,"22630",,terminal_output +3143,4535665,"TERMINAL",0,0,"3371",,terminal_output +3144,4536717,"TERMINAL",0,0,"4482",,terminal_output +3145,4537776,"TERMINAL",0,0,"66304",,terminal_output +3146,4538855,"TERMINAL",0,0,"7715",,terminal_output +3147,4539874,"TERMINAL",0,0,"8826",,terminal_output +3148,4540928,"TERMINAL",0,0,"9937",,terminal_output +3149,4541984,"TERMINAL",0,0,"203048",,terminal_output +3150,4543077,"TERMINAL",0,0,"1159",,terminal_output +3151,4544088,"TERMINAL",0,0,"22640",,terminal_output +3152,4545138,"TERMINAL",0,0,"3371",,terminal_output +3153,4546180,"TERMINAL",0,0,"4482",,terminal_output +3154,4547232,"TERMINAL",0,0,"5593",,terminal_output +3155,4548286,"TERMINAL",0,0,"66404",,terminal_output +3156,4549333,"TERMINAL",0,0,"7715",,terminal_output +3157,4550534,"TERMINAL",0,0,"8826",,terminal_output +3158,4551577,"TERMINAL",0,0,"9937",,terminal_output +3159,4552619,"TERMINAL",0,0,"304048",,terminal_output +3160,4553673,"TERMINAL",0,0,"1159",,terminal_output +3161,4554735,"TERMINAL",0,0,"23751",,terminal_output +3162,4555792,"TERMINAL",0,0,"4482",,terminal_output +3163,4556844,"TERMINAL",0,0,"5593",,terminal_output +3164,4557892,"TERMINAL",0,0,"66504",,terminal_output +3165,4558945,"TERMINAL",0,0,"7715",,terminal_output +3166,4559999,"TERMINAL",0,0,"8826",,terminal_output +3167,4561089,"TERMINAL",0,0,"9937",,terminal_output +3168,4562102,"TERMINAL",0,0,"405048",,terminal_output +3169,4563153,"TERMINAL",0,0,"1159",,terminal_output +3170,4564289,"TERMINAL",0,0,"22630:00",,terminal_output +3171,4565267,"TERMINAL",0,0,"3371",,terminal_output +3172,4566318,"TERMINAL",0,0,"4482",,terminal_output +3173,4567368,"TERMINAL",0,0,"5593",,terminal_output +3174,4568413,"TERMINAL",0,0,"6630:004",,terminal_output +3175,4569451,"TERMINAL",0,0,"7715",,terminal_output +3176,4570497,"TERMINAL",0,0,"8826",,terminal_output +3177,4572558,"TERMINAL",0,0,"930:0048",,terminal_output +3178,4573619,"TERMINAL",0,0,"51159",,terminal_output +3179,4574661,"TERMINAL",0,0,"22610",,terminal_output +3180,4575711,"TERMINAL",0,0,"3371",,terminal_output +3181,4576765,"TERMINAL",0,0,"4593",,terminal_output +3182,4577797,"TERMINAL",0,0,"66104",,terminal_output +3183,4578849,"TERMINAL",0,0,"7715",,terminal_output +3184,4579890,"TERMINAL",0,0,"8826",,terminal_output +3185,4580940,"TERMINAL",0,0,"9937",,terminal_output +3186,4581996,"TERMINAL",0,0,"8:001048",,terminal_output +3187,4583048,"TERMINAL",0,0,"1159",,terminal_output +3188,4584130,"TERMINAL",0,0,"22620",,terminal_output +3189,4585146,"TERMINAL",0,0,"3371",,terminal_output +3190,4586198,"TERMINAL",0,0,"4482",,terminal_output +3191,4587246,"TERMINAL",0,0,"5593",,terminal_output +3192,4588297,"TERMINAL",0,0,"66204",,terminal_output +3193,4589344,"TERMINAL",0,0,"7715",,terminal_output +3194,4590384,"TERMINAL",0,0,"8826",,terminal_output +3195,4591436,"TERMINAL",0,0,"9937",,terminal_output +3196,4592485,"TERMINAL",0,0,"102048",,terminal_output +3197,4593538,"TERMINAL",0,0,"1159",,terminal_output +3198,4594575,"TERMINAL",0,0,"22630",,terminal_output +3199,4595626,"TERMINAL",0,0,"3371",,terminal_output +3200,4596661,"TERMINAL",0,0,"4482",,terminal_output +3201,4597698,"TERMINAL",0,0,"5593",,terminal_output +3202,4598746,"TERMINAL",0,0,"67315",,terminal_output +3203,4599794,"TERMINAL",0,0,"8826",,terminal_output +3204,4600835,"TERMINAL",0,0,"9937",,terminal_output +3205,4601882,"TERMINAL",0,0,"203048",,terminal_output +3206,4602939,"TERMINAL",0,0,"1159",,terminal_output +3207,4603959,"TERMINAL",0,0,"22640",,terminal_output +3208,4605003,"TERMINAL",0,0,"3371",,terminal_output +3209,4606056,"TERMINAL",0,0,"4482",,terminal_output +3210,4607111,"TERMINAL",0,0,"5593",,terminal_output +3211,4608154,"TERMINAL",0,0,"66404",,terminal_output +3212,4610473,"TERMINAL",0,0,"7715",,terminal_output +3213,4610816,"TERMINAL",0,0,"8826",,terminal_output +3214,4612117,"TERMINAL",0,0,"9937",,terminal_output +3215,4612421,"TERMINAL",0,0,"304048",,terminal_output +3216,4613369,"TERMINAL",0,0,"1159",,terminal_output +3217,4614421,"TERMINAL",0,0,"22650",,terminal_output +3218,4615490,"TERMINAL",0,0,"3371",,terminal_output +3219,4616527,"TERMINAL",0,0,"4482",,terminal_output +3220,4617583,"TERMINAL",0,0,"5593",,terminal_output +3221,4618628,"TERMINAL",0,0,"66504",,terminal_output +3222,4619688,"TERMINAL",0,0,"7715",,terminal_output +3223,4620773,"TERMINAL",0,0,"8937",,terminal_output +3224,4621777,"TERMINAL",0,0,"405048",,terminal_output +3225,4622827,"TERMINAL",0,0,"1159",,terminal_output +3226,4623871,"TERMINAL",0,0,"2261:00",,terminal_output +3227,4624928,"TERMINAL",0,0,"3371",,terminal_output +3228,4625979,"TERMINAL",0,0,"4482",,terminal_output +3229,4627022,"TERMINAL",0,0,"5593",,terminal_output +3230,4628109,"TERMINAL",0,0,"661:004",,terminal_output +3231,4629130,"TERMINAL",0,0,"7715",,terminal_output +3232,4630165,"TERMINAL",0,0,"8826",,terminal_output +3233,4631215,"TERMINAL",0,0,"9937",,terminal_output +3234,4632262,"TERMINAL",0,0,"501:0048",,terminal_output +3235,4633314,"TERMINAL",0,0,"1159",,terminal_output +3236,4634360,"TERMINAL",0,0,"22610",,terminal_output +3237,4635408,"TERMINAL",0,0,"3371",,terminal_output +3238,4636448,"TERMINAL",0,0,"4482",,terminal_output +3239,4637498,"TERMINAL",0,0,"5593",,terminal_output +3240,4638517,"TERMINAL",0,0,"66104",,terminal_output +3241,4639551,"TERMINAL",0,0,"7715",,terminal_output +3242,4640608,"TERMINAL",0,0,"8826",,terminal_output +3243,4641648,"TERMINAL",0,0,"9937",,terminal_output +3244,4642699,"TERMINAL",0,0,"9:001048",,terminal_output +3245,4643734,"TERMINAL",0,0,"12620",,terminal_output +3246,4644777,"TERMINAL",0,0,"3371",,terminal_output +3247,4645810,"TERMINAL",0,0,"4482",,terminal_output +3248,4646862,"TERMINAL",0,0,"5593",,terminal_output +3249,4647913,"TERMINAL",0,0,"66204",,terminal_output +3250,4648947,"TERMINAL",0,0,"7715",,terminal_output +3251,4649999,"TERMINAL",0,0,"8826",,terminal_output +3252,4651045,"TERMINAL",0,0,"9937",,terminal_output +3253,4652113,"TERMINAL",0,0,"102048",,terminal_output +3254,4653128,"TERMINAL",0,0,"1159",,terminal_output +3255,4654172,"TERMINAL",0,0,"22630",,terminal_output +3256,4655219,"TERMINAL",0,0,"3371",,terminal_output +3257,4656269,"TERMINAL",0,0,"4482",,terminal_output +3258,4657311,"TERMINAL",0,0,"5593",,terminal_output +3259,4658357,"TERMINAL",0,0,"66304",,terminal_output +3260,4659406,"TERMINAL",0,0,"7715",,terminal_output +3261,4660463,"TERMINAL",0,0,"8826",,terminal_output +3262,4661532,"TERMINAL",0,0,"9937",,terminal_output +3263,4662552,"TERMINAL",0,0,"203048",,terminal_output +3264,4663605,"TERMINAL",0,0,"1159",,terminal_output +3265,4664656,"TERMINAL",0,0,"22640",,terminal_output +3266,4665677,"TERMINAL",0,0,"3371",,terminal_output +3267,4666751,"TERMINAL",0,0,"4593",,terminal_output +3268,4667779,"TERMINAL",0,0,"66404",,terminal_output +3269,4668834,"TERMINAL",0,0,"7715",,terminal_output +3270,4669868,"TERMINAL",0,0,"8826",,terminal_output +3271,4670913,"TERMINAL",0,0,"9937",,terminal_output +3272,4671956,"TERMINAL",0,0,"304048",,terminal_output +3273,4673004,"TERMINAL",0,0,"1159",,terminal_output +3274,4674055,"TERMINAL",0,0,"22650",,terminal_output +3275,4675154,"TERMINAL",0,0,"3371",,terminal_output +3276,4676160,"TERMINAL",0,0,"4482",,terminal_output +3277,4677218,"TERMINAL",0,0,"5593",,terminal_output +3278,4678261,"TERMINAL",0,0,"66504",,terminal_output +3279,4679338,"TERMINAL",0,0,"7715",,terminal_output +3280,4680382,"TERMINAL",0,0,"8826",,terminal_output +3281,4681434,"TERMINAL",0,0,"9937",,terminal_output +3282,4682481,"TERMINAL",0,0,"405048",,terminal_output +3283,4683531,"TERMINAL",0,0,"1159",,terminal_output +3284,4684581,"TERMINAL",0,0,"2262:00",,terminal_output +3285,4685617,"TERMINAL",0,0,"3371",,terminal_output +3286,4686665,"TERMINAL",0,0,"4482",,terminal_output +3287,4687715,"TERMINAL",0,0,"5593",,terminal_output +3288,4688765,"TERMINAL",0,0,"772:015",,terminal_output +3289,4689804,"TERMINAL",0,0,"8826",,terminal_output +3290,4690849,"TERMINAL",0,0,"9937",,terminal_output +3291,4691894,"TERMINAL",0,0,"502:0048",,terminal_output +3292,4693519,"TERMINAL",0,0,"1159",,terminal_output +3293,4694570,"TERMINAL",0,0,"22610",,terminal_output +3294,4695634,"TERMINAL",0,0,"3371",,terminal_output +3295,4696675,"TERMINAL",0,0,"4482",,terminal_output +3296,4697706,"TERMINAL",0,0,"5593",,terminal_output +3297,4698750,"TERMINAL",0,0,"67115",,terminal_output +3298,4699797,"TERMINAL",0,0,"8826",,terminal_output +3299,4700843,"TERMINAL",0,0,"9937",,terminal_output +3300,4701891,"TERMINAL",0,0,"4:00:001048",,terminal_output +3301,4702939,"TERMINAL",0,0,"1159",,terminal_output +3302,4703994,"TERMINAL",0,0,"22620",,terminal_output +3303,4705048,"TERMINAL",0,0,"3371",,terminal_output +3304,4706090,"TERMINAL",0,0,"4482",,terminal_output +3305,4707194,"TERMINAL",0,0,"5593",,terminal_output +3306,4708245,"TERMINAL",0,0,"66204",,terminal_output +3307,4709286,"TERMINAL",0,0,"7715",,terminal_output +3308,4710346,"TERMINAL",0,0,"8826",,terminal_output +3309,4711401,"TERMINAL",0,0,"9937",,terminal_output +3310,4712448,"TERMINAL",0,0,"102048",,terminal_output +3311,4713505,"TERMINAL",0,0,"1159",,terminal_output +3312,4714550,"TERMINAL",0,0,"22630",,terminal_output +3313,4715751,"TERMINAL",0,0,"3371",,terminal_output +3314,4716637,"TERMINAL",0,0,"4482",,terminal_output +3315,4717993,"TERMINAL",0,0,"5593",,terminal_output +3316,4718743,"TERMINAL",0,0,"67315",,terminal_output +3317,4719779,"TERMINAL",0,0,"8826",,terminal_output +3318,4720822,"TERMINAL",0,0,"9937",,terminal_output +3319,4721876,"TERMINAL",0,0,"203048",,terminal_output +3320,4722924,"TERMINAL",0,0,"1159",,terminal_output +3321,4724186,"TERMINAL",0,0,"22640",,terminal_output +3322,4725020,"TERMINAL",0,0,"3371",,terminal_output +3323,4726078,"TERMINAL",0,0,"4482",,terminal_output +3324,4727160,"TERMINAL",0,0,"5593",,terminal_output +3325,4728178,"TERMINAL",0,0,"66404",,terminal_output +3326,4729233,"TERMINAL",0,0,"7715",,terminal_output +3327,4730282,"TERMINAL",0,0,"8826",,terminal_output +3328,4731329,"TERMINAL",0,0,"9937",,terminal_output +3329,4732390,"TERMINAL",0,0,"304048",,terminal_output +3330,4733425,"TERMINAL",0,0,"1159",,terminal_output +3331,4734473,"TERMINAL",0,0,"22650",,terminal_output +3332,4735519,"TERMINAL",0,0,"3371",,terminal_output +3333,4736570,"TERMINAL",0,0,"4482",,terminal_output +3334,4737615,"TERMINAL",0,0,"5593",,terminal_output +3335,4738787,"TERMINAL",0,0,"66504",,terminal_output +3336,4739026,"train_tokenizer.py",0,0,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n init_lr: float = 0.0\n max_lr: float = 3e-4\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 20000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n ffn_dim: int = 2048\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 4\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\n\ndef tokenizer_loss_fn(\n model: TokenizerVQVAE, inputs: dict\n) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n # --- Compute loss ---\n # FIXME (f.srambical): Can we even do native int8 training without casting the video at all?\n # FIXME (f.srambical): If the tokenizer is the reason for the dynamics model being memory-bound,\n # should we at least train the tokenizer natively in int8?\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n model.train()\n outputs = model(inputs, training=True)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n mse = jnp.square(gt - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n\n@nnx.jit\ndef train_step(\n tokenizer: TokenizerVQVAE, optimizer: nnx.Optimizer, inputs: dict\n) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(model: TokenizerVQVAE) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n return tokenizer_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n tokenizer\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n ffn_dim=args.ffn_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n\n _, params, _ = nnx.split(tokenizer, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n optimizer = nnx.Optimizer(tokenizer, tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in grain_iterator\n )\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n inputs = dict(videos=videos)\n loss, recon, metrics = train_step(tokenizer, optimizer, inputs)\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n grain_iterator # type: ignore\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +3337,4739718,"TERMINAL",0,0,"7715",,terminal_output +3338,4740705,"train_tokenizer.py",597,0,"",python,selection_mouse +3339,4740723,"train_tokenizer.py",596,0,"",python,selection_command +3340,4740776,"TERMINAL",0,0,"8937",,terminal_output +3341,4741838,"TERMINAL",0,0,"405048",,terminal_output +3342,4742870,"TERMINAL",0,0,"1159",,terminal_output +3343,4743909,"TERMINAL",0,0,"2263:00",,terminal_output +3344,4744971,"TERMINAL",0,0,"3371",,terminal_output +3345,4746013,"TERMINAL",0,0,"4482",,terminal_output +3346,4747070,"TERMINAL",0,0,"5593",,terminal_output +3347,4748159,"TERMINAL",0,0,"663:004",,terminal_output +3348,4749191,"TERMINAL",0,0,"7715",,terminal_output +3349,4750197,"TERMINAL",0,0,"8826",,terminal_output +3350,4751240,"TERMINAL",0,0,"9937",,terminal_output +3351,4752292,"TERMINAL",0,0,"503:0048",,terminal_output +3352,4753390,"TERMINAL",0,0,"1159",,terminal_output +3353,4754414,"TERMINAL",0,0,"22610",,terminal_output +3354,4755437,"TERMINAL",0,0,"3371",,terminal_output +3355,4756483,"train_dynamics.py",0,0,"",python,tab +3356,4756985,"TERMINAL",0,0,"4593",,terminal_output +3357,4758116,"TERMINAL",0,0,"66104",,terminal_output +3358,4759056,"TERMINAL",0,0,"7715",,terminal_output +3359,4760088,"TERMINAL",0,0,"8826",,terminal_output +3360,4760834,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +3361,4762438,"TERMINAL",0,0,"idling",,terminal_command +3362,4762537,"TERMINAL",0,0,"]633;E;2025-09-01 14:01:00 idling;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Mon Sep 1 14:01:00 2025Partition dev_cpuonly: 12 nodes idle\rPartition cpuonly: 107 nodes idle\rPartition dev_accelerated:\t 0 nodes idle\rPartition accelerated:\t 8 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 0 nodes idle\rPartition accelerated-h200:\t 0 nodes idle",,terminal_output +3363,4763559,"TERMINAL",0,0,"1",,terminal_output +3364,4764602,"TERMINAL",0,0,"2",,terminal_output +3365,4765642,"TERMINAL",0,0,"3",,terminal_output +3366,4766687,"TERMINAL",0,0,"4",,terminal_output +3367,4767505,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +3368,4799115,"train_dynamics.py",0,0,"",python,tab +3369,4800180,"train_dynamics.py",1155,0,"",python,selection_mouse +3370,4800681,"train_dynamics.py",1213,0,"",python,selection_mouse +3371,4802891,"train_tokenizer.py",0,0,"",python,tab +3372,4804170,"train_tokenizer.py",1198,0,"",python,selection_mouse +3373,4805914,"train_tokenizer.py",1214,0,"\n darkness_threshold: float = 0.0",python,content +3374,4805917,"train_tokenizer.py",1219,0,"",python,selection_command +3375,4807447,"train_dynamics.py",0,0,"",python,tab +3376,4816990,"train_dynamics.py",10510,0,"",python,selection_mouse +3377,4820047,"train_tokenizer.py",0,0,"",python,tab +3378,4828118,"train_tokenizer.py",9320,0,"",python,selection_mouse +3379,4828824,"train_tokenizer.py",9326,0,"\n darkness_threshold=args.darkness_threshold,",python,content +3380,4828862,"train_tokenizer.py",9335,0,"",python,selection_command +3381,4831959,"train_tokenizer.py",9098,0,"",python,selection_mouse +3382,4832444,"utils/dataloader.py",0,0,"import jax\nimport numpy as np\nimport grain\nfrom typing import Any\nimport pickle\n\n\nclass EpisodeLengthFilter(grain.transforms.Filter):\n """"""\n A Grain Filter that keeps only episodes with sufficient length.\n """"""\n\n def __init__(self, seq_len: int, image_h: int, image_w: int, image_c: int):\n """"""Initializes the filter with sequence length requirements.""""""\n self.seq_len = seq_len\n self.image_h = image_h\n self.image_w = image_w\n self.image_c = image_c\n\n def filter(self, element: Any) -> bool:\n """"""\n Filters episodes based on length.\n\n Args:\n element: A dictionary representing one record from the DataSource.\n Expected to contain 'raw_video' (bytes) and 'sequence_length' (int)\n\n Returns:\n True if the episode has sufficient length, False otherwise.\n """"""\n assert isinstance(element, bytes)\n element = pickle.loads(element)\n\n current_episode_len = element[""sequence_length""]\n if current_episode_len < self.seq_len:\n print(\n f""Filtering out episode with length {current_episode_len}, which is ""\n f""shorter than the requested sequence length {self.seq_len}.""\n )\n return False\n\n return True\n\n\nclass ProcessEpisodeAndSlice(grain.transforms.RandomMap):\n """"""\n A Grain Transformation that combines parsing, slicing, and normalizing.\n """"""\n\n def __init__(self, seq_len: int, image_h: int, image_w: int, image_c: int):\n """"""Initializes the transformation with processing parameters.""""""\n self.seq_len = seq_len\n self.image_h = image_h\n self.image_w = image_w\n self.image_c = image_c\n\n def random_map(self, element: dict, rng: np.random.Generator) -> Any:\n """"""\n Processes a single raw episode from the data source.\n\n Args:\n element: A dictionary representing one record from the DataSource.\n Expected to contain 'raw_video' (bytes) and 'sequence_length' (int)\n rng: A per-record random number generator provided by the Grain sampler.\n\n Returns:\n A processed video sequence as a NumPy array with shape\n (seq_len, height, width, channels) and dtype float32.\n """"""\n assert isinstance(element, bytes)\n element = pickle.loads(element)\n\n video_shape = (\n element[""sequence_length""],\n self.image_h,\n self.image_w,\n self.image_c,\n )\n episode_tensor = np.frombuffer(element[""raw_video""], dtype=np.uint8)\n episode_tensor = episode_tensor.reshape(video_shape)\n\n current_episode_len = episode_tensor.shape[0]\n if current_episode_len < self.seq_len:\n raise ValueError(\n f""Episode length {current_episode_len} is shorter than ""\n f""requested sequence length {self.seq_len}. This should ""\n f""have been filtered out.""\n )\n\n max_start_idx = current_episode_len - self.seq_len\n\n start_idx = rng.integers(0, max_start_idx + 1)\n\n seq = episode_tensor[start_idx : start_idx + self.seq_len]\n\n return seq\n\n\nclass DarknessFilter(grain.transforms.Filter):\n """"""\n A Grain Filter that filters out sequences with images that are too dark.\n """"""\n\n def __init__(self, darkness_threshold: float):\n """"""Initializes the filter with darkness threshold.""""""\n self.darkness_threshold = darkness_threshold\n\n def filter(self, element: Any) -> bool:\n """"""\n Filters sequences based on darkness.\n\n Args:\n element: A NumPy array representing a processed video sequence.\n\n Returns:\n True if the sequence is not too dark, False otherwise.\n """"""\n # Convert the RGB image to grayscale using numpy\n element_greyscale = np.dot(element[...,:3], [0.2989, 0.5870, 0.1140])\n average_brightness = np.mean(element_greyscale)\n if average_brightness < self.darkness_threshold:\n print(\n f""Filtering out sequence with average brightness {average_brightness}, ""\n f""which is below the darkness threshold {self.darkness_threshold}.""\n )\n return False\n\n return True\n\n\ndef get_dataloader(\n array_record_paths: list[str],\n seq_len: int,\n global_batch_size: int,\n image_h: int,\n image_w: int,\n image_c: int,\n darkness_threshold: float = 0.,\n num_workers: int = 1,\n prefetch_buffer_size: int = 1,\n seed: int = 42,\n):\n """"""\n Creates a data loading pipeline using Grain.\n """"""\n if not array_record_paths:\n raise ValueError(""array_record_paths list cannot be empty."")\n\n num_processes = jax.process_count()\n\n if global_batch_size % num_processes != 0:\n raise ValueError(\n f""Global batch size {global_batch_size} must be divisible by ""\n f""the number of JAX processes {num_processes} for proper sharding.""\n )\n per_process_batch_size = global_batch_size // num_processes\n\n source = grain.sources.ArrayRecordDataSource(array_record_paths)\n\n sampler = grain.samplers.IndexSampler(\n num_records=len(source),\n shard_options=grain.sharding.ShardByJaxProcess(drop_remainder=True),\n shuffle=True,\n num_epochs=None,\n seed=seed,\n )\n\n operations = [\n EpisodeLengthFilter(\n seq_len=seq_len, image_h=image_h, image_w=image_w, image_c=image_c\n ),\n ProcessEpisodeAndSlice(\n seq_len=seq_len, image_h=image_h, image_w=image_w, image_c=image_c\n ),\n DarknessFilter(\n darkness_threshold=darkness_threshold\n ),\n grain.transforms.Batch(batch_size=per_process_batch_size, drop_remainder=True),\n ]\n\n read_options = grain.ReadOptions(\n prefetch_buffer_size=prefetch_buffer_size,\n num_threads=1,\n )\n dataloader = grain.DataLoader(\n data_source=source,\n sampler=sampler,\n operations=operations,\n worker_count=num_workers,\n worker_buffer_size=1,\n read_options=read_options,\n )\n\n return dataloader\n",python,tab +3383,4896534,"utils/dataloader.py",75,0,"",python,selection_mouse +3384,4897049,"utils/dataloader.py",80,0,"",python,selection_mouse +3385,4898863,"utils/dataloader.py",80,0,"\n",python,content +3386,4899279,"utils/dataloader.py",81,0,"RGB_TO_GRAYSCALE_WEIGHTS",python,content +3387,4900432,"utils/dataloader.py",105,0," ",python,content +3388,4900433,"utils/dataloader.py",106,0,"",python,selection_keyboard +3389,4900607,"utils/dataloader.py",106,0,"=",python,content +3390,4900608,"utils/dataloader.py",107,0,"",python,selection_keyboard +3391,4900717,"utils/dataloader.py",107,0," ",python,content +3392,4900718,"utils/dataloader.py",108,0,"",python,selection_keyboard +3393,4905137,"utils/dataloader.py",3967,0,"",python,selection_mouse +3394,4905306,"utils/dataloader.py",3967,1,"[",python,selection_mouse +3395,4905306,"utils/dataloader.py",3967,2,"[0",python,selection_mouse +3396,4905308,"utils/dataloader.py",3967,4,"[0.2",python,selection_mouse +3397,4905330,"utils/dataloader.py",3967,6,"[0.298",python,selection_mouse +3398,4905340,"utils/dataloader.py",3967,7,"[0.2989",python,selection_mouse +3399,4905373,"utils/dataloader.py",3967,10,"[0.2989, 0",python,selection_mouse +3400,4905374,"utils/dataloader.py",3967,11,"[0.2989, 0.",python,selection_mouse +3401,4905407,"utils/dataloader.py",3967,13,"[0.2989, 0.58",python,selection_mouse +3402,4905407,"utils/dataloader.py",3967,15,"[0.2989, 0.5870",python,selection_mouse +3403,4905449,"utils/dataloader.py",3967,16,"[0.2989, 0.5870,",python,selection_mouse +3404,4905480,"utils/dataloader.py",3967,17,"[0.2989, 0.5870, ",python,selection_mouse +3405,4905481,"utils/dataloader.py",3967,18,"[0.2989, 0.5870, 0",python,selection_mouse +3406,4905514,"utils/dataloader.py",3967,19,"[0.2989, 0.5870, 0.",python,selection_mouse +3407,4905549,"utils/dataloader.py",3967,20,"[0.2989, 0.5870, 0.1",python,selection_mouse +3408,4905583,"utils/dataloader.py",3967,21,"[0.2989, 0.5870, 0.11",python,selection_mouse +3409,4905622,"utils/dataloader.py",3967,22,"[0.2989, 0.5870, 0.114",python,selection_mouse +3410,4905662,"utils/dataloader.py",3967,23,"[0.2989, 0.5870, 0.1140",python,selection_mouse +3411,4905824,"utils/dataloader.py",3967,24,"[0.2989, 0.5870, 0.1140]",python,selection_mouse +3412,4908040,"utils/dataloader.py",3967,24,"",python,content +3413,4908042,"utils/dataloader.py",3967,0,"",python,selection_keyboard +3414,4910195,"utils/dataloader.py",108,0,"",python,selection_mouse +3415,4910901,"utils/dataloader.py",108,0,"[0.2989, 0.5870, 0.1140]",python,content +3416,4913345,"utils/dataloader.py",132,0,")",python,content +3417,4913345,"utils/dataloader.py",108,0,"np.array(",python,content +3418,4917942,"utils/dataloader.py",4001,0,"RGB_TO_GRAYSCALE_WEIGHTS",python,content +3419,4917943,"utils/dataloader.py",4025,0,"",python,selection_command +3420,4925122,"utils/dataloader.py",80,0,"",python,selection_mouse +3421,4925851,"utils/dataloader.py",143,0,"",python,selection_mouse +3422,4926606,"utils/dataloader.py",115,0,"",python,selection_mouse +3423,4926760,"utils/dataloader.py",111,5,"array",python,selection_mouse +3424,4926942,"utils/dataloader.py",110,6,".array",python,selection_mouse +3425,4926983,"utils/dataloader.py",108,8,"np.array",python,selection_mouse +3426,4927064,"utils/dataloader.py",107,9," np.array",python,selection_mouse +3427,4928821,"utils/dataloader.py",108,8,"np.array",python,selection_mouse +3428,4931605,"utils/dataloader.py",108,8,"",python,content +3429,4931747,"utils/dataloader.py",108,1,"",python,content +3430,4932887,"utils/dataloader.py",133,0,"",python,selection_mouse +3431,4933189,"utils/dataloader.py",132,1,"",python,content +3432,4938425,"utils/dataloader.py",4288,0,"",python,selection_mouse +3433,4939111,"utils/dataloader.py",4202,0,"",python,selection_mouse +3434,4958128,"utils/dataloader.py",108,0,"",python,selection_mouse +3435,4959120,"utils/dataloader.py",132,0,")",python,content +3436,4959492,"utils/dataloader.py",108,0,"np.array(",python,content +3437,5211908,"train_tokenizer.py",0,0,"",python,tab +3438,5220759,"TERMINAL",0,0,"idling",,terminal_command +3439,5220831,"TERMINAL",0,0,"]633;E;2025-09-01 14:08:39 idling;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Mon Sep 1 14:08:39 2025Partition dev_cpuonly: 12 nodes idle\rPartition cpuonly: 108 nodes idle\rPartition dev_accelerated:\t 0 nodes idle\rPartition accelerated:\t 1 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 0 nodes idle\rPartition accelerated-h200:\t 0 nodes idle",,terminal_output +3440,5221866,"TERMINAL",0,0,"40",,terminal_output +3441,5222917,"TERMINAL",0,0,"1",,terminal_output +3442,5223432,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +3443,5224311,"TERMINAL",0,0,"queue",,terminal_command +3444,5224395,"TERMINAL",0,0,"]633;E;2025-09-01 14:08:42 queue;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 14:08:42 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454956 accelerat train_la tum_cte0 R\t4:54\t 8 hkn[0410,0530-0533,0536,0604,0813]3454955 accelerat train_la tum_cte0 R40:52\t 8 hkn[0504,0520,0720,0722-0724,0728,0731]3454954 accelerat train_la tum_cte0 R40:56\t 8 hkn[0521-0528]3454953 accelerat train_la tum_cte0 R41:00\t 8 hkn[0703,0706-0707,0711-0715]",,terminal_output +3445,5225438,"TERMINAL",0,0,"35371",,terminal_output +3446,5226485,"TERMINAL",0,0,"46482",,terminal_output +3447,5227530,"TERMINAL",0,0,"57593",,terminal_output +3448,5228467,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs",,terminal_output +3449,5276839,"TERMINAL",0,0,"idling",,terminal_command +3450,5276948,"TERMINAL",0,0,"]633;E;2025-09-01 14:09:35 idling;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Mon Sep 1 14:09:35 2025Partition dev_cpuonly: 12 nodes idle\rPartition cpuonly: 108 nodes idle\rPartition dev_accelerated:\t 0 nodes idle\rPartition accelerated:\t 1 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 0 nodes idle\rPartition accelerated-h200:\t 0 nodes idle",,terminal_output +3451,5277969,"TERMINAL",0,0,"6",,terminal_output +3452,5278900,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output +3453,5279757,"TERMINAL",0,0,"queue",,terminal_command +3454,5279845,"TERMINAL",0,0,"]633;E;2025-09-01 14:09:38 queue;2bb8f881-ac5e-40fe-80c9-f1201978c519]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Mon Sep 1 14:09:38 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3454956 accelerat train_la tum_cte0 R\t5:50\t 8 hkn[0410,0530-0533,0536,0604,0813]3454955 accelerat train_la tum_cte0 R41:48\t 8 hkn[0504,0520,0720,0722-0724,0728,0731]3454954 accelerat train_la tum_cte0 R41:52\t 8 hkn[0521-0528]3454953 accelerat train_la tum_cte0 R41:56\t 8 hkn[0703,0706-0707,0711-0715]",,terminal_output +3455,5280885,"TERMINAL",0,0,"91937",,terminal_output +3456,5281936,"TERMINAL",0,0,"4025048",,terminal_output +3457,5282983,"TERMINAL",0,0,"13159",,terminal_output +3458,5284037,"TERMINAL",0,0,"24262:00",,terminal_output +3459,5285087,"TERMINAL",0,0,"35371",,terminal_output +3460,5286135,"TERMINAL",0,0,"46482",,terminal_output +3461,5287192,"TERMINAL",0,0,"57593",,terminal_output +3462,5288114,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar_jobs]633;D;0",,terminal_output diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-50f57e52-47d1-4d1e-826b-d55baff9760f1752091399803-2025_07_09-22.05.03.836/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-50f57e52-47d1-4d1e-826b-d55baff9760f1752091399803-2025_07_09-22.05.03.836/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..4e05af01f2e88835dce553040f6300c04da24ec1 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-50f57e52-47d1-4d1e-826b-d55baff9760f1752091399803-2025_07_09-22.05.03.836/source.csv @@ -0,0 +1,844 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +2,5413,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:05:03 PM [info] Activating crowd-code\n10:05:03 PM [info] Recording started\n10:05:03 PM [info] Initializing git provider using file system watchers...\n10:05:04 PM [info] Git repository found\n10:05:04 PM [info] Git provider initialized successfully\n10:05:04 PM [info] Initial git state: [object Object]\n",Log,tab +3,5417,"TERMINAL",0,0,"bash",,terminal_focus +4,5417,"TERMINAL",0,0,"bash",,terminal_focus +5,6124,"TERMINAL",0,0,"cd $ws_dir",,terminal_command +6,6170,"TERMINAL",0,0,"]633;E;2025-07-09 22:05:09 cd $ws_dir;af8e09eb-6e10-4bc6-b809-4360d3d8a533]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared]633;D;0",,terminal_output +7,6750,"TERMINAL",0,0,"ls",,terminal_command +8,6793,"TERMINAL",0,0,"]633;E;2025-07-09 22:05:10 ls;af8e09eb-6e10-4bc6-b809-4360d3d8a533]633;C",,terminal_output +9,6877,"TERMINAL",0,0,"checkpoints count_items.sh data data_new huggingface logs scripts\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared]633;D;0",,terminal_output +10,9152,"TERMINAL",0,0,"/bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt",,terminal_command +11,9197,"TERMINAL",0,0,"]633;E;2025-07-09 22:05:12 /bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt;cf945349-79a2-42c6-b5b1-8428ad73779e]633;C]0;tum_cte0515@hkn1993:/hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash]633;D;0",,terminal_output +12,14140,"TERMINAL",0,0,"cd logs/logs_mihir/",,terminal_command +13,14654,"TERMINAL",0,0,"ls",,terminal_command +14,14676,"TERMINAL",0,0,"]633;E;2025-07-09 22:05:18 ls;af8e09eb-6e10-4bc6-b809-4360d3d8a533]633;Ctrain_dyn_yolorun_3333026.log\r\ntrain_dyn_yolorun_3333448.log\r\ntrain_lam_action_space_scaling_10_3320179.log\r\ntrain_lam_action_space_scaling_10_3321529.log\r\ntrain_lam_action_space_scaling_10_3329786.log\r\ntrain_lam_action_space_scaling_10_3329801.log\r\ntrain_lam_action_space_scaling_10_3331283.log\r\ntrain_lam_action_space_scaling_12_3318546.log\r\ntrain_lam_action_space_scaling_12_3320177.log\r\ntrain_lam_action_space_scaling_12_3321527.log\r\ntrain_lam_action_space_scaling_12_3329787.log\r\ntrain_lam_action_space_scaling_12_3329802.log\r\ntrain_lam_action_space_scaling_12_3331284.log\r\ntrain_lam_action_space_scaling_20_3318547.log\r\ntrain_lam_action_space_scaling_20_3329788.log\r\ntrain_lam_action_space_scaling_20_3329803.log\r\ntrain_lam_action_space_scaling_20_3331285.log\r\ntrain_lam_action_space_scaling_50_3320180.log\r\ntrain_lam_action_space_scaling_50_3329789.log\r\ntrain_lam_action_space_scaling_50_3329804.log\r\ntrain_lam_action_space_scaling_50_3331286.log\r\ntrain_lam_action_space_scaling_6_3318549.log\r\ntrain_lam_action_space_scaling_6_3320178.log\r\ntrain_lam_action_space_scaling_6_3321528.log\r\ntrain_lam_action_space_scaling_6_3329790.log\r\ntrain_lam_action_space_scaling_6_3329805.log\r\ntrain_lam_action_space_scaling_6_3331287.log\r\ntrain_lam_action_space_scaling_8_3318550.log\r\ntrain_lam_action_space_scaling_8_3329791.log\r\ntrain_lam_action_space_scaling_8_3329806.log\r\ntrain_lam_action_space_scaling_8_3331288.log\r\ntrain_lam_minecraft_overfit_sample_3309655.log\r\ntrain_lam_model_size_scaling_38M_3317098.log\r\ntrain_lam_model_size_scaling_38M_3317115.log\r\ntrain_lam_model_size_scaling_38M_3317231.log\r\ntrain_tokenizer_batch_size_scaling_16_node_3321526.log\r\ntrain_tokenizer_batch_size_scaling_1_node_3318551.log\r\ntrain_tokenizer_batch_size_scaling_2_node_3318552.log\r\ntrain_tokenizer_batch_size_scaling_2_node_3330806.log\r\ntrain_tokenizer_batch_size_scaling_2_node_3330848.log\r\ntrain_tokenizer_batch_size_scaling_2_node_3331282.log\r\ntrain_tokenizer_batch_size_scaling_4_node_3318553.log\r\ntrain_tokenizer_batch_size_scaling_4_node_3320175.log\r\ntrain_tokenizer_batch_size_scaling_4_node_3321524.log\r\ntrain_tokenizer_batch_size_scaling_8_node_3320176.log\r\ntrain_tokenizer_batch_size_scaling_8_node_3321525.log\r\ntrain_tokenizer_minecraft_overfit_sample_3309656.log\r\ntrain_tokenizer_model_size_scaling_127M_3317233.log\r\ntrain_tokenizer_model_size_scaling_127M_3318554.log\r\ntrain_tokenizer_model_size_scaling_140M_3313562.log\r\ntrain_tokenizer_model_size_scaling_140M_3316019.log\r\ntrain_tokenizer_model_size_scaling_200M_3313563.log\r\ntrain_tokenizer_model_size_scaling_200M_3316020.log\r\ntrain_tokenizer_model_size_scaling_227M_3317234.log\r\ntrain_tokenizer_model_size_scaling_227M_3318555.log\r\ntrain_tokenizer_model_size_scaling_227M_3320173.log\r\ntrain_tokenizer_model_size_scaling_227M_3321523.log\r\ntrain_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_tokenizer_model_size_scaling_80M_3316026.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +15,19274,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\n# array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\ntf_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/open_ai_minecraft_tfrecord\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/train_lam_action_space_scaling_20/3318547/lam_1751657975_200000/\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $tf_records_dir \\n --lam_checkpoint=$lam_ckpt_dir\nSLURM_JOB_USER=tum_cte0515\nSLURM_TASKS_PER_NODE=4(x2)\nSLURM_JOB_UID=999226\nSLURM_TASK_PID=2954262\nSLURM_JOB_GPUS=0,1,2,3\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\nSLURMD_NODENAME=hkn0423\nSLURM_JOB_START_TIME=1752084560\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1752120560\nSLURM_CPUS_ON_NODE=24\nSLURM_JOB_CPUS_PER_NODE=24(x2)\nSLURM_GPUS_ON_NODE=4\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=accelerated\nSLURM_TRES_PER_TASK=cpu=5\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=2\nSLURM_JOBID=3333448\nSLURM_JOB_QOS=normal\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=5\nSLURM_NTASKS=8\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e12.hkn0423\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_NODELIST=hkn[0423,0806]\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=8\nSLURM_NNODES=2\nSLURM_SUBMIT_HOST=hkn1991.localdomain\nSLURM_JOB_ID=3333448\nSLURM_NODEID=0\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_dyn_yolorun\nSLURM_NTASKS_PER_NODE=4\nSLURM_JOB_GID=502226\nSLURM_JOB_NODELIST=hkn[0423,0806]\nSLURM_JOB_USER=tum_cte0515\nSLURM_TASKS_PER_NODE=4(x2)\nSLURM_JOB_UID=999226\nSLURM_TASK_PID=2954262\nSLURM_JOB_GPUS=0,1,2,3\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\nSLURMD_NODENAME=hkn0423\nSLURM_JOB_START_TIME=1752084560\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1752120560\nSLURM_CPUS_ON_NODE=24\nSLURM_JOB_CPUS_PER_NODE=24(x2)\nSLURM_GPUS_ON_NODE=4\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=accelerated\nSLURM_TRES_PER_TASK=cpu=5\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=2\nSLURM_JOBID=3333448\nSLURM_JOB_QOS=normal\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=5\nSLURM_NTASKS=8\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e12.hkn0423\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_NODELIST=hkn[0423,0806]\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=8\nSLURM_NNODES=2\nSLURM_SUBMIT_HOST=hkn1991.localdomain\nSLURM_JOB_ID=3333448\nSLURM_NODEID=0\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_dyn_yolorun\nSLURM_NTASKS_PER_NODE=4\nSLURM_JOB_GID=502226\nSLURM_JOB_NODELIST=hkn[0423,0806]\nGpuFreq=control_disabled\nGpuFreq=control_disabled\n2025-07-09 20:19:02.647121: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-07-09 20:19:02.647117: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-07-09 20:19:02.648691: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-07-09 20:19:02.648812: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1752085142.716728 2954324 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1752085142.716563 2954325 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1752085142.716532 2954326 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1752085142.716725 2954327 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nE0000 00:00:1752085142.737751 2954324 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1752085142.737711 2954325 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1752085142.737715 2954326 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1752085142.738027 2954327 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nW0000 00:00:1752085142.856993 2954324 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.857013 2954324 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.857016 2954324 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.857019 2954324 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.856933 2954325 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.856963 2954325 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.856965 2954325 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.856967 2954325 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.856931 2954326 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.856960 2954326 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.856962 2954326 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.856964 2954326 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.856992 2954327 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.857013 2954327 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.857016 2954327 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085142.857019 2954327 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\n2025-07-09 20:22:41.392553: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-07-09 20:22:41.392553: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-07-09 20:22:41.392657: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-07-09 20:22:41.392551: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1752085361.421720 2486505 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1752085361.421857 2486506 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1752085361.421738 2486507 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1752085361.421746 2486508 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nE0000 00:00:1752085361.431849 2486505 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1752085361.431899 2486506 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1752085361.431838 2486507 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1752085361.431848 2486508 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nW0000 00:00:1752085440.068327 2486505 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068381 2486505 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068383 2486505 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068385 2486505 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068450 2486507 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068482 2486507 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068484 2486507 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068486 2486507 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068452 2486506 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068517 2486506 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068524 2486506 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068529 2486506 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068578 2486508 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068647 2486508 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068654 2486508 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752085440.068659 2486508 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1752086325.429163 2954325 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1752086325.429223 2954326 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1752086325.429137 2954327 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1752086325.429380 2954324 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1752086619.300899 2486505 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1752086619.300888 2486506 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1752086619.300918 2486507 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1752086619.300903 2486508 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\n2025-07-09 20:46:13.254855: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:46:13.628110: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:46:16.135680: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:46:16.561003: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:46:17.956943: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:46:19.857695: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:46:21.437934: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:46:23.111821: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\nwandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\nwandb: Tracking run with wandb version 0.19.11\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250709_204631-otadl0hf\nwandb: Run `wandb offline` to turn off syncing.\nwandb: Syncing run dynamics-yolorun-tf-records-3333448\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/otadl0hf\n2025-07-09 20:48:53.135539: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:48:53.716202: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:48:56.159316: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:48:56.602919: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:48:57.910995: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:48:58.401831: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:49:01.090595: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:49:01.653781: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-07-09 20:52:32.339391: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1730] Use error polling to propagate the following error to all tasks: UNAVAILABLE: The following tasks are unhealthy (stopped sending heartbeats):\n/job:jax_worker/replica:0/task:3\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:52:32.339939: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: DEADLINE_EXCEEDED: Deadline Exceeded\n\nRPC: /tensorflow.CoordinationService/Heartbeat\n2025-07-09 20:52:32.355006: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.354928: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.354975: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.355098: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.355597: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: The following tasks are unhealthy (stopped sending heartbeats):\n/job:jax_worker/replica:0/task:3\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:52:32.355511: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: The following tasks are unhealthy (stopped sending heartbeats):\n/job:jax_worker/replica:0/task:3\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:52:32.355721: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: The following tasks are unhealthy (stopped sending heartbeats):\n/job:jax_worker/replica:0/task:3\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:52:32.355778: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: The following tasks are unhealthy (stopped sending heartbeats):\n/job:jax_worker/replica:0/task:3\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:53:06.248355: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:53:06.250291: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:53:06.251049: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:53:06.251219: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: The following tasks are unhealthy (stopped sending heartbeats):\n/job:jax_worker/replica:0/task:3\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:53:06.251366: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: The following tasks are unhealthy (stopped sending heartbeats):\n/job:jax_worker/replica:0/task:3\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:53:06.253667: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: The following tasks are unhealthy (stopped sending heartbeats):\n/job:jax_worker/replica:0/task:3\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nsrun: error: hkn0423: task 3: Aborted (core dumped)\nsrun: error: hkn0423: task 1: Aborted (core dumped)\nsrun: error: hkn0423: task 0: Aborted (core dumped)\nsrun: error: hkn0423: task 2: Aborted (core dumped)\nsrun: error: hkn0806: task 6: Aborted (core dumped)\nsrun: error: hkn0806: task 7: Aborted (core dumped)\nsrun: error: hkn0806: task 4: Aborted (core dumped)\nsrun: error: hkn0806: task 5: Aborted (core dumped)\n\n============================= JOB FEEDBACK =============================\n\nJob ID: 3333448\nCluster: hk\nUser/Group: tum_cte0515/hk-project-p0023960\nAccount: hk-project-p0023960\nState: FAILED (exit code 134)\nPartition: accelerated\nNodes: 2\nCores per node: 24\nNodelist: hkn[0423,0806]\nCPU Utilized: 00:34:42\nCPU Efficiency: 1.58% of 1-12:40:00 core-walltime\nJob Wall-clock time: 00:45:50\nStarttime: Wed Jul 9 20:09:20 2025\nEndtime: Wed Jul 9 20:55:10 2025\nMemory Utilized: 39.02 GB (estimated maximum)\nMemory Efficiency: 0.00% of 0.00 MB (0.00 MB/node)\nEnergy Consumed: 2748325 Joule / 763.423611111111 Watthours\nAverage node power draw: 999.390909090909 Watt\n",log,tab +16,20276,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",2028,0,"",log,selection_mouse +17,20301,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",2027,0,"",log,selection_command +18,21932,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",31703,0,"",log,selection_command +19,43289,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23977,0,"",log,selection_mouse +20,44225,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24021,0,"",log,selection_mouse +21,44976,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24081,0,"",log,selection_mouse +22,45779,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24126,0,"",log,selection_mouse +23,46581,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24162,0,"",log,selection_mouse +24,47228,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24149,0,"",log,selection_mouse +25,47384,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24138,17,"DEADLINE_EXCEEDED",log,selection_mouse +26,48169,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23787,0,"",log,selection_mouse +27,48200,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23786,0,"",log,selection_command +28,48313,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23787,0,"",log,selection_mouse +29,48340,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23786,0,"",log,selection_command +30,48554,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23595,193,"The tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n",log,selection_mouse +31,48567,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23596,192,"he tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n",log,selection_command +32,49181,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24048,0,"",log,selection_mouse +33,49364,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24045,6,"Python",log,selection_mouse +34,50492,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24127,0,"",log,selection_mouse +35,51508,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24149,0,"",log,selection_mouse +36,51683,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24138,17,"DEADLINE_EXCEEDED",log,selection_mouse +37,52515,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24146,0,"",log,selection_mouse +38,52516,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24138,17,"DEADLINE_EXCEEDED",log,selection_mouse +39,69294,"TERMINAL",0,0,"bash",,terminal_focus +40,80583,"TERMINAL",0,0,"queue",,terminal_command +41,80635,"TERMINAL",0,0,"]633;E;2025-07-09 22:06:24 queue;16c08be6-3885-420f-ac53-f5272aed6e54]633;C",,terminal_output +42,80736,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:24 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON) 3331283 accelerat train_la tum_cte0 R 21:28:11\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:11\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:11\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:11\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:11\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:11\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:52\t 2 hkn[0513,0516]",,terminal_output +43,81868,"TERMINAL",0,0,"52222223",,terminal_output +44,82788,"TERMINAL",0,0,"63333334",,terminal_output +45,83881,"TERMINAL",0,0,"74444445",,terminal_output +46,84876,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:28 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON) 3331283 accelerat train_la tum_cte0 R 21:28:15\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:15\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:15\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:15\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:15\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:15\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:56\t 2 hkn[0513,0516]",,terminal_output +47,84985,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:28 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:15\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:15\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:15\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:15\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:15\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:15\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:56\t 2 hkn[0513,0516]",,terminal_output +48,85932,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:29 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:57\t 2 hkn[0513,0516]",,terminal_output +49,86060,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:29 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:57\t 2 hkn[0513,0516]",,terminal_output +50,86129,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:29 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:57\t 2 hkn[0513,0516]",,terminal_output +51,86248,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:29 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:16\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:57\t 2 hkn[0513,0516]",,terminal_output +52,86693,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:30 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:58\t 2 hkn[0513,0516]",,terminal_output +53,86823,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:30 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:58\t 2 hkn[0513,0516]\t",,terminal_output +54,87156,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:30 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:17\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:58\t 2 hkn[0513,0516]",,terminal_output +55,87417,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:31 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:59\t 2 hkn[0513,0516]",,terminal_output +56,87520,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:31 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:59\t 2 hkn[0513,0516]",,terminal_output +57,88247,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:31 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:18\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:28:59\t 2 hkn[0513,0516] ",,terminal_output +58,89480,"TERMINAL",0,0,"22020202020209:01",,terminal_output +59,89604,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:33 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:20\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:20\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:20\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:20\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:20\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:20\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:29:01\t 2 hkn[0513,0516] ",,terminal_output +60,90130,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:33 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:20\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:20\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:20\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:20\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:20\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:20\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:29:01\t 2 hkn[0513,0516]",,terminal_output +61,90990,"TERMINAL",0,0,"41111112",,terminal_output +62,92227,"TERMINAL",0,0,"52222223",,terminal_output +63,93240,"TERMINAL",0,0,"63333334",,terminal_output +64,93295,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:06:36 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:28:23\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:28:23\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:28:23\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:28:23\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:28:23\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:28:23\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:29:04\t 2 hkn[0513,0516] ",,terminal_output +65,94297,"TERMINAL",0,0,"75555556",,terminal_output +66,95475,"TERMINAL",0,0,"96666667",,terminal_output +67,96416,"TERMINAL",0,0,"407777778",,terminal_output +68,97444,"TERMINAL",0,0,"18888889",,terminal_output +69,98517,"TERMINAL",0,0,"299999910",,terminal_output +70,99545,"TERMINAL",0,0,"33030303030301",,terminal_output +71,100638,"TERMINAL",0,0,"41111112",,terminal_output +72,101870,"TERMINAL",0,0,"52222223",,terminal_output +73,102894,"TERMINAL",0,0,"63333334",,terminal_output +74,103743,"TERMINAL",0,0,"74444445",,terminal_output +75,104794,"TERMINAL",0,0,"85555556",,terminal_output +76,105845,"TERMINAL",0,0,"96666667",,terminal_output +77,106889,"TERMINAL",0,0,"507777778",,terminal_output +78,107927,"TERMINAL",0,0,"18888889",,terminal_output +79,108981,"TERMINAL",0,0,"299999920",,terminal_output +80,110032,"TERMINAL",0,0,"34040404040401",,terminal_output +81,111083,"TERMINAL",0,0,"41111112",,terminal_output +82,112123,"TERMINAL",0,0,"52222223",,terminal_output +83,113193,"TERMINAL",0,0,"63333334",,terminal_output +84,114244,"TERMINAL",0,0,"74444445",,terminal_output +85,115295,"TERMINAL",0,0,"86666667",,terminal_output +86,116353,"TERMINAL",0,0,"7:007777778",,terminal_output +87,117392,"TERMINAL",0,0,"18888889",,terminal_output +88,118447,"TERMINAL",0,0,"299999930",,terminal_output +89,119493,"TERMINAL",0,0,"35050505050501",,terminal_output +90,120568,"TERMINAL",0,0,"41111112",,terminal_output +91,121613,"TERMINAL",0,0,"52222223",,terminal_output +92,122873,"TERMINAL",0,0,"63333334",,terminal_output +93,123818,"TERMINAL",0,0,"74444445",,terminal_output +94,124776,"TERMINAL",0,0,"85555556",,terminal_output +95,125992,"TERMINAL",0,0,"96666667",,terminal_output +96,126858,"TERMINAL",0,0,"107777778",,terminal_output +97,127938,"TERMINAL",0,0,"18888889",,terminal_output +98,128975,"TERMINAL",0,0,"299999940",,terminal_output +99,130016,"TERMINAL",0,0,"39:009:009:009:009:009:001",,terminal_output +100,131074,"TERMINAL",0,0,"41111112",,terminal_output +101,132080,"TERMINAL",0,0,"52222223",,terminal_output +102,133179,"TERMINAL",0,0,"63333334",,terminal_output +103,134209,"TERMINAL",0,0,"74444445",,terminal_output +104,135221,"TERMINAL",0,0,"85555556",,terminal_output +105,136275,"TERMINAL",0,0,"97777778",,terminal_output +106,137407,"TERMINAL",0,0,"218888889",,terminal_output +107,138442,"TERMINAL",0,0,"299999950",,terminal_output +108,139490,"TERMINAL",0,0,"31010101010101",,terminal_output +109,140472,"TERMINAL",0,0,"41111112",,terminal_output +110,141535,"TERMINAL",0,0,"52222223",,terminal_output +111,142675,"TERMINAL",0,0,"63333334",,terminal_output +112,143722,"TERMINAL",0,0,"74444445",,terminal_output +113,144782,"TERMINAL",0,0,"85555556",,terminal_output +114,145860,"TERMINAL",0,0,"96666667",,terminal_output +115,146884,"TERMINAL",0,0,"307777778",,terminal_output +116,147938,"TERMINAL",0,0,"18888889",,terminal_output +117,148973,"TERMINAL",0,0,"299999930:00",,terminal_output +118,150041,"TERMINAL",0,0,"32020202020201",,terminal_output +119,151080,"TERMINAL",0,0,"41111112",,terminal_output +120,152148,"TERMINAL",0,0,"52222223",,terminal_output +121,153178,"TERMINAL",0,0,"63333334",,terminal_output +122,154235,"TERMINAL",0,0,"74444445",,terminal_output +123,155292,"TERMINAL",0,0,"86666667",,terminal_output +124,156366,"TERMINAL",0,0,"407777778",,terminal_output +125,157372,"TERMINAL",0,0,"18888889",,terminal_output +126,158494,"TERMINAL",0,0,"299999910",,terminal_output +127,159593,"TERMINAL",0,0,"33030303030301",,terminal_output +128,160535,"TERMINAL",0,0,"41111112",,terminal_output +129,161634,"TERMINAL",0,0,"52222223",,terminal_output +130,162706,"TERMINAL",0,0,"63333334",,terminal_output +131,163683,"TERMINAL",0,0,"74444445",,terminal_output +132,164835,"TERMINAL",0,0,"85555556",,terminal_output +133,165818,"TERMINAL",0,0,"96666667",,terminal_output +134,166868,"TERMINAL",0,0,"507777778",,terminal_output +135,167974,"TERMINAL",0,0,"18888889",,terminal_output +136,168978,"TERMINAL",0,0,"299999920",,terminal_output +137,170008,"TERMINAL",0,0,"34040404040401",,terminal_output +138,171057,"TERMINAL",0,0,"41111112",,terminal_output +139,172111,"TERMINAL",0,0,"52222223",,terminal_output +140,173164,"TERMINAL",0,0,"63333334",,terminal_output +141,174215,"TERMINAL",0,0,"74444445",,terminal_output +142,175286,"TERMINAL",0,0,"85555556",,terminal_output +143,176340,"TERMINAL",0,0,"97777778",,terminal_output +144,177390,"TERMINAL",0,0,"8:018888889",,terminal_output +145,178419,"TERMINAL",0,0,"299999930",,terminal_output +146,179464,"TERMINAL",0,0,"35050505050501",,terminal_output +147,180519,"TERMINAL",0,0,"41111112",,terminal_output +148,181541,"TERMINAL",0,0,"52222223",,terminal_output +149,182811,"TERMINAL",0,0,"63333334",,terminal_output +150,183724,"TERMINAL",0,0,"74444445",,terminal_output +151,184715,"TERMINAL",0,0,"85555556",,terminal_output +152,185745,"TERMINAL",0,0,"96666667",,terminal_output +153,186823,"TERMINAL",0,0,"107777778",,terminal_output +154,187920,"TERMINAL",0,0,"18888889",,terminal_output +155,188948,"TERMINAL",0,0,"299999940",,terminal_output +156,189949,"TERMINAL",0,0,"330:0030:0030:0030:0030:0030:001",,terminal_output +157,191020,"TERMINAL",0,0,"41111112",,terminal_output +158,192046,"TERMINAL",0,0,"52222223",,terminal_output +159,193195,"TERMINAL",0,0,"63333334",,terminal_output +160,194204,"TERMINAL",0,0,"74444445",,terminal_output +161,195229,"TERMINAL",0,0,"85555556",,terminal_output +162,196327,"TERMINAL",0,0,"96666667",,terminal_output +163,196483,"TERMINAL",0,0,"bash",,terminal_focus +164,197352,"TERMINAL",0,0,"207777778",,terminal_output +165,198501,"TERMINAL",0,0,"299999950",,terminal_output +166,199488,"TERMINAL",0,0,"31010101010101",,terminal_output +167,200538,"TERMINAL",0,0,"41111112",,terminal_output +168,201513,"TERMINAL",0,0,"52222223",,terminal_output +169,202621,"TERMINAL",0,0,"63333334",,terminal_output +170,203574,"TERMINAL",0,0,"74444445",,terminal_output +171,204616,"TERMINAL",0,0,"85555556",,terminal_output +172,205666,"TERMINAL",0,0,"96666667",,terminal_output +173,206726,"TERMINAL",0,0,"307777778",,terminal_output +174,207760,"TERMINAL",0,0,"18888889",,terminal_output +175,208813,"TERMINAL",0,0,"29999991:00",,terminal_output +176,209867,"TERMINAL",0,0,"32020202020201",,terminal_output +177,210928,"TERMINAL",0,0,"41111112",,terminal_output +178,211979,"TERMINAL",0,0,"52222223",,terminal_output +179,213023,"TERMINAL",0,0,"63333334",,terminal_output +180,214070,"TERMINAL",0,0,"74444445",,terminal_output +181,215122,"TERMINAL",0,0,"85555556",,terminal_output +182,216170,"TERMINAL",0,0,"96666667",,terminal_output +183,217223,"TERMINAL",0,0,"407777778",,terminal_output +184,218272,"TERMINAL",0,0,"18888889",,terminal_output +185,219337,"TERMINAL",0,0,"330303030303011",,terminal_output +186,220369,"TERMINAL",0,0,"41111112",,terminal_output +187,221420,"TERMINAL",0,0,"52222223",,terminal_output +188,222464,"TERMINAL",0,0,"63333334",,terminal_output +189,223499,"TERMINAL",0,0,"74444445",,terminal_output +190,224540,"TERMINAL",0,0,"85555556",,terminal_output +191,225590,"TERMINAL",0,0,"96666667",,terminal_output +192,226642,"TERMINAL",0,0,"507777778",,terminal_output +193,227691,"TERMINAL",0,0,"18888889",,terminal_output +194,228745,"TERMINAL",0,0,"299999920",,terminal_output +195,229801,"TERMINAL",0,0,"34040404040401",,terminal_output +196,230843,"TERMINAL",0,0,"41111112",,terminal_output +197,231901,"TERMINAL",0,0,"52222223",,terminal_output +198,232970,"TERMINAL",0,0,"63333334",,terminal_output +199,234001,"TERMINAL",0,0,"74444445",,terminal_output +200,235051,"TERMINAL",0,0,"85555556",,terminal_output +201,236099,"TERMINAL",0,0,"96666667",,terminal_output +202,237148,"TERMINAL",0,0,"9:007777778",,terminal_output +203,238203,"TERMINAL",0,0,"18888889",,terminal_output +204,239255,"TERMINAL",0,0,"299999930",,terminal_output +205,240310,"TERMINAL",0,0,"35151515151512",,terminal_output +206,241428,"TERMINAL",0,0,"52222223",,terminal_output +207,242434,"TERMINAL",0,0,"63333334",,terminal_output +208,243701,"TERMINAL",0,0,"74444445",,terminal_output +209,244581,"TERMINAL",0,0,"85555556",,terminal_output +210,245722,"TERMINAL",0,0,"96666667",,terminal_output +211,246745,"TERMINAL",0,0,"107777778",,terminal_output +212,247732,"TERMINAL",0,0,"18888889",,terminal_output +213,248750,"TERMINAL",0,0,"299999940",,terminal_output +214,249819,"TERMINAL",0,0,"31:001:001:001:001:001:001",,terminal_output +215,250884,"TERMINAL",0,0,"41111112",,terminal_output +216,251892,"TERMINAL",0,0,"52222223",,terminal_output +217,252948,"TERMINAL",0,0,"63333334",,terminal_output +218,254005,"TERMINAL",0,0,"74444445",,terminal_output +219,255003,"TERMINAL",0,0,"85555556",,terminal_output +220,256052,"TERMINAL",0,0,"96666667",,terminal_output +221,257107,"TERMINAL",0,0,"207777778",,terminal_output +222,258163,"TERMINAL",0,0,"18888889",,terminal_output +223,259205,"TERMINAL",0,0,"299999950",,terminal_output +224,260256,"TERMINAL",0,0,"31010101010101",,terminal_output +225,261310,"TERMINAL",0,0,"42222223",,terminal_output +226,262367,"TERMINAL",0,0,"63333334",,terminal_output +227,263417,"TERMINAL",0,0,"74444445",,terminal_output +228,264469,"TERMINAL",0,0,"85555556",,terminal_output +229,265546,"TERMINAL",0,0,"96666667",,terminal_output +230,266587,"TERMINAL",0,0,"307777778",,terminal_output +231,267675,"TERMINAL",0,0,"18888889",,terminal_output +232,268682,"TERMINAL",0,0,"29999992:00",,terminal_output +233,269735,"TERMINAL",0,0,"32020202020201",,terminal_output +234,270789,"TERMINAL",0,0,"41111112",,terminal_output +235,271840,"TERMINAL",0,0,"52222223",,terminal_output +236,272890,"TERMINAL",0,0,"63333334",,terminal_output +237,273943,"TERMINAL",0,0,"74444445",,terminal_output +238,274985,"TERMINAL",0,0,"85555556",,terminal_output +239,276035,"TERMINAL",0,0,"96666667",,terminal_output +240,277084,"TERMINAL",0,0,"407777778",,terminal_output +241,278131,"TERMINAL",0,0,"18888889",,terminal_output +242,279202,"TERMINAL",0,0,"299999910",,terminal_output +243,280231,"TERMINAL",0,0,"33030303030301",,terminal_output +244,281286,"TERMINAL",0,0,"42222223",,terminal_output +245,282334,"TERMINAL",0,0,"63333334",,terminal_output +246,283393,"TERMINAL",0,0,"74444445",,terminal_output +247,284447,"TERMINAL",0,0,"85555556",,terminal_output +248,285476,"TERMINAL",0,0,"96666667",,terminal_output +249,286519,"TERMINAL",0,0,"507777778",,terminal_output +250,287569,"TERMINAL",0,0,"18888889",,terminal_output +251,288665,"TERMINAL",0,0,"299999920",,terminal_output +252,289671,"TERMINAL",0,0,"34040404040401",,terminal_output +253,290726,"TERMINAL",0,0,"41111112",,terminal_output +254,291794,"TERMINAL",0,0,"52222223",,terminal_output +255,292823,"TERMINAL",0,0,"63333334",,terminal_output +256,293874,"TERMINAL",0,0,"74444445",,terminal_output +257,294924,"TERMINAL",0,0,"85555556",,terminal_output +258,295973,"TERMINAL",0,0,"96666667",,terminal_output +259,297211,"TERMINAL",0,0,"10:007777778",,terminal_output +260,298072,"TERMINAL",0,0,"18888889",,terminal_output +261,299148,"TERMINAL",0,0,"299999930",,terminal_output +262,300171,"TERMINAL",0,0,"35050505050501",,terminal_output +263,301231,"TERMINAL",0,0,"41111112",,terminal_output +264,302271,"TERMINAL",0,0,"53333334",,terminal_output +265,303324,"TERMINAL",0,0,"74444445",,terminal_output +266,304479,"TERMINAL",0,0,"85555556",,terminal_output +267,305462,"TERMINAL",0,0,"96666667",,terminal_output +268,306491,"TERMINAL",0,0,"107777778",,terminal_output +269,307512,"TERMINAL",0,0,"18888889",,terminal_output +270,308577,"TERMINAL",0,0,"299999940",,terminal_output +271,309695,"TERMINAL",0,0,"32:002:002:002:002:002:001",,terminal_output +272,310707,"TERMINAL",0,0,"41111112",,terminal_output +273,311740,"TERMINAL",0,0,"52222223",,terminal_output +274,312770,"TERMINAL",0,0,"63333334",,terminal_output +275,313876,"TERMINAL",0,0,"74444445",,terminal_output +276,314867,"TERMINAL",0,0,"85555556",,terminal_output +277,315965,"TERMINAL",0,0,"96666667",,terminal_output +278,317030,"TERMINAL",0,0,"207777778",,terminal_output +279,318037,"TERMINAL",0,0,"18888889",,terminal_output +280,319065,"TERMINAL",0,0,"299999950",,terminal_output +281,320120,"TERMINAL",0,0,"31010101010101",,terminal_output +282,321166,"TERMINAL",0,0,"41111112",,terminal_output +283,322278,"TERMINAL",0,0,"52222223",,terminal_output +284,323360,"TERMINAL",0,0,"64444445",,terminal_output +285,324370,"TERMINAL",0,0,"85555556",,terminal_output +286,325444,"TERMINAL",0,0,"96666667",,terminal_output +287,326468,"TERMINAL",0,0,"307777778",,terminal_output +288,327524,"TERMINAL",0,0,"18888889",,terminal_output +289,328573,"TERMINAL",0,0,"29999993:00",,terminal_output +290,329627,"TERMINAL",0,0,"32020202020201",,terminal_output +291,330678,"TERMINAL",0,0,"41111112",,terminal_output +292,331740,"TERMINAL",0,0,"52222223",,terminal_output +293,332814,"TERMINAL",0,0,"63333334",,terminal_output +294,333860,"TERMINAL",0,0,"74444445",,terminal_output +295,334920,"TERMINAL",0,0,"85555556",,terminal_output +296,335971,"TERMINAL",0,0,"96666667",,terminal_output +297,337041,"TERMINAL",0,0,"407777778",,terminal_output +298,338090,"TERMINAL",0,0,"18888889",,terminal_output +299,339138,"TERMINAL",0,0,"299999910",,terminal_output +300,340192,"TERMINAL",0,0,"33030303030301",,terminal_output +301,341249,"TERMINAL",0,0,"41111112",,terminal_output +302,342291,"TERMINAL",0,0,"53333334",,terminal_output +303,343342,"TERMINAL",0,0,"74444445",,terminal_output +304,344400,"TERMINAL",0,0,"85555556",,terminal_output +305,345447,"TERMINAL",0,0,"96666667",,terminal_output +306,346503,"TERMINAL",0,0,"507777778",,terminal_output +307,347576,"TERMINAL",0,0,"18888889",,terminal_output +308,348614,"TERMINAL",0,0,"299999920",,terminal_output +309,349683,"TERMINAL",0,0,"34040404040401",,terminal_output +310,350701,"TERMINAL",0,0,"41111112",,terminal_output +311,351747,"TERMINAL",0,0,"52222223",,terminal_output +312,352818,"TERMINAL",0,0,"63333334",,terminal_output +313,353862,"TERMINAL",0,0,"74444445",,terminal_output +314,354902,"TERMINAL",0,0,"85555556",,terminal_output +315,355957,"TERMINAL",0,0,"96666667",,terminal_output +316,357001,"TERMINAL",0,0,"1:007777778",,terminal_output +317,358040,"TERMINAL",0,0,"18888889",,terminal_output +318,359084,"TERMINAL",0,0,"299999930",,terminal_output +319,360131,"TERMINAL",0,0,"35050505050501",,terminal_output +320,361185,"TERMINAL",0,0,"41111112",,terminal_output +321,362239,"TERMINAL",0,0,"52222223",,terminal_output +322,363282,"TERMINAL",0,0,"64444445",,terminal_output +323,364335,"TERMINAL",0,0,"85555556",,terminal_output +324,365402,"TERMINAL",0,0,"96666667",,terminal_output +325,366443,"TERMINAL",0,0,"107777778",,terminal_output +326,367488,"TERMINAL",0,0,"18888889",,terminal_output +327,368538,"TERMINAL",0,0,"299999940",,terminal_output +328,369658,"TERMINAL",0,0,"33:003:003:003:003:003:001",,terminal_output +329,370724,"TERMINAL",0,0,"41111112",,terminal_output +330,371689,"TERMINAL",0,0,"52222223",,terminal_output +331,372744,"TERMINAL",0,0,"63333334",,terminal_output +332,373797,"TERMINAL",0,0,"74444445",,terminal_output +333,374876,"TERMINAL",0,0,"85555556",,terminal_output +334,375895,"TERMINAL",0,0,"96666667",,terminal_output +335,376953,"TERMINAL",0,0,"207777778",,terminal_output +336,378022,"TERMINAL",0,0,"18888889",,terminal_output +337,379051,"TERMINAL",0,0,"299999950",,terminal_output +338,380110,"TERMINAL",0,0,"31010101010101",,terminal_output +339,381151,"TERMINAL",0,0,"41111112",,terminal_output +340,382201,"TERMINAL",0,0,"52222223",,terminal_output +341,383258,"TERMINAL",0,0,"63333334",,terminal_output +342,384302,"TERMINAL",0,0,"75555556",,terminal_output +343,385385,"TERMINAL",0,0,"96666667",,terminal_output +344,386433,"TERMINAL",0,0,"307777778",,terminal_output +345,387480,"TERMINAL",0,0,"18888889",,terminal_output +346,388539,"TERMINAL",0,0,"29999994:00",,terminal_output +347,389619,"TERMINAL",0,0,"32020202020201",,terminal_output +348,390725,"TERMINAL",0,0,"41111112",,terminal_output +349,391738,"TERMINAL",0,0,"52222223",,terminal_output +350,392742,"TERMINAL",0,0,"63333334",,terminal_output +351,393794,"TERMINAL",0,0,"74444445",,terminal_output +352,394836,"TERMINAL",0,0,"85555556",,terminal_output +353,395884,"TERMINAL",0,0,"96666667",,terminal_output +354,396941,"TERMINAL",0,0,"407777778",,terminal_output +355,397989,"TERMINAL",0,0,"18888889",,terminal_output +356,399033,"TERMINAL",0,0,"299999910",,terminal_output +357,400083,"TERMINAL",0,0,"33030303030301",,terminal_output +358,401138,"TERMINAL",0,0,"41111112",,terminal_output +359,402183,"TERMINAL",0,0,"52222223",,terminal_output +360,403234,"TERMINAL",0,0,"63333334",,terminal_output +361,404287,"TERMINAL",0,0,"75555556",,terminal_output +362,405329,"TERMINAL",0,0,"96666667",,terminal_output +363,406379,"TERMINAL",0,0,"507777778",,terminal_output +364,407435,"TERMINAL",0,0,"18888889",,terminal_output +365,408494,"TERMINAL",0,0,"299999920",,terminal_output +366,409533,"TERMINAL",0,0,"34040404040401",,terminal_output +367,410579,"TERMINAL",0,0,"41111112",,terminal_output +368,411632,"TERMINAL",0,0,"52222223",,terminal_output +369,412732,"TERMINAL",0,0,"63333334",,terminal_output +370,413778,"TERMINAL",0,0,"74444445",,terminal_output +371,414815,"TERMINAL",0,0,"85555556",,terminal_output +372,415868,"TERMINAL",0,0,"96666667",,terminal_output +373,417008,"TERMINAL",0,0,"2:007777778",,terminal_output +374,418191,"TERMINAL",0,0,"18888889",,terminal_output +375,419160,"TERMINAL",0,0,"299999930",,terminal_output +376,420087,"TERMINAL",0,0,"35050505050501",,terminal_output +377,421163,"TERMINAL",0,0,"41111112",,terminal_output +378,422186,"TERMINAL",0,0,"52222223",,terminal_output +379,423378,"TERMINAL",0,0,"64444445",,terminal_output +380,424413,"TERMINAL",0,0,"85555556",,terminal_output +381,425459,"TERMINAL",0,0,"96666667",,terminal_output +382,426505,"TERMINAL",0,0,"107777778",,terminal_output +383,427565,"TERMINAL",0,0,"18888889",,terminal_output +384,428618,"TERMINAL",0,0,"299999940",,terminal_output +385,429665,"TERMINAL",0,0,"34:004:004:004:004:004:001",,terminal_output +386,430741,"TERMINAL",0,0,"41111112",,terminal_output +387,431756,"TERMINAL",0,0,"52222223",,terminal_output +388,432848,"TERMINAL",0,0,"63333334",,terminal_output +389,433896,"TERMINAL",0,0,"74444445",,terminal_output +390,434949,"TERMINAL",0,0,"85555556",,terminal_output +391,435999,"TERMINAL",0,0,"96666667",,terminal_output +392,437042,"TERMINAL",0,0,"207777778",,terminal_output +393,438095,"TERMINAL",0,0,"18888889",,terminal_output +394,439140,"TERMINAL",0,0,"299999950",,terminal_output +395,440208,"TERMINAL",0,0,"31010101010101",,terminal_output +396,441249,"TERMINAL",0,0,"41111112",,terminal_output +397,442295,"TERMINAL",0,0,"53333334",,terminal_output +398,443351,"TERMINAL",0,0,"74444445",,terminal_output +399,444393,"TERMINAL",0,0,"85555556",,terminal_output +400,445446,"TERMINAL",0,0,"96666667",,terminal_output +401,446490,"TERMINAL",0,0,"307777778",,terminal_output +402,447539,"TERMINAL",0,0,"18888889",,terminal_output +403,448594,"TERMINAL",0,0,"29999995:00",,terminal_output +404,449642,"TERMINAL",0,0,"32020202020201",,terminal_output +405,450710,"TERMINAL",0,0,"41111112",,terminal_output +406,451784,"TERMINAL",0,0,"52222223",,terminal_output +407,452776,"TERMINAL",0,0,"63333334",,terminal_output +408,453890,"TERMINAL",0,0,"74444445",,terminal_output +409,454922,"TERMINAL",0,0,"85555556",,terminal_output +410,455944,"TERMINAL",0,0,"96666667",,terminal_output +411,456983,"TERMINAL",0,0,"407777778",,terminal_output +412,458043,"TERMINAL",0,0,"18888889",,terminal_output +413,459145,"TERMINAL",0,0,"299999910",,terminal_output +414,460145,"TERMINAL",0,0,"33030303030301",,terminal_output +415,461168,"TERMINAL",0,0,"41111112",,terminal_output +416,462225,"TERMINAL",0,0,"52222223",,terminal_output +417,463282,"TERMINAL",0,0,"64444445",,terminal_output +418,464333,"TERMINAL",0,0,"85555556",,terminal_output +419,465424,"TERMINAL",0,0,"96666667",,terminal_output +420,466476,"TERMINAL",0,0,"507777778",,terminal_output +421,467520,"TERMINAL",0,0,"18888889",,terminal_output +422,468569,"TERMINAL",0,0,"299999920",,terminal_output +423,469614,"TERMINAL",0,0,"34040404040401",,terminal_output +424,470686,"TERMINAL",0,0,"41111112",,terminal_output +425,471752,"TERMINAL",0,0,"52222223",,terminal_output +426,472772,"TERMINAL",0,0,"63333334",,terminal_output +427,473811,"TERMINAL",0,0,"74444445",,terminal_output +428,474854,"TERMINAL",0,0,"85555556",,terminal_output +429,475983,"TERMINAL",0,0,"96666667",,terminal_output +430,477057,"TERMINAL",0,0,"3:007777778",,terminal_output +431,478145,"TERMINAL",0,0,"18888889",,terminal_output +432,479100,"TERMINAL",0,0,"299999930",,terminal_output +433,480124,"TERMINAL",0,0,"35050505050501",,terminal_output +434,481193,"TERMINAL",0,0,"41111112",,terminal_output +435,482320,"TERMINAL",0,0,"52222223",,terminal_output +436,483296,"TERMINAL",0,0,"64444445",,terminal_output +437,484368,"TERMINAL",0,0,"85555556",,terminal_output +438,485397,"TERMINAL",0,0,"96666667",,terminal_output +439,486444,"TERMINAL",0,0,"107777778",,terminal_output +440,487499,"TERMINAL",0,0,"18888889",,terminal_output +441,488568,"TERMINAL",0,0,"299999940",,terminal_output +442,489595,"TERMINAL",0,0,"35:005:005:005:005:005:001",,terminal_output +443,490643,"TERMINAL",0,0,"41111112",,terminal_output +444,491696,"TERMINAL",0,0,"52222223",,terminal_output +445,492787,"TERMINAL",0,0,"63333334",,terminal_output +446,493802,"TERMINAL",0,0,"74444445",,terminal_output +447,494850,"TERMINAL",0,0,"85555556",,terminal_output +448,495890,"TERMINAL",0,0,"96666667",,terminal_output +449,496937,"TERMINAL",0,0,"207777778",,terminal_output +450,497989,"TERMINAL",0,0,"18888889",,terminal_output +451,499038,"TERMINAL",0,0,"299999950",,terminal_output +452,500095,"TERMINAL",0,0,"31010101010101",,terminal_output +453,501154,"TERMINAL",0,0,"41111112",,terminal_output +454,502210,"TERMINAL",0,0,"52222223",,terminal_output +455,503265,"TERMINAL",0,0,"63333334",,terminal_output +456,504308,"TERMINAL",0,0,"75555556",,terminal_output +457,505377,"TERMINAL",0,0,"96666667",,terminal_output +458,506410,"TERMINAL",0,0,"307777778",,terminal_output +459,507480,"TERMINAL",0,0,"18888889",,terminal_output +460,508532,"TERMINAL",0,0,"29999996:00",,terminal_output +461,509907,"TERMINAL",0,0,"32020202020201",,terminal_output +462,510950,"TERMINAL",0,0,"41111112",,terminal_output +463,511987,"TERMINAL",0,0,"52222223",,terminal_output +464,513043,"TERMINAL",0,0,"63333334",,terminal_output +465,514093,"TERMINAL",0,0,"74444445",,terminal_output +466,515190,"TERMINAL",0,0,"85555556",,terminal_output +467,516195,"TERMINAL",0,0,"96666667",,terminal_output +468,517242,"TERMINAL",0,0,"407777778",,terminal_output +469,518310,"TERMINAL",0,0,"199999910",,terminal_output +470,519347,"TERMINAL",0,0,"33030303030301",,terminal_output +471,520401,"TERMINAL",0,0,"41111112",,terminal_output +472,521450,"TERMINAL",0,0,"52222223",,terminal_output +473,522503,"TERMINAL",0,0,"63333334",,terminal_output +474,523559,"TERMINAL",0,0,"74444445",,terminal_output +475,524590,"TERMINAL",0,0,"85555556",,terminal_output +476,525640,"TERMINAL",0,0,"96666667",,terminal_output +477,526688,"TERMINAL",0,0,"507777778",,terminal_output +478,527788,"TERMINAL",0,0,"18888889",,terminal_output +479,528810,"TERMINAL",0,0,"299999920",,terminal_output +480,529852,"TERMINAL",0,0,"34040404040401",,terminal_output +481,530899,"TERMINAL",0,0,"41111112",,terminal_output +482,531986,"TERMINAL",0,0,"52222223",,terminal_output +483,533014,"TERMINAL",0,0,"63333334",,terminal_output +484,534045,"TERMINAL",0,0,"74444445",,terminal_output +485,535091,"TERMINAL",0,0,"85555556",,terminal_output +486,536142,"TERMINAL",0,0,"96666667",,terminal_output +487,537201,"TERMINAL",0,0,"4:007777778",,terminal_output +488,538246,"TERMINAL",0,0,"18888889",,terminal_output +489,539315,"TERMINAL",0,0,"250505050505031",,terminal_output +490,540104,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",0,0,"",log,tab +491,540356,"TERMINAL",0,0,"41111112",,terminal_output +492,540960,"TERMINAL",0,0,"watch",,terminal_focus +493,541434,"TERMINAL",0,0,"52222223",,terminal_output +494,542235,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +495,544198,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",0,0,"",log,tab +496,565148,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30546,0,"",log,selection_mouse +497,565341,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30534,19,"CoordinationService",log,selection_mouse +498,565485,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30517,111,"RPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n",log,selection_mouse +499,566182,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30546,0,"",log,selection_mouse +500,566183,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30534,19,"CoordinationService",log,selection_mouse +501,567229,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30527,0,"",log,selection_mouse +502,567421,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30523,10,"tensorflow",log,selection_mouse +503,568542,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30562,0,"",log,selection_mouse +504,568719,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30554,12,"PollForError",log,selection_mouse +505,568905,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30553,13,"/PollForError",log,selection_mouse +506,568966,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30554,111,"PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nsrun: error: hkn0423: task 3: Aborted",log,selection_mouse +507,568967,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30554,103,"PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nsrun: error: hkn0423: task 3:",log,selection_mouse +508,568968,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30554,101,"PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nsrun: error: hkn0423: task ",log,selection_mouse +509,568968,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30554,100,"PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nsrun: error: hkn0423: task",log,selection_mouse +510,569006,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30554,95,"PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nsrun: error: hkn0423:",log,selection_mouse +511,569028,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30554,94,"PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nsrun: error: hkn0423",log,selection_mouse +512,569064,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30534,32,"CoordinationService/PollForError",log,selection_mouse +513,569155,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30533,33,".CoordinationService/PollForError",log,selection_mouse +514,569187,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30523,43,"tensorflow.CoordinationService/PollForError",log,selection_mouse +515,570163,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",30528,0,"",log,selection_mouse +516,574272,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23605,0,"",log,selection_mouse +517,574442,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23605,4,"have",log,selection_mouse +518,574709,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23605,5,"have ",log,selection_mouse +519,575079,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23610,0,"",log,selection_mouse +520,575472,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23586,0,"",log,selection_mouse +521,575665,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23586,1,"0",log,selection_mouse +522,575840,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",23562,33,"/job:jax_worker/replica:0/task:3\n",log,selection_mouse +523,585293,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",25820,0,"",log,selection_mouse +524,585585,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",25709,111," [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:52:32.355511: F external/xla/xla/pj",log,selection_mouse +525,585634,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",25659,161,"\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:52:32.355511: F external/xla/xla/pj",log,selection_mouse +526,585634,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",25578,242,"earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:52:32.355511: F external/xla/xla/pj",log,selection_mouse +527,585635,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",25526,294,"\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:52:32.355511: F external/xla/xla/pj",log,selection_mouse +528,585716,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",25118,702,"/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: The following tasks are unhealthy (stopped sending heartbeats):\n/job:jax_worker/replica:0/task:3\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:52:32.355511: F external/xla/xla/pj",log,selection_mouse +529,585719,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24696,1124,"distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.355098: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.355597: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: The following tasks are unhealthy (stopped sending heartbeats):\n/job:jax_worker/replica:0/task:3\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:52:32.355511: F external/xla/xla/pj",log,selection_mouse +530,585720,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24273,1547,"/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.354928: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.354975: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.355098: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.355597: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: The following tasks are unhealthy (stopped sending heartbeats):\n/job:jax_worker/replica:0/task:3\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:52:32.355511: F external/xla/xla/pj",log,selection_mouse +531,585757,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24175,1645,"\nRPC: /tensorflow.CoordinationService/Heartbeat\n2025-07-09 20:52:32.355006: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.354928: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.354975: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.355098: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-09 20:52:32.355597: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: UNAVAILABLE: The following tasks are unhealthy (stopped sending heartbeats):\n/job:jax_worker/replica:0/task:3\nThe tasks have crashed. Check the task logs for an earlier error, or scheduler events (e.g. preemption, eviction) to debug further.\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-09 20:52:32.355511: F external/xla/xla/pj",log,selection_mouse +532,585938,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3333448.log",24175,0,"",log,selection_mouse +533,660171,"TERMINAL",0,0,"bash",,terminal_focus +534,668014,"TERMINAL",0,0,"queue",,terminal_command +535,668035,"TERMINAL",0,0,"]633;E;2025-07-09 22:16:11 queue;af8e09eb-6e10-4bc6-b809-4360d3d8a533]633;C",,terminal_output +536,668113,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:16:11 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:37:58\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:37:58\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:37:58\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:37:58\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:37:58\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:37:58\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:38:39\t 2 hkn[0513,0516]",,terminal_output +537,669185,"TERMINAL",0,0,"299999940",,terminal_output +538,670175,"TERMINAL",0,0,"38:008:008:008:008:008:001",,terminal_output +539,670925,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:16:14 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:38:01\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:38:01\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:38:01\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:38:01\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:38:01\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:38:01\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:38:42\t 2 hkn[0513,0516]",,terminal_output +540,671937,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:16:15 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:38:02\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:38:02\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:38:02\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:38:02\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:38:02\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:38:02\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:38:43\t 2 hkn[0513,0516]",,terminal_output +541,672333,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:16:16 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:38:44\t 2 hkn[0513,0516]",,terminal_output +542,672817,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:16:16 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:38:44\t 2 hkn[0513,0516]",,terminal_output +543,673023,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:16:16 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:38:44\t 2 hkn[0513,0516]",,terminal_output +544,673333,"TERMINAL",0,0,"Every 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:16:16 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3331283 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:38:03\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:38:44\t 2 hkn[0513,0516]",,terminal_output +545,674406,"TERMINAL",0,0,"75555556",,terminal_output +546,675389,"TERMINAL",0,0,"96666667",,terminal_output +547,676401,"TERMINAL",0,0,"207777778",,terminal_output +548,677455,"TERMINAL",0,0,"18888889",,terminal_output +549,678501,"TERMINAL",0,0,"299999950",,terminal_output +550,679553,"TERMINAL",0,0,"31010101010101",,terminal_output +551,680613,"TERMINAL",0,0,"41111112",,terminal_output +552,681656,"TERMINAL",0,0,"52222223",,terminal_output +553,682707,"TERMINAL",0,0,"63333334",,terminal_output +554,683770,"TERMINAL",0,0,"74444445",,terminal_output +555,684814,"TERMINAL",0,0,"85555556",,terminal_output +556,685909,"TERMINAL",0,0,"96666667",,terminal_output +557,686924,"TERMINAL",0,0,"307777778",,terminal_output +558,687963,"TERMINAL",0,0,"18888889",,terminal_output +559,689054,"TERMINAL",0,0,"29999999:00",,terminal_output +560,690100,"TERMINAL",0,0,"32020202020201",,terminal_output +561,691106,"TERMINAL",0,0,"41111112",,terminal_output +562,692160,"TERMINAL",0,0,"52222223",,terminal_output +563,693214,"TERMINAL",0,0,"63333334",,terminal_output +564,694256,"TERMINAL",0,0,"74444445",,terminal_output +565,694836,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +566,697584,"TERMINAL",0,0,"idling",,terminal_command +567,697634,"TERMINAL",0,0,"]633;E;2025-07-09 22:16:41 idling;af8e09eb-6e10-4bc6-b809-4360d3d8a533]633;C",,terminal_output +568,697745,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Wed Jul 9 22:16:41 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly:\t 9 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated:\t 1 nodes idle\rPartition dev_accelerated-h100 :\t 1 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 7 nodes idle",,terminal_output +569,698740,"TERMINAL",0,0,"2\t",,terminal_output +570,699788,"TERMINAL",0,0,"3\t",,terminal_output +571,700802,"TERMINAL",0,0,"4\t",,terminal_output +572,701881,"TERMINAL",0,0,"5\t",,terminal_output +573,702916,"TERMINAL",0,0,"6\t",,terminal_output +574,703934,"TERMINAL",0,0,"7\t",,terminal_output +575,704993,"TERMINAL",0,0,"8\t",,terminal_output +576,706012,"TERMINAL",0,0,"9\t",,terminal_output +577,706736,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +578,863469,"TERMINAL",0,0,"salloc --time=24:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=5 --gres=gpu:1 --cpus-per-task=5",,terminal_command +579,863544,"TERMINAL",0,0,"]633;E;2025-07-09 22:19:27 salloc --time=24:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=5 --gres=gpu:1 --cpus-per-task=5;af8e09eb-6e10-4bc6-b809-4360d3d8a533]633;Csalloc: Pending job allocation 3333583\r\nsalloc: job 3333583 queued and waiting for resources\r\n",,terminal_output +580,869813,"TERMINAL",0,0,"^Csalloc: Job allocation 3333583 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;1",,terminal_output +581,878939,"TERMINAL",0,0,"salloc --time=24:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=5 --gres=gpu:1 --cpus-per-task=5^C",,terminal_command +582,878972,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]633;E;;af8e09eb-6e10-4bc6-b809-4360d3d8a533]633;C]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D",,terminal_output +583,882335,"TERMINAL",0,0,"screen",,terminal_command +584,882414,"TERMINAL",0,0,"]633;E;2025-07-09 22:19:46 screen;af8e09eb-6e10-4bc6-b809-4360d3d8a533]633;Cbash: screen: command not found...\r\n",,terminal_output +585,883624,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;130",,terminal_output +586,885644,"TERMINAL",0,0,"tmux",,terminal_command +587,885758,"TERMINAL",0,0,"]633;E;2025-07-09 22:19:49 tmux;af8e09eb-6e10-4bc6-b809-4360d3d8a533]633;C[?1049h[?1h=[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[>c[>q[?25l\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[0] 0:tmux* ""hkn1993.localdomain"" 22:19 09-Jul-25(B[?12l[?25h(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[?25l\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[0] 0:tmux* ""hkn1993.localdomain"" 22:19 09-Jul-25(B[?12l[?25h",,terminal_output +588,885836,"TERMINAL",0,0,"[?7727h",,terminal_output +589,888714,"TERMINAL",0,0,"jafar[tum_cte0515@hkn1993 logs_mihir]$ [?2004h[?25l[0] 0:bash* ""hkn1993.localdomain"" 22:19 09-Jul-25(B[?12l[?25h",,terminal_output +590,894603,"TERMINAL",0,0,"\r\nlogout\r\n[?2004l(B[?1l>[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l[?7727l[?1004l[?1049l[exited]\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +591,895553,"TERMINAL",0,0,"bash",,terminal_focus +592,896773,"TERMINAL",0,0,"tmux",,terminal_command +593,896829,"TERMINAL",0,0,"]633;E;2025-07-09 22:20:00 tmux;249d0687-2136-49d1-ad97-c40acd813191]633;C[?1049h[?1h=[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[>c[>q[?25l\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[0] 0:tmux* ""hkn1993.localdomain"" 22:20 09-Jul-25(B[?12l[?25h(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[?25l\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[0] 0:tmux* ""hkn1993.localdomain"" 22:20 09-Jul-25(B[?12l[?25h",,terminal_output +594,896957,"TERMINAL",0,0,"[?7727h",,terminal_output +595,898822,"TERMINAL",0,0,"jafar[tum_cte0515@hkn1993 jafar]$ [?2004h[?25l[0] 0:bash* ""hkn1993.localdomain"" 22:20 09-Jul-25(B[?12l[?25h",,terminal_output +596,899101,"TERMINAL",0,0,"\r\n[?2004ljafar[tum_cte0515@hkn1993 jafar]$ [?2004h",,terminal_output +597,900391,"TERMINAL",0,0,"salloc --time=24:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=5 --gres=gpu:1 --cpus-per-task=5(B",,terminal_output +598,902613,"TERMINAL",0,0,"salloc --time=24:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=5 --gres=gpu:1 --cpus-per-task=5\r\n[?2004l",,terminal_output +599,902669,"TERMINAL",0,0,"salloc: Granted job allocation 3333584\r\n",,terminal_output +600,902812,"TERMINAL",0,0,"salloc: Waiting for resource configuration\r\n",,terminal_output +601,906747,"TERMINAL",0,0,"bash",,terminal_focus +602,907970,"TERMINAL",0,0,"queue",,terminal_command +603,908100,"TERMINAL",0,0,"]633;E;2025-07-09 22:20:11 queue;af8e09eb-6e10-4bc6-b809-4360d3d8a533]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:20:11 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3333584 accelerat interact tum_cte0 R\t0:05\t 1 hkn05233331283 accelerat train_la tum_cte0 R 21:41:58\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:41:58\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:41:58\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:41:58\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:41:58\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:41:58\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:42:39\t 2 hkn[0513,0516]",,terminal_output +604,908544,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +605,909517,"TERMINAL",0,0,"idling",,terminal_command +606,909598,"TERMINAL",0,0,"]633;E;2025-07-09 22:20:13 idling;af8e09eb-6e10-4bc6-b809-4360d3d8a533]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Wed Jul 9 22:20:13 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly:\t 9 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated:\t 1 nodes idle\rPartition dev_accelerated-h100 :\t 1 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 7 nodes idle",,terminal_output +607,910626,"TERMINAL",0,0,"4",,terminal_output +608,911671,"TERMINAL",0,0,"5",,terminal_output +609,911825,"TERMINAL",0,0,"[?25l[0] 0:salloc* ""hkn1993.localdomain"" 22:20 09-Jul-25(B[?12l[?25h",,terminal_output +610,912718,"TERMINAL",0,0,"6",,terminal_output +611,913758,"TERMINAL",0,0,"7",,terminal_output +612,914799,"TERMINAL",0,0,"8",,terminal_output +613,915837,"TERMINAL",0,0,"9",,terminal_output +614,916893,"TERMINAL",0,0,"20",,terminal_output +615,917916,"TERMINAL",0,0,"1",,terminal_output +616,918965,"TERMINAL",0,0,"2",,terminal_output +617,920009,"TERMINAL",0,0,"3",,terminal_output +618,921051,"TERMINAL",0,0,"4",,terminal_output +619,922126,"TERMINAL",0,0,"5",,terminal_output +620,923137,"TERMINAL",0,0,"6",,terminal_output +621,924185,"TERMINAL",0,0,"7",,terminal_output +622,925224,"TERMINAL",0,0,"8",,terminal_output +623,925737,"TERMINAL",0,0,"tmux",,terminal_focus +624,926268,"TERMINAL",0,0,"9",,terminal_output +625,927321,"TERMINAL",0,0,"31",,terminal_output +626,928326,"TERMINAL",0,0,"(B[?1l>[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l[?7727l[?1004l[?1049l[detached (from session 0)]\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +627,928375,"TERMINAL",0,0,"2",,terminal_output +628,929396,"TERMINAL",0,0,"3",,terminal_output +629,930430,"TERMINAL",0,0,"4",,terminal_output +630,930598,"TERMINAL",0,0,"tmux",,terminal_command +631,930647,"TERMINAL",0,0,"]633;E;2025-07-09 22:20:34 tmux;249d0687-2136-49d1-ad97-c40acd813191]633;C[?1049h[?1h=[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[>c[>q[?25l\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[1] 0:tmux* ""hkn1993.localdomain"" 22:20 09-Jul-25(B[?12l[?25h(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[?25l\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[1] 0:tmux* ""hkn1993.localdomain"" 22:20 09-Jul-25(B[?12l[?25h",,terminal_output +632,930731,"TERMINAL",0,0,"[?7727h",,terminal_output +633,931483,"TERMINAL",0,0,"5",,terminal_output +634,932500,"TERMINAL",0,0,"6",,terminal_output +635,932807,"TERMINAL",0,0,"jafar[tum_cte0515@hkn1993 jafar]$ [?2004h[?25l[1] 0:bash* ""hkn1993.localdomain"" 22:20 09-Jul-25(B[?12l[?25h",,terminal_output +636,933538,"TERMINAL",0,0,"7",,terminal_output +637,933956,"TERMINAL",0,0,"salloc --time=24:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=5 --gres=gpu:1 --cpus-per-task=5(B",,terminal_output +638,934579,"TERMINAL",0,0,"8",,terminal_output +639,935622,"TERMINAL",0,0,"9",,terminal_output +640,936137,"TERMINAL",0,0,"salloc --time=24:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=5 --gres=gpu:1 --cpus-per-task=5",,terminal_output +641,936670,"TERMINAL",0,0,"40",,terminal_output +642,936827,"TERMINAL",0,0,"",,terminal_output +643,937012,"TERMINAL",0,0,"",,terminal_output +644,937278,"TERMINAL",0,0,"",,terminal_output +645,937526,"TERMINAL",0,0,"",,terminal_output +646,937707,"TERMINAL",0,0,"1",,terminal_output +647,937989,"TERMINAL",0,0,"",,terminal_output +648,938268,"TERMINAL",0,0,"",,terminal_output +649,938634,"TERMINAL",0,0,"",,terminal_output +650,938759,"TERMINAL",0,0,"2",,terminal_output +651,938809,"TERMINAL",0,0,"",,terminal_output +652,939435,"TERMINAL",0,0," --cpus-per-task=5",,terminal_output +653,939577,"TERMINAL",0,0,"4 --cpus-per-task=5",,terminal_output +654,939788,"TERMINAL",0,0,"3",,terminal_output +655,940010,"TERMINAL",0,0,"",,terminal_output +656,940231,"TERMINAL",0,0,"",,terminal_output +657,940407,"TERMINAL",0,0,"",,terminal_output +658,940627,"TERMINAL",0,0,"",,terminal_output +659,940764,"TERMINAL",0,0,"",,terminal_output +660,940831,"TERMINAL",0,0,"4",,terminal_output +661,940899,"TERMINAL",0,0,"",,terminal_output +662,941165,"TERMINAL",0,0,"",,terminal_output +663,941624,"TERMINAL",0,0,"",,terminal_output +664,941879,"TERMINAL",0,0,"5",,terminal_output +665,942102,"TERMINAL",0,0,"",,terminal_output +666,942922,"TERMINAL",0,0,"6",,terminal_output +667,943103,"TERMINAL",0,0,"",,terminal_output +668,943207,"TERMINAL",0,0,"[1@2",,terminal_output +669,943958,"TERMINAL",0,0,"7",,terminal_output +670,944997,"TERMINAL",0,0,"8",,terminal_output +671,946031,"TERMINAL",0,0,"9",,terminal_output +672,946336,"TERMINAL",0,0,"",,terminal_output +673,946566,"TERMINAL",0,0,"",,terminal_output +674,946914,"TERMINAL",0,0,"",,terminal_output +675,947084,"TERMINAL",0,0,"50",,terminal_output +676,947391,"TERMINAL",0,0,"",,terminal_output +677,948125,"TERMINAL",0,0,"1",,terminal_output +678,948372,"TERMINAL",0,0,"",,terminal_output +679,948536,"TERMINAL",0,0,"[1@4",,terminal_output +680,949187,"TERMINAL",0,0,"2",,terminal_output +681,950202,"TERMINAL",0,0,"3",,terminal_output +682,951244,"TERMINAL",0,0,"4",,terminal_output +683,952285,"TERMINAL",0,0,"5",,terminal_output +684,953325,"TERMINAL",0,0,"7",,terminal_output +685,954366,"TERMINAL",0,0,"8",,terminal_output +686,955404,"TERMINAL",0,0,"9",,terminal_output +687,956448,"TERMINAL",0,0,"1:00",,terminal_output +688,956587,"TERMINAL",0,0,"\r\n[?2004lsalloc: Pending job allocation 3333586\r\nsalloc: job 3333586 queued and waiting for resources\r\n",,terminal_output +689,957491,"TERMINAL",0,0,"1",,terminal_output +690,958537,"TERMINAL",0,0,"2",,terminal_output +691,958852,"TERMINAL",0,0,"[?25l[1] 0:salloc* ""hkn1993.localdomain"" 22:21 09-Jul-25(B[?12l[?25h",,terminal_output +692,959576,"TERMINAL",0,0,"3",,terminal_output +693,960613,"TERMINAL",0,0,"4",,terminal_output +694,961662,"TERMINAL",0,0,"5",,terminal_output +695,962722,"TERMINAL",0,0,"6",,terminal_output +696,963747,"TERMINAL",0,0,"7",,terminal_output +697,964829,"TERMINAL",0,0,"8",,terminal_output +698,965630,"TERMINAL",0,0,"tmux --reattach",,terminal_command +699,965666,"TERMINAL",0,0,"]633;E;2025-07-09 22:21:09 tmux --reattach;249d0687-2136-49d1-ad97-c40acd813191]633;Cusage: tmux [-2CDlNuvV] [-c shell-command] [-f file] [-L socket-name]\r\n [-S socket-path] [-T features] [command [flags]]\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;1",,terminal_output +700,965831,"TERMINAL",0,0,"9",,terminal_output +701,966875,"TERMINAL",0,0,"10",,terminal_output +702,967928,"TERMINAL",0,0,"1",,terminal_output +703,968963,"TERMINAL",0,0,"2",,terminal_output +704,970013,"TERMINAL",0,0,"3",,terminal_output +705,971040,"TERMINAL",0,0,"4",,terminal_output +706,972132,"TERMINAL",0,0,"5",,terminal_output +707,972281,"TERMINAL",0,0,"tmux --help",,terminal_command +708,972326,"TERMINAL",0,0,"]633;E;2025-07-09 22:21:16 tmux --help;249d0687-2136-49d1-ad97-c40acd813191]633;Cusage: tmux [-2CDlNuvV] [-c shell-command] [-f file] [-L socket-name]\r\n [-S socket-path] [-T features] [command [flags]]\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;1",,terminal_output +709,973157,"TERMINAL",0,0,"6",,terminal_output +710,974218,"TERMINAL",0,0,"7",,terminal_output +711,975223,"TERMINAL",0,0,"8",,terminal_output +712,976284,"TERMINAL",0,0,"9",,terminal_output +713,976802,"TERMINAL",0,0,"tmux -h",,terminal_command +714,976841,"TERMINAL",0,0,"]633;E;2025-07-09 22:21:20 tmux -h;249d0687-2136-49d1-ad97-c40acd813191]633;Ctmux: unknown option -- h\r\nusage: tmux [-2CDlNuvV] [-c shell-command] [-f file] [-L socket-name]\r\n [-S socket-path] [-T features] [command [flags]]\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;1",,terminal_output +715,977310,"TERMINAL",0,0,"20",,terminal_output +716,978348,"TERMINAL",0,0,"2",,terminal_output +717,979400,"TERMINAL",0,0,"3",,terminal_output +718,980443,"TERMINAL",0,0,"4",,terminal_output +719,981484,"TERMINAL",0,0,"5",,terminal_output +720,982530,"TERMINAL",0,0,"6",,terminal_output +721,983571,"TERMINAL",0,0,"7",,terminal_output +722,984632,"TERMINAL",0,0,"8",,terminal_output +723,985659,"TERMINAL",0,0,"9",,terminal_output +724,986704,"TERMINAL",0,0,"30",,terminal_output +725,987753,"TERMINAL",0,0,"1",,terminal_output +726,988787,"TERMINAL",0,0,"2",,terminal_output +727,989831,"TERMINAL",0,0,"3",,terminal_output +728,990879,"TERMINAL",0,0,"4",,terminal_output +729,991925,"TERMINAL",0,0,"5",,terminal_output +730,993006,"TERMINAL",0,0,"6",,terminal_output +731,994052,"TERMINAL",0,0,"7",,terminal_output +732,995047,"TERMINAL",0,0,"8",,terminal_output +733,995193,"TERMINAL",0,0,"tmux a",,terminal_command +734,995272,"TERMINAL",0,0,"]633;E;2025-07-09 22:21:38 tmux a ;249d0687-2136-49d1-ad97-c40acd813191]633;C[?1049h[?1h=[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[>c[>q[?25ljafar[tum_cte0515@hkn1993 jafar]$ salloc --time=24:00:00 --partition=accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5\r\nsalloc: Pending job allocation 3333586\r\nsalloc: job 3333586 queued and waiting for resources\r\n\r\n\r\n\r\n\r\n\r\n\r\n[1] 0:salloc* ""hkn1993.localdomain"" 22:21 09-Jul-25(B[?12l[?25h(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[?25ljafar[tum_cte0515@hkn1993 jafar]$ salloc --time=24:00:00 --partition=accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5\r\nsalloc: Pending job allocation 3333586\r\nsalloc: job 3333586 queued and waiting for resources\r\n\r\n\r\n\r\n\r\n\r\n\r\n[1] 0:salloc* ""hkn1993.localdomain"" 22:21 09-Jul-25(B[?12l[?25h",,terminal_output +735,995364,"TERMINAL",0,0,"[?7727h",,terminal_output +736,996137,"TERMINAL",0,0,"9",,terminal_output +737,997155,"TERMINAL",0,0,"40",,terminal_output +738,998197,"TERMINAL",0,0,"1",,terminal_output +739,999242,"TERMINAL",0,0,"2",,terminal_output +740,1000378,"TERMINAL",0,0,"313",,terminal_output +741,1001339,"TERMINAL",0,0,"5",,terminal_output +742,1001548,"TERMINAL",0,0,"(B[?1l>[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l[?7727l[?1004l[?1049l[detached (from session 1)]\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +743,1002376,"TERMINAL",0,0,"6",,terminal_output +744,1003423,"TERMINAL",0,0,"7",,terminal_output +745,1004467,"TERMINAL",0,0,"8",,terminal_output +746,1005512,"TERMINAL",0,0,"9",,terminal_output +747,1006645,"TERMINAL",0,0,"50",,terminal_output +748,1007424,"TERMINAL",0,0,"tmux a 2",,terminal_command +749,1007611,"TERMINAL",0,0,"1",,terminal_output +750,1008655,"TERMINAL",0,0,"2",,terminal_output +751,1009698,"TERMINAL",0,0,"3",,terminal_output +752,1010742,"TERMINAL",0,0,"4",,terminal_output +753,1011801,"TERMINAL",0,0,"5",,terminal_output +754,1012839,"TERMINAL",0,0,"6",,terminal_output +755,1013903,"TERMINAL",0,0,"7",,terminal_output +756,1014950,"TERMINAL",0,0,"8",,terminal_output +757,1015966,"TERMINAL",0,0,"9",,terminal_output +758,1017014,"TERMINAL",0,0,"2:00",,terminal_output +759,1018048,"TERMINAL",0,0,"1",,terminal_output +760,1019146,"TERMINAL",0,0,"2",,terminal_output +761,1020152,"TERMINAL",0,0,"3",,terminal_output +762,1021182,"TERMINAL",0,0,"4",,terminal_output +763,1022219,"TERMINAL",0,0,"5",,terminal_output +764,1023263,"TERMINAL",0,0,"6",,terminal_output +765,1024167,"TERMINAL",0,0,"watch",,terminal_focus +766,1024307,"TERMINAL",0,0,"7",,terminal_output +767,1024924,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +768,1030504,"TERMINAL",0,0,"queue",,terminal_command +769,1030572,"TERMINAL",0,0,"]633;E;2025-07-09 22:22:14 queue;af8e09eb-6e10-4bc6-b809-4360d3d8a533]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Wed Jul 9 22:22:14 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3333586 accelerat interact tum_cte0 PD\t0:00\t 2 (Resources)3333584 accelerat interact tum_cte0 R\t2:08\t 1 hkn05233331283 accelerat train_la tum_cte0 R 21:44:01\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 21:44:01\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 21:44:01\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 21:44:01\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 21:44:01\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 21:44:01\t 2 hkn[0809,0811]3331282 accelerat train_to tum_cte0 R 21:44:42\t 2 hkn[0513,0516]",,terminal_output +770,1031622,"TERMINAL",0,0,"592222223",,terminal_output +771,1032672,"TERMINAL",0,0,"6103333334",,terminal_output +772,1033751,"TERMINAL",0,0,"714444445",,terminal_output +773,1034804,"TERMINAL",0,0,"825555556",,terminal_output +774,1035815,"TERMINAL",0,0,"936666667",,terminal_output +775,1036882,"TERMINAL",0,0,"2047777778",,terminal_output +776,1037903,"TERMINAL",0,0,"158888889",,terminal_output +777,1038951,"TERMINAL",0,0,"2699999950",,terminal_output +778,1040019,"TERMINAL",0,0,"371010101010101",,terminal_output +779,1041043,"TERMINAL",0,0,"481111112",,terminal_output +780,1042127,"TERMINAL",0,0,"592222223",,terminal_output +781,1043141,"TERMINAL",0,0,"6203333334",,terminal_output +782,1044227,"TERMINAL",0,0,"714444445",,terminal_output +783,1045341,"TERMINAL",0,0,"825555556",,terminal_output +784,1046304,"TERMINAL",0,0,"947777778",,terminal_output +785,1047361,"TERMINAL",0,0,"3158888889",,terminal_output +786,1048382,"TERMINAL",0,0,"269999995:00",,terminal_output +787,1049431,"TERMINAL",0,0,"372020202020201",,terminal_output +788,1050477,"TERMINAL",0,0,"481111112",,terminal_output +789,1051537,"TERMINAL",0,0,"592222223",,terminal_output +790,1052575,"TERMINAL",0,0,"6303333334",,terminal_output +791,1053623,"TERMINAL",0,0,"714444445",,terminal_output +792,1054694,"TERMINAL",0,0,"825555556",,terminal_output +793,1055765,"TERMINAL",0,0,"936666667",,terminal_output +794,1056788,"TERMINAL",0,0,"4047777778",,terminal_output +795,1057839,"TERMINAL",0,0,"158888889",,terminal_output +796,1058895,"TERMINAL",0,0,"2699999910",,terminal_output +797,1059956,"TERMINAL",0,0,"373030303030301",,terminal_output +798,1061006,"TERMINAL",0,0,"481111112",,terminal_output +799,1062046,"TERMINAL",0,0,"592222223",,terminal_output +800,1063150,"TERMINAL",0,0,"6403333334",,terminal_output +801,1064185,"TERMINAL",0,0,"714444445",,terminal_output +802,1065219,"TERMINAL",0,0,"825555556",,terminal_output +803,1066257,"TERMINAL",0,0,"936666667",,terminal_output +804,1067302,"TERMINAL",0,0,"5058888889",,terminal_output +805,1068352,"TERMINAL",0,0,"2699999920",,terminal_output +806,1069394,"TERMINAL",0,0,"374040404040401",,terminal_output +807,1070443,"TERMINAL",0,0,"481111112",,terminal_output +808,1071502,"TERMINAL",0,0,"592222223",,terminal_output +809,1072553,"TERMINAL",0,0,"6503333334",,terminal_output +810,1073588,"TERMINAL",0,0,"714444445",,terminal_output +811,1074637,"TERMINAL",0,0,"825555556",,terminal_output +812,1075784,"TERMINAL",0,0,"936666667",,terminal_output +813,1076736,"TERMINAL",0,0,"3:0047777778",,terminal_output +814,1077789,"TERMINAL",0,0,"158888889",,terminal_output +815,1078840,"TERMINAL",0,0,"2699999930",,terminal_output +816,1079898,"TERMINAL",0,0,"375050505050501",,terminal_output +817,1080964,"TERMINAL",0,0,"481111112",,terminal_output +818,1082004,"TERMINAL",0,0,"592222223",,terminal_output +819,1083063,"TERMINAL",0,0,"63:003333334",,terminal_output +820,1084103,"TERMINAL",0,0,"714444445",,terminal_output +821,1085196,"TERMINAL",0,0,"825555556",,terminal_output +822,1086202,"TERMINAL",0,0,"936666667",,terminal_output +823,1087240,"TERMINAL",0,0,"1047777778",,terminal_output +824,1088278,"TERMINAL",0,0,"1699999940",,terminal_output +825,1089331,"TERMINAL",0,0,"375:005:005:005:005:005:001",,terminal_output +826,1090379,"TERMINAL",0,0,"481111112",,terminal_output +827,1091442,"TERMINAL",0,0,"592222223",,terminal_output +828,1092492,"TERMINAL",0,0,"6103333334",,terminal_output +829,1093542,"TERMINAL",0,0,"714444445",,terminal_output +830,1094593,"TERMINAL",0,0,"825555556",,terminal_output +831,1095640,"TERMINAL",0,0,"936666667",,terminal_output +832,1096688,"TERMINAL",0,0,"2047777778",,terminal_output +833,1097743,"TERMINAL",0,0,"158888889",,terminal_output +834,1098794,"TERMINAL",0,0,"2699999950",,terminal_output +835,1099838,"TERMINAL",0,0,"371010101010101",,terminal_output +836,1100906,"TERMINAL",0,0,"481111112",,terminal_output +837,1102010,"TERMINAL",0,0,"592222223",,terminal_output +838,1102982,"TERMINAL",0,0,"6203333334",,terminal_output +839,1104045,"TERMINAL",0,0,"714444445",,terminal_output +840,1105138,"TERMINAL",0,0,"825555556",,terminal_output +841,1106200,"TERMINAL",0,0,"936666667",,terminal_output +842,1107197,"TERMINAL",0,0,"3047777778",,terminal_output +843,1107898,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +844,2647719,"genie.py",0,0,"from typing import Dict, Any\n\nimport optax\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nfrom jax import NamedSharding\nfrom flax.training.train_state import TrainState\nfrom flax.training import orbax_utils\nfrom orbax.checkpoint import PyTreeCheckpointer\n\nfrom models.dynamics import DynamicsMaskGIT\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\n\nclass Genie(nn.Module):\n """"""Genie model""""""\n\n # --- Tokenizer ---\n in_dim: int\n tokenizer_dim: int\n latent_patch_dim: int\n num_patch_latents: int\n patch_size: int\n tokenizer_num_blocks: int\n tokenizer_num_heads: int\n # --- LAM ---\n lam_dim: int\n latent_action_dim: int\n num_latent_actions: int\n lam_patch_size: int\n lam_num_blocks: int\n lam_num_heads: int\n # --- Dynamics ---\n dyna_dim: int\n dyna_num_blocks: int\n dyna_num_heads: int\n dropout: float = 0.0\n mask_limit: float = 0.0\n\n def setup(self):\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n )\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_latent_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n )\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(tokenizer_outputs[""indices""]),\n latent_actions=jax.lax.stop_gradient(lam_outputs[""z_q""]),\n )\n outputs[""mask_rng""] = batch[""mask_rng""]\n dyna_outputs = self.dynamics(outputs, training)\n outputs.update(dyna_outputs)\n mle_indices = jnp.argmax(outputs[""token_logits""], axis=-1)\n outputs[""recon""] = self.tokenizer.decode(\n mle_indices, batch[""videos""].shape[2:4]\n )\n return outputs\n\n @nn.compact\n def sample_mihir(\n self,\n batch: Dict[str, Any],\n seq_len: int,\n steps: int,\n temperature: float,\n sample_argmax: bool,\n ) -> Any:\n # B == batch_size\n # T == num_frames (input)\n # N == num_patches\n # S == seq_len\n # A == action_space\n # D == latent_dim\n # --- Encode videos and actions ---\n tokenizer_out = self.tokenizer.vq_encode(batch[""videos""], training=False)\n token_idxs = tokenizer_out[""indices""] # (B, T, N)\n B, T, N = token_idxs.shape\n S = seq_len\n print(""token_idxs shape:"", token_idxs.shape)\n pad_shape = (B, S - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs.dtype)\n token_idxs = jnp.concatenate([token_idxs, pad], axis=1) # shape (B, S, N)\n action_tokens = self.lam.vq.get_codes(batch[""latent_actions""])\n\n # begin potential forloop (from T to S)\n initial_T = T\n for T in range(initial_T, S):\n # Create a mask that is 1 (True) where we just padded\n # token_idxs shape: (B, S, N), T = original length, S = seq_len\n # mask is True for padded positions (i.e., t >= T)\n mask = (jnp.arange(S)[None, :, None] >= T) # shape (1, S, 1)\n mask = jnp.broadcast_to(mask, (B, S, N)) # shape (B, S, N)\n init_mask = mask.astype(bool)\n token_idxs *= ~init_mask\n #print(""token_idxs[0,:,0]:"", token_idxs[0,:,0])\n #print(""init_mask[0,:,0]:"", init_mask[0,:,0])\n\n assert init_mask.shape == (B, S, N), ""Wrong mask shape""\n\n # --- Initialize MaskGIT ---\n init_carry = (\n batch[""rng""],\n init_mask,\n token_idxs,\n action_tokens,\n )\n MaskGITLoop = nn.scan(\n MaskGITStepMihir,\n variable_broadcast=""params"",\n split_rngs={""params"": False},\n in_axes=0,\n out_axes=0,\n length=steps,\n )\n\n # --- Run MaskGIT loop ---\n loop_fn = MaskGITLoop(\n dynamics=self.dynamics,\n tokenizer=self.tokenizer,\n temperature=temperature,\n sample_argmax=sample_argmax,\n steps=steps,\n )\n final_carry, _ = loop_fn(init_carry, jnp.arange(steps))\n token_idxs = final_carry[2]\n\n new_frame_pixels = self.tokenizer.decode(\n token_idxs,\n video_hw=batch[""videos""].shape[2:4],\n )\n return new_frame_pixels\n\n @nn.compact\n def sample(\n self,\n batch: Dict[str, Any],\n steps: int = 25,\n temperature: int = 1,\n sample_argmax: bool = False,\n ) -> Any:\n # B == batch_size\n # T == num_frames (input)\n # N == num_patches\n # S == seq_len\n # A == action_space\n # D == latent_dim\n # --- Encode videos and actions ---\n tokenizer_out = self.tokenizer.vq_encode(batch[""videos""], training=False)\n token_idxs = tokenizer_out[""indices""]# (B, T, N)\n new_frame_idxs = jnp.zeros_like(token_idxs)[:, 0] # (B, N) \n action_tokens = self.lam.vq.get_codes(batch[""latent_actions""]) # (B, S, A, D)\n\n # --- Initialize MaskGIT ---\n init_mask = jnp.ones_like(token_idxs, dtype=bool)[:, 0] # (B, N)\n init_carry = (\n batch[""rng""],\n new_frame_idxs,\n init_mask,\n token_idxs,\n action_tokens,\n )\n MaskGITLoop = nn.scan(\n MaskGITStep,\n variable_broadcast=""params"",\n split_rngs={""params"": False},\n in_axes=0,\n out_axes=0,\n length=steps,\n )\n\n # --- Run MaskGIT loop ---\n loop_fn = MaskGITLoop(\n dynamics=self.dynamics,\n tokenizer=self.tokenizer,\n temperature=temperature,\n sample_argmax=sample_argmax,\n steps=steps,\n )\n final_carry, _ = loop_fn(init_carry, jnp.arange(steps))\n new_frame_idxs = final_carry[1]\n new_frame_pixels = self.tokenizer.decode(\n jnp.expand_dims(new_frame_idxs, 1),\n video_hw=batch[""videos""].shape[2:4],\n )\n return new_frame_pixels\n\n def vq_encode(self, batch, training) -> Dict[str, Any]:\n # --- Preprocess videos ---\n lam_output = self.lam.vq_encode(batch[""videos""], training=training)\n return lam_output[""indices""]\n\n\nclass MaskGITStepMihir(nn.Module):\n dynamics: nn.Module\n tokenizer: nn.Module\n temperature: float\n sample_argmax: bool\n steps: int\n\n @nn.compact\n def __call__(self, carry, x):\n rng, mask, token_idxs, action_tokens = carry\n step = x\n B, S, N = token_idxs.shape[:3]\n\n # --- Construct + encode video ---\n vid_embed = self.dynamics.patch_embed(token_idxs)\n # Mask vid_embed: set to mask_token where mask==1, else keep vid_embed\n # mask: (B, S, N), vid_embed: (B, S, N, D), mask_token: (D,)\n mask_token = self.dynamics.mask_token # (1,1, 1, D,)\n # Expand mask to (B, S, N, 1) for broadcasting\n mask_expanded = mask[..., None]\n vid_embed = jnp.where(mask_expanded, mask_token, vid_embed)\n\n # --- Predict transition ---\n act_embed = self.dynamics.action_up(action_tokens)\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (self.steps * 2))\n step_temp = self.temperature * (1.0 - unmasked_ratio)\n final_logits = self.dynamics.dynamics(vid_embed)[:, -1] / step_temp\n\n # --- Sample new tokens for final frame ---\n if self.sample_argmax:\n sampled_token_idxs = jnp.argmax(final_logits, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs = jnp.where(\n step == self.steps - 1,\n jnp.argmax(final_logits, axis=-1),\n jax.random.categorical(_rng, final_logits),\n )\n gather_fn = jax.vmap(jax.vmap(lambda x, y: x[y]))\n final_token_probs = gather_fn(jax.nn.softmax(final_logits), sampled_token_idxs)\n final_token_probs += ~mask\n # Update masked tokens only\n jax.debug.print(""maskgit-sampled-token_idxs[0,:,0]: {}"", token_idxs[0,:,0])\n token_idxs = jnp.where(mask, sampled_token_idxs, token_idxs)\n jax.debug.print(""maskgit-token_idxs[0,:,0]: {}"", token_idxs[0,:,0])\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n idx_mask = jnp.arange(final_token_probs.shape[-1]) > num_unmasked_tokens\n sorted_idxs = jnp.argsort(final_token_probs, axis=-1, descending=True)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask))\n new_mask = mask_update_fn(mask, sorted_idxs)\n\n new_carry = (rng, new_mask, token_idxs, action_tokens)\n return new_carry, None\n\n\nclass MaskGITStep(nn.Module):\n dynamics: nn.Module\n tokenizer: nn.Module\n temperature: float\n sample_argmax: bool\n steps: int\n\n @nn.compact\n def __call__(self, carry, x):\n rng, final_token_idxs, mask, token_idxs, action_tokens = carry\n step = x\n B, T, N = token_idxs.shape[:3]\n\n # --- Construct + encode video ---\n vid_token_idxs = jnp.concatenate(\n (token_idxs, jnp.expand_dims(final_token_idxs, 1)), axis=1\n ) # (B, T+1, N)\n vid_embed = self.dynamics.patch_embed(vid_token_idxs) # (B, T+1, N, D)\n curr_masked_frame = jnp.where(\n jnp.expand_dims(mask, -1), # (B, N, 1)\n self.dynamics.mask_token[0], # (B, 1, D)\n vid_embed[:, -1], # (B, N, D)\n ) # (B, N, D)\n vid_embed = vid_embed.at[:, -1].set(curr_masked_frame)\n\n # --- Predict transition ---\n act_embed = self.dynamics.action_up(action_tokens)\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (self.steps * 2))\n step_temp = self.temperature * (1.0 - unmasked_ratio)\n final_logits = self.dynamics.dynamics(vid_embed)[:, -1] / step_temp\n\n # --- Sample new tokens for final frame ---\n if self.sample_argmax:\n sampled_token_idxs = jnp.argmax(final_logits, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs = jnp.where(\n step == self.steps - 1,\n jnp.argmax(final_logits, axis=-1),\n jax.random.categorical(_rng, final_logits),\n )\n gather_fn = jax.vmap(jax.vmap(lambda x, y: x[y]))\n final_token_probs = gather_fn(jax.nn.softmax(final_logits), sampled_token_idxs)\n final_token_probs += ~mask\n # Update masked tokens only\n new_token_idxs = jnp.where(mask, sampled_token_idxs, final_token_idxs)\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n idx_mask = jnp.arange(final_token_probs.shape[-1]) > num_unmasked_tokens\n sorted_idxs = jnp.argsort(final_token_probs, axis=-1, descending=True)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask))\n new_mask = mask_update_fn(mask, sorted_idxs)\n\n new_carry = (rng, new_token_idxs, new_mask, token_idxs, action_tokens)\n return new_carry, None\n\n\ndef restore_genie_components(\n train_state: TrainState,\n sharding: NamedSharding,\n inputs: Dict[str, jax.Array],\n rng: jax.Array,\n args,\n):\n """"""Restore pre-trained Genie components""""""\n rng, _rng = jax.random.split(rng)\n\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_latent_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n )\n tokenizer_init_params = dummy_tokenizer.init(_rng, inputs)\n lam_init_params = dummy_lam.init(_rng, inputs)\n\n # dummy values since we only use tx to initialize the dummy train states\n dummy_tx = optax.adamw(\n learning_rate=optax.constant_schedule(args.max_lr),\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n )\n\n dummy_tokenizer_train_state = TrainState.create(\n apply_fn=dummy_tokenizer.apply, params=tokenizer_init_params, tx=dummy_tx\n )\n dummy_lam_train_state = TrainState.create(\n apply_fn=dummy_lam.apply, params=lam_init_params, tx=dummy_tx\n )\n\n def create_abstract_sharded_pytree(pytree_template, sharding_spec):\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)\n\n abstract_sharded_tokenizer_state = create_abstract_sharded_pytree(\n dummy_tokenizer_train_state, sharding\n )\n abstract_sharded_lam_state = create_abstract_sharded_pytree(\n dummy_lam_train_state, sharding\n )\n\n tokenizer_restore_target = {""model"": abstract_sharded_tokenizer_state}\n lam_restore_target = {""model"": abstract_sharded_lam_state}\n\n tokenizer_restore_args = orbax_utils.restore_args_from_target(\n tokenizer_restore_target\n )\n lam_restore_args = orbax_utils.restore_args_from_target(lam_restore_target)\n\n restored_tokenizer_params = (\n PyTreeCheckpointer()\n .restore(\n args.tokenizer_checkpoint,\n item=tokenizer_restore_target,\n restore_args=tokenizer_restore_args,\n )[""model""]\n .params[""params""]\n )\n restored_lam_params = (\n PyTreeCheckpointer()\n .restore(\n args.lam_checkpoint, item=lam_restore_target, restore_args=lam_restore_args\n )[""model""]\n .params[""params""]\n )\n # Genie does not initialize all LAM modules, thus we omit those extra modules during restoration\n # (f.srambical) FIXME: Currently, this is a small HBM memory crunch since the LAM's decoder is loaded into HBM and immediately dicarded.\n # A workaround would be to restore to host memory first, and only move the weights to HBM after pruning the decoder\n restored_lam_params = {\n k: v\n for k, v in restored_lam_params.items()\n if k in train_state.params[""params""][""lam""]\n }\n\n train_state.params[""params""][""tokenizer""].update(restored_tokenizer_params)\n train_state.params[""params""][""lam""].update(restored_lam_params)\n\n return train_state\n",python,tab diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-5612308a-14dd-4f05-ae65-c6cd496f68351752499707410-2025_07_14-15.29.02.547/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-5612308a-14dd-4f05-ae65-c6cd496f68351752499707410-2025_07_14-15.29.02.547/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..2330b57218282aa5f45b812dfce9366c34fb4f9e --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-5612308a-14dd-4f05-ae65-c6cd496f68351752499707410-2025_07_14-15.29.02.547/source.csv @@ -0,0 +1,127 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +2,206,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"3:29:02 PM [info] Activating crowd-code\n3:29:02 PM [info] Recording started\n3:29:02 PM [info] Initializing git provider using file system watchers...\n",Log,tab +3,365,"extension-output-pdoom-org.crowd-code-#1-crowd-code",150,0,"3:29:02 PM [info] Git repository found\n3:29:02 PM [info] Git provider initialized successfully\n3:29:02 PM [info] Initial git state: [object Object]\n",Log,content +4,2862,"TERMINAL",0,0,"/bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt",,terminal_command +5,2909,"TERMINAL",0,0,"]633;E;2025-07-14 15:29:05 /bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt;96649176-7412-48ad-98ea-7639505dc671]633;C",,terminal_output +6,2952,"TERMINAL",0,0,"]0;tum_cte0515@hkn1990:/hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash]633;D;0",,terminal_output +7,7671,"TERMINAL",0,0,"watch",,terminal_focus +8,24040,"TERMINAL",0,0,"salloc --time=10:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5",,terminal_command +9,24098,"TERMINAL",0,0,"]633;E;2025-07-14 15:29:26 salloc --time=10:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5;ba61a590-3751-4127-93b1-5716be1c7bd0]633;Csalloc: Pending job allocation 3344502\r\nsalloc: job 3344502 queued and waiting for resources\r\n",,terminal_output +10,25688,"TERMINAL",0,0,"^Csalloc: Job allocation 3344502 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;1]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +11,30368,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5",,terminal_command +12,30400,"TERMINAL",0,0,"]633;E;2025-07-14 15:29:32 salloc --time=01:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5;ba61a590-3751-4127-93b1-5716be1c7bd0]633;Csalloc: Pending job allocation 3344503\r\nsalloc: job 3344503 queued and waiting for resources\r\n",,terminal_output +13,31244,"TERMINAL",0,0,"^Csalloc: Job allocation 3344503 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;1",,terminal_output +14,35141,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5",,terminal_command +15,35153,"TERMINAL",0,0,"]633;E;2025-07-14 15:29:37 salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5;ba61a590-3751-4127-93b1-5716be1c7bd0]633;Csalloc: Granted job allocation 3344505\r\n",,terminal_output +16,35274,"TERMINAL",0,0,"salloc: Waiting for resource configuration\r\n",,terminal_output +17,37228,"TERMINAL",0,0,"bash",,terminal_focus +18,38854,"TERMINAL",0,0,"queue",,terminal_command +19,38926,"TERMINAL",0,0,"]633;E;2025-07-14 15:29:41 queue;450fbccd-48d1-432b-b44b-06394616d158]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1990.localdomain: Mon Jul 14 15:29:41 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3341079 accelerat train_to tum_cte0 R 1-04:59:15\t 2 hkn[0628-0629]3341080 accelerat train_to tum_cte0 R 1-04:59:15\t 2 hkn[0631-0632]3344246 accelerat interact tum_cte0 R 1:13:32\t 2 hkn[0626,0630]3344505 dev_accel interact tum_cte0 R\t0:04\t 1 hkn0402",,terminal_output +20,39972,"TERMINAL",0,0,"26635",,terminal_output +21,41004,"TERMINAL",0,0,"37746",,terminal_output +22,42041,"TERMINAL",0,0,"48857",,terminal_output +23,43101,"TERMINAL",0,0,"59968",,terminal_output +24,44132,"TERMINAL",0,0,"6202079",,terminal_output +25,45165,"TERMINAL",0,0,"711810",,terminal_output +26,46236,"TERMINAL",0,0,"82291",,terminal_output +27,46520,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;0",,terminal_output +28,49617,"TERMINAL",0,0,"idling",,terminal_command +29,49671,"TERMINAL",0,0,"]633;E;2025-07-14 15:29:52 idling;450fbccd-48d1-432b-b44b-06394616d158]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1990.localdomain: Mon Jul 14 15:29:52 2025Partition dev_cpuonly: 10 nodes idle\rPartition cpuonly: 42 nodes idle\rPartition dev_accelerated:\t 1 nodes idle\rPartition accelerated: 40 nodes idle\rPartition dev_accelerated-h100 :\t 1 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 7 nodes idle",,terminal_output +30,50700,"TERMINAL",0,0,"3",,terminal_output +31,51749,"TERMINAL",0,0,"4",,terminal_output +32,52802,"TERMINAL",0,0,"5",,terminal_output +33,53446,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;0",,terminal_output +34,62317,"TERMINAL",0,0,"salloc: Nodes hkn0402 are ready for job\r\n",,terminal_output +35,63118,"TERMINAL",0,0,"]0;tum_cte0515@hkn0402:~/Projects/jafar[?2004h[tum_cte0515@hkn0402 jafar]$ ",,terminal_output +36,84352,"TERMINAL",0,0,"srun",,terminal_focus +37,85382,"TERMINAL",0,0,"[?25lso[?25h[?25lo[?25h",,terminal_output +38,85520,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +39,85615,"TERMINAL",0,0,"[?25lr[?25h",,terminal_output +40,85840,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +41,85947,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +42,86009,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +43,86141,"TERMINAL",0,0,"[?25l.[?25h[?25lv[?25h",,terminal_output +44,86295,"TERMINAL",0,0,"env/",,terminal_output +45,86587,"TERMINAL",0,0,"",,terminal_output +46,86945,"TERMINAL",0,0,"[?25lb[?25h",,terminal_output +47,87054,"TERMINAL",0,0,"in/",,terminal_output +48,87522,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +49,87637,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +50,87784,"TERMINAL",0,0,"tivate",,terminal_output +51,88103,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0402:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0402 jafar]$ ",,terminal_output +52,94820,"sample.py",0,0,"from dataclasses import dataclass\nimport time\nimport os\n\nimport dm_pix as pix\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nimport numpy as np\nfrom orbax.checkpoint import PyTreeCheckpointer\nfrom flax.training.train_state import TrainState\nimport grain\nimport orbax.checkpoint as ocp\nimport optax\nfrom PIL import Image, ImageDraw\nimport tyro\n\nfrom genie import Genie\nfrom utils.dataloader import get_dataloader\n\n\n@dataclass\nclass Args:\n # Experiment\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data/coinrun_episodes""\n checkpoint: str = """"\n # Sampling\n batch_size: int = 1\n maskgit_steps: int = 25\n temperature: float = 1.0\n sample_argmax: bool = True\n start_frame: int = 0\n # Tokenizer checkpoint\n tokenizer_dim: int = 512\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 8\n tokenizer_num_heads: int = 8\n # LAM checkpoint\n lam_dim: int = 512\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 8\n lam_num_heads: int = 8\n # Dynamics checkpoint\n dyna_dim: int = 512\n dyna_num_blocks: int = 12\n dyna_num_heads: int = 8\n\n\nargs = tyro.cli(Args)\nrng = jax.random.PRNGKey(args.seed)\n\n# --- Load Genie checkpoint ---\ngenie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=True,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n)\nrng, _rng = jax.random.split(rng)\nimage_shape = (args.image_height, args.image_width, args.image_channels)\ndummy_inputs = dict(\n videos=jnp.zeros((args.batch_size, args.seq_len, *image_shape), dtype=jnp.float32),\n mask_rng=_rng,\n)\nrng, _rng = jax.random.split(rng)\nparams = genie.init(_rng, dummy_inputs)\n\n# ckpt = PyTreeCheckpointer().restore(args.checkpoint)[""model""][""params""][""params""]\n# Create a dummy TrainState for checkpoint structure\nlr_schedule = optax.warmup_cosine_decay_schedule(\n 0.0, 3e-4, 1000, 300000\n)\ntx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\ndummy_train_state = TrainState.create(\n apply_fn=genie.apply,\n params=params, # or params if your model expects that\n tx=tx, # No optimizer needed for eval\n)\n\nhandler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\nhandler_registry.add('model_state', ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler)\nhandler_registry.add('model_state', ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler)\n\n\n# Set up Orbax CheckpointManager\ncheckpoint_manager = ocp.CheckpointManager(\n args.checkpoint, # Directory containing the checkpoint\n options=ocp.CheckpointManagerOptions(max_to_keep=1, step_format_fixed_length=6),\n handler_registry=handler_registry\n)\n\n# Prepare abstract state for restore\nabstract_train_state = jax.tree_util.tree_map(\n ocp.utils.to_shape_dtype_struct, dummy_train_state\n)\n\n# Restore the checkpoint (only model_state needed for sampling)\nrestored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_train_state),\n ),\n)\nrestored_train_state = restored[""model_state""]\n\n# Extract model parameters for inference\nparams = restored_train_state.params\n\n\n# params[""params""].update(ckpt)\n\n\ndef _sampling_wrapper(module, batch):\n return module.sample(batch, args.seq_len, args.maskgit_steps, args.temperature, args.sample_argmax)\n\n# --- Define autoregressive sampling loop ---\ndef _autoreg_sample(rng, video_batch, action_batch):\n vid = video_batch[:, : args.start_frame + 1]\n sampling_fn = jax.jit(nn.apply(_sampling_wrapper, genie)) \n rng, _rng = jax.random.split(rng)\n batch = dict(videos=vid, latent_actions=action_batch, rng=_rng)\n generated_vid = sampling_fn(\n params,\n batch\n )\n return generated_vid\n\n# # --- Get video + latent actions ---\narray_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n]\ngrain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n)\ninitial_state = grain_dataloader._create_initial_state()\ngrain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\njax.debug.breakpoint()\nvideo_batch = next(grain_iterator)\njax.debug.breakpoint()\n# video_batch = np.load(""overfit_dir/single_sample_corner.npy"")\n# Get latent actions for all videos in the batch\nbatch = dict(videos=video_batch)\naction_batch = genie.apply(params, batch, False, method=Genie.vq_encode)\naction_batch = action_batch.reshape(video_batch.shape[0], args.seq_len - 1, 1)\n\n# --- Sample + evaluate video ---\nvid = _autoreg_sample(rng, video_batch, action_batch)\ngt = video_batch[:, : vid.shape[1]].clip(0, 1).reshape(-1, *video_batch.shape[2:])\nrecon = vid.clip(0, 1).reshape(-1, *vid.shape[2:])\nssim = pix.ssim(gt[:, args.start_frame + 1 :], recon[:, args.start_frame + 1 :]).mean()\nprint(f""SSIM: {ssim}"")\n\n# --- Construct video ---\ntrue_videos = (video_batch * 255).astype(np.uint8)\npred_videos = (vid * 255).astype(np.uint8)\nvideo_comparison = np.zeros((2, *vid.shape), dtype=np.uint8)\nvideo_comparison[0] = true_videos[:, :args.seq_len]\nvideo_comparison[1] = pred_videos\nframes = einops.rearrange(video_comparison, ""n b t h w c -> t (b h) (n w) c"")\n\n# --- Save video --- \nimgs = [Image.fromarray(img) for img in frames]\n# Write actions on each frame, on each row (i.e., for each video in the batch, on the GT row)\nfor t, img in enumerate(imgs[1:]):\n d = ImageDraw.Draw(img)\n for row in range(action_batch.shape[0]):\n action = action_batch[row, t, 0]\n y_offset = row * video_batch.shape[2] + 2\n d.text((2, y_offset), f""{action}"", fill=255)\nimgs[0].save(\n f""generation_{time.time()}.gif"",\n save_all=True,\n append_images=imgs[1:],\n duration=250,\n loop=0,\n)\n",python,tab +53,94915,"TERMINAL",0,0,"\r(jafar) [tum_cte0515@hkn0402 jafar]$ \r(jafar) [tum_cte0515@hkn0402 jafar]$ ",,terminal_output +54,148636,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_output +55,148874,"TERMINAL",0,0,"idling",,terminal_output +56,149041,"TERMINAL",0,0,"python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +57,160599,"TERMINAL",0,0,"bash",,terminal_focus +58,164468,"TERMINAL",0,0,"tmux a",,terminal_command +59,164479,"TERMINAL",0,0,"]633;E;2025-07-14 15:31:46 tmux a;450fbccd-48d1-432b-b44b-06394616d158]633;Cno sessions\r\n]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;1",,terminal_output +60,176523,"TERMINAL",0,0,"srun",,terminal_focus +61,181253,"TERMINAL",0,0,"idling\r\n\r\r\n\r\r\n\r",,terminal_output +62,182032,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_output +63,182304,"TERMINAL",0,0,"",,terminal_output +64,182514,"TERMINAL",0,0,"",,terminal_output +65,183774,"TERMINAL",0,0,"\r(reverse-i-search)`': ",,terminal_output +66,184082,"TERMINAL",0,0,"m': python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +67,184283,"TERMINAL",0,0,"\ro': python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +68,184445,"TERMINAL",0,0,"\rd': python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +69,184620,"TERMINAL",0,0,"[?25lm\ru': module unload devel/cuda/12.4\r\n\r\r\n\r\r\n\r[?25h",,terminal_output +70,184813,"TERMINAL",0,0,"[1@l': modul",,terminal_output +71,184920,"TERMINAL",0,0,"[?25lm[1@e': module[?25h",,terminal_output +72,186275,"TERMINAL",0,0,"[?25l\r[9@jafar) [tum_cte0515@hkn0402 jafar]$ module\r\n[?2004l\r[?25h",,terminal_output +73,186368,"TERMINAL",0,0,"Lmod Warning: \r\n----------------------------------------------------------------------------------------------------\r\nThe following dependent module(s) are not currently loaded: devel/cuda/12.4 (required by:\r\nmpi/openmpi/5.0)\r\n----------------------------------------------------------------------------------------------------\r\n\r\n\r\n\r\n]0;tum_cte0515@hkn0402:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0402 jafar]$ ",,terminal_output +74,187458,"TERMINAL",0,0,"m",,terminal_output +75,188318,"TERMINAL",0,0,"\r(reverse-i-search)`': ",,terminal_output +76,188537,"TERMINAL",0,0,"[?25lm': module unload devel/cuda/12.4\r[?25h",,terminal_output +77,188712,"TERMINAL",0,0,"[?25lm[1@o': mo[?25h",,terminal_output +78,188827,"TERMINAL",0,0,"[?25lm[1@d': mod[?25h",,terminal_output +79,188885,"TERMINAL",0,0,"[?25lme': python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked[?25h",,terminal_output +80,189124,"TERMINAL",0,0,"\rfailed reverse-i-search)`modeu': python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsizescaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +81,189331,"TERMINAL",0,0,"[?25lm\rl': python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked[?25h",,terminal_output +82,190338,"TERMINAL",0,0,"^C\r\n\r\n\r[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0402:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0402 jafar]$ ",,terminal_output +83,190778,"TERMINAL",0,0,"\r(reverse-i-search)`': ",,terminal_output +84,190999,"TERMINAL",0,0,"m': module unload devel/cuda/12.4\r",,terminal_output +85,191308,"TERMINAL",0,0,"[?25lm[1@o': mo[?25h",,terminal_output +86,192625,"TERMINAL",0,0,"python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +87,193357,"TERMINAL",0,0,"modelsize-scaling/train_dynamics_modelsize_scaling_3",,terminal_output +88,193987,"TERMINAL",0,0,"modelsize-scaling/train_dynamics_modelsize_scaling_3087000 --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +89,194447,"TERMINAL",0,0,"modelsize-scaling/train_dynamics_modelsize_scaling_3",,terminal_output +90,194807,"TERMINAL",0,0,"\rmodule unload devel/cuda/12.4\r\n\r\r\n\r\r\n\r",,terminal_output +91,195611,"TERMINAL",0,0,"mpi/openmpi/5.0\r",,terminal_output +92,196151,"TERMINAL",0,0,"[?25l\r[13@jafar) [tum_cte0515@hkn0402 jafar]$ mo\r\n[?2004l\r[?25h",,terminal_output +93,196276,"TERMINAL",0,0,"]0;tum_cte0515@hkn0402:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0402 jafar]$ ",,terminal_output +94,197041,"TERMINAL",0,0,"module unload mpi/openmpi/5.0",,terminal_output +95,197244,"TERMINAL",0,0,"devel/cuda/12.4",,terminal_output +96,197809,"TERMINAL",0,0,"[?25l[?2004l\r[?25h\r\nNote: the module ""devel/cuda/12.4"" cannot be unloaded because it was not loaded.\r\n\r\n]0;tum_cte0515@hkn0402:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0402 jafar]$ ",,terminal_output +97,202765,"TERMINAL",0,0,"module unload devel/cuda/12.4",,terminal_output +98,202990,"TERMINAL",0,0,"mpi/openmpi/5.0",,terminal_output +99,203154,"TERMINAL",0,0,"devel/cuda/12.4",,terminal_output +100,203381,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_output +101,203782,"TERMINAL",0,0,"idling",,terminal_output +102,204328,"TERMINAL",0,0,"python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/ --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +103,205187,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +104,206994,"sample.py",0,0,"",python,tab +105,206997,"sample.py",5254,0,"",python,selection_mouse +106,207025,"sample.py",5253,0,"",python,selection_command +107,218430,"TERMINAL",0,0,"2025-07-14 15:32:40.851942: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +108,231780,"TERMINAL",0,0,"2025-07-14 15:32:54.216693: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +109,248304,"TERMINAL",0,0,"2025-07-14 15:33:10.775327: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +110,252736,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 91000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/091000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 89000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/089000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 60000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/060000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 40000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/040000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 90000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/090000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 80000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_36M_2_node/080000/metrics/metrics not found.\r\n",,terminal_output +111,273123,"sample.py",5019,0,"",python,selection_mouse +112,273126,"sample.py",5018,0,"",python,selection_command +113,273749,"sample.py",5254,0,"",python,selection_mouse +114,273751,"sample.py",5253,0,"",python,selection_command +115,277876,"TERMINAL",0,0,"bash",,terminal_focus +116,283470,"TERMINAL",0,0,"cd $ws_dir",,terminal_command +117,284396,"TERMINAL",0,0,"cd logs/",,terminal_command +118,284816,"TERMINAL",0,0,"ls",,terminal_command +119,284828,"TERMINAL",0,0,"]633;E;2025-07-14 15:33:47 ls;450fbccd-48d1-432b-b44b-06394616d158]633;C3306965 logs_alfred logs_franz logs_mihir\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs]633;D;0",,terminal_output +120,287852,"TERMINAL",0,0,"Entering jdb:\r\n(jdb) ",,terminal_output +121,287863,"TERMINAL",0,0,"cd logs_mihir/",,terminal_command +122,287879,"TERMINAL",0,0,"]633;E;2025-07-14 15:33:50 cd logs_mihir/;450fbccd-48d1-432b-b44b-06394616d158]633;C]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +123,288184,"TERMINAL",0,0,"ls",,terminal_command +124,288235,"TERMINAL",0,0,"]633;E;2025-07-14 15:33:50 ls;450fbccd-48d1-432b-b44b-06394616d158]633;C",,terminal_output +125,288365,"TERMINAL",0,0,"big_run train_lam_action_space_scaling_20_3329788.log train_lam_model_size_scaling_38M_3317098.log train_tokenizer_model_size_scaling_140M_3316019.log\r\nbig-runs train_lam_action_space_scaling_20_3329803.log train_lam_model_size_scaling_38M_3317115.log train_tokenizer_model_size_scaling_200M_3313563.log\r\ntrain_dyn_yolorun_3333026.log train_lam_action_space_scaling_20_3331285.log train_lam_model_size_scaling_38M_3317231.log train_tokenizer_model_size_scaling_200M_3316020.log\r\ntrain_dyn_yolorun_3333448.log train_lam_action_space_scaling_50_3320180.log train_tokenizer_batch_size_scaling_16_node_3321526.log train_tokenizer_model_size_scaling_227M_3317234.log\r\ntrain_dyn_yolorun_3335345.log train_lam_action_space_scaling_50_3329789.log train_tokenizer_batch_size_scaling_1_node_3318551.log train_tokenizer_model_size_scaling_227M_3318555.log\r\ntrain_dyn_yolorun_3335362.log train_lam_action_space_scaling_50_3329804.log train_tokenizer_batch_size_scaling_2_node_3318552.log train_tokenizer_model_size_scaling_227M_3320173.log\r\ntrain_lam_action_space_scaling_10_3320179.log train_lam_action_space_scaling_50_3331286.log train_tokenizer_batch_size_scaling_2_node_3330806.log train_tokenizer_model_size_scaling_227M_3321523.log\r\ntrain_lam_action_space_scaling_10_3321529.log train_lam_action_space_scaling_6_3318549.log train_tokenizer_batch_size_scaling_2_node_3330848.log train_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_lam_action_space_scaling_10_3329786.log train_lam_action_space_scaling_6_3320178.log train_tokenizer_batch_size_scaling_2_node_3331282.log train_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_lam_action_space_scaling_10_3329801.log train_lam_action_space_scaling_6_3321528.log train_tokenizer_batch_size_scaling_4_node_3318553.log train_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_lam_action_space_scaling_10_3331283.log train_lam_action_space_scaling_6_3329790.log train_tokenizer_batch_size_scaling_4_node_3320175.log train_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_lam_action_space_scaling_12_3318546.log train_lam_action_space_scaling_6_3329805.log train_tokenizer_batch_size_scaling_4_node_3321524.log train_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_lam_action_space_scaling_12_3320177.log train_lam_action_space_scaling_6_3331287.log train_tokenizer_batch_size_scaling_8_node_3320176.log train_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_lam_action_space_scaling_12_3321527.log train_lam_action_space_scaling_8_3318550.log train_tokenizer_batch_size_scaling_8_node_3321525.log train_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_lam_action_space_scaling_12_3329787.log train_lam_action_space_scaling_8_3329791.log train_tokenizer_minecraft_overfit_sample_3309656.log train_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_lam_action_space_scaling_12_3329802.log train_lam_action_space_scaling_8_3329806.log train_tokenizer_model_size_scaling_127M_3317233.log train_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_lam_action_space_scaling_12_3331284.log train_lam_action_space_scaling_8_3331288.log train_tokenizer_model_size_scaling_127M_3318554.log train_tokenizer_model_size_scaling_80M_3316026.log\r\ntrain_lam_action_space_scaling_20_3318547.log train_lam_minecraft_overfit_sample_3309655.log train_tokenizer_model_size_scaling_140M_3313562.log\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +126,291553,"TERMINAL",0,0,"cd big-runs/",,terminal_command +127,292418,"TERMINAL",0,0,"ls",,terminal_command diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-619817ef-f0fa-4ff3-8f61-fe4646f7e2971752667688154-2025_07_16-14.09.00.461/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-619817ef-f0fa-4ff3-8f61-fe4646f7e2971752667688154-2025_07_16-14.09.00.461/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..4dc8c4b5a63aa986b76dd3e7edb6716c0072fe53 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-619817ef-f0fa-4ff3-8f61-fe4646f7e2971752667688154-2025_07_16-14.09.00.461/source.csv @@ -0,0 +1,2382 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +2,93,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,tab +3,329,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"2:09:00 PM [info] Activating crowd-code\n2:09:00 PM [info] Recording started\n2:09:00 PM [info] Initializing git provider using file system watchers...\n2:09:00 PM [info] Git repository found\n2:09:00 PM [info] Git provider initialized successfully\n",Log,content +4,380,"extension-output-pdoom-org.crowd-code-#1-crowd-code",245,0,"2:09:00 PM [info] Initial git state: [object Object]\n",Log,content +5,2739,"TERMINAL",0,0,"",,terminal_focus +6,3206,"TERMINAL",0,0,"/bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt",,terminal_command +7,3235,"TERMINAL",0,0,"]633;E;2025-07-16 14:09:03 /bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt;f275eb64-c8e0-41e9-a028-3428eb3f91db]633;C]0;tum_cte0515@hkn1990:/hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash]633;D;0",,terminal_output +8,6981,"TERMINAL",0,0,"bash",,terminal_focus +9,9682,"TERMINAL",0,0,"queue",,terminal_command +10,9732,"TERMINAL",0,0,"]633;E;2025-07-16 14:09:10 queue;97c203bb-2de3-4bf0-b19e-fa122ab0b933]633;C",,terminal_output +11,9800,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1990.localdomain: Wed Jul 16 14:09:10 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3348592 accelerat train_dy tum_cte0 R 14:08:48\t 2 hkn[0416,0421]3348397 accelerat train_dy tum_cte0 R 19:30:30\t 2 hkn[0810,0815]3348399 accelerat train_dy tum_cte0 R 19:30:30\t 2 hkn[0601,0603]3348400 accelerat train_dy tum_cte0 R 19:30:30\t 2 hkn[0604,0608]3350245 accelerat interact tum_cte0 R17:42\t 2 hkn[0423-0424]3345116 accelerat train_dy tum_cte0 R 1-20:14:15\t 2 hkn[0503,0506]",,terminal_output +12,10526,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;0",,terminal_output +13,11309,"TERMINAL",0,0,"idling",,terminal_command +14,11376,"TERMINAL",0,0,"]633;E;2025-07-16 14:09:11 idling;97c203bb-2de3-4bf0-b19e-fa122ab0b933]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1990.localdomain: Wed Jul 16 14:09:11 2025Partition dev_cpuonly:\t 9 nodes idle\rPartition cpuonly:\t 4 nodes idle\rPartition dev_accelerated:\t 1 nodes idle\rPartition accelerated: 10 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 7 nodes idle",,terminal_output +15,12159,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;0",,terminal_output +16,15453,"TERMINAL",0,0,"bs/mihir/horeka/modelsize_scaling/runner.sh",,terminal_command +17,15496,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]633;E;;97c203bb-2de3-4bf0-b19e-fa122ab0b933]633;C]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D",,terminal_output +18,16915,"TERMINAL",0,0,"salloc --time=10:00:00 --partition=accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5",,terminal_command +19,16975,"TERMINAL",0,0,"]633;E;2025-07-16 14:09:17 salloc --time=10:00:00 --partition=accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5;97c203bb-2de3-4bf0-b19e-fa122ab0b933]633;Csalloc: Granted job allocation 3350302\r\n",,terminal_output +20,17083,"TERMINAL",0,0,"salloc: Waiting for resource configuration\r\n",,terminal_output +21,18085,"TERMINAL",0,0,"bash",,terminal_focus +22,19221,"TERMINAL",0,0,"queu",,terminal_command +23,19262,"TERMINAL",0,0,"]633;E;2025-07-16 14:09:19 queu;dea9d5fc-91fd-447c-886d-4b0240ae057d]633;Cbash: queu: command not found...\r\n",,terminal_output +24,20011,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;130",,terminal_output +25,21463,"TERMINAL",0,0,"queue",,terminal_command +26,21514,"TERMINAL",0,0,"]633;E;2025-07-16 14:09:21 queue;dea9d5fc-91fd-447c-886d-4b0240ae057d]633;C",,terminal_output +27,21579,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1990.localdomain: Wed Jul 16 14:09:21 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3348592 accelerat train_dy tum_cte0 R 14:08:59\t 2 hkn[0416,0421]3348397 accelerat train_dy tum_cte0 R 19:30:41\t 2 hkn[0810,0815]3348399 accelerat train_dy tum_cte0 R 19:30:41\t 2 hkn[0601,0603]3348400 accelerat train_dy tum_cte0 R 19:30:41\t 2 hkn[0604,0608]3350302 accelerat interact tum_cte0 R\t0:04\t 2 hkn[0509,0511]3350245 accelerat interact tum_cte0 R17:53\t 2 hkn[0423-0424]3345116 accelerat train_dy tum_cte0 R 1-20:14:26\t 2 hkn[0503,0506]",,terminal_output +28,22616,"TERMINAL",0,0,"29:01333658",,terminal_output +29,23678,"TERMINAL",0,0,"42444769",,terminal_output +30,24469,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;0",,terminal_output +31,26669,"TERMINAL",0,0,"scancel 3350245",,terminal_command +32,26698,"TERMINAL",0,0,"]633;E;2025-07-16 14:09:27 scancel 3350245;dea9d5fc-91fd-447c-886d-4b0240ae057d]633;C]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +33,28074,"TERMINAL",0,0,"salloc",,terminal_focus +34,32584,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=""debug""\nslurm_job_id=""debug-mihir""\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4\n\nenv | grep SLURM\n\nXLA_FLAGS=--xla_gpu_autotune_level=0 srun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --grad_clip_threshold=10 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-debug-run-$slurm_job_id \\n --tags dynamics debug \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n ",shellscript,tab +35,34735,"utils/nn.py",0,0,"import math\nfrom typing import Dict, Tuple\n\nfrom flax import linen as nn\nimport jax\nimport jax.numpy as jnp\n\n\nclass PositionalEncoding(nn.Module):\n """"""https://uvadlc-notebooks.readthedocs.io/en/latest/tutorial_notebooks/JAX/tutorial6/Transformers_and_MHAttention.html""""""\n\n d_model: int # Hidden dimensionality of the input.\n max_len: int = 5000 # Maximum length of a sequence to expect.\n\n def setup(self):\n # Create matrix of [SeqLen, HiddenDim] representing the positional encoding for max_len inputs\n self.pe = jnp.zeros((self.max_len, self.d_model))\n position = jnp.arange(0, self.max_len, dtype=jnp.float32)[:, None]\n div_term = jnp.exp(\n jnp.arange(0, self.d_model, 2) * (-math.log(10000.0) / self.d_model)\n )\n self.pe = self.pe.at[:, 0::2].set(jnp.sin(position * div_term))\n self.pe = self.pe.at[:, 1::2].set(jnp.cos(position * div_term))\n\n def __call__(self, x):\n x = x + self.pe[: x.shape[2]]\n return x\n\nclass CausalTransformerBlock(nn.Module):\n model_dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x\n\nclass CausalTransformer(nn.Module):\n model_dim: int\n out_dim: int\n num_blocks: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # Input projection and normalization\n x = nn.Sequential(\n [\n nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.Dense(self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n ]\n )(x)\n # Causal transformer blocks\n for _ in range(self.num_blocks):\n x = CausalTransformerBlock(\n model_dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # Output projection\n x = nn.Dense(\n self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x\n\n\nclass STTransformer(nn.Module):\n model_dim: int\n out_dim: int\n num_blocks: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n x = nn.Sequential(\n [\n nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.Dense(self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n ]\n )(x)\n for _ in range(self.num_blocks):\n x = STBlock(\n dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n x = nn.Dense(\n self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\ndef normalize(x):\n return x / (jnp.linalg.norm(x, ord=2, axis=-1, keepdims=True) + 1e-8)\n\n\nclass VectorQuantizer(nn.Module):\n latent_dim: int\n num_latents: int\n dropout: float\n\n def setup(self):\n self.codebook = normalize(\n self.param(\n ""codebook"",\n nn.initializers.lecun_uniform(),\n (self.num_latents, self.latent_dim),\n )\n )\n self.drop = nn.Dropout(self.dropout, deterministic=False)\n\n def __call__(\n self, x: jax.Array, training: bool\n ) -> Tuple[jax.Array, jax.Array, jax.Array, jax.Array]:\n # --- Compute distances ---\n x = normalize(x)\n codebook = normalize(self.codebook)\n distance = -jnp.matmul(x, codebook.T)\n if training:\n dropout_key = self.make_rng(""dropout"")\n distance = self.drop(distance, rng=dropout_key)\n\n # --- Get indices and embeddings ---\n indices = jnp.argmin(distance, axis=-1)\n z = self.codebook[indices]\n\n # --- Straight through estimator ---\n z_q = x + jax.lax.stop_gradient(z - x)\n return z_q, z, x, indices\n\n def get_codes(self, indices: jax.Array):\n return self.codebook[indices]\n",python,tab +36,44575,"TERMINAL",0,0,"salloc: Nodes hkn[0509,0511] are ready for job\r\n",,terminal_output +37,45793,"TERMINAL",0,0,"]0;tum_cte0515@hkn0509:~/Projects/jafar[?2004h[tum_cte0515@hkn0509 jafar]$ ",,terminal_output +38,61665,"utils/nn.py",1286,0,"",python,selection_mouse +39,61679,"utils/nn.py",1285,0,"",python,selection_command +40,64584,"TERMINAL",0,0,"s",,terminal_output +41,64687,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +42,64795,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +43,64980,"TERMINAL",0,0,"[?25lr[?25h",,terminal_output +44,65173,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +45,65291,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +46,65353,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +47,65416,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +48,65504,"TERMINAL",0,0,"[?25lv[?25h",,terminal_output +49,65713,"TERMINAL",0,0,"env/",,terminal_output +50,65938,"TERMINAL",0,0,"[?25lb[?25h",,terminal_output +51,66001,"TERMINAL",0,0,"in/",,terminal_output +52,66262,"TERMINAL",0,0,"[?25la[?25h[?25lc[?25h",,terminal_output +53,66474,"TERMINAL",0,0,"tivate",,terminal_output +54,66740,"TERMINAL",0,0,"[?25l[?2004l\r[?25h]0;tum_cte0515@hkn0509:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0509 jafar]$ ",,terminal_output +55,68208,"genie.py",0,0,"from typing import Dict, Any\n\nimport optax\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nfrom flax.training.train_state import TrainState\nimport orbax.checkpoint as ocp\n\nfrom models.dynamics import DynamicsMaskGIT\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\nimport os\nimport grain\n\n\nclass Genie(nn.Module):\n """"""Genie model""""""\n\n # --- Tokenizer ---\n in_dim: int\n tokenizer_dim: int\n latent_patch_dim: int\n num_patch_latents: int\n patch_size: int\n tokenizer_num_blocks: int\n tokenizer_num_heads: int\n # --- LAM ---\n lam_dim: int\n latent_action_dim: int\n num_latent_actions: int\n lam_patch_size: int\n lam_num_blocks: int\n lam_num_heads: int\n lam_co_train: bool\n # --- Dynamics ---\n dyna_dim: int\n dyna_num_blocks: int\n dyna_num_heads: int\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n dropout: float = 0.0\n mask_limit: float = 0.0\n\n def setup(self):\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_latent_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\n latent_actions = jax.lax.cond(\n self.lam_co_train,\n lambda: lam_outputs[""z_q""],\n lambda: jax.lax.stop_gradient(lam_outputs[""z_q""])\n )\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(tokenizer_outputs[""indices""]),\n latent_actions=latent_actions,\n )\n outputs[""mask_rng""] = batch[""mask_rng""]\n dyna_outputs = self.dynamics(outputs, training)\n outputs.update(dyna_outputs)\n mle_indices = jnp.argmax(outputs[""token_logits""], axis=-1)\n outputs[""recon""] = self.tokenizer.decode(\n mle_indices, batch[""videos""].shape[2:4]\n )\n outputs[""lam_indices""] = lam_outputs[""indices""]\n return outputs\n\n @nn.compact\n def sample(\n self,\n batch: Dict[str, Any],\n seq_len: int,\n steps: int = 25,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> Any:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by \n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size \n T: number of input (conditioning) frames \n N: patches per frame \n S: sequence length \n A: action space \n D: model latent dimension\n """"""\n # --- Encode videos and actions ---\n tokenizer_out = self.tokenizer.vq_encode(batch[""videos""], training=False)\n token_idxs = tokenizer_out[""indices""] # (B, T, N)\n B, T, N = token_idxs.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs.dtype)\n token_idxs = jnp.concatenate([token_idxs, pad], axis=1) # (B, S, N)\n action_tokens = self.lam.vq.get_codes(batch[""latent_actions""])\n\n MaskGITLoop = nn.scan(\n MaskGITStep,\n variable_broadcast=""params"",\n split_rngs={""params"": False},\n in_axes=0,\n out_axes=0,\n length=steps,\n )\n \n loop_fn = MaskGITLoop(\n dynamics=self.dynamics,\n tokenizer=self.tokenizer,\n temperature=temperature,\n sample_argmax=sample_argmax,\n steps=steps,\n )\n\n def generation_step_fn(carry, step_t):\n rng, current_token_idxs = carry\n rng, step_rng = jax.random.split(rng)\n\n # Mask current and future frames (i.e., t >= step_t)\n mask = jnp.arange(seq_len) >= step_t # (S,)\n mask = jnp.broadcast_to(mask[None, :, None], (B, seq_len, N)) # (B, S, N)\n mask = mask.astype(bool)\n masked_token_idxs = current_token_idxs * ~mask\n\n # --- Initialize and run MaskGIT loop ---\n init_carry_maskgit = (\n step_rng,\n masked_token_idxs,\n mask,\n action_tokens,\n )\n final_carry_maskgit, _ = loop_fn(init_carry_maskgit, jnp.arange(steps))\n updated_token_idxs = final_carry_maskgit[1]\n new_carry = (rng, updated_token_idxs)\n return new_carry, None\n\n # --- Run the autoregressive generation using scan ---\n initial_carry = (batch[""rng""], token_idxs)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry, _ = jax.lax.scan(\n generation_step_fn,\n initial_carry,\n timesteps_to_scan\n )\n final_token_idxs = final_carry[1]\n\n # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n final_token_idxs,\n video_hw=batch[""videos""].shape[2:4],\n )\n return final_frames\n\n def vq_encode(self, batch, training) -> Dict[str, Any]:\n # --- Preprocess videos ---\n lam_output = self.lam.vq_encode(batch[""videos""], training=training)\n return lam_output[""indices""]\n\n\nclass MaskGITStep(nn.Module):\n dynamics: nn.Module\n tokenizer: nn.Module\n temperature: float\n sample_argmax: bool\n steps: int\n\n @nn.compact\n def __call__(self, carry, x):\n rng, token_idxs, mask, action_tokens = carry\n step = x\n N = token_idxs.shape[2]\n\n # --- Construct + encode video ---\n vid_embed = self.dynamics.patch_embed(token_idxs) # (B, S, N, D)\n mask_token = self.dynamics.mask_token # (1, 1, 1, D,)\n mask_expanded = mask[..., None] # (B, S, N, 1) \n vid_embed = jnp.where(mask_expanded, mask_token, vid_embed)\n\n # --- Predict transition ---\n act_embed = self.dynamics.action_up(action_tokens)\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (self.steps * 2))\n step_temp = self.temperature * (1.0 - unmasked_ratio)\n final_logits = self.dynamics.dynamics(vid_embed) / step_temp\n\n # --- Sample new tokens for final frame ---\n if self.sample_argmax:\n sampled_token_idxs = jnp.argmax(final_logits, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs = jax.random.categorical(_rng, final_logits)\n gather_fn = jax.vmap(jax.vmap(jax.vmap(lambda x, y: x[y])))\n final_token_probs = gather_fn(jax.nn.softmax(final_logits), sampled_token_idxs)\n final_token_probs += ~mask\n # Update masked tokens only\n token_idxs = jnp.where(mask, sampled_token_idxs, token_idxs)\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n idx_mask = jnp.arange(final_token_probs.shape[-1]) > num_unmasked_tokens\n sorted_idxs = jnp.argsort(final_token_probs, axis=-1, descending=True)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask))\n new_mask = mask_update_fn(mask, sorted_idxs)\n\n new_carry = (rng, token_idxs, new_mask, action_tokens)\n return new_carry, None\n\ndef restore_genie_components(\n train_state: TrainState,\n sharding: jax.sharding.NamedSharding,\n grain_iterator: grain.DataLoaderIterator,\n inputs: Dict[str, jax.Array],\n rng: jax.Array,\n args,\n):\n """"""Restore pre-trained Genie components""""""\n rng, _rng = jax.random.split(rng)\n\n # dummy values since we only use tx to initialize the dummy train states\n dummy_tx = optax.adamw(\n learning_rate=optax.constant_schedule(args.max_lr),\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n )\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add('model_state', ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler)\n handler_registry.add('dataloader_state', grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler)\n \n\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n tokenizer_checkpoint_manager = ocp.CheckpointManager(\n directory=args.tokenizer_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n )\n tokenizer_init_params = dummy_tokenizer.init(_rng, inputs)\n dummy_tokenizer_train_state = TrainState.create(\n apply_fn=dummy_tokenizer.apply, params=tokenizer_init_params, tx=dummy_tx\n )\n abstract_sharded_tokenizer_state = _create_abstract_sharded_pytree(\n dummy_tokenizer_train_state, sharding\n )\n restored_tokenizer = tokenizer_checkpoint_manager.restore(\n step=tokenizer_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_sharded_tokenizer_state),\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator),\n ),\n )[""model_state""]\n restored_tokenizer_params = restored_tokenizer.params[""params""]\n train_state.params[""params""][""tokenizer""].update(restored_tokenizer_params)\n tokenizer_checkpoint_manager.close()\n\n if args.lam_checkpoint:\n lam_checkpoint_manager = ocp.CheckpointManager(\n directory=args.lam_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_latent_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n )\n lam_init_params = dummy_lam.init(_rng, inputs)\n dummy_lam_train_state = TrainState.create(\n apply_fn=dummy_lam.apply, params=lam_init_params, tx=dummy_tx\n )\n abstract_sharded_lam_state = _create_abstract_sharded_pytree(\n dummy_lam_train_state, sharding\n )\n restored_lam = lam_checkpoint_manager.restore(\n step=lam_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_sharded_lam_state),\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator),\n ),\n )[""model_state""]\n restored_lam_params = restored_lam.params[""params""]\n # Genie does not initialize all LAM modules, thus we omit those extra modules during restoration\n # (f.srambical) FIXME: Currently, this is a small HBM memory crunch since the LAM's decoder is loaded into HBM and immediately dicarded.\n # A workaround would be to restore to host memory first, and only move the weights to HBM after pruning the decoder\n restored_lam_params = {\n k: v\n for k, v in restored_lam_params.items()\n if k in train_state.params[""params""][""lam""]\n }\n train_state.params[""params""][""lam""].update(restored_lam_params)\n lam_checkpoint_manager.close()\n\n return train_state\n\ndef _create_abstract_sharded_pytree(pytree_template, sharding_spec):\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)",python,tab +56,71445,"genie.py",1936,0,"",python,selection_mouse +57,71450,"genie.py",1935,0,"",python,selection_command +58,72209,"genie.py",1962,0,"",python,selection_mouse +59,72748,"genie.py",1952,0,"",python,selection_mouse +60,73513,"genie.py",1937,40," self.dynamics = DynamicsMaskGIT(",python,selection_command +61,73697,"genie.py",1937,77," self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,",python,selection_command +62,73850,"genie.py",1937,125," self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,",python,selection_command +63,73999,"genie.py",1937,170," self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,\n num_blocks=self.dyna_num_blocks,",python,selection_command +64,74154,"genie.py",1937,213," self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,",python,selection_command +65,74333,"genie.py",1937,247," self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,",python,selection_command +66,74493,"genie.py",1937,287," self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,",python,selection_command +67,74650,"genie.py",1937,329," self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,",python,selection_command +68,74782,"genie.py",1937,359," self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,",python,selection_command +69,74929,"genie.py",1937,369," self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )",python,selection_command +70,75176,"genie.py",1945,0,"",python,selection_command +71,76155,"genie.py",2305,0,"#",python,content +72,76156,"genie.py",2279,0,"#",python,content +73,76156,"genie.py",2237,0,"#",python,content +74,76156,"genie.py",2197,0,"#",python,content +75,76156,"genie.py",2163,0,"#",python,content +76,76156,"genie.py",2120,0,"#",python,content +77,76156,"genie.py",2075,0,"#",python,content +78,76156,"genie.py",2027,0,"#",python,content +79,76156,"genie.py",1990,0,"#",python,content +80,76157,"genie.py",1945,0,"#",python,content +81,76160,"genie.py",1946,0,"",python,selection_keyboard +82,76331,"genie.py",2315,0," ",python,content +83,76332,"genie.py",2288,0," ",python,content +84,76332,"genie.py",2245,0," ",python,content +85,76332,"genie.py",2204,0," ",python,content +86,76332,"genie.py",2169,0," ",python,content +87,76332,"genie.py",2125,0," ",python,content +88,76332,"genie.py",2079,0," ",python,content +89,76333,"genie.py",2030,0," ",python,content +90,76333,"genie.py",1992,0," ",python,content +91,76333,"genie.py",1946,0," ",python,content +92,76334,"genie.py",1947,0,"",python,selection_keyboard +93,76593,"genie.py",1946,0,"",python,selection_command +94,76718,"genie.py",1935,0,"",python,selection_command +95,76899,"genie.py",1906,0,"",python,selection_command +96,77245,"genie.py",1935,0,"",python,selection_command +97,77506,"genie.py",1936,0,"\n ",python,content +98,78857,"genie.py",1945,0,"s",python,content +99,78861,"genie.py",1946,0,"",python,selection_keyboard +100,79111,"genie.py",1946,0,"e",python,content +101,79113,"genie.py",1947,0,"",python,selection_keyboard +102,79156,"genie.py",1947,0,"l",python,content +103,79158,"genie.py",1948,0,"",python,selection_keyboard +104,79243,"genie.py",1948,0,"f",python,content +105,79245,"genie.py",1949,0,"",python,selection_keyboard +106,79353,"genie.py",1949,0,".",python,content +107,79354,"genie.py",1950,0,"",python,selection_keyboard +108,79710,"genie.py",1950,0,"d",python,content +109,79712,"genie.py",1951,0,"",python,selection_keyboard +110,79943,"genie.py",1951,0,"y",python,content +111,79945,"genie.py",1952,0,"",python,selection_keyboard +112,80028,"genie.py",1952,0,"n",python,content +113,80030,"genie.py",1953,0,"",python,selection_keyboard +114,81277,"genie.py",1953,0,"a",python,content +115,81280,"genie.py",1954,0,"",python,selection_keyboard +116,81385,"genie.py",1954,0,"m",python,content +117,81386,"genie.py",1955,0,"",python,selection_keyboard +118,81531,"genie.py",1955,0,"i",python,content +119,81532,"genie.py",1956,0,"",python,selection_keyboard +120,81629,"genie.py",1956,0,"c",python,content +121,81630,"genie.py",1957,0,"",python,selection_keyboard +122,81745,"genie.py",1957,0,"s",python,content +123,81747,"genie.py",1958,0,"",python,selection_keyboard +124,81963,"genie.py",1958,0," ",python,content +125,81964,"genie.py",1959,0,"",python,selection_keyboard +126,82183,"genie.py",1959,0,"=",python,content +127,82184,"genie.py",1960,0,"",python,selection_keyboard +128,82283,"genie.py",1960,0," ",python,content +129,82284,"genie.py",1961,0,"",python,selection_keyboard +130,85672,"genie.py",1961,0,"N",python,content +131,85674,"genie.py",1962,0,"",python,selection_keyboard +132,85874,"genie.py",1962,0,"o",python,content +133,85876,"genie.py",1963,0,"",python,selection_keyboard +134,85997,"genie.py",1963,0,"n",python,content +135,85999,"genie.py",1964,0,"",python,selection_keyboard +136,86109,"genie.py",1964,0,"e",python,content +137,86110,"genie.py",1965,0,"",python,selection_keyboard +138,86438,"genie.py",1964,0,"",python,selection_command +139,88763,"genie.py",267,0,"",python,selection_mouse +140,88765,"genie.py",266,0,"",python,selection_command +141,89247,"genie.py",226,0,"",python,selection_mouse +142,89259,"genie.py",225,0,"",python,selection_command +143,90226,"genie.py",226,0,"",python,selection_command +144,90410,"genie.py",226,0,",",python,content +145,90412,"genie.py",227,0,"",python,selection_keyboard +146,90505,"genie.py",227,0," ",python,content +147,90506,"genie.py",228,0,"",python,selection_keyboard +148,90719,"genie.py",228,0,"C",python,content +149,90720,"genie.py",229,0,"",python,selection_keyboard +150,91429,"genie.py",228,1,"CausalTransformer",python,content +151,94577,"genie.py",1955,0,"",python,selection_mouse +152,95205,"genie.py",1984,0,"",python,selection_mouse +153,95939,"genie.py",1980,4,"",python,content +154,96353,"genie.py",1980,0,"C",python,content +155,96354,"genie.py",1981,0,"",python,selection_keyboard +156,96571,"genie.py",1981,0,"a",python,content +157,96573,"genie.py",1982,0,"",python,selection_keyboard +158,97070,"genie.py",1980,2,"CausalTransformer",python,content +159,98850,"models/dynamics.py",0,0,"from typing import Dict, Any\n\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\n\nfrom utils.nn import STTransformer, CausalTransformer\n\n\nclass DynamicsMaskGIT(nn.Module):\n """"""MaskGIT dynamics model""""""\n\n model_dim: int\n num_latents: int\n num_blocks: int\n num_heads: int\n dropout: float\n mask_limit: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n def setup(self):\n self.dynamics = STTransformer(\n self.model_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n )\n self.patch_embed = nn.Embed(self.num_latents, self.model_dim)\n self.mask_token = self.param(\n ""mask_token"",\n nn.initializers.lecun_uniform(),\n (1, 1, 1, self.model_dim),\n )\n self.action_up = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n # --- Mask videos ---\n vid_embed = self.patch_embed(batch[""video_tokens""])\n if training:\n rng1, rng2 = jax.random.split(batch[""mask_rng""])\n mask_prob = jax.random.uniform(rng1, minval=self.mask_limit)\n mask = jax.random.bernoulli(rng2, mask_prob, vid_embed.shape[:-1])\n mask = mask.at[:, 0].set(False)\n vid_embed = jnp.where(jnp.expand_dims(mask, -1), self.mask_token, vid_embed)\n else:\n mask = None\n\n # --- Predict transition ---\n act_embed = self.action_up(batch[""latent_actions""])\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n logits = self.dynamics(vid_embed)\n return dict(token_logits=logits, mask=mask)\n\n\nclass DynamicsAutoregressive(nn.Module):\n """"""Autoregressive (causal) dynamics model""""""\n\n model_dim: int\n num_latents: int\n num_blocks: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n def setup(self):\n # Use a causal transformer instead of STTransformer\n self.dynamics = CausalTransformer(\n self.model_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n )\n self.patch_embed = nn.Embed(self.num_latents, self.model_dim)\n self.action_up = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n # No random masking, just embed the tokens\n vid_embed = self.patch_embed(batch[""video_tokens""])\n act_embed = self.action_up(batch[""latent_actions""])\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n logits = self.dynamics(vid_embed, causal=True) # Pass causal=True if needed\n return dict(token_logits=logits, mask=None)\n",python,tab +160,102366,"models/dynamics.py",0,0,"",python,tab +161,104695,"genie.py",0,0,"",python,tab +162,104696,"genie.py",1955,0,"",python,selection_mouse +163,105080,"genie.py",1997,0,"",python,selection_mouse +164,106023,"genie.py",1997,0,"()",python,content +165,106025,"genie.py",1998,0,"",python,selection_keyboard +166,106251,"genie.py",1998,0,"\n \n ",python,content +167,116272,"genie.py",2190,0,"",python,selection_mouse +168,116952,"genie.py",1996,0,"",python,selection_mouse +169,124670,"genie.py",1990,0,"",python,selection_mouse +170,124830,"genie.py",1980,17,"CausalTransformer",python,selection_mouse +171,125439,"genie.py",1999,12,"",python,content +172,125440,"genie.py",1980,17,"",python,content +173,127325,"genie.py",1980,0,"D",python,content +174,127326,"genie.py",1981,0,"",python,selection_keyboard +175,127860,"genie.py",1981,0,"y",python,content +176,127862,"genie.py",1982,0,"",python,selection_keyboard +177,128463,"genie.py",1982,0,"n",python,content +178,128464,"genie.py",1983,0,"",python,selection_keyboard +179,129281,"genie.py",1983,0,"a",python,content +180,129282,"genie.py",1984,0,"",python,selection_keyboard +181,129359,"genie.py",1984,0,"m",python,content +182,129361,"genie.py",1985,0,"",python,selection_keyboard +183,129537,"genie.py",1985,0,"i",python,content +184,129538,"genie.py",1986,0,"",python,selection_keyboard +185,129610,"genie.py",1986,0,"c",python,content +186,129612,"genie.py",1987,0,"",python,selection_keyboard +187,129787,"genie.py",1987,0,"s",python,content +188,129788,"genie.py",1988,0,"",python,selection_keyboard +189,130293,"genie.py",1988,0,"A",python,content +190,130294,"genie.py",1989,0,"",python,selection_keyboard +191,130460,"genie.py",1989,0,"u",python,content +192,130461,"genie.py",1990,0,"",python,selection_keyboard +193,132024,"genie.py",1990,0,"t",python,content +194,132026,"genie.py",1991,0,"",python,selection_keyboard +195,133252,"genie.py",1991,0,"o",python,content +196,133254,"genie.py",1992,0,"",python,selection_keyboard +197,138030,"models/dynamics.py",0,0,"",python,tab +198,138031,"models/dynamics.py",89,0,"",python,selection_mouse +199,138660,"models/dynamics.py",130,0,"",python,selection_mouse +200,138805,"models/dynamics.py",126,17,"CausalTransformer",python,selection_mouse +201,139386,"models/dynamics.py",112,0,"",python,selection_mouse +202,139588,"models/dynamics.py",111,13,"STTransformer",python,selection_mouse +203,141603,"models/dynamics.py",1889,0,"",python,selection_mouse +204,141802,"models/dynamics.py",1879,22,"DynamicsAutoregressive",python,selection_mouse +205,144435,"genie.py",0,0,"",python,tab +206,144437,"genie.py",1987,0,"",python,selection_mouse +207,144438,"genie.py",1980,12,"DynamicsAuto",python,selection_mouse +208,145011,"genie.py",1980,12,"",python,content +209,145206,"genie.py",1980,0,"DynamicsAutoregressive",python,content +210,148409,"genie.py",237,0,"",python,selection_mouse +211,148603,"genie.py",228,17,"CausalTransformer",python,selection_mouse +212,148904,"genie.py",228,17,"",python,content +213,149222,"genie.py",228,0,"DynamicsAutoregressive",python,content +214,152461,"genie.py",2009,0,"",python,selection_mouse +215,153270,"genie.py",2009,0," ",python,content +216,160091,"genie.py",2070,0,"",python,selection_mouse +217,160886,"genie.py",2069,0,"",python,selection_command +218,161420,"genie.py",2112,0,"",python,selection_command +219,162461,"genie.py",2075,38," # model_dim=self.dyna_dim,",python,selection_command +220,162905,"genie.py",2075,88," # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,",python,selection_command +221,163783,"genie.py",2075,135," # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,\n # num_blocks=self.dyna_num_blocks,",python,selection_command +222,164936,"genie.py",2075,180," # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,\n # num_blocks=self.dyna_num_blocks,\n # num_heads=self.dyna_num_heads,",python,selection_command +223,165895,"genie.py",2075,216," # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,\n # num_blocks=self.dyna_num_blocks,\n # num_heads=self.dyna_num_heads,\n # dropout=self.dropout,",python,selection_command +224,168465,"genie.py",2075,0,"",python,selection_command +225,168614,"genie.py",2032,0,"",python,selection_command +226,168802,"genie.py",2022,0,"",python,selection_command +227,168979,"genie.py",2009,0,"",python,selection_command +228,169264,"genie.py",2021,0,"\n # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,\n # num_blocks=self.dyna_num_blocks,\n # num_heads=self.dyna_num_heads,\n # dropout=self.dropout,",python,content +229,169264,"genie.py",2009,12,"",python,content +230,169319,"genie.py",2022,0,"",python,selection_command +231,169910,"genie.py",2009,0,"",python,selection_command +232,170060,"genie.py",2022,0,"",python,selection_command +233,170621,"genie.py",2061,0,"",python,selection_command +234,170622,"genie.py",2111,0,"",python,selection_command +235,170668,"genie.py",2158,0,"",python,selection_command +236,170697,"genie.py",2203,0,"",python,selection_command +237,170697,"genie.py",2235,0,"",python,selection_command +238,170707,"genie.py",2249,0,"",python,selection_command +239,170744,"genie.py",2292,0,"",python,selection_command +240,170790,"genie.py",2331,0,"",python,selection_command +241,170816,"genie.py",2381,0,"",python,selection_command +242,170917,"genie.py",2428,0,"",python,selection_command +243,171094,"genie.py",2473,0,"",python,selection_command +244,171254,"genie.py",2509,0,"",python,selection_command +245,171378,"genie.py",2551,0,"",python,selection_command +246,171848,"genie.py",2509,0,"",python,selection_command +247,171998,"genie.py",2473,0,"",python,selection_command +248,172308,"genie.py",2428,0,"",python,selection_command +249,172646,"genie.py",2381,0,"",python,selection_command +250,172807,"genie.py",2331,0,"",python,selection_command +251,172950,"genie.py",2292,0,"",python,selection_command +252,173122,"genie.py",2249,0,"",python,selection_command +253,173255,"genie.py",2235,0,"",python,selection_command +254,173678,"genie.py",2203,0,"",python,selection_command +255,173874,"genie.py",2226,0,"\n # param_dtype=self.param_dtype,\n # dtype=self.dtype,",python,content +256,173940,"genie.py",2239,0,"",python,selection_command +257,174092,"genie.py",2203,0,"",python,selection_command +258,174278,"genie.py",2158,0,"",python,selection_command +259,174404,"genie.py",2111,0,"",python,selection_command +260,174551,"genie.py",2061,0,"",python,selection_command +261,174705,"genie.py",2022,0,"",python,selection_command +262,174888,"genie.py",2009,0,"",python,selection_command +263,175436,"genie.py",2009,1,"",python,content +264,175477,"genie.py",2021,0,"",python,selection_command +265,175684,"genie.py",2060,0,"",python,selection_command +266,175918,"genie.py",2021,0,"",python,selection_command +267,176444,"genie.py",2009,38," # model_dim=self.dyna_dim,",python,selection_command +268,176665,"genie.py",2009,88," # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,",python,selection_command +269,177177,"genie.py",2009,135," # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,\n # num_blocks=self.dyna_num_blocks,",python,selection_command +270,177256,"genie.py",2009,180," # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,\n # num_blocks=self.dyna_num_blocks,\n # num_heads=self.dyna_num_heads,",python,selection_command +271,177257,"genie.py",2009,216," # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,\n # num_blocks=self.dyna_num_blocks,\n # num_heads=self.dyna_num_heads,\n # dropout=self.dropout,",python,selection_command +272,177286,"genie.py",2009,260," # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,\n # num_blocks=self.dyna_num_blocks,\n # num_heads=self.dyna_num_heads,\n # dropout=self.dropout,\n # param_dtype=self.param_dtype,",python,selection_command +273,177351,"genie.py",2009,292," # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,\n # num_blocks=self.dyna_num_blocks,\n # num_heads=self.dyna_num_heads,\n # dropout=self.dropout,\n # param_dtype=self.param_dtype,\n # dtype=self.dtype,",python,selection_command +274,177494,"genie.py",2009,302," # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,\n # num_blocks=self.dyna_num_blocks,\n # num_heads=self.dyna_num_heads,\n # dropout=self.dropout,\n # param_dtype=self.param_dtype,\n # dtype=self.dtype,\n )",python,selection_command +275,177770,"genie.py",2009,292," # model_dim=self.dyna_dim,\n # num_latents=self.num_patch_latents,\n # num_blocks=self.dyna_num_blocks,\n # num_heads=self.dyna_num_heads,\n # dropout=self.dropout,\n # param_dtype=self.param_dtype,\n # dtype=self.dtype,",python,selection_command +276,178058,"genie.py",2021,0,"",python,selection_command +277,178504,"genie.py",2282,1,"",python,content +278,178505,"genie.py",2238,1,"",python,content +279,178505,"genie.py",2202,1,"",python,content +280,178505,"genie.py",2157,1,"",python,content +281,178505,"genie.py",2110,1,"",python,content +282,178505,"genie.py",2060,1,"",python,content +283,178505,"genie.py",2021,1,"",python,content +284,178664,"genie.py",2276,1,"",python,content +285,178665,"genie.py",2233,1,"",python,content +286,178665,"genie.py",2198,1,"",python,content +287,178665,"genie.py",2154,1,"",python,content +288,178665,"genie.py",2108,1,"",python,content +289,178665,"genie.py",2059,1,"",python,content +290,178665,"genie.py",2021,1,"",python,content +291,178909,"genie.py",2020,0,"",python,selection_command +292,180156,"genie.py",2215,0,"",python,selection_mouse +293,180165,"genie.py",2214,0,"",python,selection_command +294,180720,"genie.py",2287,0,"",python,selection_mouse +295,180727,"genie.py",2286,0,"",python,selection_command +296,181279,"genie.py",2297,0,"",python,selection_mouse +297,181281,"genie.py",2296,0,"",python,selection_command +298,207064,"models/dynamics.py",0,0,"",python,tab +299,207064,"models/dynamics.py",2655,0,"",python,selection_mouse +300,207959,"models/dynamics.py",2023,0,"",python,selection_mouse +301,207962,"models/dynamics.py",2022,0,"",python,selection_command +302,208522,"models/dynamics.py",1982,0,"",python,selection_mouse +303,208528,"models/dynamics.py",1981,0,"",python,selection_command +304,209153,"models/dynamics.py",2110,0,"",python,selection_mouse +305,209639,"models/dynamics.py",2086,0,"",python,selection_mouse +306,210234,"models/dynamics.py",2110,0,"",python,selection_mouse +307,210942,"models/dynamics.py",2321,0,"",python,selection_mouse +308,210953,"models/dynamics.py",2320,0,"",python,selection_command +309,211544,"models/dynamics.py",2405,0,"",python,selection_mouse +310,211562,"models/dynamics.py",2404,0,"",python,selection_command +311,212218,"models/dynamics.py",2219,0,"",python,selection_mouse +312,212893,"models/dynamics.py",2292,0,"",python,selection_mouse +313,212896,"models/dynamics.py",2291,0,"",python,selection_command +314,213510,"models/dynamics.py",2262,0,"",python,selection_mouse +315,213514,"models/dynamics.py",2261,0,"",python,selection_command +316,220664,"models/dynamics.py",2644,0,"",python,selection_mouse +317,220679,"models/dynamics.py",2643,0,"",python,selection_command +318,221316,"models/dynamics.py",2439,0,"",python,selection_mouse +319,221318,"models/dynamics.py",2438,0,"",python,selection_command +320,221875,"models/dynamics.py",2535,0,"",python,selection_mouse +321,222528,"models/dynamics.py",2642,0,"",python,selection_mouse +322,223612,"models/dynamics.py",2655,0,"",python,selection_mouse +323,255998,"models/dynamics.py",2558,0,"",python,selection_mouse +324,256993,"models/dynamics.py",2756,0,"",python,selection_mouse +325,257575,"models/dynamics.py",2872,0,"",python,selection_mouse +326,258137,"models/dynamics.py",2820,0,"",python,selection_mouse +327,259360,"models/dynamics.py",2766,0,"",python,selection_mouse +328,260298,"models/dynamics.py",2744,51,"",python,content +329,260326,"models/dynamics.py",2752,0,"",python,selection_command +330,261355,"models/dynamics.py",2664,0,"",python,selection_command +331,261577,"models/dynamics.py",2655,0,"",python,selection_command +332,261814,"models/dynamics.py",2664,0,"",python,selection_command +333,267360,"genie.py",0,0,"",python,tab +334,267361,"genie.py",3353,0,"",python,selection_mouse +335,267407,"genie.py",3350,8,"dynamics",python,selection_mouse +336,269532,"genie.py",3363,0,"",python,selection_mouse +337,269688,"genie.py",3359,7,"outputs",python,selection_mouse +338,270358,"genie.py",3306,0,"",python,selection_mouse +339,270493,"genie.py",3304,5,"batch",python,selection_mouse +340,271107,"genie.py",3354,0,"",python,selection_mouse +341,271257,"genie.py",3350,8,"dynamics",python,selection_mouse +342,275148,"genie.py",3371,0,"",python,selection_mouse +343,275382,"genie.py",3368,8,"training",python,selection_mouse +344,276205,"genie.py",3363,0,"",python,selection_mouse +345,276367,"genie.py",3359,7,"outputs",python,selection_mouse +346,277226,"genie.py",3353,0,"",python,selection_mouse +347,277502,"genie.py",3350,8,"dynamics",python,selection_mouse +348,294495,"models/dynamics.py",0,0,"",python,tab +349,294496,"models/dynamics.py",2171,0,"",python,selection_mouse +350,295703,"models/dynamics.py",2132,60,"",python,content +351,295743,"models/dynamics.py",2140,0,"",python,selection_command +352,295911,"models/dynamics.py",2183,0,"",python,selection_command +353,296407,"models/dynamics.py",2211,0,"",python,selection_command +354,296489,"models/dynamics.py",2241,0,"",python,selection_command +355,296497,"models/dynamics.py",2270,0,"",python,selection_command +356,296525,"models/dynamics.py",2298,0,"",python,selection_command +357,296546,"models/dynamics.py",2324,0,"",python,selection_command +358,296570,"models/dynamics.py",2354,0,"",python,selection_command +359,296644,"models/dynamics.py",2378,0,"",python,selection_command +360,296676,"models/dynamics.py",2388,0,"",python,selection_command +361,296707,"models/dynamics.py",2458,0,"",python,selection_command +362,296739,"models/dynamics.py",2493,0,"",python,selection_command +363,296740,"models/dynamics.py",2521,0,"",python,selection_command +364,296856,"models/dynamics.py",2563,0,"",python,selection_command +365,297066,"models/dynamics.py",2593,0,"",python,selection_command +366,297241,"models/dynamics.py",2595,0,"",python,selection_command +367,297425,"models/dynamics.py",2604,0,"",python,selection_command +368,297769,"models/dynamics.py",2692,0,"",python,selection_command +369,300149,"models/dynamics.py",2752,0,"",python,selection_command +370,303683,"models/dynamics.py",2812,0,"",python,selection_command +371,312238,"models/dynamics.py",2886,0,"",python,selection_command +372,312571,"models/dynamics.py",2887,0,"",python,selection_command +373,313074,"models/dynamics.py",2888,0,"",python,selection_command +374,313105,"models/dynamics.py",2889,0,"",python,selection_command +375,313146,"models/dynamics.py",2890,0,"",python,selection_command +376,313165,"models/dynamics.py",2891,0,"",python,selection_command +377,313221,"models/dynamics.py",2892,0,"",python,selection_command +378,313222,"models/dynamics.py",2893,0,"",python,selection_command +379,313252,"models/dynamics.py",2894,0,"",python,selection_command +380,313324,"models/dynamics.py",2895,0,"",python,selection_command +381,313334,"models/dynamics.py",2896,0,"",python,selection_command +382,313722,"models/dynamics.py",2895,0,"",python,selection_command +383,320003,"models/dynamics.py",2962,0,"",python,selection_command +384,320361,"models/dynamics.py",2956,6,"",python,content +385,320566,"models/dynamics.py",2953,3,"",python,content +386,320786,"models/dynamics.py",2948,5,"",python,content +387,321002,"models/dynamics.py",2947,1,"",python,content +388,321215,"models/dynamics.py",2941,6,"",python,content +389,321443,"models/dynamics.py",2936,5,"",python,content +390,321646,"models/dynamics.py",2934,2,"",python,content +391,322028,"models/dynamics.py",2933,1,"",python,content +392,322194,"models/dynamics.py",2932,1,"",python,content +393,322423,"models/dynamics.py",2932,0,"^",python,content +394,322424,"models/dynamics.py",2933,0,"",python,selection_keyboard +395,322974,"models/dynamics.py",2932,1,"",python,content +396,322975,"models/dynamics.py",2932,0,"",python,selection_keyboard +397,323169,"models/dynamics.py",2931,0,"",python,selection_command +398,323519,"models/dynamics.py",2983,0,"",python,selection_command +399,324616,"models/dynamics.py",2931,0,"",python,selection_command +400,324838,"models/dynamics.py",2983,0,"",python,selection_command +401,324955,"models/dynamics.py",2931,0,"",python,selection_command +402,325060,"models/dynamics.py",2983,0,"",python,selection_command +403,325154,"models/dynamics.py",2931,0,"",python,selection_command +404,325243,"models/dynamics.py",2983,0,"",python,selection_command +405,325360,"models/dynamics.py",2931,0,"",python,selection_command +406,325478,"models/dynamics.py",2983,0,"",python,selection_command +407,325575,"models/dynamics.py",2931,0,"",python,selection_command +408,325648,"models/dynamics.py",2983,0,"",python,selection_command +409,325781,"models/dynamics.py",2931,0,"",python,selection_command +410,325848,"models/dynamics.py",2983,0,"",python,selection_command +411,326034,"models/dynamics.py",2931,0,"",python,selection_command +412,326125,"models/dynamics.py",2983,0,"",python,selection_command +413,326413,"models/dynamics.py",2931,0,"",python,selection_command +414,326491,"models/dynamics.py",2983,0,"",python,selection_command +415,326687,"models/dynamics.py",2931,0,"",python,selection_command +416,326744,"models/dynamics.py",2983,0,"",python,selection_command +417,326940,"models/dynamics.py",2931,0,"",python,selection_command +418,327011,"models/dynamics.py",2983,0,"",python,selection_command +419,327202,"models/dynamics.py",2931,0,"",python,selection_command +420,327247,"models/dynamics.py",2983,0,"",python,selection_command +421,327433,"models/dynamics.py",2931,0,"",python,selection_command +422,327568,"models/dynamics.py",2983,0,"",python,selection_command +423,327658,"models/dynamics.py",2931,0,"",python,selection_command +424,327736,"models/dynamics.py",2983,0,"",python,selection_command +425,327902,"models/dynamics.py",2931,0,"",python,selection_command +426,327977,"models/dynamics.py",2983,0,"",python,selection_command +427,328151,"models/dynamics.py",2931,0,"",python,selection_command +428,328196,"models/dynamics.py",2983,0,"",python,selection_command +429,328356,"models/dynamics.py",2931,0,"",python,selection_command +430,328438,"models/dynamics.py",2983,0,"",python,selection_command +431,328588,"models/dynamics.py",2931,0,"",python,selection_command +432,328651,"models/dynamics.py",2983,0,"",python,selection_command +433,328791,"models/dynamics.py",2931,0,"",python,selection_command +434,329335,"models/dynamics.py",2983,0,"",python,selection_command +435,329613,"models/dynamics.py",2931,0,"",python,selection_command +436,329678,"models/dynamics.py",2983,0,"",python,selection_command +437,329849,"models/dynamics.py",2931,0,"",python,selection_command +438,329931,"models/dynamics.py",2983,0,"",python,selection_command +439,330057,"models/dynamics.py",2931,0,"",python,selection_command +440,330091,"models/dynamics.py",2983,0,"",python,selection_command +441,337032,"genie.py",0,0,"",python,tab +442,337107,"genie.py",3350,0,"",python,selection_mouse +443,337478,"genie.py",3405,0,"",python,selection_mouse +444,337661,"genie.py",3401,12,"dyna_outputs",python,selection_mouse +445,342202,"genie.py",3337,0,"",python,selection_mouse +446,342407,"genie.py",3330,12,"dyna_outputs",python,selection_mouse +447,344402,"genie.py",3345,0,"",python,selection_mouse +448,344925,"genie.py",3349,0,"",python,selection_mouse +449,345914,"genie.py",3429,0,"",python,selection_mouse +450,346080,"genie.py",3423,11,"mle_indices",python,selection_mouse +451,347035,"genie.py",3447,0,"",python,selection_mouse +452,399320,"models/dynamics.py",0,0,"",python,tab +453,399321,"models/dynamics.py",2613,0,"",python,selection_mouse +454,401612,"genie.py",0,0,"",python,tab +455,401614,"genie.py",2722,0,"",python,selection_mouse +456,402760,"genie.py",2776,0,"\n ",python,content +457,403283,"genie.py",2785,0,"j",python,content +458,403284,"genie.py",2786,0,"",python,selection_keyboard +459,403375,"genie.py",2786,0,"a",python,content +460,403376,"genie.py",2787,0,"",python,selection_keyboard +461,404098,"genie.py",2785,2,"jax",python,content +462,404395,"genie.py",2788,0,"-",python,content +463,404396,"genie.py",2789,0,"",python,selection_keyboard +464,404724,"genie.py",2789,0,"d",python,content +465,404725,"genie.py",2790,0,"",python,selection_keyboard +466,405044,"genie.py",2789,1,"",python,content +467,405182,"genie.py",2788,1,"",python,content +468,405791,"genie.py",2788,0,".",python,content +469,405792,"genie.py",2789,0,"",python,selection_keyboard +470,406013,"genie.py",2789,0,"d",python,content +471,406015,"genie.py",2790,0,"",python,selection_keyboard +472,406109,"genie.py",2790,0,"e",python,content +473,406110,"genie.py",2791,0,"",python,selection_keyboard +474,406261,"genie.py",2791,0,"b",python,content +475,406263,"genie.py",2792,0,"",python,selection_keyboard +476,406369,"genie.py",2792,0,"u",python,content +477,406369,"genie.py",2793,0,"",python,selection_keyboard +478,406479,"genie.py",2793,0,"g",python,content +479,406480,"genie.py",2794,0,"",python,selection_keyboard +480,406655,"genie.py",2794,0,".",python,content +481,406656,"genie.py",2795,0,"",python,selection_keyboard +482,406881,"genie.py",2795,0,"b",python,content +483,406883,"genie.py",2796,0,"",python,selection_keyboard +484,406958,"genie.py",2796,0,"r",python,content +485,406959,"genie.py",2797,0,"",python,selection_keyboard +486,407155,"genie.py",2797,0,"e",python,content +487,407157,"genie.py",2798,0,"",python,selection_keyboard +488,407451,"genie.py",2795,3,"breakpoint",python,content +489,408224,"genie.py",2805,0,"()",python,content +490,408225,"genie.py",2806,0,"",python,selection_keyboard +491,408322,"genie.py",2806,1,")",python,content +492,408323,"genie.py",2807,0,"",python,selection_keyboard +493,408520,"genie.py",2806,0,"",python,selection_command +494,410565,"models/dynamics.py",0,0,"",python,tab +495,410566,"models/dynamics.py",2595,0,"",python,selection_mouse +496,410907,"models/dynamics.py",2612,0,"",python,selection_mouse +497,411481,"models/dynamics.py",2683,0,"\n jax.debug.breakpoint()",python,content +498,411497,"models/dynamics.py",2692,0,"",python,selection_command +499,415546,"models/dynamics.py",2858,0,"",python,selection_mouse +500,416402,"models/dynamics.py",2908,0,"\n jax.debug.breakpoint()",python,content +501,416448,"models/dynamics.py",2917,0,"",python,selection_command +502,418295,"models/dynamics.py",2165,0,"",python,selection_mouse +503,418628,"utils/nn.py",0,0,"",python,tab +504,420679,"utils/nn.py",2391,0,"",python,selection_mouse +505,421216,"utils/nn.py",2415,0,"\n jax.debug.breakpoint()",python,content +506,421228,"utils/nn.py",2424,0,"",python,selection_command +507,439987,"utils/nn.py",3425,0,"",python,selection_mouse +508,440031,"utils/nn.py",3424,0,"",python,selection_command +509,440880,"utils/nn.py",3425,0,"\n jax.debug.breakpoint()",python,content +510,440915,"utils/nn.py",3434,0,"",python,selection_command +511,441906,"utils/nn.py",3281,0,"",python,selection_mouse +512,442178,"utils/nn.py",3292,0,"\n jax.debug.breakpoint()",python,content +513,442213,"utils/nn.py",3301,0,"",python,selection_command +514,445681,"utils/nn.py",2945,0,"",python,selection_mouse +515,446207,"utils/nn.py",2964,0,"\n jax.debug.breakpoint()",python,content +516,446234,"utils/nn.py",2973,0,"",python,selection_command +517,477050,"TERMINAL",0,0,"\r(jafar) [tum_cte0515@hkn0509 jafar]$ ",,terminal_output +518,483057,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +519,492259,"TERMINAL",0,0,"sh",,terminal_output +520,492543,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +521,492746,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +522,493713,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +523,494668,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",720,0,"",shellscript,selection_mouse +524,495327,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",756,0,"",shellscript,selection_mouse +525,495895,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",746,31,"",shellscript,content +526,495905,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",750,0,"",shellscript,selection_command +527,501332,"TERMINAL",0,0,"bash",,terminal_focus +528,502633,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +529,506033,"TERMINAL",0,0,"srun",,terminal_focus +530,507058,"TERMINAL",0,0,"[?25l\rslurm/jobs/mihir/horeka/yolo-runs/tester.sh\r\n[?2004l\r[?25h\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\n\r\njob_name=""debug""\r\nslurm_job_id=""debug-mihir""\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4\r\n\r\nenv | grep SLURM\r\n\r\nXLA_FLAGS=--xla_gpu_autotune_level=0 srun python train_dynamics.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --min_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --log_checkpoint_interval=1000 \\r\n --name=dynamics-debug-run-$slurm_job_id \\r\n --tags dynamics debug \\r\n --entity instant-uv \\r\n --project jafar \\r\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\r\n --data_dir $array_records_dir\r\n ",,terminal_output +531,507214,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=692048\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\r\nSLURMD_NODENAME=hkn0509\r\nSLURM_JOB_START_TIME=1752667757\r\nSLURM_STEP_NODELIST=hkn0509\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752703757\r\nSLURM_PMI2_SRUN_PORT=40311\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3350302\r\nSLURM_PTY_PORT=34583\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.198\r\nSLURM_PTY_WIN_ROW=43\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0509\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.198\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=300\r\nSLURM_NODELIST=hkn[0509,0511]\r\nSLURM_SRUN_COMM_PORT=35053\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1990.localdomain\r\nSLURM_JOB_ID=3350302\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0509\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_STEP_LAUNCHER_PORT=35053\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0509,0511]\r\n",,terminal_output +532,507345,"TERMINAL",0,0,"GpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +533,534008,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n╭─ Unrecognized options ──────────────────────────╮\r\n│ Unrecognized options: --min-lr=0 │\r\n│ ─────────────────────────────────────────────── │\r\n│ For full helptext, run train_dynamics.py --help │\r\n╰─────────────────────────────────────────────────╯\r\n╭─ Unrecognized options ──────────────────────────╮\r\n│ Unrecognized options: --min-lr=0 │\r\n│ ─────────────────────────────────────────────── │\r\n│ For full helptext, run train_dynamics.py --help │\r\n╰─────────────────────────────────────────────────╯\r\n╭─ Unrecognized options ──────────────────────────╮\r\n│ Unrecognized options: --min-lr=0 │\r\n│ ─────────────────────────────────────────────── │\r\n│ For full helptext, run train_dynamics.py --help │\r\n╰─────────────────────────────────────────────────╯\r\n╭─ Unrecognized options ──────────────────────────╮\r\n│ Unrecognized options: --min-lr=0 │\r\n│ ─────────────────────────────────────────────── │\r\n│ For full helptext, run train_dynamics.py --help │\r\n╰─────────────────────────────────────────────────╯\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n╭─ Unrecognized options ──────────────────────────╮\r\n│ Unrecognized options: --min-lr=0 │\r\n│ ─────────────────────────────────────────────── │\r\n│ For full helptext, run train_dynamics.py --help │\r\n╰─────────────────────────────────────────────────╯\r\n╭─ Unrecognized options ──────────────────────────╮\r\n│ Unrecognized options: --min-lr=0 │\r\n│ ─────────────────────────────────────────────── │\r\n│ For full helptext, run train_dynamics.py --help │\r\n╰─────────────────────────────────────────────────╯\r\n╭─ Unrecognized options ──────────────────────────╮\r\n│ Unrecognized options: --min-lr=0 │\r\n│ ─────────────────────────────────────────────── │\r\n│ For full helptext, run train_dynamics.py --help │\r\n╰─────────────────────────────────────────────────╯\r\n╭─ Unrecognized options ──────────────────────────╮\r\n│ Unrecognized options: --min-lr=0 │\r\n│ ─────────────────────────────────────────────── │\r\n│ For full helptext, run train_dynamics.py --help │\r\n╰─────────────────────────────────────────────────╯\r\n",,terminal_output +534,534244,"TERMINAL",0,0,"srun: error: hkn0511: tasks 4-7: Exited with exit code 2\r\n",,terminal_output +535,534318,"TERMINAL",0,0,"srun: error: hkn0509: tasks 0-3: Exited with exit code 2\r\n]0;tum_cte0515@hkn0509:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0509 jafar]$ ",,terminal_output +536,548393,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +537,549773,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",733,0,"",shellscript,selection_command +538,550689,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",729,17,"",shellscript,content +539,550747,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",733,0,"",shellscript,selection_command +540,551731,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",734,0,"",shellscript,selection_command +541,552292,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",735,0,"",shellscript,selection_command +542,552337,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",736,0,"",shellscript,selection_command +543,552338,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",737,0,"",shellscript,selection_command +544,552413,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",738,0,"",shellscript,selection_command +545,552414,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",739,0,"",shellscript,selection_command +546,552443,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",740,0,"",shellscript,selection_command +547,552444,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",741,0,"",shellscript,selection_command +548,552478,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",742,0,"",shellscript,selection_command +549,552535,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",743,0,"",shellscript,selection_command +550,552545,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",744,0,"",shellscript,selection_command +551,552565,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",745,0,"",shellscript,selection_command +552,552608,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",746,0,"",shellscript,selection_command +553,552609,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",747,0,"",shellscript,selection_command +554,552944,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",746,0,"",shellscript,selection_command +555,553453,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",745,0,"",shellscript,selection_command +556,554408,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",736,0,"",shellscript,selection_mouse +557,554522,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",735,6,"max_lr",shellscript,selection_mouse +558,560232,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +559,561487,"TERMINAL",0,0,"\r\n[?2004l\r\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\n\r\njob_name=""debug""\r\nslurm_job_id=""debug-mihir""\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4\r\n\r\nenv | grep SLURM\r\n\r\nXLA_FLAGS=--xla_gpu_autotune_level=0 srun python train_dynamics.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --log_checkpoint_interval=1000 \\r\n --name=dynamics-debug-run-$slurm_job_id \\r\n --tags dynamics debug \\r\n --entity instant-uv \\r\n --project jafar \\r\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\r\n --data_dir $array_records_dir\r\n ",,terminal_output +560,561603,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=692048\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\r\nSLURMD_NODENAME=hkn0509\r\nSLURM_JOB_START_TIME=1752667757\r\nSLURM_STEP_NODELIST=hkn0509\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752703757\r\nSLURM_PMI2_SRUN_PORT=40311\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3350302\r\nSLURM_PTY_PORT=34583\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.198\r\nSLURM_PTY_WIN_ROW=43\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0509\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.198\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=300\r\nSLURM_NODELIST=hkn[0509,0511]\r\nSLURM_SRUN_COMM_PORT=35053\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1990.localdomain\r\nSLURM_JOB_ID=3350302\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0509\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_STEP_LAUNCHER_PORT=35053\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0509,0511]\r\n",,terminal_output +561,561725,"TERMINAL",0,0,"GpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +562,563734,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +563,563821,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +564,567415,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +565,589805,"genie.py",0,0,"",python,tab +566,599111,"genie.py",0,0,"",python,tab +567,599698,"genie.py",1372,0,"",python,selection_mouse +568,599701,"genie.py",1371,0,"",python,selection_command +569,602582,"genie.py",1372,0,"",python,selection_mouse +570,602584,"genie.py",1371,0,"",python,selection_command +571,604660,"models/dynamics.py",0,0,"",python,tab +572,608559,"train_dynamics.py",0,0,"from dataclasses import dataclass, field\nimport os\n\nimport einops\nfrom flax.training.train_state import TrainState\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n warmup_steps: int = 5000\n lr_schedule : str = ""wsd"" # supported options: wsd, cos\n # Tokenizer\n tokenizer_dim: int = 512\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 8\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 8\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_dim: int = 512\n dyna_num_blocks: int = 12\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n param_dtype: jnp.dtype = jnp.float32\n dtype: jnp.dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n wandb_id: str = """"\n\n\nargs = tyro.cli(Args)\n\n\ndef dynamics_loss_fn(params, state, inputs):\n """"""Compute masked dynamics loss""""""\n inputs[""videos""] = inputs[""videos""].astype(args.dtype) / 255.0\n outputs = state.apply_fn(\n params,\n inputs,\n training=True,\n rngs={""params"": inputs[""rng""], ""dropout"": inputs[""dropout_rng""]},\n )\n mask = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n gt = inputs[""videos""].clip(0, 1).reshape(-1, *inputs[""videos""].shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = pix.psnr(gt, recon).mean() # type: ignore\n ssim = pix.ssim(gt, recon).mean() # type: ignore\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]), size=args.num_latent_actions, fill_value=0\n )\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]), size=args.num_patch_latents, fill_value=0\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n psnr=psnr,\n ssim=ssim,\n codebook_usage_lam=codebook_usage_lam,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n """"""Update state and compute metrics""""""\n grad_fn = jax.value_and_grad(dynamics_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n\n # --- Initialize model ---\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_height, args.image_width, args.image_channels)\n dummy_inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=args.dtype,\n ),\n action=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len), dtype=args.dtype\n ),\n mask_rng=_rng,\n )\n rng, _rng = jax.random.split(rng)\n init_params = genie.init(_rng, dummy_inputs)\n\n param_counts = count_parameters_by_component(init_params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(args.lr_schedule, \n args.init_lr, \n args.max_lr, \n args.decay_end, \n args.num_steps, \n args.warmup_steps, \n args.wsd_decay_steps)\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4, mu_dtype=args.dtype)\n train_state = TrainState.create(apply_fn=genie.apply, params=init_params, tx=tx)\n\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler\n )\n handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler) # type: ignore\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n # Restore full dynamics model\n abstract_train_state = jax.tree_util.tree_map(\n ocp.utils.to_shape_dtype_struct, train_state\n )\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_train_state),\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator),\n ),\n )\n train_state = restored[""model_state""]\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n train_state = restore_genie_components(\n train_state, replicated_sharding, grain_iterator, dummy_inputs, rng, args\n )\n\n # --- TRAIN LOOP ---\n dataloader = (jax.make_array_from_process_local_data(videos_sharding, elem) for elem in grain_iterator) # type: ignore\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _rng_dropout, _rng_mask = jax.random.split(rng, 4)\n\n inputs = dict(\n videos=videos,\n rng=_rng,\n dropout_rng=_rng_dropout,\n mask_rng=_rng_mask,\n )\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.StandardSave(train_state),\n dataloader_state=grain.checkpoint.CheckpointSave(\n grain_iterator\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +573,611943,"train_dynamics.py",1962,0,"",python,selection_mouse +574,611965,"train_dynamics.py",1961,0,"",python,selection_command +575,612803,"train_dynamics.py",2275,0,"",python,selection_mouse +576,612815,"train_dynamics.py",2274,0,"",python,selection_command +577,613417,"train_dynamics.py",2251,0,"",python,selection_mouse +578,613432,"train_dynamics.py",2250,0,"",python,selection_command +579,664891,"genie.py",0,0,"",python,tab +580,665641,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +581,669567,"TERMINAL",0,0,"l",,terminal_output +582,669634,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py(94)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\r\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\r\n latent_actions = jax.lax.cond(\r\n self.lam_co_train,\r\n lambda: lam_outputs[""z_q""],\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py(94)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\r\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\r\n latent_actions = jax.lax.cond(\r\n self.lam_co_train,\r\n lambda: lam_outputs[""z_q""],\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py(94)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\r\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\r\n latent_actions = jax.lax.cond(\r\n self.lam_co_train,\r\n lambda: lam_outputs[""z_q""],\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py(94)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\r\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\r\n latent_actions = jax.lax.cond(\r\n self.lam_co_train,\r\n lambda: lam_outputs[""z_q""],\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py(94)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\r\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\r\n latent_actions = jax.lax.cond(\r\n self.lam_co_train,\r\n lambda: lam_outputs[""z_q""],\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py(94)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\r\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\r\n latent_actions = jax.lax.cond(\r\n self.lam_co_train,\r\n lambda: lam_outputs[""z_q""],\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py(94)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\r\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\r\n latent_actions = jax.lax.cond(\r\n self.lam_co_train,\r\n lambda: lam_outputs[""z_q""],\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py(94)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\r\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\r\n latent_actions = jax.lax.cond(\r\n self.lam_co_train,\r\n lambda: lam_outputs[""z_q""],\r\n",,terminal_output +583,676493,"genie.py",0,0,"",python,tab +584,677388,"genie.py",0,0,"",python,tab +585,678421,"utils/nn.py",0,0,"",python,tab +586,679025,"utils/nn.py",0,0,"",python,tab +587,681584,"models/dynamics.py",0,0,"",python,tab +588,683073,"models/dynamics.py",2310,0,"",python,selection_mouse +589,685956,"models/dynamics.py",2202,0,"",python,selection_mouse +590,685958,"models/dynamics.py",2201,0,"",python,selection_command +591,688934,"models/dynamics.py",2110,0,"",python,selection_mouse +592,694241,"genie.py",0,0,"",python,tab +593,696704,"genie.py",2795,0,"",python,selection_command +594,705850,"TERMINAL",0,0,"b",,terminal_output +595,705996,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +596,706058,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +597,706273,"TERMINAL",0,0,"[?25lc[?25h[?25lh[?25h",,terminal_output +598,707023,"TERMINAL",0,0,"[?25l[[?25h",,terminal_output +599,707373,"TERMINAL",0,0,"[?25l""[?25h",,terminal_output +600,707633,"TERMINAL",0,0,"[?25lv[?25h",,terminal_output +601,707695,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +602,707876,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +603,708003,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +604,708080,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +605,708266,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +606,710169,"TERMINAL",0,0,"[?25l""[?25h",,terminal_output +607,711021,"TERMINAL",0,0,"[?25l][?25h",,terminal_output +608,711757,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +609,712001,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +610,712119,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +611,712235,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +612,712296,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +613,712472,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +614,712663,"TERMINAL",0,0,"\r\n(jdb) (12, 16, 90, 160, 3)\r\n(jdb) (12, 16, 90, 160, 3)\r\n(jdb) (12, 16, 90, 160, 3)\r\n(jdb) (12, 16, 90, 160, 3)\r\n(jdb) (12, 16, 90, 160, 3)\r\n(jdb) (12, 16, 90, 160, 3)\r\n(jdb) (12, 16, 90, 160, 3)\r\n(jdb) (12, 16, 90, 160, 3)\r\n",,terminal_output +615,718913,"TERMINAL",0,0,"c",,terminal_output +616,719773,"TERMINAL",0,0,"\r\n",,terminal_output +617,720917,"TERMINAL",0,0,"2025-07-16 14:21:01.263644: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:21:01.274352: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:21:01.282749: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:21:01.304237: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +618,720971,"TERMINAL",0,0,"2025-07-16 14:21:01.332260: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:21:01.332261: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +619,721072,"TERMINAL",0,0,"2025-07-16 14:21:01.381194: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:21:01.394754: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +620,730283,"genie.py",0,0,"",python,tab +621,730283,"genie.py",2818,0,"",python,selection_mouse +622,730545,"genie.py",2816,17,"tokenizer_outputs",python,selection_mouse +623,734720,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n",,terminal_output +624,734841,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n",,terminal_output +625,735141,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +626,736058,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n",,terminal_output +627,736233,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n",,terminal_output +628,737074,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n",,terminal_output +629,737384,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n",,terminal_output +630,739373,"genie.py",3394,0,"",python,selection_mouse +631,740393,"genie.py",3384,0,"",python,selection_mouse +632,743693,"genie.py",1994,0,"",python,selection_mouse +633,744090,"models/dynamics.py",0,0,"",python,tab +634,749659,"TERMINAL",0,0,"l",,terminal_output +635,749837,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py(92)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n vid_embed = self.patch_embed(batch[""video_tokens""])\r\n act_embed = self.action_up(batch[""latent_actions""])\r\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\r\n jax.debug.breakpoint()\r\n logits = self.dynamics(vid_embed, causal=True)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py(92)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n vid_embed = self.patch_embed(batch[""video_tokens""])\r\n act_embed = self.action_up(batch[""latent_actions""])\r\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\r\n jax.debug.breakpoint()\r\n logits = self.dynamics(vid_embed, causal=True)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py(92)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n vid_embed = self.patch_embed(batch[""video_tokens""])\r\n act_embed = self.action_up(batch[""latent_actions""])\r\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\r\n jax.debug.breakpoint()\r\n logits = self.dynamics(vid_embed, causal=True)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py(92)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n vid_embed = self.patch_embed(batch[""video_tokens""])\r\n act_embed = self.action_up(batch[""latent_actions""])\r\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\r\n jax.debug.breakpoint()\r\n logits = self.dynamics(vid_embed, causal=True)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py(92)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n vid_embed = self.patch_embed(batch[""video_tokens""])\r\n act_embed = self.action_up(batch[""latent_actions""])\r\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\r\n jax.debug.breakpoint()\r\n logits = self.dynamics(vid_embed, causal=True)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py(92)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n vid_embed = self.patch_embed(batch[""video_tokens""])\r\n act_embed = self.action_up(batch[""latent_actions""])\r\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\r\n jax.debug.breakpoint()\r\n logits = self.dynamics(vid_embed, causal=True)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py(92)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n vid_embed = self.patch_embed(batch[""video_tokens""])\r\n act_embed = self.action_up(batch[""latent_actions""])\r\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\r\n jax.debug.breakpoint()\r\n logits = self.dynamics(vid_embed, causal=True)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py(92)\r\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\r\n-> jax.debug.breakpoint()\r\n vid_embed = self.patch_embed(batch[""video_tokens""])\r\n act_embed = self.action_up(batch[""latent_actions""])\r\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\r\n jax.debug.breakpoint()\r\n logits = self.dynamics(vid_embed, causal=True)\r\n",,terminal_output +636,755438,"TERMINAL",0,0,"^[[A",,terminal_output +637,755935,"TERMINAL",0,0,"^[[A",,terminal_output +638,756888,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +639,757282,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +640,757918,"TERMINAL",0,0,"b",,terminal_output +641,757980,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +642,758124,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +643,758263,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +644,758319,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +645,758772,"TERMINAL",0,0,"[?25l[[?25h",,terminal_output +646,759142,"TERMINAL",0,0,"[?25l""[?25h",,terminal_output +647,759688,"TERMINAL",0,0,"[?25lv[?25h",,terminal_output +648,759825,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +649,759965,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +650,760124,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +651,760187,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +652,760399,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +653,760812,"TERMINAL",0,0,"[?25l""[?25h",,terminal_output +654,761198,"TERMINAL",0,0,"[?25l][?25h",,terminal_output +655,761546,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +656,761721,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +657,761782,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +658,761945,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +659,762060,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +660,762122,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +661,762249,"TERMINAL",0,0,"\r\n(jdb) *** KeyError: 'videos'\r\n(jdb) *** KeyError: 'videos'\r\n(jdb) *** KeyError: 'videos'\r\n(jdb) *** KeyError: 'videos'\r\n(jdb) *** KeyError: 'videos'\r\n(jdb) *** KeyError: 'videos'\r\n(jdb) *** KeyError: 'videos'\r\n(jdb) *** KeyError: 'videos'\r\n",,terminal_output +662,765928,"genie.py",0,0,"",python,tab +663,770635,"genie.py",3433,0,"",python,selection_mouse +664,773618,"genie.py",3193,0,"",python,selection_mouse +665,773754,"genie.py",3186,12,"video_tokens",python,selection_mouse +666,776878,"TERMINAL",0,0,"b",,terminal_output +667,776985,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +668,777133,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +669,777237,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +670,777299,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +671,778208,"TERMINAL",0,0,"[?25l[[?25h",,terminal_output +672,778774,"TERMINAL",0,0,"[?25l""[?25h",,terminal_output +673,779741,"TERMINAL",0,0,"[?25lvideo_tokens[?25h",,terminal_output +674,780347,"TERMINAL",0,0,"[?25l""[?25h",,terminal_output +675,780795,"TERMINAL",0,0,"[?25l][?25h",,terminal_output +676,781091,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +677,781268,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +678,781405,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +679,781467,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +680,781618,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +681,781679,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +682,781918,"TERMINAL",0,0,"\r\n(jdb) (12, 16, 920)\r\n(jdb) (12, 16, 920)\r\n(jdb) (12, 16, 920)\r\n(jdb) (12, 16, 920)\r\n(jdb) (12, 16, 920)\r\n(jdb) (12, 16, 920)\r\n(jdb) (12, 16, 920)\r\n(jdb) (12, 16, 920)\r\n",,terminal_output +683,809411,"genie.py",0,0,"",python,tab +684,809412,"genie.py",3680,0,"",python,selection_mouse +685,809515,"genie.py",3679,0,"",python,selection_command +686,810951,"utils/nn.py",0,0,"",python,tab +687,826122,"models/dynamics.py",0,0,"",python,tab +688,827981,"genie.py",0,0,"",python,tab +689,828926,"models/dynamics.py",0,0,"",python,tab +690,832268,"TERMINAL",0,0,"c",,terminal_output +691,832847,"TERMINAL",0,0,"\r\n",,terminal_output +692,833527,"TERMINAL",0,0,"2025-07-16 14:22:53.868052: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:22:53.876120: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:22:53.886693: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:22:53.887821: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:22:53.891457: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:22:53.915671: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:22:53.915639: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:22:53.917118: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +693,833662,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +694,833732,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +695,841177,"TERMINAL",0,0,"b",,terminal_output +696,841322,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +697,841378,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +698,841586,"TERMINAL",0,0,"[?25lch[?25h",,terminal_output +699,842445,"TERMINAL",0,0,"[?25l[[?25h",,terminal_output +700,842819,"TERMINAL",0,0,"[?25l""[?25h",,terminal_output +701,843341,"TERMINAL",0,0,"[?25lv[?25h",,terminal_output +702,843399,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +703,843581,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +704,844106,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +705,844402,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +706,844464,"TERMINAL",0,0,"[?25lm[?25h",,terminal_output +707,844778,"TERMINAL",0,0,"[?25lb[?25h",,terminal_output +708,844840,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +709,844963,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +710,845250,"TERMINAL",0,0,"[?25l""[?25h",,terminal_output +711,845639,"TERMINAL",0,0,"[?25l][?25h",,terminal_output +712,845974,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +713,846480,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +714,846733,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +715,847276,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +716,847337,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +717,847518,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +718,847579,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +719,847787,"TERMINAL",0,0,"\r\n(jdb) *** KeyError: 'vid_embed'\r\n(jdb) *** KeyError: 'vid_embed'\r\n(jdb) *** KeyError: 'vid_embed'\r\n(jdb) *** KeyError: 'vid_embed'\r\n(jdb) *** KeyError: 'vid_embed'\r\n(jdb) *** KeyError: 'vid_embed'\r\n(jdb) *** KeyError: 'vid_embed'\r\n(jdb) *** KeyError: 'vid_embed'\r\n",,terminal_output +720,852375,"TERMINAL",0,0,"[?25lvi[?25h[?25li[?25h",,terminal_output +721,852609,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +722,852790,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +723,853032,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +724,853137,"TERMINAL",0,0,"[?25lm[?25h",,terminal_output +725,853409,"TERMINAL",0,0,"[?25lb[?25h",,terminal_output +726,853472,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +727,853574,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +728,853635,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +729,853869,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +730,853958,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +731,854071,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +732,854133,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +733,854388,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +734,854408,"TERMINAL",0,0,"\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n",,terminal_output +735,858933,"models/dynamics.py",0,0,"",python,tab +736,858934,"models/dynamics.py",2952,0,"",python,selection_mouse +737,859456,"models/dynamics.py",2928,0,"",python,selection_mouse +738,859970,"models/dynamics.py",2917,0,"",python,selection_mouse +739,860482,"models/dynamics.py",2922,0,"",python,selection_mouse +740,860621,"models/dynamics.py",2921,5,"debug",python,selection_mouse +741,860826,"models/dynamics.py",2921,6,"debug.",python,selection_mouse +742,860826,"models/dynamics.py",2921,16,"debug.breakpoint",python,selection_mouse +743,860902,"models/dynamics.py",2921,18,"debug.breakpoint()",python,selection_mouse +744,861258,"models/dynamics.py",2939,0,"",python,selection_mouse +745,861297,"models/dynamics.py",2938,0,"",python,selection_command +746,861598,"models/dynamics.py",2939,0,"",python,selection_mouse +747,861599,"models/dynamics.py",2938,0,"",python,selection_command +748,861747,"models/dynamics.py",2938,1,")",python,selection_mouse +749,861749,"models/dynamics.py",2939,0,"",python,selection_command +750,861788,"models/dynamics.py",2927,12,"breakpoint()",python,selection_mouse +751,861855,"models/dynamics.py",2852,87," += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n jax.debug.breakpoint()",python,selection_mouse +752,861856,"models/dynamics.py",2843,96,"vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n jax.debug.breakpoint()",python,selection_mouse +753,862155,"models/dynamics.py",2917,22,"jax.debug.breakpoint()",python,selection_mouse +754,862620,"models/dynamics.py",2918,0,"",python,selection_mouse +755,862621,"models/dynamics.py",2917,3,"jax",python,selection_mouse +756,862792,"models/dynamics.py",2917,9,"jax.debug",python,selection_mouse +757,862827,"models/dynamics.py",2917,20,"jax.debug.breakpoint",python,selection_mouse +758,862892,"models/dynamics.py",2917,21,"jax.debug.breakpoint(",python,selection_mouse +759,862912,"models/dynamics.py",2917,22,"jax.debug.breakpoint()",python,selection_mouse +760,863217,"models/dynamics.py",2939,0,"",python,selection_mouse +761,863234,"models/dynamics.py",2938,0,"",python,selection_command +762,863364,"models/dynamics.py",2939,0,"",python,selection_mouse +763,863378,"models/dynamics.py",2938,0,"",python,selection_command +764,863519,"models/dynamics.py",2938,1,")",python,selection_mouse +765,863519,"models/dynamics.py",2939,0,"",python,selection_command +766,863584,"models/dynamics.py",2937,2,"()",python,selection_mouse +767,863585,"models/dynamics.py",2927,12,"breakpoint()",python,selection_mouse +768,863685,"models/dynamics.py",2926,13,".breakpoint()",python,selection_mouse +769,863705,"models/dynamics.py",2921,18,"debug.breakpoint()",python,selection_mouse +770,863784,"models/dynamics.py",2920,19,".debug.breakpoint()",python,selection_mouse +771,863818,"models/dynamics.py",2917,22,"jax.debug.breakpoint()",python,selection_mouse +772,864279,"models/dynamics.py",2919,0,"",python,selection_mouse +773,864280,"models/dynamics.py",2917,3,"jax",python,selection_mouse +774,864455,"models/dynamics.py",2917,9,"jax.debug",python,selection_mouse +775,864531,"models/dynamics.py",2853,67,"+= jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n jax",python,selection_mouse +776,864531,"models/dynamics.py",2856,64,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n jax",python,selection_mouse +777,864532,"models/dynamics.py",2860,60,"pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n jax",python,selection_mouse +778,864532,"models/dynamics.py",2863,57,"(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n jax",python,selection_mouse +779,864575,"models/dynamics.py",2864,56,"act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n jax",python,selection_mouse +780,864989,"models/dynamics.py",2869,0,"",python,selection_mouse +781,865488,"models/dynamics.py",2967,0,"",python,selection_mouse +782,869712,"models/dynamics.py",2167,0,"",python,selection_mouse +783,870038,"utils/nn.py",0,0,"",python,tab +784,871380,"utils/nn.py",1190,0,"",python,selection_mouse +785,872380,"genie.py",0,0,"",python,tab +786,873999,"models/dynamics.py",0,0,"",python,tab +787,875464,"models/dynamics.py",2980,0,"",python,selection_mouse +788,875639,"models/dynamics.py",2980,2,", ",python,selection_mouse +789,875640,"models/dynamics.py",2980,3,", c",python,selection_mouse +790,875641,"models/dynamics.py",2980,5,", cau",python,selection_mouse +791,875706,"models/dynamics.py",2980,7,", causa",python,selection_mouse +792,875706,"models/dynamics.py",2980,8,", causal",python,selection_mouse +793,875738,"models/dynamics.py",2980,9,", causal=",python,selection_mouse +794,875738,"models/dynamics.py",2980,10,", causal=T",python,selection_mouse +795,875792,"models/dynamics.py",2980,11,", causal=Tr",python,selection_mouse +796,875804,"models/dynamics.py",2980,12,", causal=Tru",python,selection_mouse +797,875828,"models/dynamics.py",2980,13,", causal=True",python,selection_mouse +798,876863,"models/dynamics.py",2980,13,"",python,content +799,878124,"utils/nn.py",0,0,"",python,tab +800,879364,"utils/nn.py",1193,0,"",python,selection_mouse +801,880683,"utils/nn.py",1307,0,"",python,selection_mouse +802,881583,"utils/nn.py",1608,0,"",python,selection_mouse +803,882377,"utils/nn.py",1273,0,"",python,selection_mouse +804,883265,"utils/nn.py",1816,0,"",python,selection_mouse +805,884763,"TERMINAL",0,0,"c",,terminal_output +806,886111,"TERMINAL",0,0,"\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 107, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 107, in __call__\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 107, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 107, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 107, in __call__\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 107, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 107, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 107, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed)\r\nTypeError: CausalTransformer.__call__() got an unexpected keyword argument 'causal'\r\n logits = self.dynamics(vid_embed)\r\nTypeError: CausalTransformer.__call__() got an unexpected keyword argument 'causal'\r\n logits = self.dynamics(vid_embed)\r\nTypeError: CausalTransformer.__call__() got an unexpected keyword argument 'causal'\r\n logits = self.dynamics(vid_embed)\r\nTypeError: CausalTransformer.__call__() got an unexpected keyword argument 'causal'\r\n logits = self.dynamics(vid_embed)\r\nTypeError: CausalTransformer.__call__() got an unexpected keyword argument 'causal'\r\n logits = self.dynamics(vid_embed)\r\nTypeError: CausalTransformer.__call__() got an unexpected keyword argument 'causal'\r\n logits = self.dynamics(vid_embed)\r\nTypeError: CausalTransformer.__call__() got an unexpected keyword argument 'causal'\r\n logits = self.dynamics(vid_embed)\r\nTypeError: CausalTransformer.__call__() got an unexpected keyword argument 'causal'\r\n",,terminal_output +807,887417,"TERMINAL",0,0,"(jdb) (jdb) (jdb) (jdb) (jdb) (jdb) (jdb) (jdb) ",,terminal_output +808,887814,"TERMINAL",0,0,"srun: error: hkn0511: tasks 4-5: Exited with exit code 1\r\nsrun: error: hkn0509: tasks 1-2: Exited with exit code 1\r\n",,terminal_output +809,887921,"TERMINAL",0,0,"srun: error: hkn0509: task 0: Exited with exit code 1\r\nsrun: error: hkn0511: task 7: Exited with exit code 1\r\n",,terminal_output +810,888156,"TERMINAL",0,0,"srun: error: hkn0511: task 6: Exited with exit code 1\r\nsrun: error: hkn0509: task 3: Exited with exit code 1\r\n]0;tum_cte0515@hkn0509:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0509 jafar]$ ",,terminal_output +811,888241,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0509:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0509 jafar]$ ",,terminal_output +812,890190,"TERMINAL",0,0,"[?25lls[?25h[?25ls[?25h",,terminal_output +813,891194,"TERMINAL",0,0,"[?25ls[?25h[?25lm[?25h",,terminal_output +814,891314,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +815,891504,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +816,893039,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: nvidia-smihkn0509.localdomain: Wed Jul 16 14:23:51 2025Wed Jul 16 14:23:51 2025\r+-----------------------------------------------------------------------------------------+\r| NVIDIA-SMI 570.133.20Driver Version: 570.133.20 CUDA Version: 12.8 |\r|-----------------------------------------+------------------------+----------------------+\r| GPU NamePersistence-M | Bus-IdDisp.A | Volatile Uncorr. ECC |\r| Fan Temp PerfPwr:Usage/Cap |Memory-Usage | GPU-Util Compute M. |\r|||MIG M. |\r|=========================================+========================+======================|\r| 0 NVIDIA A100-SXM4-40GBOn | 00000000:31:00.0 Off |0 |\r| N/A 46C P057W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r| 1 NVIDIA A100-SXM4-40GBOn | 00000000:4B:00.0 Off |0 |\r| N/A 46C P063W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r| 2 NVIDIA A100-SXM4-40GBOn | 00000000:CA:00.0 Off |0 |\r| N/A 45C P053W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r| 3 NVIDIA A100-SXM4-40GBOn | 00000000:E3:00.0 Off |0 |\r| N/A 43C P054W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r+-----------------------------------------------------------------------------------------+\r| Processes:|\r| GPU GI CIPID Type Process nameGPU Memory |\r|ID IDUsage\t |\r|=========================================================================================|\r| 0 N/A N/A2668G /usr/libexec/Xorg17MiB |\r| 1 N/A N/A2668G /usr/libexec/Xorg17MiB |\r| 2 N/A N/A2668G /usr/libexec/Xorg17MiB |\r| 3 N/A N/A2668G /usr/libexec/Xorg17MiB |\r+-----------------------------------------------------------------------------------------+",,terminal_output +817,894058,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0509:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0509 jafar]$ ",,terminal_output +818,895173,"genie.py",0,0,"",python,tab +819,897158,"genie.py",2797,0,"",python,selection_mouse +820,897846,"genie.py",2777,31,"",python,content +821,897883,"genie.py",2785,0,"",python,selection_command +822,899449,"models/dynamics.py",0,0,"",python,tab +823,900816,"models/dynamics.py",2707,0,"",python,selection_mouse +824,901357,"models/dynamics.py",2684,31,"",python,content +825,901419,"models/dynamics.py",2692,0,"",python,selection_command +826,901942,"models/dynamics.py",2752,0,"",python,selection_command +827,902134,"models/dynamics.py",2812,0,"",python,selection_command +828,902300,"models/dynamics.py",2886,0,"",python,selection_command +829,902451,"models/dynamics.py",2917,0,"",python,selection_command +830,902688,"models/dynamics.py",2886,0,"",python,selection_command +831,905085,"models/dynamics.py",2878,31,"",python,content +832,905185,"models/dynamics.py",2886,0,"",python,selection_command +833,906299,"models/dynamics.py",2904,0,"",python,selection_mouse +834,907614,"models/dynamics.py",2164,0,"",python,selection_mouse +835,907857,"utils/nn.py",0,0,"",python,tab +836,908726,"utils/nn.py",1186,0,"",python,selection_mouse +837,910292,"utils/nn.py",1216,0,"\n jax.debug.breakpoint()",python,content +838,910324,"utils/nn.py",1225,0,"",python,selection_command +839,911049,"utils/nn.py",1256,0,"",python,selection_command +840,911286,"utils/nn.py",1225,0,"",python,selection_command +841,911802,"utils/nn.py",1256,0,"",python,selection_command +842,911923,"utils/nn.py",1300,0,"",python,selection_command +843,912093,"utils/nn.py",1326,0,"",python,selection_command +844,912234,"utils/nn.py",1368,0,"",python,selection_command +845,912357,"utils/nn.py",1398,0,"",python,selection_command +846,912972,"utils/nn.py",1411,0,"",python,selection_command +847,913385,"utils/nn.py",1431,0,"\n jax.debug.breakpoint()",python,content +848,913423,"utils/nn.py",1440,0,"",python,selection_command +849,913719,"utils/nn.py",1471,0,"",python,selection_command +850,913932,"utils/nn.py",1519,0,"",python,selection_command +851,914028,"utils/nn.py",1592,0,"",python,selection_command +852,914166,"utils/nn.py",1627,0,"",python,selection_command +853,914351,"utils/nn.py",1665,0,"",python,selection_command +854,914458,"utils/nn.py",1706,0,"",python,selection_command +855,914662,"utils/nn.py",1745,0,"",python,selection_command +856,914820,"utils/nn.py",1787,0,"",python,selection_command +857,914942,"utils/nn.py",1817,0,"",python,selection_command +858,915134,"utils/nn.py",1848,0,"",python,selection_command +859,915859,"utils/nn.py",1857,0,"\n jax.debug.breakpoint()",python,content +860,915910,"utils/nn.py",1866,0,"",python,selection_command +861,916296,"utils/nn.py",1848,0,"",python,selection_command +862,916708,"utils/nn.py",1817,0,"",python,selection_command +863,916898,"utils/nn.py",1839,0,"\n jax.debug.breakpoint()",python,content +864,916909,"utils/nn.py",1848,0,"",python,selection_command +865,917137,"utils/nn.py",1879,0,"",python,selection_command +866,917301,"utils/nn.py",1897,0,"",python,selection_command +867,917437,"utils/nn.py",1920,0,"",python,selection_command +868,917587,"utils/nn.py",1929,0,"",python,selection_command +869,917767,"utils/nn.py",1951,0,"",python,selection_command +870,918042,"utils/nn.py",1977,0,"",python,selection_command +871,918398,"utils/nn.py",2019,0,"",python,selection_command +872,918709,"utils/nn.py",2049,0,"",python,selection_command +873,919178,"utils/nn.py",2053,0,"\n jax.debug.breakpoint()",python,content +874,919214,"utils/nn.py",2062,0,"",python,selection_command +875,919416,"utils/nn.py",2093,0,"",python,selection_command +876,919590,"utils/nn.py",2115,0,"",python,selection_command +877,919746,"utils/nn.py",2143,0,"",python,selection_command +878,919880,"utils/nn.py",2185,0,"",python,selection_command +879,920086,"utils/nn.py",2215,0,"",python,selection_command +880,920258,"utils/nn.py",2228,0,"",python,selection_command +881,920597,"utils/nn.py",2215,0,"",python,selection_command +882,920798,"utils/nn.py",2219,0,"\n jax.debug.breakpoint()",python,content +883,920849,"utils/nn.py",2228,0,"",python,selection_command +884,920992,"utils/nn.py",2259,0,"",python,selection_command +885,921179,"utils/nn.py",2282,0,"",python,selection_command +886,921526,"utils/nn.py",2259,0,"",python,selection_command +887,921726,"utils/nn.py",2273,0,"\n jax.debug.breakpoint()",python,content +888,921755,"utils/nn.py",2282,0,"",python,selection_command +889,921877,"utils/nn.py",2313,0,"",python,selection_command +890,922299,"utils/nn.py",2322,0,"\n jax.debug.breakpoint()",python,content +891,922306,"utils/nn.py",2331,0,"",python,selection_command +892,925613,"TERMINAL",0,0,"smi",,terminal_output +893,925993,"TERMINAL",0,0,"h slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +894,926808,"TERMINAL",0,0,"\r\n[?2004l\r\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\n\r\njob_name=""debug""\r\nslurm_job_id=""debug-mihir""\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4\r\n\r\nenv | grep SLURM\r\n\r\nXLA_FLAGS=--xla_gpu_autotune_level=0 srun python train_dynamics.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --log_checkpoint_interval=1000 \\r\n --name=dynamics-debug-run-$slurm_job_id \\r\n --tags dynamics debug \\r\n --entity instant-uv \\r\n --project jafar \\r\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\r\n --data_dir $array_records_dir\r\n ",,terminal_output +895,926922,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=692048\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\r\nSLURMD_NODENAME=hkn0509\r\nSLURM_JOB_START_TIME=1752667757\r\nSLURM_STEP_NODELIST=hkn0509\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752703757\r\nSLURM_PMI2_SRUN_PORT=40311\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3350302\r\nSLURM_PTY_PORT=34583\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.198\r\nSLURM_PTY_WIN_ROW=43\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0509\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.198\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=300\r\nSLURM_NODELIST=hkn[0509,0511]\r\nSLURM_SRUN_COMM_PORT=35053\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1990.localdomain\r\nSLURM_JOB_ID=3350302\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0509\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_STEP_LAUNCHER_PORT=35053\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0509,0511]\r\n",,terminal_output +896,927072,"TERMINAL",0,0,"GpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +897,929214,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +898,932881,"TERMINAL",0,0,"2025-07-16 14:24:33.189282: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:24:33.191967: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:24:33.192346: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:24:33.203020: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:24:33.220334: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:24:33.225770: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:24:33.225768: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:24:33.229109: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +899,947412,"TERMINAL",0,0,"2025-07-16 14:24:47.774012: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +900,947586,"TERMINAL",0,0,"2025-07-16 14:24:47.947507: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:24:47.975268: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +901,947643,"TERMINAL",0,0,"2025-07-16 14:24:47.982137: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +902,947857,"TERMINAL",0,0,"2025-07-16 14:24:48.115747: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:24:48.119287: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\nRunning on 8 devices.\r\nEntering jdb:\r\n2025-07-16 14:24:48.180127: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:24:48.180058: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +903,947962,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +904,948022,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +905,1029957,"TERMINAL",0,0,"x",,terminal_output +906,1030122,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +907,1030346,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +908,1030482,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +909,1030645,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +910,1030781,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +911,1030886,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +912,1031103,"TERMINAL",0,0,"\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n",,terminal_output +913,1034920,"TERMINAL",0,0,"c",,terminal_output +914,1035143,"TERMINAL",0,0,"\r\n",,terminal_output +915,1035607,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +916,1041147,"TERMINAL",0,0,"s",,terminal_output +917,1041980,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +918,1042152,"TERMINAL",0,0,"[?25lq[?25h",,terminal_output +919,1042429,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +920,1042679,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +921,1042819,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +922,1042880,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +923,1043103,"TERMINAL",0,0,"\r\n(jdb) *** NameError: name 'seq_len' is not defined\r\n(jdb) *** NameError: name 'seq_len' is not defined\r\n(jdb) *** NameError: name 'seq_len' is not defined\r\n(jdb) *** NameError: name 'seq_len' is not defined\r\n(jdb) *** NameError: name 'seq_len' is not defined\r\n(jdb) *** NameError: name 'seq_len' is not defined\r\n(jdb) *** NameError: name 'seq_len' is not defined\r\n(jdb) *** NameError: name 'seq_len' is not defined\r\n",,terminal_output +924,1044906,"TERMINAL",0,0,"l",,terminal_output +925,1045087,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(108)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n ),\r\n ]\r\n )(x)\r\n # Causal transformer blocks\r\n-> jax.debug.breakpoint()\r\n for _ in range(self.num_blocks):\r\n x = CausalTransformerBlock(\r\n model_dim=self.model_dim,\r\n num_heads=self.num_heads,\r\n dropout=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(108)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n ),\r\n ]\r\n )(x)\r\n # Causal transformer blocks\r\n-> jax.debug.breakpoint()\r\n for _ in range(self.num_blocks):\r\n x = CausalTransformerBlock(\r\n model_dim=self.model_dim,\r\n num_heads=self.num_heads,\r\n dropout=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(108)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n ),\r\n ]\r\n )(x)\r\n # Causal transformer blocks\r\n-> jax.debug.breakpoint()\r\n for _ in range(self.num_blocks):\r\n x = CausalTransformerBlock(\r\n model_dim=self.model_dim,\r\n num_heads=self.num_heads,\r\n dropout=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(108)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n ),\r\n ]\r\n )(x)\r\n # Causal transformer blocks\r\n-> jax.debug.breakpoint()\r\n for _ in range(self.num_blocks):\r\n x = CausalTransformerBlock(\r\n model_dim=self.model_dim,\r\n num_heads=self.num_heads,\r\n dropout=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(108)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n ),\r\n ]\r\n )(x)\r\n # Causal transformer blocks\r\n-> jax.debug.breakpoint()\r\n for _ in range(self.num_blocks):\r\n x = CausalTransformerBlock(\r\n model_dim=self.model_dim,\r\n num_heads=self.num_heads,\r\n dropout=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(108)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n ),\r\n ]\r\n )(x)\r\n # Causal transformer blocks\r\n-> jax.debug.breakpoint()\r\n for _ in range(self.num_blocks):\r\n x = CausalTransformerBlock(\r\n model_dim=self.model_dim,\r\n num_heads=self.num_heads,\r\n dropout=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(108)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n ),\r\n ]\r\n )(x)\r\n # Causal transformer blocks\r\n-> jax.debug.breakpoint()\r\n for _ in range(self.num_blocks):\r\n x = CausalTransformerBlock(\r\n model_dim=self.model_dim,\r\n num_heads=self.num_heads,\r\n dropout=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(108)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n ),\r\n ]\r\n )(x)\r\n # Causal transformer blocks\r\n-> jax.debug.breakpoint()\r\n for _ in range(self.num_blocks):\r\n x = CausalTransformerBlock(\r\n model_dim=self.model_dim,\r\n num_heads=self.num_heads,\r\n dropout=self.dropout,\r\n",,terminal_output +926,1051890,"utils/nn.py",0,0,"",python,tab +927,1051891,"utils/nn.py",1517,0,"",python,selection_mouse +928,1052606,"utils/nn.py",1853,0,"",python,selection_mouse +929,1060259,"utils/nn.py",1235,0,"",python,selection_mouse +930,1060431,"utils/nn.py",1235,10,"breakpoint",python,selection_mouse +931,1060566,"utils/nn.py",1217,31," jax.debug.breakpoint()\n",python,selection_mouse +932,1081669,"TERMINAL",0,0,"x",,terminal_output +933,1081836,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +934,1082058,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +935,1082218,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +936,1082405,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +937,1083331,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +938,1083441,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +939,1083578,"TERMINAL",0,0,"\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n",,terminal_output +940,1092010,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +941,1093839,"TERMINAL",0,0,"\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +942,1096510,"utils/nn.py",0,0,"",python,tab +943,1096511,"utils/nn.py",1630,0,"",python,selection_mouse +944,1098665,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +945,1098855,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(38)\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n-> jax.debug.breakpoint()\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(38)\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n-> jax.debug.breakpoint()\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(38)\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n-> jax.debug.breakpoint()\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(38)\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n-> jax.debug.breakpoint()\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(38)\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n-> jax.debug.breakpoint()\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(38)\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n-> jax.debug.breakpoint()\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(38)\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n-> jax.debug.breakpoint()\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(38)\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n-> jax.debug.breakpoint()\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n",,terminal_output +946,1107685,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +947,1107810,"TERMINAL",0,0,"\r\n",,terminal_output +948,1107862,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +949,1109529,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +950,1109757,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +951,1110631,"TERMINAL",0,0,"[?25lq[?25h",,terminal_output +952,1110916,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +953,1111192,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +954,1111435,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +955,1111489,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +956,1112032,"TERMINAL",0,0,"\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n",,terminal_output +957,1116819,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +958,1116963,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(45)\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n-> jax.debug.breakpoint()\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(45)\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n-> jax.debug.breakpoint()\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(45)\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n-> jax.debug.breakpoint()\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(45)\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n-> jax.debug.breakpoint()\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(45)\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n-> jax.debug.breakpoint()\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(45)\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n-> jax.debug.breakpoint()\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(45)\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n-> jax.debug.breakpoint()\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(45)\r\n # LayerNorm + Causal Self-Attention\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n-> jax.debug.breakpoint()\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n",,terminal_output +959,1119212,"TERMINAL",0,0,"[?25lz[?25h",,terminal_output +960,1119599,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +961,1119932,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +962,1119993,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +963,1120262,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +964,1121427,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +965,1121490,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +966,1121585,"TERMINAL",0,0,"\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n",,terminal_output +967,1145764,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +968,1145939,"TERMINAL",0,0,"\r\n",,terminal_output +969,1146524,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 95, in __call__\r\n logits = self.dynamics(vid_embed)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 110, in __call__\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 48, in __call__\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 95, in __call__\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n logits = self.dynamics(vid_embed)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 110, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 95, in __call__\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 48, in __call__\r\n logits = self.dynamics(vid_embed)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 110, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 48, in __call__\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 140, in dot_product_attention_weights\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 140, in dot_product_attention_weights\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n attn_weights = jnp.where(mask, attn_weights, big_neg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/lax_numpy.py"", line 2821, in where\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n attn_weights = jnp.where(mask, attn_weights, big_neg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/lax_numpy.py"", line 2821, in where\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 95, in __call__\r\n logits = self.dynamics(vid_embed)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 110, in __call__\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 48, in __call__\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 140, in dot_product_attention_weights\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n attn_weights = jnp.where(mask, attn_weights, big_neg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/lax_numpy.py"", line 2821, in where\r\n return util._where(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 311, in _where\r\n condition, x_arr, y_arr = _broadcast_arrays(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 264, in _broadcast_arrays\r\n result_shape = lax.broadcast_shapes(*shapes)\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n return util._where(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 311, in _where\r\n condition, x_arr, y_arr = _broadcast_arrays(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 264, in _broadcast_arrays\r\n result_shape = lax.broadcast_shapes(*shapes)\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 140, in dot_product_attention_weights\r\n return util._where(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 311, in _where\r\n condition, x_arr, y_arr = _broadcast_arrays(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 264, in _broadcast_arrays\r\n result_shape = lax.broadcast_shapes(*shapes)\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n attn_weights = jnp.where(mask, attn_weights, big_neg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/lax_numpy.py"", line 2821, in where\r\n return util._where(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 311, in _where\r\n condition, x_arr, y_arr = _broadcast_arrays(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 264, in _broadcast_arrays\r\n result_shape = lax.broadcast_shapes(*shapes)\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 95, in __call__\r\n logits = self.dynamics(vid_embed)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 110, in __call__\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 48, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 95, in __call__\r\n logits = self.dynamics(vid_embed)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 110, in __call__\r\n x = CausalTransformerBlock(\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 48, in __call__\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (1, 1, 1, 16, 16), (12, 16, 8, 920, 920), (1, 1, 1, 1, 1).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 95, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 95, in __call__\r\n logits = self.dynamics(vid_embed)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 110, in __call__\r\n logits = self.dynamics(vid_embed)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 110, in __call__\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 48, in __call__\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 48, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 140, in dot_product_attention_weights\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 140, in dot_product_attention_weights\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n attn_weights = jnp.where(mask, attn_weights, big_neg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/lax_numpy.py"", line 2821, in where\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 140, in dot_product_attention_weights\r\n attn_weights = jnp.where(mask, attn_weights, big_neg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/lax_numpy.py"", line 2821, in where\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 140, in dot_product_attention_weights\r\n attn_weights = jnp.where(mask, attn_weights, big_neg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/lax_numpy.py"", line 2821, in where\r\n return util._where(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 311, in _where\r\n return util._where(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 311, in _where\r\n attn_weights = jnp.where(mask, attn_weights, big_neg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/lax_numpy.py"", line 2821, in where\r\n condition, x_arr, y_arr = _broadcast_arrays(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 264, in _broadcast_arrays\r\n condition, x_arr, y_arr = _broadcast_arrays(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 264, in _broadcast_arrays\r\n result_shape = lax.broadcast_shapes(*shapes)\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n result_shape = lax.broadcast_shapes(*shapes)\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n return util._where(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 311, in _where\r\n condition, x_arr, y_arr = _broadcast_arrays(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 264, in _broadcast_arrays\r\n return util._where(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 311, in _where\r\n condition, x_arr, y_arr = _broadcast_arrays(condition, x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 264, in _broadcast_arrays\r\n result_shape = lax.broadcast_shapes(*shapes)\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n result_shape = lax.broadcast_shapes(*shapes)\r\nValueError: Incompatible shapes for broadcasting: shapes=[(16, 16), (12, 16, 8, 920, 920), ()]\r\n",,terminal_output +970,1147501,"TERMINAL",0,0,"(jdb) (jdb) (jdb) (jdb) (jdb) ",,terminal_output +971,1147575,"TERMINAL",0,0,"(jdb) (jdb) (jdb) ",,terminal_output +972,1147867,"TERMINAL",0,0,"srun: error: hkn0509: tasks 0,2: Exited with exit code 1\r\n",,terminal_output +973,1147973,"TERMINAL",0,0,"srun: error: hkn0511: tasks 5,7: Exited with exit code 1\r\n",,terminal_output +974,1148080,"TERMINAL",0,0,"srun: error: hkn0509: task 1: Exited with exit code 1\r\nsrun: error: hkn0511: task 6: Exited with exit code 1\r\n",,terminal_output +975,1148206,"TERMINAL",0,0,"srun: error: hkn0509: task 3: Exited with exit code 1\r\n",,terminal_output +976,1148332,"TERMINAL",0,0,"srun: error: hkn0511: task 4: Exited with exit code 1\r\n]0;tum_cte0515@hkn0509:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0509 jafar]$ ",,terminal_output +977,1164752,"utils/nn.py",0,0,"",python,tab +978,1164753,"utils/nn.py",1616,0,"",python,selection_mouse +979,1165465,"utils/nn.py",1538,0,"",python,selection_mouse +980,1166883,"utils/nn.py",1446,0,"",python,selection_mouse +981,1167387,"utils/nn.py",1477,0,"",python,selection_command +982,1167546,"utils/nn.py",1525,0,"",python,selection_command +983,1167879,"utils/nn.py",1583,0,"\n jax.debug.breakpoint()",python,content +984,1167921,"utils/nn.py",1592,0,"",python,selection_command +985,1199064,"utils/nn.py",1870,0,"",python,selection_mouse +986,1199094,"utils/nn.py",1869,0,"",python,selection_command +987,1204515,"utils/nn.py",0,0,"",python,tab +988,1204515,"utils/nn.py",1242,0,"",python,selection_mouse +989,1206179,"utils/nn.py",1532,0,"",python,selection_mouse +990,1206888,"utils/nn.py",1605,0,"",python,selection_mouse +991,1207859,"utils/nn.py",1553,0,"",python,selection_mouse +992,1208677,"utils/nn.py",1548,0,"",python,selection_mouse +993,1209455,"utils/nn.py",1647,0,"",python,selection_mouse +994,1247122,"utils/nn.py",1606,0,"",python,selection_mouse +995,1247798,"utils/nn.py",1455,0,"",python,selection_mouse +996,1248712,"utils/nn.py",1432,31,"",python,content +997,1248786,"utils/nn.py",1440,0,"",python,selection_command +998,1248825,"utils/nn.py",1411,0,"",python,selection_command +999,1248975,"utils/nn.py",1398,0,"",python,selection_command +1000,1249106,"utils/nn.py",1368,0,"",python,selection_command +1001,1249261,"utils/nn.py",1326,0,"",python,selection_command +1002,1249372,"utils/nn.py",1300,0,"",python,selection_command +1003,1249506,"utils/nn.py",1256,0,"",python,selection_command +1004,1249671,"utils/nn.py",1225,0,"",python,selection_command +1005,1250082,"utils/nn.py",1217,31,"",python,content +1006,1250095,"utils/nn.py",1225,0,"",python,selection_command +1007,1250206,"utils/nn.py",1269,0,"",python,selection_command +1008,1250406,"utils/nn.py",1295,0,"",python,selection_command +1009,1250547,"utils/nn.py",1337,0,"",python,selection_command +1010,1250707,"utils/nn.py",1367,0,"",python,selection_command +1011,1250859,"utils/nn.py",1380,0,"",python,selection_command +1012,1251024,"utils/nn.py",1409,0,"",python,selection_command +1013,1251159,"utils/nn.py",1457,0,"",python,selection_command +1014,1251330,"utils/nn.py",1530,0,"",python,selection_command +1015,1251833,"utils/nn.py",1561,0,"",python,selection_command +1016,1251879,"utils/nn.py",1596,0,"",python,selection_command +1017,1253976,"utils/nn.py",2683,0,"",python,selection_mouse +1018,1254668,"utils/nn.py",2652,0,"",python,selection_command +1019,1254913,"utils/nn.py",2633,31,"",python,content +1020,1255002,"utils/nn.py",2641,0,"",python,selection_command +1021,1255037,"utils/nn.py",2668,0,"",python,selection_command +1022,1255565,"utils/nn.py",2682,0,"",python,selection_command +1023,1255567,"utils/nn.py",2712,0,"",python,selection_command +1024,1255582,"utils/nn.py",2762,0,"",python,selection_command +1025,1255587,"utils/nn.py",2800,0,"",python,selection_command +1026,1255621,"utils/nn.py",2819,0,"",python,selection_command +1027,1255648,"utils/nn.py",2860,0,"",python,selection_command +1028,1255674,"utils/nn.py",2906,0,"",python,selection_command +1029,1255764,"utils/nn.py",2940,0,"",python,selection_command +1030,1255764,"utils/nn.py",2959,0,"",python,selection_command +1031,1255775,"utils/nn.py",2989,0,"",python,selection_command +1032,1255859,"utils/nn.py",3039,0,"",python,selection_command +1033,1256058,"utils/nn.py",3077,0,"",python,selection_command +1034,1256221,"utils/nn.py",3096,0,"",python,selection_command +1035,1256344,"utils/nn.py",3110,0,"",python,selection_command +1036,1256467,"utils/nn.py",3123,0,"",python,selection_command +1037,1256605,"utils/nn.py",3159,0,"",python,selection_command +1038,1257006,"utils/nn.py",3151,31,"",python,content +1039,1257060,"utils/nn.py",3159,0,"",python,selection_command +1040,1257108,"utils/nn.py",3123,0,"",python,selection_command +1041,1258411,"utils/nn.py",3496,0,"",python,selection_mouse +1042,1258700,"utils/nn.py",3479,31,"",python,content +1043,1258758,"utils/nn.py",3487,0,"",python,selection_command +1044,1259777,"utils/nn.py",3626,0,"",python,selection_mouse +1045,1260049,"utils/nn.py",3612,31,"",python,content +1046,1260090,"utils/nn.py",3620,0,"",python,selection_command +1047,1261689,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +1048,1262287,"TERMINAL",0,0,"\r\n[?2004l\r\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\n\r\njob_name=""debug""\r\nslurm_job_id=""debug-mihir""\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4\r\n\r\nenv | grep SLURM\r\n\r\nXLA_FLAGS=--xla_gpu_autotune_level=0 srun python train_dynamics.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --log_checkpoint_interval=1000 \\r\n --name=dynamics-debug-run-$slurm_job_id \\r\n --tags dynamics debug \\r\n --entity instant-uv \\r\n --project jafar \\r\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\r\n --data_dir $array_records_dir\r\n ",,terminal_output +1049,1262405,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=692048\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\r\nSLURMD_NODENAME=hkn0509\r\nSLURM_JOB_START_TIME=1752667757\r\nSLURM_STEP_NODELIST=hkn0509\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752703757\r\nSLURM_PMI2_SRUN_PORT=40311\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3350302\r\nSLURM_PTY_PORT=34583\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.198\r\nSLURM_PTY_WIN_ROW=43\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0509\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.198\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=300\r\nSLURM_NODELIST=hkn[0509,0511]\r\nSLURM_SRUN_COMM_PORT=35053\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1990.localdomain\r\nSLURM_JOB_ID=3350302\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0509\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_STEP_LAUNCHER_PORT=35053\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0509,0511]\r\n",,terminal_output +1050,1262530,"TERMINAL",0,0,"GpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +1051,1264838,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +1052,1269211,"TERMINAL",0,0,"2025-07-16 14:30:09.517407: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:30:09.526133: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:30:09.549864: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:30:09.553191: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:30:09.554347: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:30:09.581597: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:30:09.596495: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1053,1269267,"TERMINAL",0,0,"2025-07-16 14:30:09.641337: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1054,1283700,"TERMINAL",0,0,"2025-07-16 14:30:24.076516: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1055,1283946,"TERMINAL",0,0,"2025-07-16 14:30:24.334901: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:30:24.339024: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1056,1284008,"TERMINAL",0,0,"2025-07-16 14:30:24.393498: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1057,1284063,"TERMINAL",0,0,"2025-07-16 14:30:24.440972: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1058,1284158,"TERMINAL",0,0,"2025-07-16 14:30:24.524450: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1059,1284515,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +1060,1284934,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +1061,1284981,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +1062,1285435,"TERMINAL",0,0,"2025-07-16 14:30:25.720006: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1063,1285678,"TERMINAL",0,0,"2025-07-16 14:30:26.068169: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1064,1286256,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +1065,1286549,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +1066,1287738,"genie.py",0,0,"",python,tab +1067,1292259,"genie.py",3408,0,"",python,selection_mouse +1068,1293013,"genie.py",3353,0,"",python,selection_mouse +1069,1293996,"genie.py",3409,0,"",python,selection_mouse +1070,1296433,"genie.py",3353,0,"",python,selection_mouse +1071,1297449,"genie.py",1997,0,"",python,selection_mouse +1072,1297715,"models/dynamics.py",0,0,"",python,tab +1073,1301836,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +1074,1301971,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n",,terminal_output +1075,1313467,"models/dynamics.py",0,0,"",python,tab +1076,1313468,"models/dynamics.py",2916,0,"",python,selection_mouse +1077,1315188,"genie.py",0,0,"",python,tab +1078,1321025,"utils/nn.py",0,0,"",python,tab +1079,1328589,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1080,1328828,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1081,1329058,"TERMINAL",0,0,"[?25lq[?25h",,terminal_output +1082,1329241,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +1083,1329488,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +1084,1329594,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1085,1329656,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +1086,1329828,"TERMINAL",0,0,"\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n",,terminal_output +1087,1330939,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +1088,1331152,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1089,1331393,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +1090,1331531,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1091,1331715,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1092,1331776,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +1093,1332151,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +1094,1332458,"TERMINAL",0,0,"[?25lm[?25h",,terminal_output +1095,1332520,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1096,1332583,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1097,1332644,"TERMINAL",0,0,"[?25lk[?25h",,terminal_output +1098,1332940,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +1099,1333045,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1100,1333191,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +1101,1333253,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1102,1333402,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +1103,1333469,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1104,1333701,"TERMINAL",0,0,"\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n",,terminal_output +1105,1336427,"utils/nn.py",0,0,"",python,tab +1106,1336428,"utils/nn.py",1478,0,"",python,selection_mouse +1107,1336914,"utils/nn.py",1421,0,"",python,selection_mouse +1108,1337285,"utils/nn.py",1418,4,"mask",python,selection_mouse +1109,1337534,"utils/nn.py",1418,5,"mask:",python,selection_mouse +1110,1337535,"utils/nn.py",1418,6,"mask: ",python,selection_mouse +1111,1337535,"utils/nn.py",1418,9,"mask: (1,",python,selection_mouse +1112,1337535,"utils/nn.py",1418,61,"mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril",python,selection_mouse +1113,1337566,"utils/nn.py",1418,62,"mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(",python,selection_mouse +1114,1337595,"utils/nn.py",1418,65,"mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp",python,selection_mouse +1115,1337614,"utils/nn.py",1418,70,"mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones",python,selection_mouse +1116,1337672,"utils/nn.py",1418,71,"mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones(",python,selection_mouse +1117,1337699,"utils/nn.py",1418,72,"mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((",python,selection_mouse +1118,1337719,"utils/nn.py",1418,79,"mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len",python,selection_mouse +1119,1337877,"utils/nn.py",1418,30,"mask: (1, 1, seq_len, seq_len)",python,selection_mouse +1120,1338512,"utils/nn.py",1448,0,"",python,selection_mouse +1121,1338555,"utils/nn.py",1447,0,"",python,selection_command +1122,1338881,"utils/nn.py",1448,0,"",python,selection_mouse +1123,1338896,"utils/nn.py",1447,0,"",python,selection_command +1124,1339055,"utils/nn.py",1447,1,")",python,selection_mouse +1125,1339110,"utils/nn.py",1448,0,"",python,selection_command +1126,1339111,"utils/nn.py",1440,8,"seq_len)",python,selection_mouse +1127,1339111,"utils/nn.py",1400,48,"\n # Causal mask: (1, 1, seq_len, seq_len)",python,selection_mouse +1128,1339288,"utils/nn.py",1428,20,"1, seq_len, seq_len)",python,selection_mouse +1129,1339361,"utils/nn.py",1427,21," 1, seq_len, seq_len)",python,selection_mouse +1130,1339362,"utils/nn.py",1426,22,", 1, seq_len, seq_len)",python,selection_mouse +1131,1339397,"utils/nn.py",1425,23,"1, 1, seq_len, seq_len)",python,selection_mouse +1132,1339452,"utils/nn.py",1424,24,"(1, 1, seq_len, seq_len)",python,selection_mouse +1133,1340019,"utils/nn.py",1424,0,"",python,selection_mouse +1134,1340519,"utils/nn.py",1424,1,"(",python,selection_mouse +1135,1340520,"utils/nn.py",1424,6,"(1, 1,",python,selection_mouse +1136,1340520,"utils/nn.py",1424,10,"(1, 1, seq",python,selection_mouse +1137,1340520,"utils/nn.py",1424,14,"(1, 1, seq_len",python,selection_mouse +1138,1340521,"utils/nn.py",1424,17,"(1, 1, seq_len, s",python,selection_mouse +1139,1340549,"utils/nn.py",1424,19,"(1, 1, seq_len, seq",python,selection_mouse +1140,1340622,"utils/nn.py",1424,22,"(1, 1, seq_len, seq_le",python,selection_mouse +1141,1340622,"utils/nn.py",1424,24,"(1, 1, seq_len, seq_len)",python,selection_mouse +1142,1341000,"utils/nn.py",1448,0,"",python,selection_mouse +1143,1341017,"utils/nn.py",1447,0,"",python,selection_command +1144,1341530,"utils/nn.py",1448,0,"",python,selection_mouse +1145,1341537,"utils/nn.py",1447,0,"",python,selection_command +1146,1341780,"utils/nn.py",1447,1,")",python,selection_mouse +1147,1341780,"utils/nn.py",1400,47,"\n # Causal mask: (1, 1, seq_len, seq_len",python,selection_mouse +1148,1341785,"utils/nn.py",1448,0,"",python,selection_command +1149,1341866,"utils/nn.py",1400,48,"\n # Causal mask: (1, 1, seq_len, seq_len)",python,selection_mouse +1150,1341867,"utils/nn.py",1371,77,"\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)",python,selection_mouse +1151,1341924,"utils/nn.py",1400,48,"\n # Causal mask: (1, 1, seq_len, seq_len)",python,selection_mouse +1152,1341978,"utils/nn.py",1399,49,"]\n # Causal mask: (1, 1, seq_len, seq_len)",python,selection_mouse +1153,1342035,"utils/nn.py",1398,50,"1]\n # Causal mask: (1, 1, seq_len, seq_len)",python,selection_mouse +1154,1342093,"utils/nn.py",1397,51,"[1]\n # Causal mask: (1, 1, seq_len, seq_len)",python,selection_mouse +1155,1342141,"utils/nn.py",1425,23,"1, 1, seq_len, seq_len)",python,selection_mouse +1156,1342208,"utils/nn.py",1424,24,"(1, 1, seq_len, seq_len)",python,selection_mouse +1157,1342228,"utils/nn.py",1423,25," (1, 1, seq_len, seq_len)",python,selection_mouse +1158,1342650,"utils/nn.py",1423,0,"",python,selection_mouse +1159,1344477,"TERMINAL",0,0,"[?25lz[?25h",,terminal_output +1160,1344826,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +1161,1344998,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1162,1345063,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +1163,1345167,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1164,1345375,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1165,1345929,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +1166,1345991,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1167,1346242,"TERMINAL",0,0,"\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n",,terminal_output +1168,1358227,"genie.py",0,0,"",python,tab +1169,1361413,"genie.py",3400,0,"",python,selection_mouse +1170,1362034,"genie.py",3436,0,"",python,selection_mouse +1171,1363250,"genie.py",3411,0,"",python,selection_mouse +1172,1363794,"genie.py",3407,0,"",python,selection_mouse +1173,1363962,"genie.py",3401,12,"dyna_outputs",python,selection_mouse +1174,1364525,"genie.py",3353,0,"",python,selection_mouse +1175,1365679,"models/dynamics.py",0,0,"",python,tab +1176,1367759,"models/dynamics.py",2765,0,"",python,selection_mouse +1177,1368270,"models/dynamics.py",2826,0,"",python,selection_mouse +1178,1369042,"models/dynamics.py",2877,0,"\n ",python,content +1179,1380860,"models/dynamics.py",2886,0,"o",python,content +1180,1380862,"models/dynamics.py",2887,0,"",python,selection_keyboard +1181,1380977,"models/dynamics.py",2887,0,"r",python,content +1182,1380979,"models/dynamics.py",2888,0,"",python,selection_keyboard +1183,1381106,"models/dynamics.py",2888,0,"i",python,content +1184,1381108,"models/dynamics.py",2889,0,"",python,selection_keyboard +1185,1381372,"models/dynamics.py",2889,0,"g",python,content +1186,1381373,"models/dynamics.py",2890,0,"",python,selection_keyboard +1187,1381480,"models/dynamics.py",2890,0,"i",python,content +1188,1381484,"models/dynamics.py",2891,0,"",python,selection_keyboard +1189,1382042,"models/dynamics.py",2891,0,"n",python,content +1190,1382043,"models/dynamics.py",2892,0,"",python,selection_keyboard +1191,1382235,"models/dynamics.py",2892,0,"a",python,content +1192,1382236,"models/dynamics.py",2893,0,"",python,selection_keyboard +1193,1382302,"models/dynamics.py",2893,0,"l",python,content +1194,1382304,"models/dynamics.py",2894,0,"",python,selection_keyboard +1195,1382615,"models/dynamics.py",2894,0,"_",python,content +1196,1382616,"models/dynamics.py",2895,0,"",python,selection_keyboard +1197,1382831,"models/dynamics.py",2895,0,"s",python,content +1198,1382832,"models/dynamics.py",2896,0,"",python,selection_keyboard +1199,1382936,"models/dynamics.py",2896,0,"h",python,content +1200,1382937,"models/dynamics.py",2897,0,"",python,selection_keyboard +1201,1383151,"models/dynamics.py",2897,0,"a",python,content +1202,1383152,"models/dynamics.py",2898,0,"",python,selection_keyboard +1203,1383341,"models/dynamics.py",2898,0,"p",python,content +1204,1383342,"models/dynamics.py",2899,0,"",python,selection_keyboard +1205,1383435,"models/dynamics.py",2899,0,"e",python,content +1206,1383436,"models/dynamics.py",2900,0,"",python,selection_keyboard +1207,1383641,"models/dynamics.py",2900,0," ",python,content +1208,1383643,"models/dynamics.py",2901,0,"",python,selection_keyboard +1209,1383826,"models/dynamics.py",2901,0,"=",python,content +1210,1383827,"models/dynamics.py",2902,0,"",python,selection_keyboard +1211,1383918,"models/dynamics.py",2902,0," ",python,content +1212,1383919,"models/dynamics.py",2903,0,"",python,selection_keyboard +1213,1384194,"models/dynamics.py",2903,0,"v",python,content +1214,1384194,"models/dynamics.py",2904,0,"",python,selection_keyboard +1215,1384321,"models/dynamics.py",2904,0,"i",python,content +1216,1384322,"models/dynamics.py",2905,0,"",python,selection_keyboard +1217,1384503,"models/dynamics.py",2905,0,"d",python,content +1218,1384504,"models/dynamics.py",2906,0,"",python,selection_keyboard +1219,1384761,"models/dynamics.py",2906,0,"_",python,content +1220,1384762,"models/dynamics.py",2907,0,"",python,selection_keyboard +1221,1385138,"models/dynamics.py",2903,4,"vid_embed",python,content +1222,1385705,"models/dynamics.py",2912,0,"ö",python,content +1223,1385706,"models/dynamics.py",2913,0,"",python,selection_keyboard +1224,1385881,"models/dynamics.py",2913,0,"s",python,content +1225,1385882,"models/dynamics.py",2914,0,"",python,selection_keyboard +1226,1385977,"models/dynamics.py",2914,0,"j",python,content +1227,1385978,"models/dynamics.py",2915,0,"",python,selection_keyboard +1228,1386259,"models/dynamics.py",2914,1,"",python,content +1229,1386380,"models/dynamics.py",2913,1,"",python,content +1230,1386508,"models/dynamics.py",2912,1,"",python,content +1231,1387056,"models/dynamics.py",2912,0,".",python,content +1232,1387057,"models/dynamics.py",2913,0,"",python,selection_keyboard +1233,1387250,"models/dynamics.py",2913,0,"s",python,content +1234,1387251,"models/dynamics.py",2914,0,"",python,selection_keyboard +1235,1387323,"models/dynamics.py",2914,0,"h",python,content +1236,1387324,"models/dynamics.py",2915,0,"",python,selection_keyboard +1237,1387445,"models/dynamics.py",2915,0,"a",python,content +1238,1387446,"models/dynamics.py",2916,0,"",python,selection_keyboard +1239,1387538,"models/dynamics.py",2916,0,"p",python,content +1240,1387539,"models/dynamics.py",2917,0,"",python,selection_keyboard +1241,1387616,"models/dynamics.py",2917,0,"e",python,content +1242,1387617,"models/dynamics.py",2918,0,"",python,selection_keyboard +1243,1387940,"models/dynamics.py",2913,5,"shape",python,content +1244,1388517,"models/dynamics.py",2918,0,"\n ",python,content +1245,1388959,"models/dynamics.py",2927,0,"v",python,content +1246,1388960,"models/dynamics.py",2928,0,"",python,selection_keyboard +1247,1389039,"models/dynamics.py",2928,0,"o",python,content +1248,1389041,"models/dynamics.py",2929,0,"",python,selection_keyboard +1249,1389466,"models/dynamics.py",2928,1,"",python,content +1250,1389689,"models/dynamics.py",2928,0,"i",python,content +1251,1389690,"models/dynamics.py",2929,0,"",python,selection_keyboard +1252,1389789,"models/dynamics.py",2929,0,"d",python,content +1253,1389790,"models/dynamics.py",2930,0,"",python,selection_keyboard +1254,1390424,"models/dynamics.py",2927,3,"vid_embed",python,content +1255,1390742,"models/dynamics.py",2936,0," ",python,content +1256,1390743,"models/dynamics.py",2937,0,"",python,selection_keyboard +1257,1390889,"models/dynamics.py",2937,0,"=",python,content +1258,1390889,"models/dynamics.py",2938,0,"",python,selection_keyboard +1259,1391034,"models/dynamics.py",2938,0," ",python,content +1260,1391034,"models/dynamics.py",2939,0,"",python,selection_keyboard +1261,1391295,"models/dynamics.py",2939,0,"v",python,content +1262,1391296,"models/dynamics.py",2940,0,"",python,selection_keyboard +1263,1391435,"models/dynamics.py",2940,0,"i",python,content +1264,1391436,"models/dynamics.py",2941,0,"",python,selection_keyboard +1265,1391584,"models/dynamics.py",2941,0,"d",python,content +1266,1391585,"models/dynamics.py",2942,0,"",python,selection_keyboard +1267,1391918,"models/dynamics.py",2939,3,"vid_embed",python,content +1268,1392817,"models/dynamics.py",2948,0,".",python,content +1269,1392818,"models/dynamics.py",2949,0,"",python,selection_keyboard +1270,1393557,"models/dynamics.py",2949,0,"r",python,content +1271,1393558,"models/dynamics.py",2950,0,"",python,selection_keyboard +1272,1393738,"models/dynamics.py",2950,0,"e",python,content +1273,1393739,"models/dynamics.py",2951,0,"",python,selection_keyboard +1274,1393963,"models/dynamics.py",2951,0,"s",python,content +1275,1393964,"models/dynamics.py",2952,0,"",python,selection_keyboard +1276,1394055,"models/dynamics.py",2952,0,"h",python,content +1277,1394056,"models/dynamics.py",2953,0,"",python,selection_keyboard +1278,1394222,"models/dynamics.py",2953,0,"a",python,content +1279,1394223,"models/dynamics.py",2954,0,"",python,selection_keyboard +1280,1394391,"models/dynamics.py",2954,0,"p",python,content +1281,1394392,"models/dynamics.py",2955,0,"",python,selection_keyboard +1282,1394524,"models/dynamics.py",2955,0,"e",python,content +1283,1394525,"models/dynamics.py",2956,0,"",python,selection_keyboard +1284,1394690,"models/dynamics.py",2949,7,"reshape",python,content +1285,1426820,"models/dynamics.py",2956,0,"()",python,content +1286,1426821,"models/dynamics.py",2957,0,"",python,selection_keyboard +1287,1441694,"models/dynamics.py",2957,0,"1",python,content +1288,1441695,"models/dynamics.py",2958,0,"",python,selection_keyboard +1289,1441753,"models/dynamics.py",2958,0,"2",python,content +1290,1441754,"models/dynamics.py",2959,0,"",python,selection_keyboard +1291,1442051,"models/dynamics.py",2959,0,",",python,content +1292,1442052,"models/dynamics.py",2960,0,"",python,selection_keyboard +1293,1442177,"models/dynamics.py",2960,0," ",python,content +1294,1442177,"models/dynamics.py",2961,0,"",python,selection_keyboard +1295,1442438,"models/dynamics.py",2961,0,"1",python,content +1296,1442439,"models/dynamics.py",2962,0,"",python,selection_keyboard +1297,1442754,"models/dynamics.py",2962,0,"6",python,content +1298,1442755,"models/dynamics.py",2963,0,"",python,selection_keyboard +1299,1443658,"models/dynamics.py",2963,0,",",python,content +1300,1443659,"models/dynamics.py",2964,0,"",python,selection_keyboard +1301,1443717,"models/dynamics.py",2964,0," ",python,content +1302,1443718,"models/dynamics.py",2965,0,"",python,selection_keyboard +1303,1444547,"models/dynamics.py",2965,0,"-",python,content +1304,1444548,"models/dynamics.py",2966,0,"",python,selection_keyboard +1305,1444603,"models/dynamics.py",2966,0,"1",python,content +1306,1444603,"models/dynamics.py",2967,0,"",python,selection_keyboard +1307,1446153,"models/dynamics.py",2954,0,"",python,selection_mouse +1308,1446899,"models/dynamics.py",2958,0,"",python,selection_mouse +1309,1447880,"models/dynamics.py",2958,1,"2",python,selection_mouse +1310,1447955,"models/dynamics.py",2958,2,"2,",python,selection_mouse +1311,1447955,"models/dynamics.py",2958,3,"2, ",python,selection_mouse +1312,1447955,"models/dynamics.py",2958,4,"2, 1",python,selection_mouse +1313,1447980,"models/dynamics.py",2958,5,"2, 16",python,selection_mouse +1314,1448035,"models/dynamics.py",2958,6,"2, 16,",python,selection_mouse +1315,1448043,"models/dynamics.py",2958,7,"2, 16, ",python,selection_mouse +1316,1448095,"models/dynamics.py",2958,8,"2, 16, -",python,selection_mouse +1317,1448152,"models/dynamics.py",2958,9,"2, 16, -1",python,selection_mouse +1318,1448208,"models/dynamics.py",2958,10,"2, 16, -1)",python,selection_mouse +1319,1448545,"models/dynamics.py",2968,0,"",python,selection_mouse +1320,1460497,"models/dynamics.py",2952,0,"",python,selection_mouse +1321,1467152,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3350302.3 tasks 0-7: running\r\n",,terminal_output +1322,1469263,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3350302.3 tasks 0-7: running\r\n",,terminal_output +1323,1469474,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3350302.3\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\nsrun: forcing job termination\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3350302.3 ON hkn0509 CANCELLED AT 2025-07-16T14:33:29 ***\r\n",,terminal_output +1324,1469628,"TERMINAL",0,0,"(jdb) (jdb) ",,terminal_output +1325,1469706,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3350302.3\r\n(jdb) srun: job abort in progress\r\n(jdb) (jdb) (jdb) (jdb) (jdb) ",,terminal_output +1326,1470594,"TERMINAL",0,0,"]0;tum_cte0515@hkn0509:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0509 jafar]$ ",,terminal_output +1327,1471659,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +1328,1472310,"TERMINAL",0,0,"\r\n[?2004l\r\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\n\r\njob_name=""debug""\r\nslurm_job_id=""debug-mihir""\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4\r\n\r\nenv | grep SLURM\r\n\r\nXLA_FLAGS=--xla_gpu_autotune_level=0 srun python train_dynamics.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --log_checkpoint_interval=1000 \\r\n --name=dynamics-debug-run-$slurm_job_id \\r\n --tags dynamics debug \\r\n --entity instant-uv \\r\n --project jafar \\r\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\r\n --data_dir $array_records_dir\r\n ",,terminal_output +1329,1472424,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=692048\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\r\nSLURMD_NODENAME=hkn0509\r\nSLURM_JOB_START_TIME=1752667757\r\nSLURM_STEP_NODELIST=hkn0509\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752703757\r\nSLURM_PMI2_SRUN_PORT=40311\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3350302\r\nSLURM_PTY_PORT=34583\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.198\r\nSLURM_PTY_WIN_ROW=43\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0509\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.198\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=300\r\nSLURM_NODELIST=hkn[0509,0511]\r\nSLURM_SRUN_COMM_PORT=35053\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1990.localdomain\r\nSLURM_JOB_ID=3350302\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0509\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_STEP_LAUNCHER_PORT=35053\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0509,0511]\r\n",,terminal_output +1330,1472554,"TERMINAL",0,0,"GpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +1331,1474493,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +1332,1474551,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +1333,1474614,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +1334,1478289,"TERMINAL",0,0,"2025-07-16 14:33:38.584675: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:38.585308: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:38.606283: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:38.606536: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:38.606615: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:38.606614: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:38.612271: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:38.613650: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1335,1493113,"TERMINAL",0,0,"2025-07-16 14:33:53.406000: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:53.474327: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:53.480254: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:53.484968: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:53.487862: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:53.491163: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:33:53.492678: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1336,1493219,"TERMINAL",0,0,"2025-07-16 14:33:53.600193: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1337,1494656,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +1338,1494760,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +1339,1515105,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +1340,1515354,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n",,terminal_output +1341,1518209,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +1342,1518315,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +1343,1518444,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1344,1518507,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1345,1518758,"TERMINAL",0,0,"[?25la[?25h[?25ll[?25h",,terminal_output +1346,1519987,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1347,1520047,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +1348,1520315,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1349,1520495,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1350,1520555,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +1351,1520859,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +1352,1521144,"TERMINAL",0,0,"[?25lm[?25h",,terminal_output +1353,1521197,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1354,1521254,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1355,1521315,"TERMINAL",0,0,"[?25lk[?25h",,terminal_output +1356,1521552,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +1357,1521741,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1358,1522285,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +1359,1522421,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1360,1522538,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +1361,1522599,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1362,1522719,"TERMINAL",0,0,"\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n(jdb) (16, 16)\r\n",,terminal_output +1363,1524268,"models/dynamics.py",0,0,"",python,tab +1364,1524269,"models/dynamics.py",3026,0,"",python,selection_mouse +1365,1524820,"models/dynamics.py",2895,0,"",python,selection_mouse +1366,1528276,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1367,1528493,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1368,1528659,"TERMINAL",0,0,"[?25lq[?25h",,terminal_output +1369,1528819,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +1370,1529094,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +1371,1529226,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1372,1529335,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +1373,1529512,"TERMINAL",0,0,"\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n(jdb) 16\r\n",,terminal_output +1374,1530280,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1375,1530809,"TERMINAL",0,0,"[?25lz[?25h",,terminal_output +1376,1530983,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +1377,1531127,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1378,1531309,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1379,1531899,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +1380,1532015,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1381,1532077,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +1382,1532204,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1383,1532369,"TERMINAL",0,0,"\r\n(jdb) (12, 16, 512)\r\n(jdb) (12, 16, 512)\r\n(jdb) (12, 16, 512)\r\n(jdb) (12, 16, 512)\r\n(jdb) (12, 16, 512)\r\n(jdb) (12, 16, 512)\r\n(jdb) (12, 16, 512)\r\n(jdb) (12, 16, 512)\r\n",,terminal_output +1384,1550116,"models/dynamics.py",0,0,"",python,tab +1385,1550117,"models/dynamics.py",2941,0,"",python,selection_mouse +1386,1550659,"models/dynamics.py",2939,0,"",python,selection_mouse +1387,1551201,"models/dynamics.py",2944,0,"",python,selection_mouse +1388,1551327,"models/dynamics.py",2939,9,"vid_embed",python,selection_mouse +1389,1554590,"models/dynamics.py",2939,0,"\n ",python,content +1390,1555176,"models/dynamics.py",2944,4,"",python,content +1391,1555511,"models/dynamics.py",2940,4,"",python,content +1392,1555835,"models/dynamics.py",2939,1,"",python,content +1393,1557582,"models/dynamics.py",2918,0,"",python,selection_mouse +1394,1557870,"models/dynamics.py",2918,0,"\n ",python,content +1395,1558321,"models/dynamics.py",2927,0,"v",python,content +1396,1558322,"models/dynamics.py",2928,0,"",python,selection_keyboard +1397,1558387,"models/dynamics.py",2928,0,"i",python,content +1398,1558389,"models/dynamics.py",2929,0,"",python,selection_keyboard +1399,1558558,"models/dynamics.py",2929,0,"d",python,content +1400,1558559,"models/dynamics.py",2930,0,"",python,selection_keyboard +1401,1559119,"models/dynamics.py",2927,3,"vid_embed",python,content +1402,1559607,"models/dynamics.py",2936,0," ",python,content +1403,1559609,"models/dynamics.py",2937,0,"",python,selection_keyboard +1404,1559772,"models/dynamics.py",2937,0,"=",python,content +1405,1559773,"models/dynamics.py",2938,0,"",python,selection_keyboard +1406,1559927,"models/dynamics.py",2938,0," ",python,content +1407,1559928,"models/dynamics.py",2939,0,"",python,selection_keyboard +1408,1560158,"models/dynamics.py",2939,0,"e",python,content +1409,1560159,"models/dynamics.py",2940,0,"",python,selection_keyboard +1410,1560461,"models/dynamics.py",2940,0,"i",python,content +1411,1560462,"models/dynamics.py",2941,0,"",python,selection_keyboard +1412,1560629,"models/dynamics.py",2941,0,"n",python,content +1413,1560630,"models/dynamics.py",2942,0,"",python,selection_keyboard +1414,1561911,"models/dynamics.py",2942,0,"o",python,content +1415,1561913,"models/dynamics.py",2943,0,"",python,selection_keyboard +1416,1561982,"models/dynamics.py",2943,0,"p",python,content +1417,1561984,"models/dynamics.py",2944,0,"",python,selection_keyboard +1418,1562088,"models/dynamics.py",2944,0,"s",python,content +1419,1562089,"models/dynamics.py",2945,0,"",python,selection_keyboard +1420,1562449,"models/dynamics.py",2945,0,".",python,content +1421,1562450,"models/dynamics.py",2946,0,"",python,selection_keyboard +1422,1562665,"models/dynamics.py",2946,0,"s",python,content +1423,1562666,"models/dynamics.py",2947,0,"",python,selection_keyboard +1424,1562974,"models/dynamics.py",2946,1,"",python,content +1425,1563080,"models/dynamics.py",2946,0,"r",python,content +1426,1563082,"models/dynamics.py",2947,0,"",python,selection_keyboard +1427,1563260,"models/dynamics.py",2947,0,"e",python,content +1428,1563260,"models/dynamics.py",2948,0,"",python,selection_keyboard +1429,1563464,"models/dynamics.py",2948,0,"a",python,content +1430,1563465,"models/dynamics.py",2949,0,"",python,selection_keyboard +1431,1563860,"models/dynamics.py",2949,0,"r",python,content +1432,1563861,"models/dynamics.py",2950,0,"",python,selection_keyboard +1433,1564038,"models/dynamics.py",2950,0,"r",python,content +1434,1564038,"models/dynamics.py",2951,0,"",python,selection_keyboard +1435,1564203,"models/dynamics.py",2951,0,"a",python,content +1436,1564204,"models/dynamics.py",2952,0,"",python,selection_keyboard +1437,1564534,"models/dynamics.py",2952,0,"n",python,content +1438,1564535,"models/dynamics.py",2953,0,"",python,selection_keyboard +1439,1564629,"models/dynamics.py",2953,0,"g",python,content +1440,1564631,"models/dynamics.py",2954,0,"",python,selection_keyboard +1441,1564710,"models/dynamics.py",2954,0,"e",python,content +1442,1564711,"models/dynamics.py",2955,0,"",python,selection_keyboard +1443,1565512,"models/dynamics.py",2955,0,"()",python,content +1444,1565513,"models/dynamics.py",2956,0,"",python,selection_keyboard +1445,1567247,"models/dynamics.py",64,0,"",python,selection_mouse +1446,1567623,"models/dynamics.py",64,0,"\n",python,content +1447,1568048,"models/dynamics.py",65,0,"i",python,content +1448,1568049,"models/dynamics.py",66,0,"",python,selection_keyboard +1449,1568401,"models/dynamics.py",66,0,"m",python,content +1450,1568402,"models/dynamics.py",67,0,"",python,selection_keyboard +1451,1568886,"models/dynamics.py",65,2,"import",python,content +1452,1569390,"models/dynamics.py",71,0," ",python,content +1453,1569392,"models/dynamics.py",72,0,"",python,selection_keyboard +1454,1569513,"models/dynamics.py",72,0,"e",python,content +1455,1569515,"models/dynamics.py",73,0,"",python,selection_keyboard +1456,1570011,"models/dynamics.py",73,0,"i",python,content +1457,1570013,"models/dynamics.py",74,0,"",python,selection_keyboard +1458,1570024,"models/dynamics.py",74,0,"n",python,content +1459,1570025,"models/dynamics.py",75,0,"",python,selection_keyboard +1460,1570535,"models/dynamics.py",75,0,"o",python,content +1461,1570536,"models/dynamics.py",76,0,"",python,selection_keyboard +1462,1570681,"models/dynamics.py",76,0,"p",python,content +1463,1570682,"models/dynamics.py",77,0,"",python,selection_keyboard +1464,1570817,"models/dynamics.py",77,0,"s",python,content +1465,1570818,"models/dynamics.py",78,0,"",python,selection_keyboard +1466,1571126,"models/dynamics.py",77,0,"",python,selection_command +1467,1571863,"models/dynamics.py",65,14,"",python,content +1468,1572028,"models/dynamics.py",88,0,"\nimport einops",python,content +1469,1572086,"models/dynamics.py",89,0,"",python,selection_command +1470,1573365,"genie.py",0,0,"",python,tab +1471,1573709,"genie.py",2499,0,"",python,selection_mouse +1472,1577792,"utils/nn.py",0,0,"",python,tab +1473,1578289,"utils/nn.py",1200,0,"",python,selection_mouse +1474,1586013,"genie.py",0,0,"",python,tab +1475,1589204,"genie.py",8091,0,"",python,selection_mouse +1476,1590184,"genie.py",7920,0,"",python,selection_mouse +1477,1590764,"genie.py",7925,0,"",python,selection_mouse +1478,1591699,"genie.py",2499,0,"",python,selection_command +1479,1615997,"utils/nn.py",0,0,"",python,tab +1480,1619657,"models/dynamics.py",0,0,"",python,tab +1481,1620145,"models/dynamics.py",889,0,"",python,selection_mouse +1482,1620163,"models/dynamics.py",888,0,"",python,selection_command +1483,1632052,"sample.py",0,0,"from dataclasses import dataclass\nimport time\nimport os\n\nimport dm_pix as pix\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nimport numpy as np\nfrom orbax.checkpoint import PyTreeCheckpointer\nfrom PIL import Image, ImageDraw\nimport tyro\n\nfrom genie import Genie\nfrom utils.dataloader import get_dataloader\n\n\n@dataclass\nclass Args:\n # Experiment\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data/coinrun_episodes""\n checkpoint: str = """"\n # Sampling\n batch_size: int = 1\n maskgit_steps: int = 25\n temperature: float = 1.0\n sample_argmax: bool = True\n start_frame: int = 0\n # Tokenizer checkpoint\n tokenizer_dim: int = 512\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 8\n tokenizer_num_heads: int = 8\n # LAM checkpoint\n lam_dim: int = 512\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 8\n lam_num_heads: int = 8\n # Dynamics checkpoint\n dyna_dim: int = 512\n dyna_num_blocks: int = 12\n dyna_num_heads: int = 8\n param_dtype: jnp.dtype = jnp.float32\n dtype: jnp.dtype = jnp.bfloat16\n\n\nargs = tyro.cli(Args)\nrng = jax.random.PRNGKey(args.seed)\n\n# --- Load Genie checkpoint ---\ngenie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=False,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n)\nrng, _rng = jax.random.split(rng)\nimage_shape = (args.image_height, args.image_width, args.image_channels)\ndummy_inputs = dict(\n videos=jnp.zeros((args.batch_size, args.seq_len, *image_shape), dtype=jnp.float32),\n mask_rng=_rng,\n)\nrng, _rng = jax.random.split(rng)\nparams = genie.init(_rng, dummy_inputs)\nckpt = PyTreeCheckpointer().restore(args.checkpoint)[""model""][""params""][""params""]\nparams[""params""].update(ckpt)\n\n\ndef _sampling_wrapper(module, batch):\n return module.sample(batch, args.seq_len, args.maskgit_steps, args.temperature, args.sample_argmax)\n\n# --- Define autoregressive sampling loop ---\ndef _autoreg_sample(rng, video_batch, action_batch):\n vid = video_batch[:, : args.start_frame + 1]\n sampling_fn = jax.jit(nn.apply(_sampling_wrapper, genie)) \n rng, _rng = jax.random.split(rng)\n batch = dict(videos=vid, latent_actions=action_batch, rng=_rng)\n generated_vid = sampling_fn(\n params,\n batch\n )\n return generated_vid\n\n# --- Get video + latent actions ---\narray_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n]\ndataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n args.batch_size,\n args.image_height,\n args.image_width,\n args.image_channels,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n)\nvideo_batch = next(iter(dataloader))\n# Get latent actions for all videos in the batch\nbatch = dict(videos=video_batch)\naction_batch = genie.apply(params, batch, False, method=Genie.vq_encode)\naction_batch = action_batch.reshape(video_batch.shape[0], args.seq_len - 1, 1)\n\n# --- Sample + evaluate video ---\nvid = _autoreg_sample(rng, video_batch, action_batch)\ngt = video_batch[:, : vid.shape[1]].clip(0, 1).reshape(-1, *video_batch.shape[2:])\nrecon = vid.clip(0, 1).reshape(-1, *vid.shape[2:])\nssim = pix.ssim(gt[:, args.start_frame + 1 :], recon[:, args.start_frame + 1 :]).mean()\nprint(f""SSIM: {ssim}"")\n\n# --- Construct video ---\ntrue_videos = (video_batch * 255).astype(np.uint8)\npred_videos = (vid * 255).astype(np.uint8)\nvideo_comparison = np.zeros((2, *vid.shape), dtype=np.uint8)\nvideo_comparison[0] = true_videos[:, :args.seq_len]\nvideo_comparison[1] = pred_videos\nframes = einops.rearrange(video_comparison, ""n b t h w c -> t (b h) (n w) c"")\n\n# --- Save video --- \nimgs = [Image.fromarray(img) for img in frames]\n# Write actions on each frame, on each row (i.e., for each video in the batch, on the GT row)\nfor t, img in enumerate(imgs[1:]):\n d = ImageDraw.Draw(img)\n for row in range(action_batch.shape[0]):\n action = action_batch[row, t, 0]\n y_offset = row * video_batch.shape[2] + 2\n d.text((2, y_offset), f""{action}"", fill=255)\nimgs[0].save(\n f""generation_{time.time()}.gif"",\n save_all=True,\n append_images=imgs[1:],\n duration=250,\n loop=0,\n)\n",python,tab +1484,1636581,"sample.py",4478,0,"",python,selection_mouse +1485,1636728,"sample.py",4478,1,"""",python,selection_mouse +1486,1636728,"sample.py",4478,3,"""n ",python,selection_mouse +1487,1636803,"sample.py",4478,6,"""n b t",python,selection_mouse +1488,1636806,"sample.py",4478,10,"""n b t h w",python,selection_mouse +1489,1636806,"sample.py",4433,45,"\nframes = einops.rearrange(video_comparison, ",python,selection_mouse +1490,1637035,"sample.py",4478,29,"""n b t h w c -> t (b h) (n w)",python,selection_mouse +1491,1637036,"sample.py",4478,31,"""n b t h w c -> t (b h) (n w) c",python,selection_mouse +1492,1637087,"sample.py",4478,32,"""n b t h w c -> t (b h) (n w) c""",python,selection_mouse +1493,1637087,"sample.py",4478,33,"""n b t h w c -> t (b h) (n w) c"")",python,selection_mouse +1494,1637400,"sample.py",4478,32,"""n b t h w c -> t (b h) (n w) c""",python,selection_mouse +1495,1640834,"models/dynamics.py",0,0,"",python,tab +1496,1642547,"models/dynamics.py",182,0,"",python,selection_mouse +1497,1644722,"models/dynamics.py",2953,0,"",python,selection_command +1498,1645279,"models/dynamics.py",2954,0,"",python,selection_command +1499,1645790,"models/dynamics.py",2955,0,"",python,selection_command +1500,1645809,"models/dynamics.py",2956,0,"",python,selection_command +1501,1645841,"models/dynamics.py",2957,0,"",python,selection_command +1502,1645898,"models/dynamics.py",2958,0,"",python,selection_command +1503,1645898,"models/dynamics.py",2959,0,"",python,selection_command +1504,1645951,"models/dynamics.py",2960,0,"",python,selection_command +1505,1645959,"models/dynamics.py",2961,0,"",python,selection_command +1506,1645991,"models/dynamics.py",2962,0,"",python,selection_command +1507,1646018,"models/dynamics.py",2963,0,"",python,selection_command +1508,1646055,"models/dynamics.py",2964,0,"",python,selection_command +1509,1646082,"models/dynamics.py",2965,0,"",python,selection_command +1510,1646110,"models/dynamics.py",2966,0,"",python,selection_command +1511,1646131,"models/dynamics.py",2967,0,"",python,selection_command +1512,1646168,"models/dynamics.py",2968,0,"",python,selection_command +1513,1646207,"models/dynamics.py",2969,0,"",python,selection_command +1514,1646242,"models/dynamics.py",2970,0,"",python,selection_command +1515,1650407,"models/dynamics.py",2970,0,"""n b t h w c -> t (b h) (n w) c""",python,content +1516,1654200,"models/dynamics.py",2970,0,"",python,selection_command +1517,1654949,"models/dynamics.py",2970,0,"v",python,content +1518,1654950,"models/dynamics.py",2971,0,"",python,selection_keyboard +1519,1655052,"models/dynamics.py",2971,0,"i",python,content +1520,1655054,"models/dynamics.py",2972,0,"",python,selection_keyboard +1521,1655165,"models/dynamics.py",2972,0,"d",python,content +1522,1655166,"models/dynamics.py",2973,0,"",python,selection_keyboard +1523,1655542,"models/dynamics.py",2973,0,"_",python,content +1524,1655543,"models/dynamics.py",2974,0,"",python,selection_keyboard +1525,1656119,"models/dynamics.py",2970,4,"vid_embed",python,content +1526,1657488,"models/dynamics.py",2979,0,",",python,content +1527,1657489,"models/dynamics.py",2980,0,"",python,selection_keyboard +1528,1657555,"models/dynamics.py",2980,0," ",python,content +1529,1657557,"models/dynamics.py",2981,0,"",python,selection_keyboard +1530,1658843,"models/dynamics.py",2950,0,"",python,selection_command +1531,1659199,"models/dynamics.py",2950,0,"_",python,content +1532,1659201,"models/dynamics.py",2951,0,"",python,selection_keyboard +1533,1659384,"models/dynamics.py",2951,0,"f",python,content +1534,1659385,"models/dynamics.py",2952,0,"",python,selection_keyboard +1535,1659478,"models/dynamics.py",2952,0,"l",python,content +1536,1659479,"models/dynamics.py",2953,0,"",python,selection_keyboard +1537,1659614,"models/dynamics.py",2953,0,"a",python,content +1538,1659615,"models/dynamics.py",2954,0,"",python,selection_keyboard +1539,1659826,"models/dynamics.py",2954,0,"t",python,content +1540,1659828,"models/dynamics.py",2955,0,"",python,selection_keyboard +1541,1660474,"models/dynamics.py",2954,0,"",python,selection_command +1542,1661017,"models/dynamics.py",2909,0,"",python,selection_mouse +1543,1661681,"models/dynamics.py",2892,41,"",python,content +1544,1661699,"models/dynamics.py",2900,0,"",python,selection_command +1545,1661825,"models/dynamics.py",2987,0,"",python,selection_command +1546,1662470,"models/dynamics.py",2988,0,"",python,selection_command +1547,1662948,"models/dynamics.py",2989,0,"",python,selection_command +1548,1662991,"models/dynamics.py",2990,0,"",python,selection_command +1549,1663014,"models/dynamics.py",2991,0,"",python,selection_command +1550,1663042,"models/dynamics.py",2992,0,"",python,selection_command +1551,1663073,"models/dynamics.py",2993,0,"",python,selection_command +1552,1663092,"models/dynamics.py",2994,0,"",python,selection_command +1553,1663152,"models/dynamics.py",2995,0,"",python,selection_command +1554,1663548,"models/dynamics.py",2979,50,"",python,content +1555,1663573,"models/dynamics.py",2987,0,"",python,selection_command +1556,1663915,"models/dynamics.py",2988,0,"",python,selection_command +1557,1664380,"models/dynamics.py",2989,0,"",python,selection_command +1558,1664410,"models/dynamics.py",2990,0,"",python,selection_command +1559,1664471,"models/dynamics.py",2991,0,"",python,selection_command +1560,1664474,"models/dynamics.py",2992,0,"",python,selection_command +1561,1664529,"models/dynamics.py",2993,0,"",python,selection_command +1562,1664532,"models/dynamics.py",2994,0,"",python,selection_command +1563,1664604,"models/dynamics.py",2995,0,"",python,selection_command +1564,1664605,"models/dynamics.py",2996,0,"",python,selection_command +1565,1664634,"models/dynamics.py",2997,0,"",python,selection_command +1566,1664666,"models/dynamics.py",2998,0,"",python,selection_command +1567,1664698,"models/dynamics.py",2999,0,"",python,selection_command +1568,1664742,"models/dynamics.py",3000,0,"",python,selection_command +1569,1664770,"models/dynamics.py",3001,0,"",python,selection_command +1570,1664779,"models/dynamics.py",3002,0,"",python,selection_command +1571,1664812,"models/dynamics.py",3003,0,"",python,selection_command +1572,1664843,"models/dynamics.py",3004,0,"",python,selection_command +1573,1664880,"models/dynamics.py",3005,0,"",python,selection_command +1574,1664921,"models/dynamics.py",3006,0,"",python,selection_command +1575,1664929,"models/dynamics.py",3007,0,"",python,selection_command +1576,1664960,"models/dynamics.py",3008,0,"",python,selection_command +1577,1664992,"models/dynamics.py",3009,0,"",python,selection_command +1578,1665017,"models/dynamics.py",3010,0,"",python,selection_command +1579,1665070,"models/dynamics.py",3011,0,"",python,selection_command +1580,1665080,"models/dynamics.py",3012,0,"",python,selection_command +1581,1665111,"models/dynamics.py",3013,0,"",python,selection_command +1582,1665140,"models/dynamics.py",3014,0,"",python,selection_command +1583,1665176,"models/dynamics.py",3015,0,"",python,selection_command +1584,1665288,"models/dynamics.py",3016,0,"",python,selection_command +1585,1665433,"models/dynamics.py",3017,0,"",python,selection_command +1586,1665575,"models/dynamics.py",3018,0,"",python,selection_command +1587,1665706,"models/dynamics.py",3019,0,"",python,selection_command +1588,1666596,"models/dynamics.py",3019,0,"_",python,content +1589,1666597,"models/dynamics.py",3020,0,"",python,selection_keyboard +1590,1666781,"models/dynamics.py",3020,0,"f",python,content +1591,1666782,"models/dynamics.py",3021,0,"",python,selection_keyboard +1592,1666871,"models/dynamics.py",3021,0,"l",python,content +1593,1666872,"models/dynamics.py",3022,0,"",python,selection_keyboard +1594,1667027,"models/dynamics.py",3022,0,"a",python,content +1595,1667028,"models/dynamics.py",3023,0,"",python,selection_keyboard +1596,1667182,"models/dynamics.py",3023,0,"t",python,content +1597,1667183,"models/dynamics.py",3024,0,"",python,selection_keyboard +1598,1667500,"models/dynamics.py",3023,0,"",python,selection_command +1599,1667799,"models/dynamics.py",2936,0,"",python,selection_command +1600,1668107,"models/dynamics.py",2937,0,"",python,selection_command +1601,1668629,"models/dynamics.py",2938,0,"",python,selection_command +1602,1668652,"models/dynamics.py",2939,0,"",python,selection_command +1603,1668682,"models/dynamics.py",2940,0,"",python,selection_command +1604,1668711,"models/dynamics.py",2941,0,"",python,selection_command +1605,1668741,"models/dynamics.py",2942,0,"",python,selection_command +1606,1668787,"models/dynamics.py",2943,0,"",python,selection_command +1607,1668799,"models/dynamics.py",2944,0,"",python,selection_command +1608,1668829,"models/dynamics.py",2945,0,"",python,selection_command +1609,1668946,"models/dynamics.py",2946,0,"",python,selection_command +1610,1671372,"models/dynamics.py",2947,0,"",python,selection_command +1611,1671742,"models/dynamics.py",2946,0,"",python,selection_command +1612,1672744,"models/dynamics.py",2946,1,"",python,content +1613,1672930,"models/dynamics.py",2946,1,"",python,content +1614,1673803,"models/dynamics.py",2947,0,"",python,selection_command +1615,1673990,"models/dynamics.py",2948,0,"",python,selection_command +1616,1674307,"models/dynamics.py",2949,0,"",python,selection_command +1617,1674507,"models/dynamics.py",2950,0,"",python,selection_command +1618,1674664,"models/dynamics.py",2951,0,"",python,selection_command +1619,1674860,"models/dynamics.py",2952,0,"",python,selection_command +1620,1675028,"models/dynamics.py",2953,0,"",python,selection_command +1621,1675178,"models/dynamics.py",2954,0,"",python,selection_command +1622,1675677,"models/dynamics.py",2953,0,"",python,selection_command +1623,1675836,"models/dynamics.py",2952,0,"",python,selection_command +1624,1676001,"models/dynamics.py",2951,0,"",python,selection_command +1625,1676188,"models/dynamics.py",2950,0,"",python,selection_command +1626,1681177,"models/dynamics.py",2950,1,"",python,content +1627,1681417,"models/dynamics.py",2950,1,"",python,content +1628,1681914,"models/dynamics.py",2950,1,"",python,content +1629,1684775,"models/dynamics.py",2950,0,"n",python,content +1630,1684776,"models/dynamics.py",2951,0,"",python,selection_keyboard +1631,1686249,"models/dynamics.py",2952,0,"",python,selection_command +1632,1686739,"models/dynamics.py",2953,0,"",python,selection_command +1633,1686871,"models/dynamics.py",2952,1,"",python,content +1634,1686942,"models/dynamics.py",2952,0,"e",python,content +1635,1686943,"models/dynamics.py",2953,0,"",python,selection_keyboard +1636,1689221,"models/dynamics.py",2957,1,"",python,content +1637,1690290,"models/dynamics.py",2957,0,"b",python,content +1638,1690291,"models/dynamics.py",2958,0,"",python,selection_keyboard +1639,1692307,"models/dynamics.py",2963,1,"",python,content +1640,1692614,"models/dynamics.py",2962,1,"",python,content +1641,1692961,"models/dynamics.py",2960,2,"",python,content +1642,1693599,"models/dynamics.py",2959,1,"",python,content +1643,1694279,"models/dynamics.py",2959,0,"t",python,content +1644,1694280,"models/dynamics.py",2960,0,"",python,selection_keyboard +1645,1695584,"models/dynamics.py",2967,0,"",python,selection_command +1646,1696074,"models/dynamics.py",2967,1,"",python,content +1647,1696242,"models/dynamics.py",2966,1,"",python,content +1648,1696397,"models/dynamics.py",2965,1,"",python,content +1649,1696809,"models/dynamics.py",2964,1,"",python,content +1650,1696964,"models/dynamics.py",2963,1,"",python,content +1651,1697135,"models/dynamics.py",2962,1,"",python,content +1652,1698324,"models/dynamics.py",2962,0,"n",python,content +1653,1698324,"models/dynamics.py",2963,0,"",python,selection_keyboard +1654,1698418,"models/dynamics.py",2963,0," ",python,content +1655,1698419,"models/dynamics.py",2964,0,"",python,selection_keyboard +1656,1698614,"models/dynamics.py",2964,0,"e",python,content +1657,1698615,"models/dynamics.py",2965,0,"",python,selection_keyboard +1658,1699475,"models/dynamics.py",2965,0,")",python,content +1659,1699476,"models/dynamics.py",2966,0,"",python,selection_keyboard +1660,1701277,"models/dynamics.py",3068,0,"",python,selection_mouse +1661,1701766,"models/dynamics.py",3015,0,"",python,selection_mouse +1662,1701996,"models/dynamics.py",3014,0,"",python,selection_command +1663,1715553,"models/dynamics.py",2966,0,"",python,selection_mouse +1664,1716634,"models/dynamics.py",2965,0,"",python,selection_command +1665,1716875,"models/dynamics.py",2965,1,"",python,content +1666,1717101,"models/dynamics.py",2964,0,"",python,selection_command +1667,1717291,"models/dynamics.py",2963,0,"",python,selection_command +1668,1717490,"models/dynamics.py",2962,0,"",python,selection_command +1669,1717719,"models/dynamics.py",2961,0,"",python,selection_command +1670,1717986,"models/dynamics.py",2961,1,"",python,content +1671,1718199,"models/dynamics.py",2962,0,"",python,selection_command +1672,1719248,"models/dynamics.py",2962,0,")",python,content +1673,1719249,"models/dynamics.py",2963,0,"",python,selection_keyboard +1674,1720678,"models/dynamics.py",2959,0,"(",python,content +1675,1720679,"models/dynamics.py",2960,0,"",python,selection_keyboard +1676,1740798,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3350302.4 tasks 0-7: running\r\n",,terminal_output +1677,1741311,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3350302.4\r\nsrun: forcing job termination\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) --KeyboardInterrupt--\r\nEntering jdb:\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3350302.4 ON hkn0509 CANCELLED AT 2025-07-16T14:38:01 ***\r\n",,terminal_output +1678,1741535,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3350302.4\r\n(jdb) (jdb) srun: job abort in progress\r\n(jdb) (jdb) (jdb) (jdb) (jdb) (jdb) ",,terminal_output +1679,1741700,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3350302.4\r\nsrun: job abort in progress\r\n",,terminal_output +1680,1742456,"TERMINAL",0,0,"]0;tum_cte0515@hkn0509:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0509 jafar]$ ",,terminal_output +1681,1742973,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +1682,1743663,"TERMINAL",0,0,"\r\n[?2004l\r\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\n\r\njob_name=""debug""\r\nslurm_job_id=""debug-mihir""\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4\r\n\r\nenv | grep SLURM\r\n\r\nXLA_FLAGS=--xla_gpu_autotune_level=0 srun python train_dynamics.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --log_checkpoint_interval=1000 \\r\n --name=dynamics-debug-run-$slurm_job_id \\r\n --tags dynamics debug \\r\n --entity instant-uv \\r\n --project jafar \\r\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\r\n --data_dir $array_records_dir\r\n ",,terminal_output +1683,1743777,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=692048\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\r\nSLURMD_NODENAME=hkn0509\r\nSLURM_JOB_START_TIME=1752667757\r\nSLURM_STEP_NODELIST=hkn0509\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752703757\r\nSLURM_PMI2_SRUN_PORT=40311\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3350302\r\nSLURM_PTY_PORT=34583\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.198\r\nSLURM_PTY_WIN_ROW=43\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0509\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.198\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=300\r\nSLURM_NODELIST=hkn[0509,0511]\r\nSLURM_SRUN_COMM_PORT=35053\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1990.localdomain\r\nSLURM_JOB_ID=3350302\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0509\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_STEP_LAUNCHER_PORT=35053\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0509,0511]\r\n",,terminal_output +1684,1743964,"TERMINAL",0,0,"GpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +1685,1745916,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +1686,1746068,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +1687,1748782,"genie.py",0,0,"",python,tab +1688,1750436,"utils/nn.py",0,0,"",python,tab +1689,1750623,"TERMINAL",0,0,"2025-07-16 14:38:10.896775: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:38:10.896774: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:38:10.909196: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:38:10.909198: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:38:10.919130: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:38:10.921034: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:38:10.933402: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:38:10.935568: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1690,1765187,"TERMINAL",0,0,"2025-07-16 14:38:25.545533: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1691,1765323,"TERMINAL",0,0,"2025-07-16 14:38:25.702584: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1692,1765401,"TERMINAL",0,0,"2025-07-16 14:38:25.770092: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:38:25.770093: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:38:25.783238: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1693,1765463,"TERMINAL",0,0,"2025-07-16 14:38:25.813381: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:38:25.837029: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:38:25.842421: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1694,1766516,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +1695,1766569,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +1696,1766686,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +1697,1766744,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +1698,1788212,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +1699,1788368,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(46)\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n seq_len = z.shape[1]\r\n # Causal mask: (1, 1, seq_len, seq_len)\r\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.model_dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n",,terminal_output +1700,1791774,"TERMINAL",0,0,"[?25lz[?25h",,terminal_output +1701,1792175,"TERMINAL",0,0,"[?25lz[?25h",,terminal_output +1702,1793638,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +1703,1793827,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1704,1793923,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +1705,1794101,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1706,1794163,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +1707,1794284,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1708,1794431,"TERMINAL",0,0,"\r\n(jdb) (12, 14720, 512)\r\n(jdb) (12, 14720, 512)\r\n(jdb) (12, 14720, 512)\r\n(jdb) (12, 14720, 512)\r\n(jdb) (12, 14720, 512)\r\n(jdb) (12, 14720, 512)\r\n(jdb) (12, 14720, 512)\r\n(jdb) (12, 14720, 512)\r\n",,terminal_output +1709,1795281,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1710,1795512,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1711,1795734,"TERMINAL",0,0,"[?25lq[?25h",,terminal_output +1712,1796155,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +1713,1796491,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +1714,1796644,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1715,1796780,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +1716,1797017,"TERMINAL",0,0,"\r\n(jdb) 14720\r\n(jdb) 14720\r\n(jdb) 14720\r\n(jdb) 14720\r\n(jdb) 14720\r\n(jdb) 14720\r\n(jdb) 14720\r\n(jdb) 14720\r\n",,terminal_output +1717,1806341,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +1718,1806833,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +1719,1807427,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1720,1807486,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +1721,1807674,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1722,1807814,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1723,1807875,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +1724,1808477,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +1725,1808789,"TERMINAL",0,0,"[?25l,[?25h",,terminal_output +1726,1809287,"TERMINAL",0,0,"[?25lm[?25h",,terminal_output +1727,1809347,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1728,1809413,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1729,1809517,"TERMINAL",0,0,"[?25lk[?25h",,terminal_output +1730,1809779,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +1731,1809963,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1732,1810109,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +1733,1810214,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1734,1810277,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +1735,1810424,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1736,1810567,"TERMINAL",0,0,"\r\n(jdb) (14720, 14720)\r\n(jdb) (14720, 14720)\r\n(jdb) (14720, 14720)\r\n(jdb) (14720, 14720)\r\n(jdb) (14720, 14720)\r\n(jdb) (14720, 14720)\r\n(jdb) (14720, 14720)\r\n(jdb) (14720, 14720)\r\n",,terminal_output +1737,1811211,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +1738,1812290,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +1739,1812497,"TERMINAL",0,0,"\r\n",,terminal_output +1740,1812807,"TERMINAL",0,0,"2025-07-16 14:39:13.194682: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:795] The byte size of input/output arguments (41964011520) exceeds the base limit (31805177856). This indicates an error in the calculation!\r\n2025-07-16 14:39:13.195190: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3023] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 39.08GiB (41964011520 bytes), down from 39.08GiB (41964011520 bytes) originally\r\n2025-07-16 14:39:13.195250: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:795] The byte size of input/output arguments (41964011520) exceeds the base limit (31805177856). This indicates an error in the calculation!\r\n2025-07-16 14:39:13.195419: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:795] The byte size of input/output arguments (41964011520) exceeds the base limit (31805177856). This indicates an error in the calculation!\r\n2025-07-16 14:39:13.195597: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3023] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 39.08GiB (41964011520 bytes), down from 39.08GiB (41964011520 bytes) originally\r\n2025-07-16 14:39:13.195662: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:795] The byte size of input/output arguments (41964011520) exceeds the base limit (31805177856). This indicates an error in the calculation!\r\n2025-07-16 14:39:13.195795: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3023] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 39.08GiB (41964011520 bytes), down from 39.08GiB (41964011520 bytes) originally\r\n2025-07-16 14:39:13.196040: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3023] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 39.08GiB (41964011520 bytes), down from 39.08GiB (41964011520 bytes) originally\r\n2025-07-16 14:39:13.197041: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:795] The byte size of input/output arguments (41964011520) exceeds the base limit (31805177856). This indicates an error in the calculation!\r\n2025-07-16 14:39:13.196941: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:795] The byte size of input/output arguments (41964011520) exceeds the base limit (31805177856). This indicates an error in the calculation!\r\n2025-07-16 14:39:13.197387: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3023] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 39.08GiB (41964011520 bytes), down from 39.08GiB (41964011520 bytes) originally\r\n2025-07-16 14:39:13.197372: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3023] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 39.08GiB (41964011520 bytes), down from 39.08GiB (41964011520 bytes) originally\r\n2025-07-16 14:39:13.197471: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:795] The byte size of input/output arguments (41964011520) exceeds the base limit (31805177856). This indicates an error in the calculation!\r\n2025-07-16 14:39:13.197862: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:795] The byte size of input/output arguments (41964011520) exceeds the base limit (31805177856). This indicates an error in the calculation!\r\n2025-07-16 14:39:13.197846: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3023] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 39.08GiB (41964011520 bytes), down from 39.08GiB (41964011520 bytes) originally\r\n2025-07-16 14:39:13.198228: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3023] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 39.08GiB (41964011520 bytes), down from 39.08GiB (41964011520 bytes) originally\r\n",,terminal_output +1741,1822942,"TERMINAL",0,0,"2025-07-16 14:39:23.304107: W external/xla/xla/tsl/framework/bfc_allocator.cc:501] Allocator (GPU_0_bfc) ran out of memory trying to allocate 38.75GiB (rounded to 41602252800)requested by op \r\nIf the cause is memory fragmentation maybe the environment variable 'TF_GPU_ALLOCATOR=cuda_malloc_async' will improve the situation. \r\nCurrent allocation summary follows.\r\nCurrent allocation summary follows.\r\n2025-07-16 14:39:23.304100: W external/xla/xla/tsl/framework/bfc_allocator.cc:501] Allocator (GPU_1_bfc) ran out of memory trying to allocate 38.75GiB (rounded to 41602252800)requested by op \r\nIf the cause is memory fragmentation maybe the environment variable 'TF_GPU_ALLOCATOR=cuda_malloc_async' will improve the situation. \r\nCurrent allocation summary follows.\r\nCurrent allocation summary follows.\r\n2025-07-16 14:39:23.304124: W external/xla/xla/tsl/framework/bfc_allocator.cc:501] Allocator (GPU_2_bfc) ran out of memory trying to allocate 38.75GiB (rounded to 41602252800)requested by op \r\nIf the cause is memory fragmentation maybe the environment variable 'TF_GPU_ALLOCATOR=cuda_malloc_async' will improve the situation. \r\nCurrent allocation summary follows.\r\nCurrent allocation summary follows.\r\n2025-07-16 14:39:23.304039: W external/xla/xla/tsl/framework/bfc_allocator.cc:501] Allocator (GPU_3_bfc) ran out of memory trying to allocate 38.75GiB (rounded to 41602252800)requested by op \r\nIf the cause is memory fragmentation maybe the environment variable 'TF_GPU_ALLOCATOR=cuda_malloc_async' will improve the situation. \r\nCurrent allocation summary follows.\r\nCurrent allocation summary follows.\r\n2025-07-16 14:39:23.304416: W external/xla/xla/tsl/framework/bfc_allocator.cc:512] *********___________________________________________________________________________________________\r\n2025-07-16 14:39:23.304421: W external/xla/xla/tsl/framework/bfc_allocator.cc:512] *********___________________________________________________________________________________________\r\nE0716 14:39:23.304428 705378 pjrt_stream_executor_client.cc:2916] Execution of replica 0 failed: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes. [tf-allocator-allocation-error='']\r\nE0716 14:39:23.304432 705380 pjrt_stream_executor_client.cc:2916] Execution of replica 0 failed: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes. [tf-allocator-allocation-error='']\r\n2025-07-16 14:39:23.304940: W external/xla/xla/tsl/framework/bfc_allocator.cc:512] *********___________________________________________________________________________________________\r\nE0716 14:39:23.304951 705379 pjrt_stream_executor_client.cc:2916] Execution of replica 0 failed: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes. [tf-allocator-allocation-error='']\r\n2025-07-16 14:39:23.304871: W external/xla/xla/tsl/framework/bfc_allocator.cc:512] *********___________________________________________________________________________________________\r\nE0716 14:39:23.304882 705381 pjrt_stream_executor_client.cc:2916] Execution of replica 0 failed: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes. [tf-allocator-allocation-error='']\r\n2025-07-16 14:39:23.304939: W external/xla/xla/tsl/framework/bfc_allocator.cc:501] Allocator (GPU_1_bfc) ran out of memory trying to allocate 38.75GiB (rounded to 41602252800)requested by op \r\nIf the cause is memory fragmentation maybe the environment variable 'TF_GPU_ALLOCATOR=cuda_malloc_async' will improve the situation. \r\nCurrent allocation summary follows.\r\nCurrent allocation summary follows.\r\n2025-07-16 14:39:23.304938: W external/xla/xla/tsl/framework/bfc_allocator.cc:501] Allocator (GPU_3_bfc) ran out of memory trying to allocate 38.75GiB (rounded to 41602252800)requested by op \r\nIf the cause is memory fragmentation maybe the environment variable 'TF_GPU_ALLOCATOR=cuda_malloc_async' will improve the situation. \r\nCurrent allocation summary follows.\r\nCurrent allocation summary follows.\r\n2025-07-16 14:39:23.305755: W external/xla/xla/tsl/framework/bfc_allocator.cc:512] *********___________________________________________________________________________________________\r\nE0716 14:39:23.305768 1058581 pjrt_stream_executor_client.cc:2916] Execution of replica 0 failed: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes. [tf-allocator-allocation-error='']\r\n2025-07-16 14:39:23.305758: W external/xla/xla/tsl/framework/bfc_allocator.cc:512] *********___________________________________________________________________________________________\r\n2025-07-16 14:39:23.306035: W external/xla/xla/tsl/framework/bfc_allocator.cc:501] Allocator (GPU_0_bfc) ran out of memory trying to allocate 38.75GiB (rounded to 41602252800)requested by op \r\nIf the cause is memory fragmentation maybe the environment variable 'TF_GPU_ALLOCATOR=cuda_malloc_async' will improve the situation. \r\nCurrent allocation summary follows.\r\nCurrent allocation summary follows.\r\n2025-07-16 14:39:23.306050: W external/xla/xla/tsl/framework/bfc_allocator.cc:501] Allocator (GPU_2_bfc) ran out of memory trying to allocate 38.75GiB (rounded to 41602252800)requested by op \r\nIf the cause is memory fragmentation maybe the environment variable 'TF_GPU_ALLOCATOR=cuda_malloc_async' will improve the situation. \r\nCurrent allocation summary follows.\r\nCurrent allocation summary follows.\r\nE0716 14:39:23.305769 1058583 pjrt_stream_executor_client.cc:2916] Execution of replica 0 failed: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes. [tf-allocator-allocation-error='']\r\n2025-07-16 14:39:23.306728: W external/xla/xla/tsl/framework/bfc_allocator.cc:512] *********___________________________________________________________________________________________\r\nE0716 14:39:23.306743 1058580 pjrt_stream_executor_client.cc:2916] Execution of replica 0 failed: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes. [tf-allocator-allocation-error='']\r\n2025-07-16 14:39:23.306729: W external/xla/xla/tsl/framework/bfc_allocator.cc:512] *********___________________________________________________________________________________________\r\nE0716 14:39:23.306743 1058582 pjrt_stream_executor_client.cc:2916] Execution of replica 0 failed: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes. [tf-allocator-allocation-error='']\r\n",,terminal_output +1742,1823009,"TERMINAL",0,0,"jax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 107, in __call__\r\n x = CausalTransformerBlock(\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 107, in __call__\r\n x = CausalTransformerBlock(\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 107, in __call__\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 47, in __call__\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 47, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 47, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 107, in __call__\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 47, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 132, in dot_product_attention_weights\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 132, in dot_product_attention_weights\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 132, in dot_product_attention_weights\r\n attn_weights = einsum('...qhd,...khd->...hqk', query, key)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/einsum.py"", line 316, in einsum\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 132, in dot_product_attention_weights\r\n return jit_einsum(operand_arrays, contractions, precision,\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/contextlib.py"", line 79, in inner\r\n attn_weights = einsum('...qhd,...khd->...hqk', query, key)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/einsum.py"", line 316, in einsum\r\n return jit_einsum(operand_arrays, contractions, precision,\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/contextlib.py"", line 79, in inner\r\n attn_weights = einsum('...qhd,...khd->...hqk', query, key)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/einsum.py"", line 316, in einsum\r\n return jit_einsum(operand_arrays, contractions, precision,\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/contextlib.py"", line 79, in inner\r\n attn_weights = einsum('...qhd,...khd->...hqk', query, key)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/einsum.py"", line 316, in einsum\r\n return jit_einsum(operand_arrays, contractions, precision,\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/contextlib.py"", line 79, in inner\r\n return func(*args, **kwds)\r\njaxlib._jax.XlaRuntimeError: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes.\r\n return func(*args, **kwds)\r\njaxlib._jax.XlaRuntimeError: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes.\r\n return func(*args, **kwds)\r\njaxlib._jax.XlaRuntimeError: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes.\r\n return func(*args, **kwds)\r\njaxlib._jax.XlaRuntimeError: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes.\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 107, in __call__\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 47, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 107, in __call__\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 47, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 132, in dot_product_attention_weights\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 132, in dot_product_attention_weights\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 107, in __call__\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 47, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 107, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n x = CausalTransformerBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 47, in __call__\r\n attn_weights = einsum('...qhd,...khd->...hqk', query, key)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/einsum.py"", line 316, in einsum\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n return jit_einsum(operand_arrays, contractions, precision,\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/contextlib.py"", line 79, in inner\r\n return func(*args, **kwds)\r\njaxlib._jax.XlaRuntimeError: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes.\r\n attn_weights = einsum('...qhd,...khd->...hqk', query, key)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/einsum.py"", line 316, in einsum\r\n return jit_einsum(operand_arrays, contractions, precision,\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/contextlib.py"", line 79, in inner\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 132, in dot_product_attention_weights\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 132, in dot_product_attention_weights\r\n attn_weights = einsum('...qhd,...khd->...hqk', query, key)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/einsum.py"", line 316, in einsum\r\n return jit_einsum(operand_arrays, contractions, precision,\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/contextlib.py"", line 79, in inner\r\n return func(*args, **kwds)\r\njaxlib._jax.XlaRuntimeError: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes.\r\n attn_weights = einsum('...qhd,...khd->...hqk', query, key)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/einsum.py"", line 316, in einsum\r\n return jit_einsum(operand_arrays, contractions, precision,\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/contextlib.py"", line 79, in inner\r\n return func(*args, **kwds)\r\njaxlib._jax.XlaRuntimeError: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes.\r\n return func(*args, **kwds)\r\njaxlib._jax.XlaRuntimeError: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes.\r\n",,terminal_output +1743,1823961,"TERMINAL",0,0,"(jdb) ",,terminal_output +1744,1824043,"TERMINAL",0,0,"(jdb) (jdb) (jdb) (jdb) (jdb) (jdb) (jdb) ",,terminal_output +1745,1824476,"TERMINAL",0,0,"srun: error: hkn0509: tasks 0,2: Exited with exit code 1\r\nsrun: error: hkn0511: tasks 5-6: Exited with exit code 1\r\n",,terminal_output +1746,1824529,"TERMINAL",0,0,"srun: error: hkn0509: task 1: Exited with exit code 1\r\n",,terminal_output +1747,1824589,"TERMINAL",0,0,"srun: error: hkn0511: task 7: Exited with exit code 1\r\n",,terminal_output +1748,1824742,"TERMINAL",0,0,"srun: error: hkn0509: task 3: Exited with exit code 1\r\n",,terminal_output +1749,1824806,"TERMINAL",0,0,"srun: error: hkn0511: task 4: Exited with exit code 1\r\n]0;tum_cte0515@hkn0509:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0509 jafar]$ ",,terminal_output +1750,1844976,"genie.py",0,0,"",python,tab +1751,1850719,"utils/nn.py",0,0,"",python,tab +1752,1891642,"utils/nn.py",5323,0,"",python,selection_mouse +1753,1891645,"utils/nn.py",5322,0,"",python,selection_command +1754,1892264,"utils/nn.py",5323,0,"",python,selection_mouse +1755,1892265,"utils/nn.py",5322,0,"",python,selection_command +1756,1892479,"utils/nn.py",5322,1,"x",python,selection_mouse +1757,1892481,"utils/nn.py",5323,0,"",python,selection_command +1758,1892687,"utils/nn.py",5300,23,"x + z\n\n return x",python,selection_mouse +1759,1892690,"utils/nn.py",5136,187," z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1760,1892691,"utils/nn.py",4776,547," )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1761,1892744,"utils/nn.py",4630,693," qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1762,1892786,"utils/nn.py",4514,809," causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1763,1892837,"utils/nn.py",4471,852," dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1764,1892873,"utils/nn.py",4429,894," param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1765,1892901,"utils/nn.py",4403,920," z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1766,1893058,"utils/nn.py",4429,894," param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1767,1893098,"utils/nn.py",4471,852," dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1768,1893110,"utils/nn.py",4501,822," )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1769,1893140,"utils/nn.py",4557,766," z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1770,1893168,"utils/nn.py",4704,619," param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1771,1893203,"utils/nn.py",4807,516," x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1772,1893239,"utils/nn.py",4885,438," z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1773,1893267,"utils/nn.py",4954,369," dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1774,1893320,"utils/nn.py",4998,325," # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1775,1893336,"utils/nn.py",4999,324," # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1776,1893356,"utils/nn.py",5000,323," # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1777,1893411,"utils/nn.py",4958,365," dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1778,1893452,"utils/nn.py",4854,469,"\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1779,1893489,"utils/nn.py",4632,691," qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1780,1893520,"utils/nn.py",4429,894," param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1781,1893563,"utils/nn.py",4359,964," z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1782,1893591,"utils/nn.py",4330,993," x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1783,1893691,"utils/nn.py",4293,1030," # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1784,1894002,"utils/nn.py",4274,1049," x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1785,1894003,"utils/nn.py",4261,1062," )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1786,1894004,"utils/nn.py",4150,1173," dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1787,1894041,"utils/nn.py",4115,1208," qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1788,1894213,"utils/nn.py",4029,1294," )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1789,1894214,"utils/nn.py",3999,1324," dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1790,1894214,"utils/nn.py",3931,1392," z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1791,1894215,"utils/nn.py",3851,1472," # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1792,1894284,"utils/nn.py",3770,1553," @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1793,1894335,"utils/nn.py",3851,1472," # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1794,1894560,"utils/nn.py",3800,1523," def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1795,1894663,"utils/nn.py",3784,1539," @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1796,1895435,"utils/nn.py",3770,1553," @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1797,1895436,"utils/nn.py",3769,1554,"\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1798,1895437,"utils/nn.py",3721,1602," param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1799,1895437,"utils/nn.py",3683,1640," num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1800,1895437,"utils/nn.py",3644,1679,"class STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1801,1895438,"utils/nn.py",3642,1681,"\n\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1802,1895438,"utils/nn.py",3501,1822," self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1803,1895438,"utils/nn.py",3400,1923," dtype=self.dtype,\n )(x)\n # Output projection\n x = nn.Dense(\n self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1804,1895480,"utils/nn.py",3232,2091," model_dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # Output projection\n x = nn.Dense(\n self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1805,1895525,"utils/nn.py",3088,2235," ]\n )(x)\n # Causal transformer blocks\n for _ in range(self.num_blocks):\n x = CausalTransformerBlock(\n model_dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # Output projection\n x = nn.Dense(\n self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1806,1895570,"utils/nn.py",2951,2372," nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n ]\n )(x)\n # Causal transformer blocks\n for _ in range(self.num_blocks):\n x = CausalTransformerBlock(\n model_dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # Output projection\n x = nn.Dense(\n self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1807,1895617,"utils/nn.py",2811,2512," nn.Dense(self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n ]\n )(x)\n # Causal transformer blocks\n for _ in range(self.num_blocks):\n x = CausalTransformerBlock(\n model_dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # Output projection\n x = nn.Dense(\n self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1808,1895669,"utils/nn.py",2674,2649," nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.Dense(self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n ]\n )(x)\n # Causal transformer blocks\n for _ in range(self.num_blocks):\n x = CausalTransformerBlock(\n model_dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # Output projection\n x = nn.Dense(\n self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1809,1895703,"utils/nn.py",3354,1969," param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # Output projection\n x = nn.Dense(\n self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1810,1895732,"utils/nn.py",3599,1724," )(x)\n return x # (B, T, E)\n\n\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1811,1895757,"utils/nn.py",3643,1680,"\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1812,1895794,"utils/nn.py",3670,1653," dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1813,1895816,"utils/nn.py",3702,1621," dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1814,1895838,"utils/nn.py",3721,1602," param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1815,1895859,"utils/nn.py",3769,1554,"\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1816,1895881,"utils/nn.py",3770,1553," @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1817,1895902,"utils/nn.py",3784,1539," @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1818,1895929,"utils/nn.py",3800,1523," def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1819,1895952,"utils/nn.py",3851,1472," # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1820,1895974,"utils/nn.py",3887,1436," z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1821,1896015,"utils/nn.py",3931,1392," z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1822,1896126,"utils/nn.py",3887,1436," z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1823,1896168,"utils/nn.py",3851,1472," # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1824,1896260,"utils/nn.py",3800,1523," def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1825,1896338,"utils/nn.py",3784,1539," @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1826,1896339,"utils/nn.py",3769,1554,"\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1827,1896339,"utils/nn.py",3721,1602," param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1828,1896369,"utils/nn.py",3683,1640," num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1829,1896396,"utils/nn.py",3670,1653," dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1830,1896575,"utils/nn.py",3644,1679,"class STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,selection_mouse +1831,1904810,"utils/nn.py",2340,0,"",python,selection_mouse +1832,1904829,"utils/nn.py",2339,0,"",python,selection_command +1833,1904965,"utils/nn.py",2339,1,"x",python,selection_mouse +1834,1904967,"utils/nn.py",2340,0,"",python,selection_command +1835,1905825,"utils/nn.py",2323,17,"\n return x",python,selection_mouse +1836,1905826,"utils/nn.py",2308,32,"g.breakpoint()\n\n return x",python,selection_mouse +1837,1905826,"utils/nn.py",2202,138,"ebug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1838,1905827,"utils/nn.py",1947,393," param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1839,1905928,"utils/nn.py",1528,812," jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1840,1905929,"utils/nn.py",1133,1207,"type: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1841,1905929,"utils/nn.py",1101,1239," param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1842,1905964,"utils/nn.py",1063,1277," num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1843,1906018,"utils/nn.py",1002,1338,"\nclass CausalTransformerBlock(nn.Module):\n model_dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1844,1906065,"utils/nn.py",919,1421,"\n def __call__(self, x):\n x = x + self.pe[: x.shape[2]]\n return x\n\nclass CausalTransformerBlock(nn.Module):\n model_dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1845,1906103,"utils/nn.py",765,1575," )\n self.pe = self.pe.at[:, 0::2].set(jnp.sin(position * div_term))\n self.pe = self.pe.at[:, 1::2].set(jnp.cos(position * div_term))\n\n def __call__(self, x):\n x = x + self.pe[: x.shape[2]]\n return x\n\nclass CausalTransformerBlock(nn.Module):\n model_dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1846,1906131,"utils/nn.py",919,1421,"\n def __call__(self, x):\n x = x + self.pe[: x.shape[2]]\n return x\n\nclass CausalTransformerBlock(nn.Module):\n model_dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1847,1906157,"utils/nn.py",948,1392," x = x + self.pe[: x.shape[2]]\n return x\n\nclass CausalTransformerBlock(nn.Module):\n model_dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1848,1906206,"utils/nn.py",1003,1337,"class CausalTransformerBlock(nn.Module):\n model_dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1849,1906244,"utils/nn.py",1101,1239," param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1850,1906272,"utils/nn.py",1150,1190," @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1851,1906317,"utils/nn.py",1166,1174," def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1852,1906361,"utils/nn.py",1261,1079," z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1853,1906384,"utils/nn.py",1329,1011," dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1854,1906460,"utils/nn.py",1359,981," )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1855,1906538,"utils/nn.py",1329,1011," dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1856,1906563,"utils/nn.py",1287,1053," param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1857,1906586,"utils/nn.py",1261,1079," z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1858,1906625,"utils/nn.py",1217,1123," # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1859,1906650,"utils/nn.py",1166,1174," def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1860,1906691,"utils/nn.py",1150,1190," @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1861,1906784,"utils/nn.py",1149,1191,"\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1862,1906801,"utils/nn.py",1128,1212," dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1863,1906843,"utils/nn.py",1101,1239," param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1864,1906896,"utils/nn.py",1063,1277," num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1865,1906957,"utils/nn.py",1044,1296," model_dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1866,1907003,"utils/nn.py",1003,1337,"class CausalTransformerBlock(nn.Module):\n model_dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1867,1907105,"utils/nn.py",1002,1338,"\nclass CausalTransformerBlock(nn.Module):\n model_dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x",python,selection_mouse +1868,1907717,"utils/nn.py",1002,0,"",python,selection_mouse +1869,1908604,"utils/nn.py",1002,0,"\n",python,content +1870,1908816,"utils/nn.py",1003,0,"\n",python,content +1871,1909221,"utils/nn.py",1003,0,"",python,selection_command +1872,1909791,"utils/nn.py",1003,0,"class STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x",python,content +1873,1911802,"utils/nn.py",1015,0,"",python,selection_mouse +1874,1912549,"utils/nn.py",1016,0,"",python,selection_command +1875,1912694,"utils/nn.py",1016,0,"2",python,content +1876,1912695,"utils/nn.py",1017,0,"",python,selection_keyboard +1877,1916643,"utils/nn.py",2183,0,"",python,selection_mouse +1878,1917720,"utils/nn.py",1892,0,"",python,selection_mouse +1879,1918104,"utils/nn.py",1891,0,"",python,selection_command +1880,1919774,"utils/nn.py",1872,0,"",python,selection_command +1881,1920304,"utils/nn.py",1848,0,"",python,selection_command +1882,1920316,"utils/nn.py",1806,0,"",python,selection_command +1883,1920398,"utils/nn.py",1780,0,"",python,selection_command +1884,1920399,"utils/nn.py",1736,0,"",python,selection_command +1885,1920400,"utils/nn.py",1707,0,"",python,selection_command +1886,1920438,"utils/nn.py",1670,0,"",python,selection_command +1887,1920461,"utils/nn.py",1652,0,"",python,selection_command +1888,1920516,"utils/nn.py",1650,0,"",python,selection_command +1889,1920527,"utils/nn.py",1632,0,"",python,selection_command +1890,1920572,"utils/nn.py",1608,0,"",python,selection_command +1891,1920573,"utils/nn.py",1566,0,"",python,selection_command +1892,1920955,"utils/nn.py",1608,0,"",python,selection_command +1893,1921151,"utils/nn.py",1632,0,"",python,selection_command +1894,1921303,"utils/nn.py",1650,0,"",python,selection_command +1895,1921452,"utils/nn.py",1652,0,"",python,selection_command +1896,1921614,"utils/nn.py",1670,0,"",python,selection_command +1897,1921769,"utils/nn.py",1707,0,"",python,selection_command +1898,1921921,"utils/nn.py",1736,0,"",python,selection_command +1899,1922054,"utils/nn.py",1780,0,"",python,selection_command +1900,1922259,"utils/nn.py",1806,0,"",python,selection_command +1901,1922429,"utils/nn.py",1848,0,"",python,selection_command +1902,1922537,"utils/nn.py",1872,0,"",python,selection_command +1903,1922704,"utils/nn.py",1891,0,"",python,selection_command +1904,1925720,"utils/nn.py",1910,0,"",python,selection_mouse +1905,1927484,"utils/nn.py",1401,0,"",python,selection_mouse +1906,1927486,"utils/nn.py",1400,0,"",python,selection_command +1907,1928278,"utils/nn.py",1401,0,"\n causal_mask = jnp.tri(z.shape[-2])",python,content +1908,1928339,"utils/nn.py",1410,0,"",python,selection_command +1909,1930498,"utils/nn.py",1663,0,"",python,selection_mouse +1910,1930509,"utils/nn.py",1662,0,"",python,selection_command +1911,1934017,"utils/nn.py",1675,0,"",python,selection_mouse +1912,1935938,"utils/nn.py",1675,0,",",python,content +1913,1935939,"utils/nn.py",1676,0,"",python,selection_keyboard +1914,1936040,"utils/nn.py",1676,0," ",python,content +1915,1936041,"utils/nn.py",1677,0,"",python,selection_keyboard +1916,1936239,"utils/nn.py",1677,0,"m",python,content +1917,1936240,"utils/nn.py",1678,0,"",python,selection_keyboard +1918,1936366,"utils/nn.py",1678,0,"a",python,content +1919,1936368,"utils/nn.py",1679,0,"",python,selection_keyboard +1920,1936479,"utils/nn.py",1679,0,"s",python,content +1921,1936481,"utils/nn.py",1680,0,"",python,selection_keyboard +1922,1936603,"utils/nn.py",1680,0,"k",python,content +1923,1936605,"utils/nn.py",1681,0,"",python,selection_keyboard +1924,1937128,"utils/nn.py",1681,0,"=",python,content +1925,1937130,"utils/nn.py",1682,0,"",python,selection_keyboard +1926,1937489,"utils/nn.py",1682,0,"c",python,content +1927,1937491,"utils/nn.py",1683,0,"",python,selection_keyboard +1928,1937705,"utils/nn.py",1683,0,"a",python,content +1929,1937707,"utils/nn.py",1684,0,"",python,selection_keyboard +1930,1937953,"utils/nn.py",1684,0,"u",python,content +1931,1937957,"utils/nn.py",1685,0,"",python,selection_keyboard +1932,1938117,"utils/nn.py",1685,0,"s",python,content +1933,1938118,"utils/nn.py",1686,0,"",python,selection_keyboard +1934,1938322,"utils/nn.py",1686,0,"a",python,content +1935,1938324,"utils/nn.py",1687,0,"",python,selection_keyboard +1936,1938459,"utils/nn.py",1687,0,"l",python,content +1937,1938460,"utils/nn.py",1688,0,"",python,selection_keyboard +1938,1938749,"utils/nn.py",1688,0,"_",python,content +1939,1938751,"utils/nn.py",1689,0,"",python,selection_keyboard +1940,1939034,"utils/nn.py",1689,0,"m",python,content +1941,1939036,"utils/nn.py",1690,0,"",python,selection_keyboard +1942,1939116,"utils/nn.py",1690,0,"a",python,content +1943,1939118,"utils/nn.py",1691,0,"",python,selection_keyboard +1944,1939169,"utils/nn.py",1691,0,"s",python,content +1945,1939171,"utils/nn.py",1692,0,"",python,selection_keyboard +1946,1939239,"utils/nn.py",1692,0,"k",python,content +1947,1939240,"utils/nn.py",1693,0,"",python,selection_keyboard +1948,1945236,"utils/nn.py",1771,0,"",python,selection_mouse +1949,1947276,"utils/nn.py",1912,0,"",python,selection_mouse +1950,2009357,"utils/nn.py",8243,0,"",python,selection_mouse +1951,2009498,"utils/nn.py",8241,9,"normalize",python,selection_mouse +1952,2017389,"utils/nn.py",4962,0,"",python,selection_mouse +1953,2017514,"utils/nn.py",4951,22,"CausalTransformerBlock",python,selection_mouse +1954,2018778,"utils/nn.py",4951,22,"S",python,content +1955,2018780,"utils/nn.py",4952,0,"",python,selection_keyboard +1956,2018904,"utils/nn.py",4952,0,"T",python,content +1957,2018905,"utils/nn.py",4953,0,"",python,selection_keyboard +1958,2020958,"utils/nn.py",4951,2,"STBlock2",python,content +1959,2029636,"utils/nn.py",8040,0,"",python,selection_mouse +1960,2029872,"utils/nn.py",8038,2,"e,",python,selection_mouse +1961,2029874,"utils/nn.py",8035,5,"type,",python,selection_mouse +1962,2029874,"utils/nn.py",8032,8,"f.dtype,",python,selection_mouse +1963,2030002,"utils/nn.py",8028,12,"=self.dtype,",python,selection_mouse +1964,2030002,"utils/nn.py",7979,61,"ram_dtype=self.param_dtype,\n dtype=self.dtype,",python,selection_mouse +1965,2030002,"utils/nn.py",7975,65," param_dtype=self.param_dtype,\n dtype=self.dtype,",python,selection_mouse +1966,2030003,"utils/nn.py",7935,105," dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,",python,selection_mouse +1967,2030003,"utils/nn.py",7931,109," dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,",python,selection_mouse +1968,2030003,"utils/nn.py",7929,111," dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,",python,selection_mouse +1969,2030003,"utils/nn.py",7884,156," num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,",python,selection_mouse +1970,2030026,"utils/nn.py",7882,158," num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,",python,selection_mouse +1971,2030046,"utils/nn.py",7881,159," num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,",python,selection_mouse +1972,2030098,"utils/nn.py",7845,195," dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,",python,selection_mouse +1973,2034096,"utils/nn.py",5989,0,"",python,selection_mouse +1974,2035069,"utils/nn.py",4951,8,"ST",python,content +1975,2035991,"utils/nn.py",4951,2,"CausalTransformerBlock",python,content +1976,2037948,"utils/nn.py",4972,0,"",python,selection_command +1977,2038334,"utils/nn.py",5217,0,"",python,selection_mouse +1978,2038801,"utils/nn.py",5193,0,"",python,selection_mouse +1979,2038805,"utils/nn.py",5192,0,"",python,selection_command +1980,2039487,"utils/nn.py",5177,16," )(x)",python,selection_command +1981,2039681,"utils/nn.py",5143,50," dtype=self.dtype,\n )(x)",python,selection_command +1982,2039845,"utils/nn.py",5097,96," param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_command +1983,2040009,"utils/nn.py",5059,134," dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_command +1984,2040120,"utils/nn.py",5017,176," num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_command +1985,2040262,"utils/nn.py",4975,218," model_dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_command +1986,2040377,"utils/nn.py",4935,258," x = CausalTransformerBlock(\n model_dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_command +1987,2040682,"utils/nn.py",4947,0,"",python,selection_command +1988,2041505,"utils/nn.py",5189,0,"#",python,content +1989,2041506,"utils/nn.py",5159,0,"#",python,content +1990,2041506,"utils/nn.py",5113,0,"#",python,content +1991,2041506,"utils/nn.py",5075,0,"#",python,content +1992,2041506,"utils/nn.py",5033,0,"#",python,content +1993,2041506,"utils/nn.py",4991,0,"#",python,content +1994,2041506,"utils/nn.py",4947,0,"#",python,content +1995,2041518,"utils/nn.py",4948,0,"",python,selection_keyboard +1996,2041588,"utils/nn.py",5196,0," ",python,content +1997,2041588,"utils/nn.py",5165,0," ",python,content +1998,2041589,"utils/nn.py",5118,0," ",python,content +1999,2041589,"utils/nn.py",5079,0," ",python,content +2000,2041589,"utils/nn.py",5036,0," ",python,content +2001,2041589,"utils/nn.py",4993,0," ",python,content +2002,2041589,"utils/nn.py",4948,0," ",python,content +2003,2041589,"utils/nn.py",4949,0,"",python,selection_keyboard +2004,2042077,"utils/nn.py",4948,0,"",python,selection_command +2005,2042286,"utils/nn.py",4907,0,"",python,selection_command +2006,2043477,"utils/nn.py",4934,0,"\n causal_mask = jnp.tri(z.shape[-2])",python,content +2007,2043510,"utils/nn.py",4943,0,"",python,selection_command +2008,2045145,"utils/nn.py",4943,43,"",python,content +2009,2045156,"utils/nn.py",4907,0,"",python,selection_command +2010,2045884,"utils/nn.py",4934,0,"\n ",python,content +2011,2046130,"utils/nn.py",4947,0," dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,",python,content +2012,2047602,"utils/nn.py",5141,0,"",python,selection_command +2013,2047762,"utils/nn.py",5095,0,"",python,selection_command +2014,2047968,"utils/nn.py",5057,0,"",python,selection_command +2015,2048246,"utils/nn.py",5015,0,"",python,selection_command +2016,2048446,"utils/nn.py",4967,0,"",python,selection_command +2017,2049308,"utils/nn.py",4947,208,"",python,content +2018,2049351,"utils/nn.py",4907,0,"",python,selection_command +2019,2052681,"utils/nn.py",8068,0,"",python,selection_mouse +2020,2052685,"utils/nn.py",8067,0,"",python,selection_command +2021,2053303,"utils/nn.py",8085,0,"",python,selection_mouse +2022,2053308,"utils/nn.py",8084,0,"",python,selection_command +2023,2053470,"utils/nn.py",8084,1,")",python,selection_mouse +2024,2053473,"utils/nn.py",8085,0,"",python,selection_command +2025,2053660,"utils/nn.py",8084,1,")",python,selection_mouse +2026,2053661,"utils/nn.py",8047,38," dtype=self.dtype,\n )(x)",python,selection_mouse +2027,2053661,"utils/nn.py",8044,41," dtype=self.dtype,\n )(x)",python,selection_mouse +2028,2053661,"utils/nn.py",8041,44," dtype=self.dtype,\n )(x)",python,selection_mouse +2029,2053661,"utils/nn.py",7994,91," param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_mouse +2030,2053662,"utils/nn.py",7991,94," param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_mouse +2031,2053662,"utils/nn.py",7990,95," param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_mouse +2032,2053662,"utils/nn.py",7989,96," param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_mouse +2033,2053662,"utils/nn.py",7951,134," dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_mouse +2034,2053741,"utils/nn.py",7909,176," num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_mouse +2035,2053816,"utils/nn.py",7873,212," dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_mouse +2036,2053918,"utils/nn.py",7848,237," x = STBlock(\n dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,selection_mouse +2037,2059600,"utils/nn.py",4914,0,"",python,selection_mouse +2038,2060709,"utils/nn.py",4934,0,"\n ",python,content +2039,2060979,"utils/nn.py",4947,0," x = STBlock(\n dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)",python,content +2040,2062258,"utils/nn.py",4959,0,"",python,selection_mouse +2041,2062716,"utils/nn.py",4955,4,"",python,content +2042,2062912,"utils/nn.py",4951,4,"",python,content +2043,2063306,"utils/nn.py",4947,4,"",python,content +2044,2064087,"utils/nn.py",4946,0,"",python,selection_command +2045,2070221,"utils/nn.py",4956,0,"",python,selection_mouse +2046,2071225,"utils/nn.py",4958,0,"",python,selection_mouse +2047,2072560,"utils/nn.py",4958,0,"2",python,content +2048,2072561,"utils/nn.py",4959,0,"",python,selection_keyboard +2049,2073772,"utils/nn.py",4980,0,"",python,selection_mouse +2050,2074174,"utils/nn.py",4957,0,"",python,selection_mouse +2051,2145248,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +2052,2146974,"utils/nn.py",0,0,"",python,tab +2053,2146975,"utils/nn.py",1246,0,"",python,selection_mouse +2054,2149184,"utils/nn.py",1246,0,"\n ",python,content +2055,2150568,"utils/nn.py",1255,0,"j",python,content +2056,2150569,"utils/nn.py",1256,0,"",python,selection_keyboard +2057,2150679,"utils/nn.py",1256,0,"a",python,content +2058,2150680,"utils/nn.py",1257,0,"",python,selection_keyboard +2059,2150858,"utils/nn.py",1257,0,"x",python,content +2060,2150859,"utils/nn.py",1258,0,"",python,selection_keyboard +2061,2150948,"utils/nn.py",1258,0,".",python,content +2062,2150949,"utils/nn.py",1259,0,"",python,selection_keyboard +2063,2151355,"utils/nn.py",1259,0,"d",python,content +2064,2151357,"utils/nn.py",1260,0,"",python,selection_keyboard +2065,2152014,"utils/nn.py",1259,1,"debug",python,content +2066,2152381,"utils/nn.py",1264,0,".",python,content +2067,2152382,"utils/nn.py",1265,0,"",python,selection_keyboard +2068,2152783,"utils/nn.py",1265,0,"b",python,content +2069,2152784,"utils/nn.py",1266,0,"",python,selection_keyboard +2070,2152915,"utils/nn.py",1266,0,"r",python,content +2071,2152916,"utils/nn.py",1267,0,"",python,selection_keyboard +2072,2153209,"utils/nn.py",1265,2,"breakpoint",python,content +2073,2154074,"utils/nn.py",1275,0,"()",python,content +2074,2154075,"utils/nn.py",1276,0,"",python,selection_keyboard +2075,2154133,"utils/nn.py",1276,1,")",python,content +2076,2154134,"utils/nn.py",1277,0,"",python,selection_keyboard +2077,2154257,"utils/nn.py",1276,0,"",python,selection_command +2078,2154938,"utils/nn.py",1307,0,"",python,selection_command +2079,2155075,"utils/nn.py",1346,0,"",python,selection_command +2080,2155228,"utils/nn.py",1377,0,"",python,selection_command +2081,2155361,"utils/nn.py",1418,0,"",python,selection_command +2082,2155590,"utils/nn.py",1431,0,"",python,selection_command +2083,2155810,"utils/nn.py",1432,0,"\n jax.debug.breakpoint()",python,content +2084,2155844,"utils/nn.py",1441,0,"",python,selection_command +2085,2156263,"utils/nn.py",1472,0,"",python,selection_command +2086,2156460,"utils/nn.py",1515,0,"",python,selection_command +2087,2156623,"utils/nn.py",1550,0,"",python,selection_command +2088,2156744,"utils/nn.py",1588,0,"",python,selection_command +2089,2157114,"utils/nn.py",1623,0,"",python,selection_command +2090,2157321,"utils/nn.py",1662,0,"",python,selection_command +2091,2157677,"utils/nn.py",1623,0,"",python,selection_command +2092,2157866,"utils/nn.py",1588,0,"",python,selection_command +2093,2158003,"utils/nn.py",1550,0,"",python,selection_command +2094,2158164,"utils/nn.py",1515,0,"",python,selection_command +2095,2158294,"utils/nn.py",1472,0,"",python,selection_command +2096,2158676,"utils/nn.py",1506,0,"\n jax.debug.breakpoint()",python,content +2097,2158721,"utils/nn.py",1515,0,"",python,selection_command +2098,2158886,"utils/nn.py",1472,0,"",python,selection_command +2099,2159068,"utils/nn.py",1441,0,"",python,selection_command +2100,2159388,"utils/nn.py",1433,31,"",python,content +2101,2159409,"utils/nn.py",1441,0,"",python,selection_command +2102,2159581,"utils/nn.py",1484,0,"",python,selection_command +2103,2159775,"utils/nn.py",1515,0,"",python,selection_command +2104,2159953,"utils/nn.py",1550,0,"",python,selection_command +2105,2160137,"utils/nn.py",1588,0,"",python,selection_command +2106,2160305,"utils/nn.py",1623,0,"",python,selection_command +2107,2160482,"utils/nn.py",1662,0,"",python,selection_command +2108,2160650,"utils/nn.py",1704,0,"",python,selection_command +2109,2160809,"utils/nn.py",1734,0,"",python,selection_command +2110,2160979,"utils/nn.py",1765,0,"",python,selection_command +2111,2161226,"utils/nn.py",1774,0,"\n jax.debug.breakpoint()",python,content +2112,2161268,"utils/nn.py",1783,0,"",python,selection_command +2113,2161459,"utils/nn.py",1806,0,"",python,selection_command +2114,2161631,"utils/nn.py",1815,0,"",python,selection_command +2115,2161794,"utils/nn.py",1852,0,"",python,selection_command +2116,2161956,"utils/nn.py",1881,0,"",python,selection_command +2117,2162168,"utils/nn.py",1925,0,"",python,selection_command +2118,2162327,"utils/nn.py",1951,0,"",python,selection_command +2119,2162493,"utils/nn.py",1993,0,"",python,selection_command +2120,2162667,"utils/nn.py",2023,0,"",python,selection_command +2121,2162994,"utils/nn.py",2027,0,"\n jax.debug.breakpoint()",python,content +2122,2163026,"utils/nn.py",2036,0,"",python,selection_command +2123,2163176,"utils/nn.py",2067,0,"",python,selection_command +2124,2163341,"utils/nn.py",2110,0,"",python,selection_command +2125,2163729,"utils/nn.py",2067,0,"",python,selection_command +2126,2163932,"utils/nn.py",2036,0,"",python,selection_command +2127,2164096,"utils/nn.py",2023,0,"",python,selection_command +2128,2164233,"utils/nn.py",2036,0,"",python,selection_command +2129,2164383,"utils/nn.py",2067,0,"",python,selection_command +2130,2164557,"utils/nn.py",2110,0,"",python,selection_command +2131,2164715,"utils/nn.py",2145,0,"",python,selection_command +2132,2164863,"utils/nn.py",2183,0,"",python,selection_command +2133,2165021,"utils/nn.py",2218,0,"",python,selection_command +2134,2165180,"utils/nn.py",2257,0,"",python,selection_command +2135,2165364,"utils/nn.py",2299,0,"",python,selection_command +2136,2165541,"utils/nn.py",2329,0,"",python,selection_command +2137,2165874,"utils/nn.py",2360,0,"",python,selection_command +2138,2166178,"utils/nn.py",2378,0,"",python,selection_command +2139,2166744,"utils/nn.py",2398,0,"\n jax.debug.breakpoint()",python,content +2140,2166802,"utils/nn.py",2407,0,"",python,selection_command +2141,2167143,"utils/nn.py",2378,0,"",python,selection_command +2142,2167321,"utils/nn.py",2360,0,"",python,selection_command +2143,2167509,"utils/nn.py",2369,0,"\n jax.debug.breakpoint()",python,content +2144,2167533,"utils/nn.py",2378,0,"",python,selection_command +2145,2167796,"utils/nn.py",2409,0,"",python,selection_command +2146,2167979,"utils/nn.py",2438,0,"",python,selection_command +2147,2168285,"utils/nn.py",2461,0,"",python,selection_command +2148,2168735,"utils/nn.py",2470,0,"",python,selection_command +2149,2168926,"utils/nn.py",2500,0,"",python,selection_command +2150,2169070,"utils/nn.py",2526,0,"",python,selection_command +2151,2169231,"utils/nn.py",2568,0,"",python,selection_command +2152,2169408,"utils/nn.py",2598,0,"",python,selection_command +2153,2169776,"utils/nn.py",2602,0,"\n jax.debug.breakpoint()",python,content +2154,2169802,"utils/nn.py",2611,0,"",python,selection_command +2155,2169942,"utils/nn.py",2642,0,"",python,selection_command +2156,2170129,"utils/nn.py",2782,0,"",python,selection_command +2157,2170336,"utils/nn.py",2804,0,"",python,selection_command +2158,2170514,"utils/nn.py",2826,0,"",python,selection_command +2159,2170728,"utils/nn.py",2868,0,"",python,selection_command +2160,2171173,"utils/nn.py",2898,0,"",python,selection_command +2161,2171742,"utils/nn.py",2911,0,"",python,selection_command +2162,2172107,"utils/nn.py",2934,0,"",python,selection_command +2163,2172651,"utils/nn.py",2943,0,"\n jax.debug.breakpoint()",python,content +2164,2172677,"utils/nn.py",2952,0,"",python,selection_command +2165,2174376,"TERMINAL",0,0,"\r\n[?2004l\r\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\n\r\njob_name=""debug""\r\nslurm_job_id=""debug-mihir""\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4\r\n\r\nenv | grep SLURM\r\n\r\nXLA_FLAGS=--xla_gpu_autotune_level=0 srun python train_dynamics.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --log_checkpoint_interval=1000 \\r\n --name=dynamics-debug-run-$slurm_job_id \\r\n --tags dynamics debug \\r\n --entity instant-uv \\r\n --project jafar \\r\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\r\n --data_dir $array_records_dir\r\n ",,terminal_output +2166,2174497,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=692048\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\r\nSLURMD_NODENAME=hkn0509\r\nSLURM_JOB_START_TIME=1752667757\r\nSLURM_STEP_NODELIST=hkn0509\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752703757\r\nSLURM_PMI2_SRUN_PORT=40311\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3350302\r\nSLURM_PTY_PORT=34583\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.198\r\nSLURM_PTY_WIN_ROW=43\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0509\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.198\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=300\r\nSLURM_NODELIST=hkn[0509,0511]\r\nSLURM_SRUN_COMM_PORT=35053\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1990.localdomain\r\nSLURM_JOB_ID=3350302\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0509\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_STEP_LAUNCHER_PORT=35053\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0509,0511]\r\n",,terminal_output +2167,2174648,"TERMINAL",0,0,"GpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +2168,2176703,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +2169,2176762,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +2170,2181258,"TERMINAL",0,0,"2025-07-16 14:45:21.647062: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2171,2181358,"TERMINAL",0,0,"2025-07-16 14:45:21.698199: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:45:21.706095: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:45:21.706095: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:45:21.707809: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:45:21.724705: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2172,2181462,"TERMINAL",0,0,"2025-07-16 14:45:21.848567: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2173,2181569,"TERMINAL",0,0,"2025-07-16 14:45:21.917413: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2174,2195681,"TERMINAL",0,0,"2025-07-16 14:45:36.071119: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2175,2196103,"TERMINAL",0,0,"2025-07-16 14:45:36.370808: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2176,2196148,"TERMINAL",0,0,"2025-07-16 14:45:36.509725: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2177,2196233,"TERMINAL",0,0,"2025-07-16 14:45:36.561480: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:45:36.609478: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:45:36.617283: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2178,2196945,"TERMINAL",0,0,"Running on 8 devices.\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 175, in __call__\r\n x = STBlock2(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 41, in __call__\r\n z = PositionalEncoding(self.dim)(x)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 26, in __call__\r\n x = x + self.pe[: x.shape[2]]\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 1083, in op\r\n return getattr(self.aval, f""_{name}"")(self, *args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 583, in deferring_binary_op\r\n return binary_op(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufunc_api.py"", line 182, in __call__\r\n return call(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufuncs.py"", line 1234, in add\r\n x, y = promote_args(""add"", x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 228, in promote_args\r\n return promote_shapes(fun_name, *promote_dtypes(*args))\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 64, in promote_shapes\r\n result_rank = len(lax.broadcast_shapes(*shapes))\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n",,terminal_output +2179,2197239,"TERMINAL",0,0,"Running on 8 devices.\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 175, in __call__\r\n x = STBlock2(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 41, in __call__\r\n z = PositionalEncoding(self.dim)(x)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 26, in __call__\r\n x = x + self.pe[: x.shape[2]]\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 1083, in op\r\n return getattr(self.aval, f""_{name}"")(self, *args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 583, in deferring_binary_op\r\n return binary_op(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufunc_api.py"", line 182, in __call__\r\n return call(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufuncs.py"", line 1234, in add\r\n x, y = promote_args(""add"", x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 228, in promote_args\r\n return promote_shapes(fun_name, *promote_dtypes(*args))\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 64, in promote_shapes\r\n result_rank = len(lax.broadcast_shapes(*shapes))\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n",,terminal_output +2180,2197344,"TERMINAL",0,0,"Running on 8 devices.\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 175, in __call__\r\n x = STBlock2(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 41, in __call__\r\n z = PositionalEncoding(self.dim)(x)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 26, in __call__\r\n x = x + self.pe[: x.shape[2]]\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 1083, in op\r\n return getattr(self.aval, f""_{name}"")(self, *args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 583, in deferring_binary_op\r\n return binary_op(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufunc_api.py"", line 182, in __call__\r\n return call(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufuncs.py"", line 1234, in add\r\n x, y = promote_args(""add"", x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 228, in promote_args\r\n return promote_shapes(fun_name, *promote_dtypes(*args))\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 64, in promote_shapes\r\n result_rank = len(lax.broadcast_shapes(*shapes))\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n",,terminal_output +2181,2197405,"TERMINAL",0,0,"Running on 8 devices.\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 175, in __call__\r\n x = STBlock2(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 41, in __call__\r\n z = PositionalEncoding(self.dim)(x)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 26, in __call__\r\n x = x + self.pe[: x.shape[2]]\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 1083, in op\r\n return getattr(self.aval, f""_{name}"")(self, *args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 583, in deferring_binary_op\r\n return binary_op(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufunc_api.py"", line 182, in __call__\r\n return call(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufuncs.py"", line 1234, in add\r\n x, y = promote_args(""add"", x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 228, in promote_args\r\n return promote_shapes(fun_name, *promote_dtypes(*args))\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 64, in promote_shapes\r\n result_rank = len(lax.broadcast_shapes(*shapes))\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n",,terminal_output +2182,2197479,"TERMINAL",0,0,"Running on 8 devices.\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 175, in __call__\r\n x = STBlock2(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 41, in __call__\r\n z = PositionalEncoding(self.dim)(x)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 26, in __call__\r\n x = x + self.pe[: x.shape[2]]\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 1083, in op\r\n return getattr(self.aval, f""_{name}"")(self, *args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 583, in deferring_binary_op\r\n return binary_op(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufunc_api.py"", line 182, in __call__\r\n return call(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufuncs.py"", line 1234, in add\r\n x, y = promote_args(""add"", x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 228, in promote_args\r\n return promote_shapes(fun_name, *promote_dtypes(*args))\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 64, in promote_shapes\r\n result_rank = len(lax.broadcast_shapes(*shapes))\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\nRunning on 8 devices.\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 175, in __call__\r\n x = STBlock2(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 41, in __call__\r\n z = PositionalEncoding(self.dim)(x)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 26, in __call__\r\n x = x + self.pe[: x.shape[2]]\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 1083, in op\r\n return getattr(self.aval, f""_{name}"")(self, *args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 583, in deferring_binary_op\r\n return binary_op(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufunc_api.py"", line 182, in __call__\r\n return call(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufuncs.py"", line 1234, in add\r\n x, y = promote_args(""add"", x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 228, in promote_args\r\n return promote_shapes(fun_name, *promote_dtypes(*args))\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 64, in promote_shapes\r\n result_rank = len(lax.broadcast_shapes(*shapes))\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n",,terminal_output +2183,2197738,"TERMINAL",0,0,"2025-07-16 14:45:38.128552: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2184,2198254,"TERMINAL",0,0,"2025-07-16 14:45:38.641253: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2185,2199025,"TERMINAL",0,0,"Running on 8 devices.\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 175, in __call__\r\n x = STBlock2(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 41, in __call__\r\n z = PositionalEncoding(self.dim)(x)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 26, in __call__\r\n x = x + self.pe[: x.shape[2]]\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 1083, in op\r\n return getattr(self.aval, f""_{name}"")(self, *args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 583, in deferring_binary_op\r\n return binary_op(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufunc_api.py"", line 182, in __call__\r\n return call(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufuncs.py"", line 1234, in add\r\n x, y = promote_args(""add"", x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 228, in promote_args\r\n return promote_shapes(fun_name, *promote_dtypes(*args))\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 64, in promote_shapes\r\n result_rank = len(lax.broadcast_shapes(*shapes))\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n",,terminal_output +2186,2199573,"TERMINAL",0,0,"Running on 8 devices.\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 197, in broadcast_shapes\r\n return _broadcast_shapes_cached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 299, in wrapper\r\n return cached(config.trace_context() if trace_context_in_key else _ignore(),\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 293, in cached\r\n return f(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 203, in _broadcast_shapes_cached\r\n return _broadcast_shapes_uncached(*shapes)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 222, in _broadcast_shapes_uncached\r\n raise ValueError(f""Incompatible shapes for broadcasting: shapes={list(shapes)}"") from err\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 219, in _broadcast_shapes_uncached\r\n return _try_broadcast_shapes(*rank_promoted_shapes, name='broadcast_shapes')\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/lax/lax.py"", line 136, in _try_broadcast_shapes\r\n raise TypeError(f'{name} got incompatible shapes for broadcasting: '\r\nTypeError: broadcast_shapes got incompatible shapes for broadcasting: (12, 14720, 512), (1, 512, 512).\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 199, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 106, in __call__\r\n dyna_outputs = self.dynamics(outputs, training)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n logits = self.dynamics(vid_embed_flat)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 175, in __call__\r\n x = STBlock2(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 41, in __call__\r\n z = PositionalEncoding(self.dim)(x)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 26, in __call__\r\n x = x + self.pe[: x.shape[2]]\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 1083, in op\r\n return getattr(self.aval, f""_{name}"")(self, *args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 583, in deferring_binary_op\r\n return binary_op(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufunc_api.py"", line 182, in __call__\r\n return call(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufuncs.py"", line 1234, in add\r\n x, y = promote_args(""add"", x, y)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 228, in promote_args\r\n return promote_shapes(fun_name, *promote_dtypes(*args))\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/util.py"", line 64, in promote_shapes\r\n result_rank = len(lax.broadcast_shapes(*shapes))\r\nValueError: Incompatible shapes for broadcasting: shapes=[(12, 14720, 512), (512, 512)]\r\n",,terminal_output +2187,2200907,"TERMINAL",0,0,"srun: error: hkn0511: task 5: Exited with exit code 1\r\n",,terminal_output +2188,2200960,"TERMINAL",0,0,"srun: error: hkn0509: tasks 1-2: Exited with exit code 1\r\n",,terminal_output +2189,2201020,"TERMINAL",0,0,"srun: error: hkn0511: tasks 6-7: Exited with exit code 1\r\n",,terminal_output +2190,2201077,"TERMINAL",0,0,"srun: error: hkn0509: task 0: Exited with exit code 1\r\n",,terminal_output +2191,2201261,"TERMINAL",0,0,"srun: error: hkn0511: task 4: Exited with exit code 1\r\n",,terminal_output +2192,2201326,"TERMINAL",0,0,"srun: error: hkn0509: task 3: Exited with exit code 1\r\n]0;tum_cte0515@hkn0509:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0509 jafar]$ ",,terminal_output +2193,2220229,"models/dynamics.py",0,0,"",python,tab +2194,2222682,"models/dynamics.py",2861,0,"",python,selection_mouse +2195,2223164,"models/dynamics.py",2933,0,"",python,selection_mouse +2196,2224148,"models/dynamics.py",2932,0,"",python,selection_command +2197,2224830,"models/dynamics.py",2892,77,"",python,content +2198,2224913,"models/dynamics.py",2900,0,"",python,selection_command +2199,2224932,"models/dynamics.py",2826,0,"",python,selection_command +2200,2225996,"models/dynamics.py",2900,0,"vid_embed_flat = einops.rearrange(vid_embed, ""b t n e -> b (t n) e"")\n ",python,content +2201,2226031,"models/dynamics.py",2932,0,"",python,selection_command +2202,2227260,"models/dynamics.py",2892,77,"",python,content +2203,2227357,"models/dynamics.py",2900,0,"",python,selection_command +2204,2227588,"models/dynamics.py",2901,0,"",python,selection_command +2205,2228075,"models/dynamics.py",2902,0,"",python,selection_command +2206,2228106,"models/dynamics.py",2903,0,"",python,selection_command +2207,2228142,"models/dynamics.py",2904,0,"",python,selection_command +2208,2228227,"models/dynamics.py",2905,0,"",python,selection_command +2209,2228227,"models/dynamics.py",2906,0,"",python,selection_command +2210,2228245,"models/dynamics.py",2907,0,"",python,selection_command +2211,2228297,"models/dynamics.py",2908,0,"",python,selection_command +2212,2228327,"models/dynamics.py",2909,0,"",python,selection_command +2213,2228343,"models/dynamics.py",2910,0,"",python,selection_command +2214,2228360,"models/dynamics.py",2911,0,"",python,selection_command +2215,2228377,"models/dynamics.py",2912,0,"",python,selection_command +2216,2228427,"models/dynamics.py",2913,0,"",python,selection_command +2217,2228447,"models/dynamics.py",2914,0,"",python,selection_command +2218,2228505,"models/dynamics.py",2915,0,"",python,selection_command +2219,2228514,"models/dynamics.py",2916,0,"",python,selection_command +2220,2228544,"models/dynamics.py",2917,0,"",python,selection_command +2221,2228575,"models/dynamics.py",2918,0,"",python,selection_command +2222,2228603,"models/dynamics.py",2919,0,"",python,selection_command +2223,2228640,"models/dynamics.py",2920,0,"",python,selection_command +2224,2228665,"models/dynamics.py",2921,0,"",python,selection_command +2225,2228701,"models/dynamics.py",2922,0,"",python,selection_command +2226,2228751,"models/dynamics.py",2923,0,"",python,selection_command +2227,2228752,"models/dynamics.py",2924,0,"",python,selection_command +2228,2228803,"models/dynamics.py",2925,0,"",python,selection_command +2229,2228813,"models/dynamics.py",2926,0,"",python,selection_command +2230,2228845,"models/dynamics.py",2927,0,"",python,selection_command +2231,2228878,"models/dynamics.py",2928,0,"",python,selection_command +2232,2228893,"models/dynamics.py",2929,0,"",python,selection_command +2233,2229117,"models/dynamics.py",2930,0,"",python,selection_command +2234,2229363,"models/dynamics.py",2931,0,"",python,selection_command +2235,2229545,"models/dynamics.py",2932,0,"",python,selection_command +2236,2229813,"models/dynamics.py",2932,5,"",python,content +2237,2231561,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +2238,2231768,"TERMINAL",0,0,"\r\n[?2004l\r\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\n\r\njob_name=""debug""\r\nslurm_job_id=""debug-mihir""\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4\r\n\r\nenv | grep SLURM\r\n\r\nXLA_FLAGS=--xla_gpu_autotune_level=0 srun python train_dynamics.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --log_checkpoint_interval=1000 \\r\n --name=dynamics-debug-run-$slurm_job_id \\r\n --tags dynamics debug \\r\n --entity instant-uv \\r\n --project jafar \\r\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\r\n --data_dir $array_records_dir\r\n ",,terminal_output +2239,2231883,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=692048\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\r\nSLURMD_NODENAME=hkn0509\r\nSLURM_JOB_START_TIME=1752667757\r\nSLURM_STEP_NODELIST=hkn0509\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752703757\r\nSLURM_PMI2_SRUN_PORT=40311\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3350302\r\nSLURM_PTY_PORT=34583\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.198\r\nSLURM_PTY_WIN_ROW=43\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e10.hkn0509\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.198\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=300\r\nSLURM_NODELIST=hkn[0509,0511]\r\nSLURM_SRUN_COMM_PORT=35053\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1990.localdomain\r\nSLURM_JOB_ID=3350302\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0509\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_STEP_LAUNCHER_PORT=35053\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0509,0511]\r\n",,terminal_output +2240,2232009,"TERMINAL",0,0,"GpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +2241,2233969,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +2242,2234029,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +2243,2234233,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +2244,2238593,"TERMINAL",0,0,"2025-07-16 14:46:18.928428: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:46:18.929688: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:46:18.929703: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:46:18.934923: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:46:18.943388: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:46:18.953337: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:46:18.953496: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:46:18.984253: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2245,2253081,"TERMINAL",0,0,"2025-07-16 14:46:33.470669: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2246,2253305,"TERMINAL",0,0,"2025-07-16 14:46:33.673322: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:46:33.687927: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2247,2253394,"TERMINAL",0,0,"2025-07-16 14:46:33.726612: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:46:33.735623: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2248,2253458,"TERMINAL",0,0,"2025-07-16 14:46:33.798415: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:46:33.841145: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-16 14:46:33.841169: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2249,2254051,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +2250,2254308,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +2251,2254369,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +2252,2254427,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +2253,2254521,"TERMINAL",0,0,"Running on 8 devices.\r\nEntering jdb:\r\nRunning on 8 devices.\r\nEntering jdb:\r\n",,terminal_output +2254,2273292,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +2255,2273395,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n",,terminal_output +2256,2280654,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",0,0,"import math\nfrom typing import Dict, Tuple\n\nfrom flax import linen as nn\nimport jax\nimport jax.numpy as jnp\n\n\nclass PositionalEncoding(nn.Module):\n """"""https://uvadlc-notebooks.readthedocs.io/en/latest/tutorial_notebooks/JAX/tutorial6/Transformers_and_MHAttention.html""""""\n\n d_model: int # Hidden dimensionality of the input.\n max_len: int = 5000 # Maximum length of a sequence to expect.\n\n def setup(self):\n # Create matrix of [SeqLen, HiddenDim] representing the positional encoding for max_len inputs\n self.pe = jnp.zeros((self.max_len, self.d_model))\n position = jnp.arange(0, self.max_len, dtype=jnp.float32)[:, None]\n div_term = jnp.exp(\n jnp.arange(0, self.d_model, 2) * (-math.log(10000.0) / self.d_model)\n )\n self.pe = self.pe.at[:, 0::2].set(jnp.sin(position * div_term))\n self.pe = self.pe.at[:, 1::2].set(jnp.cos(position * div_term))\n\n def __call__(self, x):\n x = x + self.pe[: x.shape[2]]\n return x\n\nclass STBlock2(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n jax.debug.breakpoint()\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n jax.debug.breakpoint()\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n jax.debug.breakpoint()\n x = x.swapaxes(1, 2)\n jax.debug.breakpoint()\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n jax.debug.breakpoint()\n\n return x\n\nclass CausalTransformerBlock(nn.Module):\n model_dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # LayerNorm + Causal Self-Attention\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n seq_len = z.shape[1]\n # Causal mask: (1, 1, seq_len, seq_len)\n causal_mask = jnp.tril(jnp.ones((seq_len, seq_len), dtype=bool))\n jax.debug.breakpoint()\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n # Feedforward\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n jax.debug.breakpoint()\n z = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n jax.debug.breakpoint()\n z = nn.gelu(z)\n jax.debug.breakpoint()\n x = x + z\n jax.debug.breakpoint()\n\n return x\n\nclass CausalTransformer(nn.Module):\n model_dim: int\n out_dim: int\n num_blocks: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # Input projection and normalization\n x = nn.Sequential(\n [\n nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.Dense(self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n ]\n )(x)\n # Causal transformer blocks\n for _ in range(self.num_blocks):\n x = STBlock2(\n dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # x = CausalTransformerBlock(\n # model_dim=self.model_dim,\n # num_heads=self.num_heads,\n # dropout=self.dropout,\n # param_dtype=self.param_dtype,\n # dtype=self.dtype,\n # )(x)\n # Output projection\n x = nn.Dense(\n self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\nclass STBlock(nn.Module):\n dim: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.remat\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n x = x + z\n\n # --- Temporal attention ---\n x = x.swapaxes(1, 2)\n z = PositionalEncoding(self.dim)(x)\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n causal_mask = jnp.tri(z.shape[-2])\n z = nn.MultiHeadAttention(\n num_heads=self.num_heads,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z, mask=causal_mask)\n x = x + z\n x = x.swapaxes(1, 2)\n\n # --- Feedforward ---\n z = nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\n z = nn.Dense(\n self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(z)\n z = nn.gelu(z)\n x = x + z\n\n return x\n\n\nclass STTransformer(nn.Module):\n model_dim: int\n out_dim: int\n num_blocks: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n @nn.compact\n def __call__(self, x: jax.Array) -> jax.Array:\n x = nn.Sequential(\n [\n nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.Dense(self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n nn.LayerNorm(\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ),\n ]\n )(x)\n for _ in range(self.num_blocks):\n x = STBlock(\n dim=self.model_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n x = nn.Dense(\n self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )(x)\n return x # (B, T, E)\n\n\ndef normalize(x):\n return x / (jnp.linalg.norm(x, ord=2, axis=-1, keepdims=True) + 1e-8)\n\n\nclass VectorQuantizer(nn.Module):\n latent_dim: int\n num_latents: int\n dropout: float\n\n def setup(self):\n self.codebook = normalize(\n self.param(\n ""codebook"",\n nn.initializers.lecun_uniform(),\n (self.num_latents, self.latent_dim),\n )\n )\n self.drop = nn.Dropout(self.dropout, deterministic=False)\n\n def __call__(\n self, x: jax.Array, training: bool\n ) -> Tuple[jax.Array, jax.Array, jax.Array, jax.Array]:\n # --- Compute distances ---\n x = normalize(x)\n codebook = normalize(self.codebook)\n distance = -jnp.matmul(x, codebook.T)\n if training:\n dropout_key = self.make_rng(""dropout"")\n distance = self.drop(distance, rng=dropout_key)\n\n # --- Get indices and embeddings ---\n indices = jnp.argmin(distance, axis=-1)\n z = self.codebook[indices]\n\n # --- Straight through estimator ---\n z_q = x + jax.lax.stop_gradient(z - x)\n return z_q, z, x, indices\n\n def get_codes(self, indices: jax.Array):\n return self.codebook[indices]\n",python,tab +2257,2283028,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",0,0,"",python,tab +2258,2288828,"TERMINAL",0,0,"[?25lx[?25h",,terminal_output +2259,2289162,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +2260,2289385,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +2261,2289477,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +2262,2289602,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +2263,2289666,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +2264,2289770,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +2265,2289955,"TERMINAL",0,0,"\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n",,terminal_output +2266,2295165,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +2267,2295995,"TERMINAL",0,0,"\r\n",,terminal_output +2268,2296227,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +2269,2297913,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +2270,2298083,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(47)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n causal_mask = jnp.tri(z.shape[-2])\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(47)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n causal_mask = jnp.tri(z.shape[-2])\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(47)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n causal_mask = jnp.tri(z.shape[-2])\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(47)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n causal_mask = jnp.tri(z.shape[-2])\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(47)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n causal_mask = jnp.tri(z.shape[-2])\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(47)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n causal_mask = jnp.tri(z.shape[-2])\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(47)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n causal_mask = jnp.tri(z.shape[-2])\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(47)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n causal_mask = jnp.tri(z.shape[-2])\r\n-> jax.debug.breakpoint()\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n",,terminal_output +2271,2301022,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +2272,2301529,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +2273,2301759,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +2274,2301895,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +2275,2302187,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +2276,2302292,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +2277,2302749,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +2278,2303507,"TERMINAL",0,0,"[?25lm[?25h",,terminal_output +2279,2303575,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +2280,2303636,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +2281,2303698,"TERMINAL",0,0,"[?25lk[?25h",,terminal_output +2282,2303946,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +2283,2304138,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +2284,2304338,"TERMINAL",0,0,"[?25lh[?25h[?25la[?25h",,terminal_output +2285,2304558,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +2286,2304698,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +2287,2305207,"TERMINAL",0,0,"[?25lp[?25h[?25le[?25h",,terminal_output +2288,2305408,"TERMINAL",0,0,"\r\n(jdb) (920, 920)\r\n(jdb) (920, 920)\r\n(jdb) (920, 920)\r\n(jdb) (920, 920)\r\n(jdb) (920, 920)\r\n(jdb) (920, 920)\r\n(jdb) (920, 920)\r\n(jdb) (920, 920)\r\n",,terminal_output +2289,2311543,"TERMINAL",0,0,"[?25lz[?25h",,terminal_output +2290,2311883,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +2291,2312086,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +2292,2312146,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +2293,2312297,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +2294,2312403,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +2295,2312465,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +2296,2312661,"TERMINAL",0,0,"\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n",,terminal_output +2297,2325447,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +2298,2325594,"TERMINAL",0,0,"\r\n",,terminal_output +2299,2325841,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +2300,2327092,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +2301,2327281,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(56)\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z, mask=causal_mask)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n # --- Temporal attention ---\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(56)\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z, mask=causal_mask)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n # --- Temporal attention ---\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(56)\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z, mask=causal_mask)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n # --- Temporal attention ---\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(56)\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z, mask=causal_mask)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n # --- Temporal attention ---\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(56)\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z, mask=causal_mask)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n # --- Temporal attention ---\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(56)\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z, mask=causal_mask)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n # --- Temporal attention ---\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(56)\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z, mask=causal_mask)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n # --- Temporal attention ---\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(56)\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z, mask=causal_mask)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n # --- Temporal attention ---\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n",,terminal_output +2302,2329562,"TERMINAL",0,0,"[?25lx[?25h",,terminal_output +2303,2329774,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +2304,2329879,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +2305,2330014,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +2306,2330198,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +2307,2330406,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +2308,2330875,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +2309,2330936,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +2310,2331101,"TERMINAL",0,0,"\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n(jdb) (12, 16, 920, 512)\r\n",,terminal_output +2311,2332440,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +2312,2332500,"TERMINAL",0,0,"\r\n",,terminal_output +2313,2332627,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +2314,2332688,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +2315,2334391,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +2316,2334754,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(65)\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n-> jax.debug.breakpoint()\r\n causal_mask = jnp.tri(z.shape[-2])\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(65)\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n-> jax.debug.breakpoint()\r\n causal_mask = jnp.tri(z.shape[-2])\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(65)\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n-> jax.debug.breakpoint()\r\n causal_mask = jnp.tri(z.shape[-2])\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(65)\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n-> jax.debug.breakpoint()\r\n causal_mask = jnp.tri(z.shape[-2])\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(65)\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n-> jax.debug.breakpoint()\r\n causal_mask = jnp.tri(z.shape[-2])\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(65)\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n-> jax.debug.breakpoint()\r\n causal_mask = jnp.tri(z.shape[-2])\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(65)\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n-> jax.debug.breakpoint()\r\n causal_mask = jnp.tri(z.shape[-2])\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(65)\r\n x = x.swapaxes(1, 2)\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n-> jax.debug.breakpoint()\r\n causal_mask = jnp.tri(z.shape[-2])\r\n z = nn.MultiHeadAttention(\r\n num_heads=self.num_heads,\r\n qkv_features=self.dim,\r\n dropout_rate=self.dropout,\r\n",,terminal_output +2317,2347408,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +2318,2347549,"TERMINAL",0,0,"\r\n",,terminal_output +2319,2347714,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +2320,2350908,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +2321,2351225,"TERMINAL",0,0,"\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +2322,2352452,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +2323,2352635,"TERMINAL",0,0,"\r\n",,terminal_output +2324,2352696,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +2325,2353602,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +2326,2353739,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(84)\r\n \r\n # --- Feedforward ---\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n-> jax.debug.breakpoint()\r\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\r\n z = nn.Dense(\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(84)\r\n \r\n # --- Feedforward ---\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n-> jax.debug.breakpoint()\r\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\r\n z = nn.Dense(\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(84)\r\n \r\n # --- Feedforward ---\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n-> jax.debug.breakpoint()\r\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\r\n z = nn.Dense(\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(84)\r\n \r\n # --- Feedforward ---\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n-> jax.debug.breakpoint()\r\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\r\n z = nn.Dense(\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(84)\r\n \r\n # --- Feedforward ---\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n-> jax.debug.breakpoint()\r\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\r\n z = nn.Dense(\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(84)\r\n \r\n # --- Feedforward ---\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n-> jax.debug.breakpoint()\r\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\r\n z = nn.Dense(\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(84)\r\n \r\n # --- Feedforward ---\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n-> jax.debug.breakpoint()\r\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\r\n z = nn.Dense(\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(84)\r\n \r\n # --- Feedforward ---\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(x)\r\n-> jax.debug.breakpoint()\r\n # FIXME (f.srambical): Here, the attention hidden dimension is the same as the FFN's. Usually, FFN hidden dimension is 4x model_dim\r\n z = nn.Dense(\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n",,terminal_output +2327,2354972,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +2328,2355128,"TERMINAL",0,0,"\r\n",,terminal_output +2329,2355243,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +2330,2355799,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +2331,2355976,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(93)\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n z = nn.gelu(z)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n return x\r\n \r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(93)\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n z = nn.gelu(z)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n return x\r\n \r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(93)\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n z = nn.gelu(z)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n return x\r\n \r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(93)\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n z = nn.gelu(z)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n return x\r\n \r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(93)\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n z = nn.gelu(z)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n return x\r\n \r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(93)\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n z = nn.gelu(z)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n return x\r\n \r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(93)\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n z = nn.gelu(z)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n return x\r\n \r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(93)\r\n self.dim,\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n z = nn.gelu(z)\r\n x = x + z\r\n-> jax.debug.breakpoint()\r\n \r\n return x\r\n \r\n",,terminal_output +2332,2356805,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +2333,2357075,"TERMINAL",0,0,"\r\n",,terminal_output +2334,2357443,"TERMINAL",0,0,"(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n(jdb) Entering jdb:\r\n",,terminal_output +2335,2358741,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +2336,2358918,"TERMINAL",0,0,"\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n(jdb) > /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py(40)\r\n @nn.remat\r\n @nn.compact\r\n def __call__(self, x: jax.Array) -> jax.Array:\r\n # --- Spatial attention ---\r\n-> jax.debug.breakpoint()\r\n z = PositionalEncoding(self.dim)(x)\r\n z = nn.LayerNorm(\r\n param_dtype=self.param_dtype,\r\n dtype=self.dtype,\r\n )(z)\r\n",,terminal_output +2337,2362132,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",0,0,"",python,tab +2338,2362133,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2962,0,"",python,selection_mouse +2339,2362858,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2944,31,"",python,content +2340,2362937,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2926,0,"",python,selection_command +2341,2363108,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2903,0,"",python,selection_command +2342,2363239,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2890,0,"",python,selection_command +2343,2363407,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2860,0,"",python,selection_command +2344,2363537,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2818,0,"",python,selection_command +2345,2363671,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2796,0,"",python,selection_command +2346,2363914,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2774,0,"",python,selection_command +2347,2364155,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2634,0,"",python,selection_command +2348,2364297,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2603,0,"",python,selection_command +2349,2364427,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2590,0,"",python,selection_command +2350,2364739,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2603,0,"",python,selection_command +2351,2365112,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2603,31,"",python,content +2352,2365155,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2611,0,"",python,selection_command +2353,2365156,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2598,0,"",python,selection_command +2354,2365335,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2568,0,"",python,selection_command +2355,2365458,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2526,0,"",python,selection_command +2356,2365598,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2500,0,"",python,selection_command +2357,2365743,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2470,0,"",python,selection_command +2358,2366065,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2461,0,"",python,selection_command +2359,2366209,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2438,0,"",python,selection_command +2360,2366604,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2430,31,"",python,content +2361,2366711,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2401,0,"",python,selection_command +2362,2366865,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2370,0,"",python,selection_command +2363,2367359,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2370,31,"",python,content +2364,2367451,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2378,0,"",python,selection_command +2365,2367452,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2360,0,"",python,selection_command +2366,2367568,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2329,0,"",python,selection_command +2367,2367748,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2299,0,"",python,selection_command +2368,2367908,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2257,0,"",python,selection_command +2369,2368057,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2218,0,"",python,selection_command +2370,2368188,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2183,0,"",python,selection_command +2371,2368333,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2145,0,"",python,selection_command +2372,2368540,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2110,0,"",python,selection_command +2373,2368634,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2067,0,"",python,selection_command +2374,2368805,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2036,0,"",python,selection_command +2375,2369184,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2028,31,"",python,content +2376,2369250,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",2036,0,"",python,selection_command +2377,2371149,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",1793,0,"",python,selection_mouse +2378,2371637,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",1775,31,"",python,content +2379,2372617,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",1493,0,"",python,selection_mouse +2380,2373099,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",1476,31,"",python,content +2381,2373123,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",1484,0,"",python,selection_command +2382,2374847,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py",1268,0,"",python,selection_mouse diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-69e8c2c1-3349-4fa0-950c-3777ceb9c3751758621585749-2025_09_23-12.01.04.60/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-69e8c2c1-3349-4fa0-950c-3777ceb9c3751758621585749-2025_09_23-12.01.04.60/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..e12104f1135996a6ad84990a74c9985daf2a8baf --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-69e8c2c1-3349-4fa0-950c-3777ceb9c3751758621585749-2025_09_23-12.01.04.60/source.csv @@ -0,0 +1,714 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +2,416,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:01:04 PM [info] Activating crowd-code\n12:01:04 PM [info] Recording started\n12:01:04 PM [info] Initializing git provider using file system watchers...\n12:01:04 PM [info] Git repository found\n12:01:04 PM [info] Git provider initialized successfully\n12:01:04 PM [info] Initial git state: [object Object]\n",Log,tab +3,1615,"extension-output-pdoom-org.crowd-code-#1-crowd-code",41,0,"",Log,selection_command +4,2200,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"",Log,selection_command +5,3789,"TERMINAL",0,0,"bash",,terminal_focus +6,4679,"TERMINAL",0,0,"bash",,terminal_focus +7,6266,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +8,6324,"TERMINAL",0,0,"]633;Csalloc: Granted job allocation 3512647\r\nsalloc: Waiting for resource configuration\r\n",,terminal_output +9,32134,"TERMINAL",0,0,"s",,terminal_output +10,32215,"TERMINAL",0,0,"o",,terminal_output +11,32340,"TERMINAL",0,0,"u",,terminal_output +12,32392,"TERMINAL",0,0,"r",,terminal_output +13,32598,"TERMINAL",0,0,"c",,terminal_output +14,32705,"TERMINAL",0,0,"e",,terminal_output +15,32864,"TERMINAL",0,0," ",,terminal_output +16,32944,"TERMINAL",0,0,".",,terminal_output +17,33133,"TERMINAL",0,0,"salloc: Nodes hkn0403 are ready for job\r\n",,terminal_output +18,33288,"TERMINAL",0,0,"source .v",,terminal_output +19,33680,"TERMINAL",0,0,"e",,terminal_output +20,33800,"TERMINAL",0,0,"n",,terminal_output +21,34039,"TERMINAL",0,0,"]0;tum_cte0515@hkn0403:~/Projects/jasmine[?2004h[tum_cte0515@hkn0403 jasmine]$ source .venv",,terminal_output +22,34813,"TERMINAL",0,0,"/",,terminal_output +23,35212,"TERMINAL",0,0,"b",,terminal_output +24,35329,"TERMINAL",0,0,"in/",,terminal_output +25,35655,"TERMINAL",0,0,"a",,terminal_output +26,35707,"TERMINAL",0,0,"c",,terminal_output +27,35925,"TERMINAL",0,0,"tivate",,terminal_output +28,36223,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0403:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine]$ ",,terminal_output +29,36824,"TERMINAL",0,0,"ls",,terminal_output +30,36935,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +31,37086,"TERMINAL",0,0,"debug frame-knoms.png frames gifs killer_partition.sh LICENSE logs overfit_dir __pycache__ requirements-franz.txt sample.py scripts_cremers test.py train_dynamics.py train_tokenizer.py wandb\r\ndiff.diff frame.png genie.py input_pipeline killer.sh log.log models overfit_dir.zip README.md requirements.txt samples slurm tests train_lam.py utils\r\n]0;tum_cte0515@hkn0403:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine]$ ",,terminal_output +32,37532,"TERMINAL",0,0,"qu",,terminal_output +33,37725,"TERMINAL",0,0,"eu",,terminal_output +34,37858,"TERMINAL",0,0,"e",,terminal_output +35,38234,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: squeue --mehkn0403.localdomain: Tue Sep 23 12:01:41 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3511158 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3511159 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3512647 dev_accel interact tum_cte0 R\t0:33\t 1 hkn0403",,terminal_output +36,40140,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0403:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine]$ ",,terminal_output +37,48816,"TERMINAL",0,0,"bash",,terminal_focus +38,49743,"TERMINAL",0,0,"ls",,terminal_command +39,53573,"TERMINAL",0,0,"cd $ws_dir",,terminal_command +40,54015,"TERMINAL",0,0,"ls",,terminal_command +41,54056,"TERMINAL",0,0,"]633;C",,terminal_output +42,54145,"TERMINAL",0,0,"checkpoints count_items.sh data data_atari data_breakout data_coinrun data_minecraft data_new huggingface logs possibly_corrupt_files_in_this_workspace.txt scripts\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared",,terminal_output +43,60566,"TERMINAL",0,0,"cd data_breakout/",,terminal_command +44,60930,"TERMINAL",0,0,"ls",,terminal_command +45,60966,"TERMINAL",0,0,"]633;C",,terminal_output +46,60983,"TERMINAL",0,0,"breakout_episodes_10m_gt_actions_split breakout_episodes_10m_gt_actions_split_2 breakout_episodes_10m_gt_actions_split_old breakout_episodes_10m_gt_actions_split_perfect breakout_episodes_10m_gt_actions_split_small breakout_episodes_perfect\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout",,terminal_output +47,65861,"TERMINAL",0,0,"cd breakout_episodes_perfect/",,terminal_command +48,66291,"TERMINAL",0,0,"ls",,terminal_command +49,66306,"TERMINAL",0,0,"]633;Cmetadata.json test train val\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect",,terminal_output +50,69943,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/metadata.json",0,0,"{""env"": ""MinAtar-Breakout"", ""num_actions"": 6, ""num_episodes_train"": 500000, ""num_episodes_val"": 25000, ""num_episodes_test"": 25000, ""avg_episode_len_train"": 16.0, ""avg_episode_len_val"": 16.0, ""avg_episode_len_test"": 16.0, ""episode_metadata_train"": [{""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0000.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0001.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0002.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0003.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0004.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0005.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0006.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0007.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0008.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0009.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0010.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0011.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0012.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0013.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0014.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0015.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0016.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0017.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0018.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0019.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0020.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0021.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0022.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0023.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0024.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0025.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0026.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0027.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0028.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0029.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0030.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0031.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0032.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0033.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0034.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0035.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0036.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0037.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0038.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0039.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0040.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0041.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0042.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0043.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0044.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0045.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0046.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0047.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0048.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0049.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0050.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0051.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0052.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0053.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0054.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0055.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0056.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0057.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0058.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0059.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0060.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0061.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0062.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0063.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0064.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0065.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0066.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0067.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0068.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0069.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0070.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0071.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0072.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0073.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0074.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0075.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0076.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0077.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0078.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0079.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0080.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0081.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0082.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0083.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0084.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0085.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0086.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0087.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0088.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0089.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0090.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0091.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0092.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0093.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0094.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0095.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0096.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0097.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0098.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0099.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0100.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0101.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0102.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0103.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0104.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0105.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0106.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0107.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0108.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0109.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0110.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0111.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0112.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0113.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0114.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0115.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0116.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0117.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0118.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0119.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0120.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0121.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0122.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0123.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0124.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0125.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0126.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0127.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0128.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0129.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0130.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0131.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0132.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0133.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0134.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0135.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0136.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0137.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0138.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0139.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0140.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0141.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0142.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0143.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0144.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0145.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0146.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0147.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0148.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0149.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0150.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0151.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0152.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0153.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0154.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0155.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0156.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0157.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0158.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0159.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0160.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0161.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0162.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0163.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0164.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0165.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0166.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0167.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0168.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0169.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0170.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0171.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0172.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0173.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0174.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0175.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0176.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0177.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0178.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0179.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0180.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0181.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0182.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0183.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0184.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0185.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0186.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0187.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0188.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0189.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0190.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0191.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0192.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0193.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0194.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0195.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0196.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0197.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0198.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0199.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0200.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0201.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0202.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0203.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0204.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0205.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0206.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0207.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0208.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0209.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0210.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0211.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0212.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0213.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0214.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0215.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0216.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0217.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0218.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0219.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0220.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0221.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0222.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0223.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0224.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0225.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0226.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0227.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0228.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0229.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0230.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0231.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0232.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0233.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0234.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0235.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0236.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0237.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0238.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0239.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0240.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0241.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0242.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0243.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0244.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0245.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0246.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0247.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0248.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0249.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0250.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0251.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0252.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0253.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0254.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0255.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0256.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0257.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0258.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0259.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0260.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0261.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0262.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0263.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0264.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0265.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0266.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0267.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0268.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0269.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0270.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0271.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0272.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0273.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0274.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0275.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0276.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0277.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0278.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0279.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0280.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0281.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0282.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0283.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0284.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0285.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0286.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0287.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0288.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0289.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0290.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0291.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0292.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0293.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0294.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0295.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0296.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0297.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0298.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0299.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0300.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0301.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0302.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0303.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0304.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0305.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0306.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0307.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0308.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0309.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0310.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0311.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0312.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0313.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0314.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0315.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0316.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0317.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0318.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0319.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0320.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0321.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0322.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0323.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0324.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0325.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0326.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0327.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0328.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0329.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0330.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0331.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0332.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0333.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0334.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0335.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0336.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0337.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0338.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0339.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0340.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0341.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0342.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0343.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0344.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0345.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0346.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0347.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0348.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0349.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0350.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0351.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0352.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0353.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0354.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0355.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0356.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0357.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0358.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0359.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0360.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0361.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0362.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0363.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0364.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0365.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0366.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0367.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0368.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0369.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0370.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0371.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0372.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0373.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0374.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0375.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0376.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0377.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0378.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0379.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0380.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0381.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0382.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0383.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0384.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0385.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0386.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0387.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0388.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0389.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0390.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0391.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0392.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0393.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0394.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0395.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0396.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0397.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0398.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0399.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0400.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0401.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0402.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0403.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0404.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0405.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0406.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0407.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0408.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0409.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0410.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0411.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0412.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0413.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0414.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0415.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0416.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0417.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0418.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0419.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0420.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0421.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0422.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0423.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0424.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0425.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0426.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0427.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0428.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0429.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0430.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0431.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0432.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0433.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0434.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0435.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0436.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0437.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0438.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0439.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0440.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0441.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0442.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0443.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0444.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0445.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0446.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0447.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0448.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0449.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0450.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0451.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0452.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0453.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0454.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0455.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0456.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0457.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0458.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0459.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0460.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0461.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0462.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0463.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0464.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0465.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0466.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0467.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0468.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0469.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0470.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0471.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0472.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0473.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0474.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0475.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0476.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0477.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0478.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0479.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0480.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0481.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0482.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0483.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0484.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0485.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0486.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0487.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0488.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0489.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0490.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0491.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0492.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0493.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0494.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0495.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0496.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0497.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0498.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0499.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0500.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0501.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0502.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0503.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0504.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0505.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0506.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0507.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0508.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0509.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0510.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0511.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0512.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0513.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0514.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0515.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0516.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0517.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0518.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0519.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0520.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0521.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0522.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0523.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0524.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0525.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0526.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0527.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0528.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0529.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0530.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0531.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0532.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0533.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0534.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0535.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0536.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0537.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0538.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0539.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0540.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0541.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0542.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0543.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0544.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0545.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0546.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0547.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0548.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0549.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0550.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0551.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0552.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0553.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0554.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0555.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0556.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0557.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0558.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0559.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0560.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0561.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0562.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0563.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0564.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0565.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0566.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0567.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0568.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0569.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0570.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0571.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0572.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0573.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0574.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0575.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0576.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0577.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0578.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0579.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0580.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0581.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0582.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0583.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0584.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0585.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0586.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0587.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0588.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0589.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0590.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0591.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0592.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0593.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0594.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0595.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0596.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0597.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0598.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0599.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0600.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0601.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0602.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0603.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0604.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0605.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0606.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0607.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0608.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0609.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0610.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0611.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0612.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0613.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0614.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0615.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0616.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0617.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0618.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0619.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0620.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0621.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0622.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0623.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0624.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0625.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0626.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0627.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0628.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0629.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0630.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0631.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0632.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0633.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0634.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0635.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0636.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0637.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0638.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0639.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0640.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0641.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0642.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0643.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0644.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0645.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0646.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0647.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0648.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0649.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0650.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0651.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0652.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0653.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0654.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0655.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0656.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0657.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0658.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0659.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0660.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0661.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0662.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0663.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0664.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0665.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0666.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0667.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0668.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0669.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0670.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0671.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0672.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0673.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0674.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0675.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0676.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0677.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0678.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0679.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0680.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0681.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0682.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0683.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0684.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0685.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0686.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0687.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0688.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0689.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0690.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0691.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0692.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0693.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0694.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0695.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0696.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0697.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0698.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0699.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0700.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0701.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0702.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0703.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0704.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0705.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0706.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0707.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0708.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0709.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0710.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0711.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0712.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0713.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0714.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0715.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0716.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0717.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0718.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0719.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0720.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0721.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0722.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0723.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0724.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0725.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0726.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0727.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0728.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0729.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0730.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0731.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0732.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0733.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0734.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0735.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0736.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0737.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0738.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0739.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0740.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0741.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0742.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0743.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0744.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0745.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0746.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0747.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0748.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0749.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0750.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0751.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0752.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0753.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0754.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0755.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0756.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0757.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0758.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0759.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0760.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0761.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0762.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0763.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0764.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0765.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0766.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0767.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0768.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0769.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0770.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0771.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0772.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0773.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0774.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0775.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0776.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0777.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0778.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0779.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0780.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0781.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0782.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0783.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0784.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0785.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0786.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0787.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0788.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0789.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0790.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0791.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0792.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0793.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0794.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0795.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0796.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0797.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0798.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0799.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0800.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0801.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0802.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0803.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0804.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0805.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0806.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0807.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0808.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0809.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0810.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0811.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0812.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0813.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0814.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0815.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0816.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0817.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0818.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0819.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0820.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0821.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0822.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0823.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0824.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0825.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0826.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0827.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0828.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0829.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0830.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0831.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0832.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0833.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0834.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0835.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0836.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0837.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0838.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0839.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0840.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0841.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0842.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0843.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0844.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0845.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0846.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0847.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0848.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0849.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0850.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0851.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0852.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0853.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0854.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0855.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0856.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0857.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0858.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0859.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0860.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0861.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0862.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0863.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0864.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0865.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0866.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0867.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0868.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0869.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0870.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0871.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0872.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0873.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0874.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0875.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0876.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0877.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0878.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0879.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0880.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0881.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0882.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0883.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0884.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0885.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0886.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0887.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0888.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0889.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0890.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0891.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0892.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0893.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0894.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0895.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0896.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0897.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0898.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0899.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0900.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0901.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0902.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0903.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0904.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0905.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0906.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0907.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0908.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0909.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0910.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0911.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0912.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0913.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0914.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0915.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0916.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0917.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0918.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0919.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0920.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0921.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0922.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0923.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0924.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0925.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0926.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0927.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0928.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0929.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0930.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0931.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0932.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0933.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0934.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0935.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0936.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0937.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0938.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0939.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0940.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0941.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0942.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0943.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0944.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0945.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0946.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0947.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0948.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0949.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0950.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0951.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0952.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0953.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0954.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0955.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0956.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0957.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0958.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0959.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0960.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0961.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0962.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0963.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0964.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0965.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0966.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0967.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0968.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0969.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0970.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0971.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0972.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0973.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0974.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0975.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0976.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0977.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0978.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0979.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0980.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0981.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0982.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0983.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0984.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0985.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0986.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0987.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0988.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0989.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0990.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0991.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0992.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0993.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0994.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0995.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0996.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0997.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0998.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0999.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1000.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1001.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1002.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1003.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1004.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1005.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1006.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1007.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1008.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1009.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1010.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1011.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1012.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1013.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1014.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1015.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1016.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1017.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1018.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1019.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1020.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1021.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1022.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1023.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1024.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1025.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1026.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1027.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1028.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1029.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1030.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1031.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1032.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1033.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1034.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1035.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1036.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1037.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1038.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1039.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1040.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1041.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1042.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1043.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1044.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1045.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1046.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1047.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1048.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1049.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1050.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1051.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1052.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1053.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1054.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1055.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1056.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1057.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1058.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1059.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1060.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1061.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1062.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1063.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1064.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1065.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1066.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1067.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1068.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1069.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1070.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1071.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1072.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1073.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1074.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1075.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1076.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1077.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1078.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1079.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1080.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1081.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1082.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1083.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1084.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1085.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1086.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1087.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1088.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1089.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1090.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1091.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1092.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1093.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1094.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1095.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1096.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1097.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1098.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1099.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1100.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1101.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1102.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1103.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1104.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1105.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1106.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1107.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1108.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1109.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1110.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1111.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1112.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1113.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1114.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1115.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1116.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1117.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1118.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1119.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1120.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1121.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1122.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1123.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1124.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1125.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1126.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1127.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1128.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1129.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1130.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1131.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1132.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1133.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1134.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1135.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1136.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1137.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1138.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1139.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1140.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1141.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1142.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1143.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1144.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1145.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1146.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1147.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1148.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1149.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1150.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1151.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1152.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1153.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1154.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1155.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1156.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1157.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1158.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1159.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1160.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1161.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1162.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1163.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1164.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1165.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1166.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1167.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1168.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1169.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1170.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1171.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1172.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1173.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1174.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1175.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1176.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1177.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1178.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1179.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1180.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1181.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1182.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1183.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1184.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1185.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1186.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1187.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1188.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1189.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1190.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1191.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1192.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1193.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1194.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1195.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1196.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1197.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1198.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1199.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1200.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1201.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1202.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1203.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1204.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1205.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1206.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1207.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1208.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1209.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1210.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1211.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1212.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1213.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1214.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1215.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1216.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1217.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1218.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1219.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1220.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1221.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1222.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1223.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1224.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1225.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1226.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1227.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1228.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1229.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1230.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1231.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1232.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1233.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1234.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1235.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1236.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1237.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1238.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1239.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1240.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1241.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1242.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1243.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1244.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1245.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1246.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1247.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1248.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1249.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1250.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1251.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1252.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1253.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1254.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1255.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1256.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1257.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1258.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1259.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1260.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1261.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1262.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1263.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1264.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1265.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1266.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1267.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1268.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1269.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1270.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1271.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1272.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1273.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1274.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1275.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1276.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1277.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1278.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1279.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1280.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1281.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1282.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1283.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1284.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1285.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1286.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1287.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1288.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1289.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1290.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1291.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1292.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1293.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1294.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1295.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1296.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1297.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1298.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1299.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1300.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1301.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1302.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1303.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1304.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1305.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1306.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1307.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1308.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1309.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1310.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1311.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1312.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1313.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1314.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1315.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1316.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1317.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1318.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1319.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1320.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1321.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1322.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1323.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1324.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1325.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1326.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1327.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1328.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1329.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1330.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1331.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1332.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1333.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1334.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1335.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1336.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1337.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1338.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1339.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1340.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1341.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1342.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1343.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1344.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1345.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1346.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1347.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1348.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1349.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1350.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1351.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1352.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1353.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1354.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1355.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1356.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1357.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1358.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1359.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1360.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1361.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1362.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1363.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1364.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1365.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1366.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1367.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1368.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1369.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1370.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1371.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1372.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1373.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1374.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1375.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1376.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1377.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1378.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1379.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1380.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1381.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1382.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1383.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1384.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1385.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1386.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1387.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1388.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1389.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1390.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1391.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1392.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1393.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1394.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1395.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1396.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1397.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1398.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1399.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1400.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1401.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1402.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1403.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1404.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1405.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1406.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1407.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1408.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1409.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1410.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1411.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1412.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1413.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1414.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1415.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1416.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1417.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1418.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1419.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1420.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1421.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1422.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1423.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1424.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1425.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1426.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1427.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1428.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1429.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1430.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1431.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1432.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1433.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1434.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1435.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1436.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1437.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1438.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1439.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1440.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1441.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1442.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1443.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1444.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1445.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1446.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1447.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1448.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1449.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1450.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1451.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1452.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1453.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1454.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1455.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1456.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1457.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1458.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1459.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1460.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1461.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1462.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1463.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1464.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1465.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1466.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1467.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1468.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1469.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1470.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1471.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1472.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1473.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1474.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1475.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1476.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1477.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1478.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1479.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1480.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1481.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1482.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1483.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1484.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1485.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1486.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1487.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1488.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1489.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1490.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1491.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1492.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1493.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1494.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1495.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1496.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1497.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1498.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1499.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1500.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1501.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1502.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1503.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1504.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1505.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1506.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1507.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1508.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1509.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1510.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1511.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1512.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1513.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1514.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1515.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1516.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1517.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1518.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1519.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1520.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1521.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1522.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1523.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1524.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1525.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1526.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1527.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1528.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1529.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1530.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1531.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1532.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1533.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1534.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1535.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1536.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1537.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1538.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1539.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1540.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1541.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1542.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1543.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1544.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1545.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1546.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1547.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1548.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1549.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1550.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1551.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1552.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1553.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1554.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1555.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1556.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1557.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1558.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1559.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1560.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1561.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1562.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1563.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1564.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1565.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1566.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1567.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1568.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1569.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1570.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1571.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1572.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1573.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1574.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1575.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1576.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1577.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1578.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1579.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1580.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1581.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1582.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1583.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1584.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1585.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1586.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1587.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1588.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1589.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1590.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1591.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1592.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1593.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1594.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1595.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1596.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1597.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1598.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1599.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1600.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1601.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1602.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1603.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1604.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1605.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1606.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1607.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1608.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1609.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1610.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1611.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1612.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1613.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1614.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1615.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1616.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1617.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1618.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1619.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1620.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1621.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1622.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1623.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1624.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1625.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1626.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1627.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1628.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1629.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1630.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1631.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1632.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1633.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1634.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1635.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1636.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1637.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1638.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1639.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1640.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1641.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1642.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1643.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1644.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1645.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1646.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1647.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1648.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1649.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1650.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1651.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1652.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1653.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1654.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1655.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1656.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1657.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1658.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1659.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1660.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1661.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1662.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1663.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1664.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1665.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1666.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1667.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1668.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1669.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1670.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1671.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1672.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1673.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1674.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1675.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1676.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1677.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1678.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1679.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1680.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1681.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1682.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1683.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1684.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1685.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1686.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1687.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1688.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1689.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1690.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1691.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1692.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1693.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1694.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1695.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1696.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1697.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1698.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1699.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1700.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1701.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1702.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1703.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1704.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1705.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1706.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1707.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1708.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1709.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1710.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1711.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1712.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1713.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1714.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1715.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1716.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1717.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1718.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1719.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1720.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1721.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1722.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1723.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1724.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1725.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1726.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1727.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1728.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1729.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1730.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1731.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1732.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1733.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1734.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1735.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1736.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1737.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1738.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1739.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1740.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1741.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1742.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1743.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1744.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1745.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1746.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1747.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1748.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1749.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1750.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1751.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1752.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1753.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1754.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1755.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1756.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1757.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1758.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1759.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1760.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1761.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1762.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1763.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1764.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1765.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1766.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1767.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1768.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1769.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1770.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1771.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1772.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1773.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1774.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1775.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1776.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1777.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1778.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1779.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1780.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1781.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1782.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1783.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1784.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1785.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1786.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1787.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1788.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1789.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1790.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1791.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1792.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1793.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1794.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1795.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1796.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1797.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1798.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1799.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1800.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1801.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1802.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1803.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1804.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1805.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1806.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1807.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1808.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1809.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1810.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1811.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1812.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1813.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1814.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1815.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1816.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1817.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1818.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1819.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1820.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1821.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1822.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1823.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1824.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1825.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1826.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1827.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1828.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1829.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1830.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1831.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1832.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1833.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1834.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1835.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1836.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1837.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1838.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1839.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1840.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1841.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1842.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1843.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1844.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1845.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1846.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1847.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1848.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1849.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1850.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1851.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1852.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1853.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1854.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1855.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1856.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1857.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1858.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1859.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1860.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1861.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1862.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1863.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1864.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1865.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1866.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1867.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1868.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1869.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1870.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1871.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1872.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1873.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1874.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1875.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1876.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1877.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1878.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1879.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1880.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1881.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1882.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1883.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1884.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1885.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1886.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1887.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1888.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1889.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1890.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1891.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1892.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1893.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1894.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1895.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1896.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1897.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1898.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1899.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1900.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1901.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1902.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1903.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1904.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1905.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1906.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1907.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1908.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1909.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1910.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1911.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1912.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1913.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1914.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1915.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1916.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1917.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1918.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1919.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1920.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1921.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1922.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1923.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1924.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1925.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1926.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1927.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1928.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1929.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1930.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1931.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1932.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1933.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1934.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1935.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1936.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1937.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1938.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1939.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1940.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1941.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1942.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1943.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1944.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1945.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1946.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1947.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1948.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1949.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1950.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1951.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1952.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1953.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1954.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1955.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1956.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1957.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1958.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1959.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1960.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1961.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1962.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1963.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1964.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1965.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1966.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1967.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1968.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1969.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1970.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1971.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1972.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1973.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1974.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1975.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1976.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1977.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1978.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1979.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1980.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1981.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1982.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1983.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1984.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1985.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1986.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1987.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1988.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1989.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1990.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1991.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1992.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1993.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1994.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1995.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1996.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1997.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1998.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1999.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2000.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2001.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2002.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2003.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2004.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2005.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2006.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2007.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2008.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2009.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2010.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2011.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2012.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2013.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2014.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2015.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2016.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2017.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2018.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2019.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2020.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2021.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2022.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2023.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2024.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2025.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2026.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2027.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2028.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2029.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2030.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2031.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2032.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2033.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2034.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2035.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2036.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2037.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2038.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2039.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2040.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2041.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2042.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2043.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2044.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2045.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2046.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2047.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2048.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2049.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2050.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2051.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2052.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2053.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2054.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2055.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2056.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2057.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2058.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2059.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2060.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2061.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2062.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2063.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2064.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2065.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2066.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2067.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2068.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2069.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2070.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2071.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2072.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2073.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2074.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2075.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2076.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2077.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2078.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2079.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2080.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2081.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2082.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2083.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2084.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2085.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2086.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2087.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2088.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2089.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2090.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2091.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2092.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2093.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2094.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2095.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2096.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2097.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2098.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2099.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2100.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2101.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2102.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2103.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2104.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2105.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2106.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2107.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2108.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2109.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2110.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2111.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2112.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2113.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2114.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2115.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2116.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2117.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2118.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2119.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2120.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2121.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2122.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2123.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2124.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2125.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2126.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2127.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2128.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2129.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2130.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2131.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2132.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2133.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2134.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2135.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2136.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2137.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2138.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2139.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2140.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2141.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2142.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2143.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2144.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2145.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2146.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2147.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2148.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2149.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2150.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2151.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2152.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2153.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2154.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2155.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2156.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2157.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2158.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2159.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2160.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2161.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2162.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2163.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2164.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2165.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2166.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2167.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2168.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2169.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2170.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2171.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2172.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2173.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2174.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2175.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2176.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2177.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2178.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2179.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2180.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2181.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2182.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2183.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2184.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2185.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2186.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2187.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2188.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2189.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2190.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2191.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2192.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2193.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2194.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2195.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2196.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2197.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2198.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2199.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2200.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2201.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2202.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2203.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2204.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2205.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2206.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2207.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2208.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2209.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2210.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2211.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2212.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2213.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2214.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2215.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2216.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2217.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2218.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2219.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2220.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2221.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2222.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2223.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2224.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2225.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2226.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2227.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2228.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2229.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2230.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2231.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2232.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2233.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2234.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2235.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2236.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2237.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2238.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2239.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2240.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2241.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2242.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2243.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2244.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2245.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2246.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2247.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2248.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2249.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2250.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2251.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2252.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2253.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2254.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2255.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2256.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2257.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2258.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2259.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2260.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2261.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2262.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2263.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2264.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2265.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2266.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2267.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2268.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2269.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2270.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2271.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2272.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2273.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2274.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2275.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2276.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2277.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2278.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2279.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2280.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2281.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2282.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2283.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2284.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2285.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2286.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2287.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2288.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2289.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2290.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2291.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2292.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2293.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2294.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2295.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2296.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2297.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2298.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2299.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2300.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2301.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2302.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2303.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2304.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2305.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2306.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2307.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2308.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2309.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2310.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2311.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2312.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2313.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2314.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2315.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2316.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2317.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2318.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2319.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2320.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2321.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2322.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2323.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2324.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2325.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2326.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2327.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2328.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2329.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2330.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2331.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2332.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2333.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2334.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2335.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2336.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2337.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2338.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2339.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2340.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2341.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2342.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2343.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2344.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2345.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2346.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2347.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2348.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2349.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2350.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2351.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2352.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2353.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2354.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2355.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2356.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2357.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2358.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2359.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2360.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2361.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2362.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2363.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2364.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2365.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2366.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2367.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2368.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2369.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2370.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2371.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2372.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2373.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2374.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2375.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2376.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2377.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2378.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2379.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2380.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2381.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2382.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2383.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2384.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2385.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2386.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2387.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2388.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2389.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2390.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2391.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2392.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2393.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2394.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2395.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2396.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2397.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2398.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2399.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2400.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2401.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2402.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2403.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2404.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2405.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2406.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2407.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2408.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2409.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2410.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2411.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2412.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2413.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2414.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2415.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2416.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2417.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2418.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2419.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2420.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2421.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2422.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2423.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2424.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2425.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2426.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2427.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2428.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2429.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2430.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2431.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2432.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2433.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2434.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2435.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2436.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2437.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2438.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2439.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2440.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2441.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2442.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2443.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2444.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2445.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2446.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2447.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2448.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2449.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2450.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2451.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2452.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2453.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2454.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2455.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2456.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2457.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2458.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2459.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2460.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2461.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2462.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2463.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2464.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2465.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2466.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2467.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2468.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2469.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2470.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2471.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2472.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2473.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2474.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2475.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2476.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2477.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2478.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2479.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2480.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2481.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2482.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2483.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2484.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2485.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2486.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2487.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2488.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2489.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2490.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2491.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2492.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2493.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2494.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2495.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2496.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2497.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2498.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2499.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2500.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2501.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2502.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2503.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2504.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2505.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2506.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2507.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2508.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2509.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2510.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2511.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2512.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2513.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2514.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2515.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2516.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2517.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2518.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2519.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2520.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2521.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2522.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2523.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2524.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2525.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2526.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2527.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2528.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2529.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2530.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2531.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2532.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2533.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2534.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2535.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2536.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2537.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2538.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2539.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2540.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2541.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2542.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2543.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2544.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2545.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2546.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2547.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2548.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2549.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2550.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2551.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2552.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2553.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2554.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2555.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2556.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2557.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2558.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2559.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2560.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2561.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2562.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2563.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2564.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2565.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2566.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2567.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2568.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2569.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2570.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2571.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2572.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2573.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2574.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2575.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2576.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2577.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2578.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2579.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2580.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2581.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2582.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2583.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2584.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2585.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2586.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2587.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2588.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2589.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2590.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2591.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2592.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2593.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2594.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2595.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2596.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2597.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2598.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2599.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2600.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2601.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2602.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2603.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2604.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2605.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2606.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2607.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2608.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2609.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2610.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2611.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2612.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2613.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2614.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2615.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2616.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2617.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2618.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2619.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2620.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2621.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2622.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2623.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2624.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2625.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2626.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2627.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2628.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2629.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2630.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2631.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2632.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2633.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2634.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2635.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2636.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2637.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2638.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2639.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2640.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2641.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2642.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2643.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2644.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2645.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2646.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2647.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2648.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2649.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2650.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2651.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2652.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2653.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2654.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2655.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2656.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2657.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2658.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2659.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2660.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2661.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2662.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2663.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2664.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2665.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2666.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2667.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2668.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2669.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2670.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2671.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2672.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2673.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2674.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2675.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2676.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2677.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2678.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2679.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2680.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2681.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2682.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2683.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2684.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2685.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2686.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2687.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2688.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2689.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2690.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2691.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2692.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2693.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2694.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2695.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2696.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2697.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2698.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2699.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2700.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2701.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2702.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2703.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2704.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2705.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2706.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2707.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2708.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2709.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2710.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2711.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2712.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2713.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2714.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2715.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2716.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2717.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2718.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2719.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2720.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2721.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2722.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2723.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2724.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2725.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2726.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2727.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2728.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2729.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2730.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2731.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2732.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2733.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2734.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2735.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2736.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2737.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2738.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2739.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2740.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2741.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2742.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2743.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2744.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2745.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2746.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2747.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2748.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2749.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2750.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2751.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2752.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2753.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2754.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2755.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2756.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2757.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2758.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2759.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2760.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2761.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2762.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2763.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2764.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2765.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2766.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2767.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2768.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2769.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2770.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2771.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2772.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2773.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2774.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2775.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2776.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2777.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2778.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2779.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2780.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2781.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2782.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2783.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2784.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2785.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2786.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2787.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2788.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2789.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2790.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2791.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2792.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2793.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2794.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2795.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2796.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2797.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2798.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2799.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2800.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2801.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2802.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2803.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2804.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2805.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2806.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2807.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2808.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2809.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2810.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2811.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2812.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2813.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2814.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2815.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2816.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2817.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2818.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2819.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2820.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2821.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2822.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2823.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2824.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2825.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2826.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2827.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2828.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2829.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2830.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2831.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2832.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2833.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2834.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2835.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2836.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2837.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2838.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2839.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2840.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2841.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2842.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2843.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2844.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2845.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2846.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2847.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2848.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2849.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2850.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2851.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2852.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2853.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2854.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2855.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2856.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2857.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2858.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2859.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2860.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2861.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2862.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2863.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2864.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2865.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2866.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2867.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2868.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2869.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2870.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2871.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2872.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2873.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2874.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2875.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2876.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2877.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2878.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2879.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2880.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2881.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2882.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2883.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2884.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2885.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2886.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2887.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2888.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2889.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2890.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2891.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2892.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2893.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2894.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2895.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2896.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2897.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2898.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2899.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2900.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2901.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2902.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2903.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2904.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2905.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2906.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2907.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2908.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2909.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2910.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2911.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2912.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2913.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2914.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2915.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2916.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2917.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2918.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2919.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2920.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2921.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2922.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2923.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2924.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2925.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2926.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2927.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2928.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2929.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2930.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2931.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2932.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2933.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2934.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2935.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2936.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2937.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2938.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2939.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2940.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2941.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2942.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2943.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2944.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2945.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2946.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2947.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2948.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2949.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2950.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2951.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2952.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2953.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2954.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2955.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2956.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2957.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2958.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2959.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2960.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2961.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2962.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2963.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2964.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2965.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2966.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2967.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2968.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2969.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2970.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2971.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2972.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2973.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2974.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2975.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2976.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2977.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2978.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2979.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2980.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2981.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2982.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2983.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2984.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2985.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2986.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2987.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2988.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2989.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2990.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2991.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2992.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2993.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2994.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2995.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2996.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2997.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2998.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2999.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3000.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3001.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3002.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3003.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3004.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3005.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3006.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3007.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3008.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3009.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3010.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3011.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3012.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3013.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3014.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3015.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3016.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3017.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3018.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3019.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3020.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3021.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3022.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3023.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3024.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3025.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3026.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3027.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3028.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3029.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3030.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3031.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3032.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3033.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3034.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3035.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3036.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3037.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3038.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3039.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3040.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3041.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3042.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3043.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3044.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3045.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3046.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3047.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3048.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3049.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3050.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3051.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3052.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3053.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3054.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3055.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3056.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3057.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3058.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3059.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3060.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3061.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3062.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3063.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3064.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3065.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3066.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3067.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3068.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3069.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3070.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3071.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3072.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3073.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3074.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3075.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3076.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3077.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3078.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3079.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3080.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3081.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3082.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3083.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3084.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3085.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3086.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3087.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3088.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3089.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3090.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3091.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3092.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3093.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3094.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3095.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3096.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3097.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3098.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3099.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3100.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3101.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3102.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3103.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3104.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3105.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3106.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3107.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3108.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3109.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3110.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3111.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3112.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3113.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3114.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3115.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3116.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3117.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3118.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3119.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3120.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3121.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3122.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3123.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3124.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3125.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3126.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3127.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3128.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3129.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3130.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3131.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3132.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3133.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3134.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3135.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3136.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3137.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3138.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3139.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3140.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3141.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3142.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3143.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3144.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3145.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3146.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3147.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3148.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3149.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3150.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3151.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3152.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3153.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3154.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3155.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3156.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3157.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3158.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3159.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3160.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3161.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3162.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3163.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3164.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3165.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3166.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3167.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3168.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3169.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3170.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3171.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3172.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3173.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3174.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3175.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3176.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3177.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3178.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3179.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3180.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3181.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3182.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3183.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3184.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3185.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3186.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3187.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3188.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3189.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3190.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3191.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3192.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3193.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3194.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3195.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3196.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3197.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3198.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3199.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3200.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3201.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3202.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3203.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3204.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3205.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3206.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3207.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3208.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3209.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3210.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3211.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3212.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3213.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3214.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3215.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3216.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3217.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3218.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3219.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3220.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3221.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3222.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3223.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3224.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3225.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3226.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3227.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3228.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3229.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3230.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3231.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3232.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3233.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3234.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3235.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3236.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3237.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3238.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3239.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3240.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3241.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3242.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3243.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3244.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3245.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3246.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3247.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3248.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3249.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3250.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3251.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3252.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3253.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3254.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3255.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3256.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3257.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3258.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3259.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3260.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3261.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3262.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3263.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3264.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3265.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3266.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3267.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3268.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3269.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3270.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3271.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3272.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3273.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3274.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3275.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3276.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3277.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3278.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3279.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3280.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3281.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3282.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3283.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3284.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3285.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3286.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3287.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3288.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3289.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3290.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3291.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3292.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3293.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3294.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3295.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3296.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3297.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3298.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3299.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3300.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3301.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3302.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3303.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3304.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3305.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3306.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3307.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3308.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3309.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3310.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3311.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3312.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3313.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3314.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3315.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3316.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3317.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3318.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3319.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3320.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3321.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3322.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3323.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3324.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3325.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3326.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3327.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3328.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3329.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3330.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3331.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3332.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3333.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3334.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3335.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3336.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3337.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3338.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3339.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3340.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3341.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3342.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3343.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3344.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3345.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3346.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3347.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3348.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3349.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3350.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3351.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3352.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3353.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3354.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3355.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3356.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3357.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3358.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3359.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3360.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3361.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3362.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3363.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3364.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3365.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3366.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3367.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3368.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3369.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3370.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3371.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3372.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3373.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3374.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3375.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3376.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3377.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3378.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3379.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3380.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3381.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3382.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3383.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3384.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3385.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3386.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3387.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3388.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3389.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3390.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3391.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3392.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3393.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3394.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3395.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3396.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3397.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3398.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3399.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3400.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3401.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3402.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3403.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3404.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3405.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3406.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3407.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3408.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3409.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3410.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3411.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3412.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3413.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3414.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3415.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3416.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3417.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3418.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3419.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3420.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3421.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3422.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3423.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3424.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3425.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3426.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3427.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3428.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3429.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3430.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3431.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3432.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3433.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3434.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3435.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3436.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3437.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3438.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3439.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3440.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3441.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3442.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3443.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3444.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3445.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3446.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3447.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3448.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3449.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3450.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3451.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3452.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3453.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3454.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3455.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3456.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3457.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3458.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3459.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3460.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3461.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3462.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3463.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3464.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3465.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3466.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3467.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3468.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3469.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3470.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3471.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3472.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3473.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3474.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3475.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3476.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3477.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3478.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3479.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3480.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3481.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3482.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3483.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3484.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3485.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3486.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3487.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3488.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3489.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3490.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3491.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3492.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3493.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3494.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3495.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3496.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3497.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3498.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3499.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3500.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3501.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3502.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3503.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3504.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3505.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3506.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3507.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3508.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3509.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3510.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3511.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3512.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3513.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3514.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3515.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3516.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3517.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3518.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3519.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3520.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3521.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3522.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3523.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3524.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3525.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3526.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3527.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3528.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3529.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3530.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3531.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3532.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3533.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3534.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3535.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3536.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3537.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3538.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3539.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3540.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3541.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3542.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3543.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3544.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3545.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3546.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3547.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3548.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3549.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3550.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3551.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3552.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3553.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3554.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3555.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3556.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3557.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3558.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3559.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3560.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3561.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3562.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3563.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3564.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3565.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3566.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3567.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3568.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3569.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3570.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3571.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3572.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3573.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3574.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3575.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3576.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3577.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3578.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3579.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3580.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3581.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3582.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3583.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3584.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3585.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3586.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3587.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3588.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3589.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3590.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3591.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3592.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3593.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3594.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3595.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3596.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3597.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3598.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3599.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3600.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3601.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3602.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3603.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3604.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3605.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3606.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3607.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3608.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3609.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3610.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3611.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3612.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3613.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3614.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3615.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3616.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3617.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3618.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3619.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3620.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3621.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3622.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3623.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3624.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3625.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3626.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3627.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3628.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3629.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3630.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3631.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3632.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3633.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3634.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3635.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3636.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3637.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3638.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3639.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3640.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3641.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3642.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3643.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3644.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3645.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3646.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3647.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3648.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3649.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3650.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3651.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3652.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3653.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3654.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3655.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3656.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3657.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3658.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3659.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3660.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3661.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3662.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3663.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3664.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3665.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3666.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3667.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3668.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3669.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3670.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3671.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3672.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3673.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3674.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3675.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3676.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3677.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3678.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3679.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3680.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3681.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3682.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3683.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3684.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3685.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3686.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3687.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3688.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3689.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3690.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3691.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3692.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3693.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3694.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3695.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3696.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3697.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3698.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3699.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3700.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3701.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3702.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3703.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3704.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3705.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3706.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3707.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3708.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3709.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3710.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3711.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3712.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3713.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3714.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3715.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3716.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3717.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3718.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3719.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3720.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3721.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3722.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3723.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3724.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3725.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3726.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3727.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3728.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3729.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3730.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3731.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3732.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3733.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3734.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3735.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3736.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3737.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3738.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3739.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3740.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3741.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3742.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3743.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3744.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3745.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3746.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3747.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3748.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3749.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3750.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3751.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3752.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3753.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3754.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3755.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3756.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3757.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3758.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3759.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3760.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3761.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3762.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3763.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3764.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3765.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3766.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3767.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3768.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3769.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3770.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3771.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3772.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3773.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3774.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3775.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3776.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3777.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3778.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3779.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3780.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3781.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3782.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3783.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3784.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3785.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3786.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3787.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3788.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3789.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3790.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3791.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3792.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3793.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3794.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3795.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3796.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3797.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3798.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3799.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3800.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3801.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3802.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3803.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3804.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3805.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3806.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3807.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3808.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3809.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3810.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3811.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3812.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3813.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3814.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3815.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3816.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3817.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3818.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3819.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3820.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3821.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3822.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3823.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3824.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3825.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3826.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3827.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3828.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3829.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3830.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3831.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3832.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3833.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3834.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3835.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3836.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3837.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3838.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3839.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3840.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3841.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3842.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3843.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3844.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3845.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3846.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3847.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3848.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3849.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3850.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3851.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3852.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3853.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3854.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3855.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3856.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3857.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3858.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3859.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3860.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3861.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3862.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3863.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3864.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3865.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3866.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3867.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3868.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3869.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3870.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3871.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3872.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3873.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3874.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3875.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3876.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3877.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3878.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3879.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3880.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3881.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3882.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3883.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3884.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3885.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3886.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3887.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3888.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3889.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3890.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3891.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3892.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3893.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3894.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3895.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3896.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3897.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3898.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3899.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3900.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3901.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3902.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3903.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3904.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3905.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3906.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3907.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3908.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3909.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3910.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3911.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3912.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3913.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3914.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3915.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3916.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3917.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3918.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3919.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3920.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3921.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3922.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3923.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3924.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3925.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3926.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3927.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3928.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3929.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3930.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3931.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3932.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3933.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3934.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3935.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3936.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3937.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3938.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3939.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3940.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3941.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3942.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3943.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3944.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3945.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3946.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3947.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3948.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3949.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3950.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3951.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3952.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3953.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3954.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3955.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3956.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3957.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3958.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3959.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3960.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3961.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3962.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3963.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3964.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3965.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3966.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3967.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3968.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3969.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3970.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3971.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3972.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3973.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3974.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3975.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3976.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3977.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3978.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3979.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3980.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3981.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3982.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3983.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3984.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3985.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3986.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3987.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3988.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3989.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3990.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3991.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3992.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3993.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3994.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3995.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3996.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3997.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3998.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3999.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4000.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4001.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4002.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4003.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4004.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4005.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4006.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4007.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4008.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4009.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4010.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4011.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4012.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4013.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4014.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4015.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4016.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4017.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4018.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4019.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4020.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4021.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4022.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4023.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4024.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4025.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4026.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4027.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4028.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4029.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4030.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4031.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4032.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4033.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4034.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4035.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4036.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4037.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4038.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4039.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4040.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4041.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4042.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4043.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4044.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4045.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4046.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4047.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4048.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4049.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4050.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4051.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4052.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4053.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4054.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4055.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4056.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4057.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4058.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4059.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4060.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4061.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4062.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4063.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4064.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4065.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4066.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4067.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4068.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4069.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4070.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4071.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4072.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4073.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4074.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4075.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4076.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4077.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4078.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4079.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4080.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4081.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4082.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4083.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4084.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4085.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4086.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4087.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4088.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4089.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4090.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4091.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4092.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4093.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4094.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4095.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4096.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4097.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4098.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4099.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4100.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4101.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4102.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4103.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4104.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4105.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4106.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4107.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4108.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4109.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4110.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4111.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4112.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4113.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4114.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4115.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4116.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4117.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4118.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4119.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4120.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4121.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4122.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4123.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4124.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4125.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4126.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4127.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4128.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4129.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4130.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4131.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4132.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4133.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4134.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4135.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4136.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4137.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4138.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4139.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4140.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4141.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4142.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4143.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4144.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4145.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4146.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4147.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4148.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4149.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4150.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4151.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4152.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4153.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4154.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4155.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4156.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4157.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4158.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4159.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4160.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4161.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4162.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4163.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4164.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4165.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4166.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4167.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4168.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4169.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4170.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4171.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4172.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4173.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4174.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4175.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4176.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4177.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4178.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4179.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4180.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4181.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4182.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4183.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4184.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4185.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4186.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4187.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4188.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4189.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4190.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4191.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4192.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4193.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4194.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4195.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4196.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4197.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4198.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4199.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4200.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4201.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4202.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4203.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4204.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4205.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4206.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4207.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4208.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4209.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4210.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4211.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4212.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4213.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4214.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4215.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4216.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4217.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4218.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4219.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4220.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4221.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4222.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4223.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4224.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4225.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4226.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4227.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4228.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4229.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4230.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4231.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4232.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4233.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4234.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4235.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4236.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4237.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4238.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4239.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4240.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4241.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4242.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4243.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4244.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4245.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4246.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4247.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4248.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4249.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4250.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4251.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4252.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4253.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4254.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4255.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4256.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4257.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4258.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4259.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4260.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4261.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4262.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4263.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4264.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4265.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4266.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4267.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4268.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4269.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4270.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4271.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4272.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4273.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4274.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4275.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4276.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4277.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4278.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4279.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4280.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4281.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4282.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4283.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4284.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4285.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4286.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4287.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4288.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4289.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4290.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4291.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4292.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4293.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4294.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4295.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4296.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4297.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4298.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4299.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4300.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4301.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4302.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4303.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4304.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4305.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4306.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4307.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4308.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4309.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4310.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4311.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4312.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4313.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4314.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4315.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4316.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4317.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4318.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4319.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4320.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4321.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4322.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4323.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4324.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4325.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4326.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4327.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4328.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4329.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4330.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4331.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4332.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4333.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4334.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4335.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4336.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4337.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4338.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4339.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4340.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4341.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4342.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4343.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4344.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4345.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4346.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4347.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4348.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4349.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4350.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4351.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4352.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4353.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4354.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4355.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4356.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4357.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4358.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4359.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4360.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4361.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4362.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4363.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4364.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4365.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4366.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4367.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4368.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4369.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4370.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4371.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4372.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4373.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4374.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4375.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4376.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4377.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4378.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4379.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4380.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4381.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4382.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4383.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4384.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4385.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4386.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4387.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4388.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4389.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4390.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4391.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4392.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4393.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4394.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4395.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4396.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4397.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4398.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4399.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4400.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4401.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4402.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4403.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4404.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4405.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4406.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4407.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4408.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4409.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4410.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4411.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4412.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4413.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4414.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4415.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4416.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4417.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4418.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4419.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4420.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4421.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4422.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4423.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4424.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4425.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4426.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4427.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4428.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4429.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4430.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4431.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4432.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4433.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4434.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4435.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4436.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4437.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4438.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4439.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4440.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4441.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4442.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4443.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4444.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4445.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4446.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4447.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4448.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4449.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4450.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4451.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4452.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4453.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4454.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4455.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4456.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4457.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4458.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4459.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4460.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4461.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4462.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4463.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4464.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4465.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4466.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4467.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4468.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4469.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4470.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4471.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4472.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4473.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4474.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4475.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4476.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4477.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4478.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4479.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4480.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4481.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4482.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4483.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4484.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4485.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4486.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4487.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4488.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4489.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4490.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4491.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4492.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4493.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4494.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4495.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4496.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4497.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4498.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4499.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4500.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4501.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4502.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4503.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4504.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4505.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4506.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4507.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4508.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4509.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4510.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4511.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4512.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4513.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4514.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4515.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4516.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4517.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4518.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4519.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4520.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4521.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4522.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4523.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4524.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4525.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4526.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4527.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4528.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4529.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4530.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4531.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4532.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4533.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4534.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4535.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4536.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4537.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4538.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4539.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4540.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4541.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4542.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4543.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4544.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4545.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4546.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4547.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4548.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4549.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4550.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4551.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4552.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4553.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4554.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4555.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4556.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4557.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4558.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4559.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4560.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4561.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4562.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4563.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4564.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4565.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4566.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4567.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4568.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4569.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4570.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4571.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4572.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4573.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4574.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4575.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4576.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4577.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4578.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4579.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4580.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4581.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4582.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4583.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4584.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4585.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4586.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4587.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4588.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4589.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4590.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4591.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4592.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4593.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4594.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4595.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4596.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4597.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4598.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4599.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4600.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4601.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4602.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4603.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4604.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4605.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4606.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4607.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4608.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4609.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4610.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4611.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4612.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4613.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4614.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4615.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4616.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4617.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4618.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4619.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4620.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4621.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4622.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4623.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4624.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4625.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4626.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4627.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4628.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4629.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4630.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4631.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4632.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4633.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4634.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4635.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4636.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4637.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4638.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4639.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4640.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4641.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4642.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4643.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4644.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4645.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4646.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4647.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4648.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4649.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4650.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4651.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4652.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4653.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4654.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4655.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4656.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4657.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4658.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4659.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4660.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4661.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4662.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4663.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4664.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4665.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4666.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4667.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4668.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4669.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4670.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4671.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4672.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4673.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4674.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4675.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4676.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4677.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4678.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4679.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4680.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4681.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4682.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4683.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4684.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4685.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4686.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4687.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4688.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4689.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4690.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4691.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4692.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4693.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4694.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4695.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4696.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4697.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4698.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4699.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4700.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4701.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4702.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4703.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4704.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4705.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4706.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4707.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4708.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4709.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4710.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4711.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4712.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4713.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4714.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4715.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4716.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4717.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4718.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4719.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4720.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4721.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4722.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4723.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4724.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4725.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4726.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4727.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4728.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4729.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4730.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4731.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4732.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4733.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4734.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4735.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4736.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4737.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4738.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4739.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4740.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4741.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4742.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4743.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4744.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4745.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4746.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4747.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4748.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4749.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4750.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4751.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4752.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4753.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4754.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4755.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4756.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4757.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4758.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4759.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4760.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4761.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4762.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4763.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4764.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4765.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4766.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4767.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4768.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4769.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4770.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4771.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4772.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4773.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4774.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4775.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4776.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4777.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4778.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4779.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4780.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4781.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4782.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4783.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4784.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4785.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4786.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4787.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4788.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4789.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4790.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4791.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4792.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4793.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4794.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4795.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4796.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4797.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4798.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4799.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4800.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4801.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4802.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4803.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4804.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4805.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4806.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4807.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4808.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4809.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4810.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4811.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4812.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4813.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4814.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4815.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4816.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4817.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4818.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4819.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4820.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4821.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4822.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4823.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4824.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4825.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4826.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4827.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4828.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4829.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4830.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4831.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4832.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4833.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4834.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4835.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4836.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4837.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4838.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4839.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4840.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4841.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4842.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4843.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4844.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4845.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4846.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4847.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4848.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4849.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4850.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4851.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4852.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4853.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4854.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4855.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4856.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4857.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4858.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4859.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4860.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4861.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4862.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4863.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4864.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4865.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4866.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4867.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4868.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4869.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4870.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4871.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4872.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4873.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4874.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4875.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4876.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4877.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4878.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4879.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4880.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4881.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4882.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4883.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4884.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4885.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4886.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4887.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4888.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4889.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4890.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4891.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4892.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4893.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4894.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4895.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4896.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4897.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4898.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4899.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4900.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4901.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4902.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4903.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4904.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4905.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4906.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4907.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4908.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4909.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4910.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4911.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4912.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4913.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4914.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4915.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4916.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4917.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4918.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4919.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4920.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4921.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4922.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4923.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4924.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4925.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4926.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4927.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4928.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4929.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4930.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4931.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4932.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4933.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4934.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4935.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4936.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4937.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4938.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4939.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4940.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4941.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4942.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4943.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4944.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4945.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4946.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4947.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4948.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4949.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4950.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4951.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4952.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4953.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4954.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4955.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4956.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4957.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4958.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4959.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4960.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4961.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4962.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4963.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4964.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4965.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4966.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4967.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4968.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4969.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4970.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4971.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4972.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4973.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4974.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4975.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4976.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4977.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4978.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4979.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4980.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4981.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4982.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4983.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4984.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4985.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4986.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4987.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4988.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4989.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4990.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4991.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4992.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4993.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4994.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4995.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4996.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4997.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4998.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4999.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}], ""episode_metadata_val"": [{""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0000.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0001.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0002.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0003.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0004.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0005.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0006.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0007.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0008.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0009.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0010.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0011.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0012.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0013.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0014.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0015.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0016.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0017.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0018.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0019.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0020.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0021.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0022.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0023.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0024.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0025.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0026.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0027.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0028.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0029.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0030.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0031.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0032.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0033.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0034.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0035.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0036.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0037.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0038.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0039.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0040.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0041.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0042.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0043.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0044.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0045.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0046.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0047.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0048.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0049.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0050.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0051.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0052.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0053.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0054.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0055.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0056.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0057.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0058.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0059.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0060.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0061.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0062.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0063.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0064.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0065.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0066.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0067.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0068.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0069.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0070.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0071.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0072.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0073.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0074.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0075.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0076.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0077.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0078.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0079.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0080.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0081.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0082.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0083.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0084.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0085.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0086.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0087.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0088.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0089.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0090.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0091.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0092.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0093.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0094.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0095.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0096.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0097.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0098.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0099.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0100.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0101.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0102.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0103.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0104.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0105.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0106.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0107.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0108.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0109.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0110.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0111.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0112.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0113.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0114.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0115.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0116.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0117.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0118.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0119.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0120.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0121.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0122.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0123.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0124.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0125.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0126.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0127.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0128.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0129.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0130.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0131.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0132.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0133.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0134.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0135.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0136.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0137.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0138.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0139.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0140.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0141.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0142.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0143.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0144.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0145.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0146.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0147.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0148.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0149.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0150.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0151.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0152.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0153.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0154.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0155.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0156.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0157.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0158.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0159.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0160.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0161.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0162.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0163.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0164.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0165.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0166.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0167.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0168.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0169.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0170.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0171.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0172.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0173.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0174.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0175.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0176.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0177.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0178.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0179.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0180.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0181.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0182.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0183.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0184.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0185.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0186.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0187.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0188.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0189.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0190.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0191.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0192.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0193.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0194.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0195.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0196.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0197.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0198.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0199.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0200.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0201.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0202.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0203.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0204.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0205.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0206.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0207.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0208.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0209.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0210.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0211.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0212.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0213.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0214.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0215.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0216.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0217.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0218.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0219.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0220.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0221.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0222.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0223.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0224.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0225.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0226.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0227.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0228.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0229.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0230.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0231.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0232.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0233.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0234.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0235.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0236.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0237.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0238.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0239.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0240.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0241.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0242.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0243.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0244.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0245.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0246.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0247.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0248.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0249.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}], ""episode_metadata_test"": [{""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0000.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0001.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0002.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0003.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0004.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0005.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0006.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0007.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0008.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0009.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0010.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0011.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0012.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0013.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0014.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0015.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0016.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0017.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0018.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0019.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0020.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0021.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0022.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0023.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0024.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0025.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0026.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0027.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0028.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0029.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0030.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0031.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0032.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0033.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0034.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0035.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0036.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0037.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0038.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0039.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0040.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0041.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0042.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0043.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0044.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0045.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0046.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0047.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0048.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0049.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0050.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0051.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0052.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0053.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0054.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0055.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0056.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0057.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0058.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0059.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0060.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0061.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0062.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0063.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0064.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0065.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0066.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0067.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0068.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0069.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0070.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0071.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0072.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0073.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0074.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0075.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0076.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0077.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0078.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0079.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0080.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0081.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0082.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0083.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0084.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0085.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0086.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0087.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0088.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0089.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0090.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0091.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0092.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0093.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0094.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0095.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0096.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0097.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0098.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0099.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0100.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0101.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0102.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0103.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0104.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0105.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0106.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0107.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0108.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0109.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0110.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0111.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0112.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0113.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0114.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0115.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0116.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0117.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0118.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0119.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0120.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0121.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0122.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0123.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0124.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0125.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0126.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0127.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0128.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0129.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0130.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0131.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0132.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0133.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0134.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0135.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0136.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0137.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0138.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0139.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0140.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0141.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0142.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0143.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0144.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0145.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0146.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0147.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0148.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0149.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0150.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0151.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0152.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0153.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0154.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0155.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0156.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0157.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0158.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0159.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0160.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0161.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0162.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0163.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0164.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0165.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0166.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0167.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0168.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0169.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0170.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0171.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0172.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0173.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0174.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0175.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0176.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0177.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0178.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0179.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0180.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0181.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0182.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0183.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0184.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0185.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0186.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0187.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0188.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0189.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0190.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0191.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0192.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0193.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0194.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0195.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0196.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0197.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0198.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0199.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0200.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0201.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0202.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0203.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0204.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0205.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0206.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0207.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0208.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0209.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0210.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0211.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0212.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0213.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0214.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0215.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0216.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0217.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0218.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0219.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0220.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0221.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0222.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0223.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0224.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0225.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0226.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0227.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0228.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0229.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0230.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0231.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0232.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0233.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0234.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0235.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0236.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0237.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0238.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0239.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0240.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0241.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0242.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0243.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0244.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0245.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0246.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0247.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0248.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}, {""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0249.array_record"", ""num_chunks"": 100, ""avg_seq_len"": 16.0}]}",json,tab +51,70628,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/metadata.json",3161,0,"",json,selection_mouse +52,71256,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/metadata.json",3394,0,"",json,selection_command +53,74797,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/metadata.json",0,978551,"{\n ""env"": ""MinAtar-Breakout"",\n ""num_actions"": 6,\n ""num_episodes_train"": 500000,\n ""num_episodes_val"": 25000,\n ""num_episodes_test"": 25000,\n ""avg_episode_len_train"": 16.0,\n ""avg_episode_len_val"": 16.0,\n ""avg_episode_len_test"": 16.0,\n ""episode_metadata_train"": [\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0000.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0001.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0002.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0003.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0004.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0005.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0006.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0007.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0008.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0009.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0010.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0011.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0012.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0013.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0014.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0015.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0016.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0017.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0018.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0019.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0020.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0021.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0022.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0023.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0024.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0025.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0026.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0027.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0028.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0029.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0030.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0031.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0032.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0033.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0034.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0035.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0036.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0037.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0038.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0039.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0040.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0041.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0042.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0043.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0044.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0045.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0046.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0047.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0048.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0049.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0050.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0051.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0052.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0053.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0054.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0055.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0056.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0057.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0058.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0059.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0060.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0061.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0062.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0063.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0064.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0065.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0066.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0067.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0068.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0069.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0070.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0071.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0072.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0073.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0074.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0075.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0076.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0077.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0078.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0079.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0080.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0081.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0082.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0083.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0084.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0085.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0086.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0087.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0088.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0089.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0090.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0091.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0092.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0093.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0094.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0095.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0096.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0097.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0098.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0099.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0100.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0101.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0102.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0103.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0104.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0105.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0106.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0107.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0108.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0109.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0110.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0111.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0112.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0113.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0114.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0115.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0116.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0117.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0118.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0119.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0120.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0121.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0122.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0123.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0124.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0125.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0126.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0127.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0128.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0129.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0130.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0131.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0132.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0133.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0134.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0135.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0136.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0137.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0138.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0139.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0140.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0141.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0142.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0143.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0144.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0145.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0146.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0147.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0148.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0149.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0150.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0151.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0152.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0153.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0154.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0155.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0156.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0157.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0158.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0159.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0160.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0161.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0162.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0163.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0164.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0165.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0166.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0167.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0168.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0169.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0170.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0171.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0172.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0173.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0174.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0175.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0176.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0177.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0178.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0179.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0180.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0181.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0182.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0183.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0184.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0185.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0186.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0187.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0188.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0189.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0190.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0191.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0192.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0193.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0194.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0195.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0196.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0197.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0198.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0199.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0200.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0201.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0202.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0203.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0204.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0205.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0206.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0207.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0208.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0209.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0210.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0211.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0212.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0213.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0214.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0215.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0216.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0217.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0218.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0219.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0220.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0221.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0222.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0223.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0224.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0225.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0226.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0227.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0228.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0229.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0230.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0231.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0232.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0233.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0234.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0235.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0236.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0237.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0238.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0239.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0240.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0241.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0242.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0243.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0244.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0245.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0246.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0247.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0248.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0249.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0250.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0251.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0252.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0253.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0254.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0255.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0256.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0257.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0258.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0259.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0260.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0261.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0262.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0263.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0264.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0265.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0266.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0267.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0268.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0269.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0270.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0271.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0272.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0273.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0274.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0275.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0276.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0277.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0278.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0279.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0280.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0281.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0282.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0283.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0284.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0285.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0286.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0287.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0288.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0289.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0290.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0291.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0292.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0293.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0294.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0295.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0296.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0297.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0298.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0299.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0300.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0301.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0302.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0303.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0304.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0305.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0306.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0307.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0308.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0309.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0310.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0311.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0312.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0313.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0314.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0315.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0316.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0317.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0318.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0319.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0320.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0321.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0322.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0323.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0324.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0325.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0326.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0327.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0328.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0329.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0330.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0331.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0332.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0333.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0334.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0335.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0336.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0337.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0338.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0339.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0340.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0341.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0342.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0343.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0344.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0345.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0346.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0347.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0348.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0349.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0350.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0351.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0352.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0353.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0354.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0355.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0356.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0357.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0358.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0359.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0360.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0361.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0362.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0363.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0364.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0365.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0366.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0367.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0368.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0369.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0370.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0371.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0372.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0373.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0374.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0375.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0376.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0377.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0378.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0379.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0380.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0381.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0382.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0383.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0384.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0385.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0386.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0387.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0388.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0389.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0390.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0391.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0392.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0393.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0394.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0395.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0396.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0397.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0398.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0399.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0400.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0401.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0402.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0403.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0404.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0405.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0406.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0407.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0408.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0409.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0410.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0411.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0412.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0413.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0414.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0415.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0416.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0417.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0418.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0419.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0420.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0421.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0422.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0423.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0424.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0425.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0426.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0427.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0428.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0429.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0430.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0431.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0432.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0433.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0434.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0435.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0436.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0437.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0438.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0439.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0440.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0441.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0442.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0443.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0444.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0445.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0446.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0447.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0448.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0449.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0450.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0451.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0452.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0453.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0454.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0455.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0456.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0457.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0458.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0459.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0460.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0461.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0462.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0463.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0464.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0465.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0466.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0467.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0468.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0469.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0470.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0471.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0472.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0473.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0474.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0475.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0476.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0477.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0478.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0479.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0480.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0481.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0482.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0483.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0484.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0485.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0486.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0487.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0488.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0489.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0490.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0491.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0492.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0493.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0494.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0495.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0496.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0497.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0498.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0499.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0500.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0501.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0502.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0503.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0504.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0505.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0506.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0507.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0508.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0509.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0510.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0511.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0512.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0513.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0514.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0515.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0516.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0517.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0518.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0519.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0520.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0521.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0522.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0523.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0524.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0525.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0526.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0527.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0528.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0529.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0530.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0531.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0532.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0533.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0534.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0535.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0536.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0537.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0538.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0539.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0540.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0541.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0542.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0543.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0544.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0545.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0546.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0547.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0548.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0549.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0550.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0551.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0552.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0553.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0554.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0555.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0556.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0557.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0558.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0559.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0560.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0561.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0562.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0563.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0564.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0565.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0566.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0567.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0568.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0569.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0570.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0571.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0572.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0573.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0574.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0575.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0576.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0577.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0578.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0579.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0580.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0581.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0582.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0583.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0584.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0585.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0586.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0587.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0588.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0589.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0590.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0591.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0592.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0593.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0594.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0595.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0596.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0597.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0598.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0599.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0600.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0601.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0602.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0603.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0604.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0605.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0606.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0607.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0608.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0609.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0610.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0611.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0612.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0613.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0614.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0615.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0616.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0617.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0618.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0619.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0620.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0621.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0622.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0623.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0624.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0625.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0626.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0627.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0628.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0629.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0630.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0631.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0632.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0633.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0634.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0635.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0636.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0637.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0638.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0639.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0640.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0641.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0642.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0643.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0644.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0645.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0646.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0647.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0648.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0649.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0650.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0651.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0652.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0653.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0654.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0655.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0656.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0657.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0658.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0659.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0660.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0661.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0662.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0663.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0664.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0665.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0666.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0667.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0668.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0669.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0670.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0671.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0672.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0673.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0674.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0675.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0676.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0677.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0678.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0679.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0680.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0681.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0682.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0683.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0684.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0685.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0686.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0687.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0688.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0689.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0690.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0691.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0692.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0693.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0694.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0695.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0696.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0697.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0698.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0699.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0700.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0701.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0702.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0703.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0704.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0705.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0706.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0707.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0708.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0709.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0710.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0711.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0712.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0713.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0714.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0715.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0716.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0717.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0718.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0719.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0720.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0721.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0722.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0723.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0724.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0725.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0726.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0727.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0728.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0729.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0730.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0731.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0732.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0733.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0734.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0735.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0736.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0737.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0738.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0739.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0740.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0741.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0742.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0743.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0744.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0745.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0746.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0747.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0748.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0749.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0750.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0751.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0752.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0753.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0754.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0755.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0756.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0757.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0758.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0759.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0760.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0761.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0762.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0763.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0764.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0765.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0766.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0767.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0768.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0769.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0770.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0771.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0772.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0773.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0774.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0775.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0776.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0777.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0778.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0779.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0780.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0781.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0782.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0783.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0784.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0785.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0786.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0787.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0788.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0789.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0790.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0791.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0792.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0793.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0794.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0795.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0796.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0797.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0798.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0799.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0800.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0801.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0802.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0803.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0804.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0805.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0806.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0807.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0808.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0809.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0810.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0811.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0812.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0813.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0814.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0815.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0816.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0817.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0818.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0819.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0820.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0821.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0822.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0823.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0824.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0825.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0826.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0827.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0828.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0829.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0830.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0831.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0832.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0833.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0834.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0835.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0836.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0837.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0838.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0839.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0840.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0841.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0842.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0843.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0844.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0845.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0846.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0847.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0848.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0849.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0850.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0851.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0852.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0853.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0854.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0855.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0856.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0857.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0858.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0859.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0860.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0861.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0862.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0863.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0864.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0865.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0866.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0867.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0868.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0869.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0870.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0871.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0872.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0873.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0874.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0875.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0876.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0877.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0878.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0879.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0880.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0881.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0882.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0883.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0884.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0885.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0886.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0887.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0888.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0889.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0890.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0891.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0892.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0893.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0894.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0895.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0896.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0897.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0898.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0899.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0900.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0901.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0902.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0903.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0904.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0905.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0906.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0907.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0908.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0909.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0910.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0911.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0912.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0913.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0914.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0915.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0916.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0917.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0918.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0919.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0920.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0921.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0922.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0923.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0924.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0925.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0926.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0927.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0928.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0929.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0930.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0931.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0932.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0933.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0934.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0935.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0936.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0937.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0938.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0939.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0940.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0941.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0942.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0943.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0944.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0945.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0946.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0947.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0948.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0949.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0950.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0951.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0952.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0953.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0954.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0955.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0956.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0957.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0958.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0959.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0960.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0961.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0962.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0963.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0964.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0965.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0966.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0967.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0968.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0969.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0970.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0971.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0972.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0973.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0974.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0975.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0976.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0977.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0978.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0979.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0980.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0981.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0982.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0983.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0984.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0985.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0986.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0987.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0988.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0989.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0990.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0991.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0992.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0993.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0994.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0995.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0996.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0997.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0998.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_0999.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1000.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1001.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1002.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1003.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1004.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1005.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1006.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1007.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1008.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1009.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1010.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1011.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1012.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1013.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1014.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1015.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1016.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1017.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1018.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1019.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1020.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1021.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1022.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1023.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1024.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1025.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1026.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1027.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1028.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1029.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1030.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1031.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1032.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1033.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1034.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1035.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1036.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1037.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1038.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1039.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1040.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1041.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1042.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1043.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1044.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1045.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1046.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1047.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1048.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1049.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1050.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1051.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1052.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1053.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1054.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1055.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1056.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1057.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1058.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1059.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1060.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1061.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1062.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1063.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1064.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1065.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1066.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1067.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1068.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1069.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1070.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1071.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1072.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1073.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1074.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1075.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1076.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1077.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1078.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1079.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1080.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1081.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1082.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1083.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1084.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1085.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1086.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1087.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1088.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1089.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1090.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1091.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1092.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1093.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1094.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1095.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1096.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1097.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1098.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1099.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1100.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1101.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1102.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1103.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1104.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1105.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1106.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1107.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1108.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1109.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1110.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1111.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1112.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1113.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1114.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1115.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1116.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1117.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1118.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1119.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1120.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1121.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1122.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1123.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1124.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1125.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1126.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1127.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1128.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1129.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1130.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1131.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1132.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1133.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1134.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1135.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1136.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1137.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1138.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1139.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1140.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1141.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1142.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1143.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1144.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1145.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1146.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1147.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1148.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1149.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1150.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1151.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1152.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1153.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1154.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1155.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1156.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1157.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1158.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1159.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1160.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1161.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1162.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1163.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1164.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1165.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1166.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1167.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1168.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1169.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1170.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1171.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1172.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1173.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1174.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1175.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1176.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1177.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1178.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1179.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1180.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1181.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1182.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1183.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1184.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1185.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1186.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1187.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1188.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1189.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1190.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1191.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1192.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1193.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1194.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1195.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1196.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1197.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1198.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1199.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1200.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1201.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1202.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1203.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1204.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1205.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1206.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1207.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1208.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1209.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1210.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1211.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1212.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1213.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1214.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1215.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1216.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1217.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1218.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1219.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1220.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1221.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1222.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1223.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1224.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1225.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1226.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1227.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1228.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1229.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1230.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1231.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1232.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1233.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1234.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1235.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1236.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1237.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1238.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1239.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1240.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1241.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1242.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1243.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1244.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1245.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1246.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1247.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1248.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1249.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1250.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1251.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1252.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1253.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1254.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1255.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1256.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1257.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1258.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1259.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1260.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1261.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1262.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1263.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1264.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1265.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1266.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1267.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1268.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1269.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1270.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1271.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1272.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1273.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1274.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1275.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1276.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1277.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1278.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1279.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1280.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1281.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1282.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1283.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1284.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1285.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1286.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1287.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1288.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1289.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1290.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1291.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1292.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1293.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1294.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1295.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1296.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1297.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1298.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1299.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1300.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1301.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1302.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1303.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1304.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1305.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1306.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1307.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1308.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1309.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1310.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1311.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1312.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1313.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1314.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1315.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1316.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1317.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1318.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1319.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1320.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1321.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1322.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1323.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1324.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1325.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1326.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1327.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1328.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1329.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1330.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1331.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1332.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1333.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1334.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1335.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1336.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1337.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1338.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1339.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1340.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1341.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1342.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1343.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1344.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1345.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1346.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1347.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1348.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1349.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1350.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1351.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1352.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1353.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1354.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1355.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1356.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1357.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1358.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1359.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1360.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1361.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1362.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1363.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1364.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1365.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1366.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1367.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1368.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1369.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1370.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1371.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1372.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1373.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1374.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1375.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1376.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1377.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1378.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1379.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1380.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1381.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1382.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1383.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1384.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1385.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1386.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1387.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1388.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1389.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1390.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1391.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1392.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1393.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1394.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1395.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1396.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1397.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1398.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1399.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1400.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1401.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1402.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1403.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1404.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1405.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1406.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1407.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1408.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1409.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1410.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1411.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1412.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1413.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1414.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1415.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1416.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1417.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1418.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1419.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1420.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1421.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1422.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1423.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1424.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1425.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1426.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1427.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1428.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1429.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1430.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1431.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1432.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1433.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1434.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1435.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1436.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1437.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1438.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1439.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1440.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1441.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1442.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1443.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1444.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1445.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1446.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1447.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1448.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1449.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1450.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1451.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1452.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1453.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1454.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1455.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1456.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1457.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1458.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1459.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1460.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1461.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1462.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1463.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1464.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1465.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1466.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1467.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1468.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1469.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1470.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1471.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1472.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1473.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1474.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1475.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1476.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1477.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1478.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1479.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1480.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1481.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1482.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1483.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1484.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1485.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1486.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1487.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1488.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1489.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1490.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1491.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1492.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1493.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1494.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1495.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1496.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1497.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1498.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1499.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1500.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1501.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1502.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1503.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1504.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1505.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1506.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1507.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1508.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1509.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1510.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1511.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1512.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1513.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1514.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1515.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1516.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1517.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1518.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1519.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1520.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1521.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1522.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1523.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1524.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1525.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1526.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1527.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1528.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1529.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1530.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1531.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1532.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1533.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1534.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1535.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1536.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1537.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1538.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1539.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1540.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1541.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1542.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1543.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1544.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1545.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1546.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1547.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1548.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1549.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1550.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1551.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1552.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1553.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1554.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1555.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1556.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1557.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1558.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1559.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1560.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1561.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1562.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1563.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1564.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1565.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1566.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1567.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1568.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1569.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1570.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1571.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1572.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1573.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1574.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1575.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1576.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1577.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1578.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1579.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1580.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1581.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1582.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1583.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1584.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1585.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1586.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1587.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1588.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1589.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1590.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1591.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1592.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1593.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1594.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1595.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1596.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1597.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1598.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1599.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1600.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1601.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1602.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1603.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1604.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1605.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1606.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1607.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1608.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1609.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1610.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1611.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1612.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1613.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1614.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1615.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1616.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1617.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1618.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1619.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1620.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1621.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1622.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1623.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1624.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1625.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1626.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1627.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1628.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1629.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1630.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1631.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1632.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1633.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1634.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1635.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1636.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1637.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1638.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1639.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1640.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1641.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1642.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1643.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1644.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1645.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1646.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1647.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1648.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1649.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1650.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1651.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1652.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1653.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1654.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1655.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1656.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1657.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1658.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1659.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1660.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1661.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1662.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1663.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1664.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1665.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1666.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1667.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1668.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1669.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1670.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1671.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1672.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1673.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1674.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1675.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1676.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1677.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1678.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1679.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1680.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1681.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1682.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1683.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1684.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1685.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1686.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1687.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1688.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1689.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1690.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1691.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1692.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1693.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1694.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1695.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1696.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1697.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1698.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1699.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1700.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1701.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1702.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1703.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1704.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1705.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1706.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1707.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1708.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1709.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1710.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1711.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1712.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1713.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1714.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1715.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1716.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1717.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1718.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1719.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1720.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1721.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1722.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1723.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1724.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1725.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1726.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1727.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1728.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1729.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1730.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1731.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1732.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1733.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1734.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1735.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1736.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1737.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1738.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1739.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1740.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1741.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1742.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1743.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1744.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1745.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1746.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1747.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1748.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1749.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1750.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1751.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1752.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1753.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1754.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1755.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1756.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1757.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1758.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1759.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1760.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1761.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1762.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1763.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1764.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1765.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1766.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1767.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1768.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1769.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1770.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1771.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1772.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1773.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1774.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1775.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1776.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1777.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1778.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1779.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1780.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1781.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1782.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1783.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1784.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1785.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1786.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1787.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1788.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1789.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1790.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1791.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1792.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1793.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1794.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1795.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1796.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1797.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1798.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1799.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1800.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1801.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1802.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1803.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1804.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1805.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1806.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1807.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1808.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1809.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1810.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1811.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1812.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1813.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1814.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1815.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1816.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1817.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1818.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1819.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1820.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1821.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1822.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1823.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1824.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1825.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1826.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1827.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1828.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1829.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1830.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1831.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1832.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1833.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1834.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1835.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1836.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1837.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1838.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1839.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1840.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1841.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1842.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1843.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1844.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1845.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1846.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1847.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1848.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1849.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1850.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1851.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1852.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1853.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1854.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1855.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1856.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1857.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1858.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1859.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1860.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1861.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1862.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1863.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1864.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1865.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1866.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1867.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1868.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1869.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1870.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1871.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1872.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1873.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1874.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1875.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1876.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1877.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1878.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1879.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1880.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1881.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1882.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1883.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1884.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1885.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1886.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1887.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1888.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1889.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1890.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1891.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1892.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1893.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1894.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1895.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1896.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1897.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1898.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1899.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1900.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1901.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1902.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1903.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1904.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1905.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1906.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1907.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1908.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1909.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1910.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1911.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1912.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1913.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1914.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1915.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1916.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1917.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1918.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1919.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1920.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1921.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1922.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1923.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1924.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1925.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1926.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1927.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1928.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1929.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1930.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1931.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1932.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1933.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1934.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1935.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1936.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1937.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1938.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1939.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1940.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1941.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1942.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1943.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1944.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1945.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1946.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1947.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1948.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1949.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1950.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1951.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1952.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1953.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1954.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1955.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1956.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1957.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1958.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1959.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1960.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1961.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1962.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1963.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1964.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1965.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1966.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1967.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1968.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1969.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1970.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1971.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1972.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1973.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1974.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1975.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1976.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1977.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1978.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1979.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1980.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1981.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1982.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1983.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1984.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1985.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1986.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1987.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1988.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1989.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1990.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1991.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1992.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1993.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1994.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1995.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1996.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1997.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1998.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_1999.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2000.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2001.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2002.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2003.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2004.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2005.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2006.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2007.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2008.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2009.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2010.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2011.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2012.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2013.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2014.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2015.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2016.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2017.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2018.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2019.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2020.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2021.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2022.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2023.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2024.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2025.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2026.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2027.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2028.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2029.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2030.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2031.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2032.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2033.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2034.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2035.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2036.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2037.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2038.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2039.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2040.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2041.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2042.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2043.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2044.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2045.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2046.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2047.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2048.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2049.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2050.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2051.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2052.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2053.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2054.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2055.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2056.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2057.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2058.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2059.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2060.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2061.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2062.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2063.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2064.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2065.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2066.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2067.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2068.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2069.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2070.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2071.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2072.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2073.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2074.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2075.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2076.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2077.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2078.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2079.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2080.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2081.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2082.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2083.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2084.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2085.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2086.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2087.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2088.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2089.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2090.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2091.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2092.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2093.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2094.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2095.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2096.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2097.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2098.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2099.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2100.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2101.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2102.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2103.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2104.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2105.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2106.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2107.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2108.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2109.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2110.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2111.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2112.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2113.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2114.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2115.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2116.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2117.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2118.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2119.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2120.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2121.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2122.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2123.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2124.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2125.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2126.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2127.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2128.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2129.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2130.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2131.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2132.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2133.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2134.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2135.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2136.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2137.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2138.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2139.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2140.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2141.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2142.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2143.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2144.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2145.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2146.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2147.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2148.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2149.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2150.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2151.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2152.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2153.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2154.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2155.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2156.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2157.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2158.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2159.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2160.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2161.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2162.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2163.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2164.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2165.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2166.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2167.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2168.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2169.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2170.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2171.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2172.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2173.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2174.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2175.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2176.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2177.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2178.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2179.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2180.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2181.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2182.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2183.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2184.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2185.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2186.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2187.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2188.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2189.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2190.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2191.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2192.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2193.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2194.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2195.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2196.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2197.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2198.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2199.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2200.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2201.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2202.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2203.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2204.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2205.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2206.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2207.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2208.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2209.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2210.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2211.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2212.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2213.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2214.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2215.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2216.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2217.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2218.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2219.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2220.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2221.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2222.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2223.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2224.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2225.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2226.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2227.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2228.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2229.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2230.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2231.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2232.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2233.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2234.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2235.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2236.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2237.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2238.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2239.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2240.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2241.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2242.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2243.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2244.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2245.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2246.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2247.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2248.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2249.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2250.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2251.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2252.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2253.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2254.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2255.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2256.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2257.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2258.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2259.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2260.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2261.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2262.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2263.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2264.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2265.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2266.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2267.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2268.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2269.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2270.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2271.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2272.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2273.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2274.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2275.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2276.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2277.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2278.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2279.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2280.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2281.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2282.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2283.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2284.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2285.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2286.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2287.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2288.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2289.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2290.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2291.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2292.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2293.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2294.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2295.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2296.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2297.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2298.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2299.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2300.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2301.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2302.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2303.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2304.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2305.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2306.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2307.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2308.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2309.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2310.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2311.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2312.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2313.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2314.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2315.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2316.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2317.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2318.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2319.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2320.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2321.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2322.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2323.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2324.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2325.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2326.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2327.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2328.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2329.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2330.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2331.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2332.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2333.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2334.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2335.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2336.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2337.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2338.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2339.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2340.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2341.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2342.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2343.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2344.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2345.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2346.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2347.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2348.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2349.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2350.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2351.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2352.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2353.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2354.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2355.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2356.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2357.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2358.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2359.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2360.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2361.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2362.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2363.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2364.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2365.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2366.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2367.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2368.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2369.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2370.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2371.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2372.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2373.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2374.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2375.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2376.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2377.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2378.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2379.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2380.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2381.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2382.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2383.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2384.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2385.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2386.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2387.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2388.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2389.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2390.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2391.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2392.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2393.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2394.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2395.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2396.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2397.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2398.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2399.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2400.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2401.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2402.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2403.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2404.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2405.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2406.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2407.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2408.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2409.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2410.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2411.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2412.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2413.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2414.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2415.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2416.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2417.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2418.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2419.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2420.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2421.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2422.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2423.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2424.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2425.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2426.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2427.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2428.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2429.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2430.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2431.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2432.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2433.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2434.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2435.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2436.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2437.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2438.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2439.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2440.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2441.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2442.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2443.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2444.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2445.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2446.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2447.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2448.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2449.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2450.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2451.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2452.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2453.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2454.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2455.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2456.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2457.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2458.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2459.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2460.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2461.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2462.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2463.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2464.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2465.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2466.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2467.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2468.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2469.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2470.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2471.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2472.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2473.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2474.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2475.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2476.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2477.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2478.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2479.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2480.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2481.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2482.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2483.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2484.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2485.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2486.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2487.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2488.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2489.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2490.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2491.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2492.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2493.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2494.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2495.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2496.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2497.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2498.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2499.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2500.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2501.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2502.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2503.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2504.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2505.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2506.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2507.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2508.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2509.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2510.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2511.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2512.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2513.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2514.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2515.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2516.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2517.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2518.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2519.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2520.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2521.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2522.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2523.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2524.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2525.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2526.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2527.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2528.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2529.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2530.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2531.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2532.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2533.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2534.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2535.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2536.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2537.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2538.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2539.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2540.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2541.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2542.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2543.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2544.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2545.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2546.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2547.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2548.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2549.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2550.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2551.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2552.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2553.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2554.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2555.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2556.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2557.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2558.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2559.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2560.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2561.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2562.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2563.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2564.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2565.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2566.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2567.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2568.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2569.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2570.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2571.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2572.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2573.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2574.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2575.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2576.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2577.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2578.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2579.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2580.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2581.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2582.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2583.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2584.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2585.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2586.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2587.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2588.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2589.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2590.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2591.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2592.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2593.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2594.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2595.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2596.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2597.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2598.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2599.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2600.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2601.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2602.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2603.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2604.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2605.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2606.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2607.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2608.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2609.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2610.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2611.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2612.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2613.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2614.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2615.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2616.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2617.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2618.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2619.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2620.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2621.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2622.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2623.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2624.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2625.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2626.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2627.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2628.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2629.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2630.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2631.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2632.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2633.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2634.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2635.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2636.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2637.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2638.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2639.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2640.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2641.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2642.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2643.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2644.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2645.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2646.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2647.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2648.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2649.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2650.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2651.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2652.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2653.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2654.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2655.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2656.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2657.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2658.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2659.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2660.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2661.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2662.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2663.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2664.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2665.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2666.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2667.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2668.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2669.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2670.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2671.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2672.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2673.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2674.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2675.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2676.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2677.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2678.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2679.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2680.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2681.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2682.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2683.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2684.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2685.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2686.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2687.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2688.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2689.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2690.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2691.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2692.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2693.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2694.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2695.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2696.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2697.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2698.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2699.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2700.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2701.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2702.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2703.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2704.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2705.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2706.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2707.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2708.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2709.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2710.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2711.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2712.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2713.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2714.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2715.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2716.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2717.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2718.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2719.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2720.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2721.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2722.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2723.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2724.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2725.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2726.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2727.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2728.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2729.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2730.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2731.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2732.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2733.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2734.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2735.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2736.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2737.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2738.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2739.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2740.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2741.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2742.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2743.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2744.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2745.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2746.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2747.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2748.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2749.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2750.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2751.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2752.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2753.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2754.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2755.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2756.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2757.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2758.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2759.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2760.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2761.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2762.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2763.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2764.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2765.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2766.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2767.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2768.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2769.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2770.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2771.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2772.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2773.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2774.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2775.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2776.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2777.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2778.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2779.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2780.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2781.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2782.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2783.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2784.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2785.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2786.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2787.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2788.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2789.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2790.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2791.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2792.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2793.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2794.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2795.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2796.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2797.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2798.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2799.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2800.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2801.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2802.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2803.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2804.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2805.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2806.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2807.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2808.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2809.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2810.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2811.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2812.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2813.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2814.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2815.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2816.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2817.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2818.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2819.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2820.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2821.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2822.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2823.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2824.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2825.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2826.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2827.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2828.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2829.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2830.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2831.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2832.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2833.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2834.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2835.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2836.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2837.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2838.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2839.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2840.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2841.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2842.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2843.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2844.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2845.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2846.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2847.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2848.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2849.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2850.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2851.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2852.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2853.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2854.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2855.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2856.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2857.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2858.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2859.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2860.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2861.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2862.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2863.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2864.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2865.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2866.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2867.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2868.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2869.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2870.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2871.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2872.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2873.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2874.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2875.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2876.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2877.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2878.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2879.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2880.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2881.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2882.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2883.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2884.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2885.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2886.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2887.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2888.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2889.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2890.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2891.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2892.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2893.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2894.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2895.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2896.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2897.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2898.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2899.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2900.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2901.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2902.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2903.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2904.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2905.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2906.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2907.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2908.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2909.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2910.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2911.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2912.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2913.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2914.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2915.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2916.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2917.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2918.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2919.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2920.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2921.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2922.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2923.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2924.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2925.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2926.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2927.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2928.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2929.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2930.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2931.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2932.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2933.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2934.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2935.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2936.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2937.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2938.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2939.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2940.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2941.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2942.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2943.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2944.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2945.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2946.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2947.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2948.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2949.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2950.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2951.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2952.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2953.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2954.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2955.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2956.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2957.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2958.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2959.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2960.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2961.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2962.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2963.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2964.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2965.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2966.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2967.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2968.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2969.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2970.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2971.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2972.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2973.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2974.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2975.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2976.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2977.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2978.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2979.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2980.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2981.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2982.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2983.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2984.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2985.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2986.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2987.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2988.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2989.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2990.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2991.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2992.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2993.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2994.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2995.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2996.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2997.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2998.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_2999.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3000.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3001.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3002.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3003.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3004.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3005.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3006.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3007.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3008.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3009.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3010.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3011.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3012.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3013.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3014.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3015.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3016.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3017.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3018.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3019.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3020.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3021.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3022.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3023.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3024.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3025.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3026.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3027.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3028.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3029.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3030.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3031.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3032.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3033.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3034.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3035.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3036.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3037.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3038.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3039.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3040.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3041.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3042.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3043.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3044.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3045.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3046.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3047.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3048.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3049.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3050.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3051.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3052.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3053.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3054.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3055.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3056.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3057.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3058.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3059.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3060.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3061.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3062.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3063.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3064.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3065.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3066.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3067.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3068.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3069.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3070.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3071.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3072.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3073.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3074.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3075.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3076.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3077.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3078.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3079.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3080.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3081.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3082.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3083.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3084.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3085.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3086.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3087.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3088.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3089.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3090.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3091.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3092.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3093.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3094.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3095.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3096.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3097.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3098.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3099.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3100.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3101.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3102.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3103.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3104.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3105.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3106.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3107.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3108.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3109.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3110.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3111.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3112.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3113.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3114.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3115.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3116.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3117.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3118.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3119.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3120.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3121.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3122.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3123.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3124.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3125.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3126.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3127.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3128.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3129.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3130.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3131.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3132.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3133.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3134.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3135.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3136.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3137.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3138.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3139.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3140.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3141.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3142.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3143.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3144.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3145.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3146.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3147.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3148.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3149.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3150.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3151.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3152.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3153.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3154.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3155.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3156.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3157.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3158.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3159.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3160.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3161.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3162.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3163.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3164.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3165.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3166.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3167.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3168.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3169.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3170.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3171.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3172.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3173.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3174.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3175.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3176.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3177.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3178.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3179.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3180.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3181.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3182.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3183.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3184.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3185.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3186.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3187.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3188.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3189.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3190.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3191.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3192.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3193.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3194.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3195.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3196.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3197.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3198.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3199.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3200.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3201.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3202.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3203.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3204.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3205.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3206.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3207.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3208.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3209.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3210.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3211.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3212.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3213.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3214.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3215.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3216.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3217.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3218.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3219.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3220.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3221.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3222.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3223.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3224.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3225.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3226.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3227.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3228.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3229.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3230.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3231.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3232.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3233.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3234.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3235.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3236.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3237.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3238.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3239.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3240.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3241.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3242.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3243.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3244.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3245.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3246.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3247.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3248.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3249.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3250.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3251.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3252.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3253.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3254.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3255.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3256.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3257.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3258.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3259.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3260.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3261.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3262.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3263.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3264.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3265.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3266.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3267.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3268.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3269.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3270.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3271.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3272.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3273.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3274.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3275.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3276.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3277.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3278.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3279.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3280.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3281.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3282.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3283.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3284.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3285.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3286.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3287.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3288.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3289.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3290.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3291.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3292.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3293.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3294.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3295.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3296.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3297.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3298.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3299.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3300.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3301.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3302.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3303.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3304.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3305.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3306.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3307.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3308.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3309.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3310.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3311.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3312.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3313.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3314.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3315.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3316.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3317.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3318.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3319.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3320.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3321.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3322.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3323.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3324.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3325.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3326.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3327.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3328.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3329.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3330.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3331.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3332.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3333.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3334.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3335.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3336.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3337.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3338.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3339.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3340.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3341.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3342.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3343.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3344.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3345.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3346.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3347.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3348.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3349.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3350.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3351.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3352.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3353.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3354.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3355.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3356.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3357.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3358.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3359.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3360.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3361.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3362.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3363.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3364.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3365.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3366.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3367.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3368.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3369.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3370.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3371.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3372.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3373.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3374.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3375.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3376.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3377.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3378.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3379.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3380.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3381.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3382.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3383.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3384.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3385.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3386.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3387.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3388.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3389.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3390.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3391.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3392.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3393.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3394.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3395.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3396.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3397.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3398.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3399.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3400.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3401.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3402.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3403.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3404.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3405.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3406.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3407.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3408.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3409.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3410.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3411.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3412.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3413.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3414.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3415.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3416.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3417.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3418.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3419.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3420.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3421.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3422.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3423.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3424.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3425.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3426.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3427.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3428.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3429.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3430.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3431.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3432.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3433.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3434.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3435.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3436.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3437.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3438.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3439.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3440.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3441.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3442.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3443.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3444.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3445.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3446.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3447.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3448.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3449.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3450.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3451.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3452.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3453.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3454.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3455.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3456.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3457.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3458.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3459.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3460.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3461.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3462.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3463.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3464.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3465.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3466.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3467.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3468.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3469.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3470.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3471.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3472.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3473.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3474.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3475.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3476.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3477.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3478.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3479.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3480.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3481.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3482.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3483.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3484.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3485.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3486.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3487.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3488.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3489.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3490.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3491.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3492.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3493.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3494.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3495.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3496.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3497.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3498.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3499.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3500.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3501.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3502.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3503.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3504.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3505.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3506.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3507.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3508.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3509.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3510.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3511.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3512.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3513.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3514.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3515.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3516.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3517.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3518.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3519.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3520.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3521.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3522.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3523.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3524.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3525.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3526.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3527.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3528.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3529.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3530.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3531.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3532.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3533.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3534.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3535.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3536.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3537.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3538.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3539.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3540.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3541.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3542.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3543.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3544.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3545.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3546.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3547.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3548.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3549.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3550.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3551.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3552.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3553.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3554.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3555.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3556.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3557.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3558.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3559.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3560.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3561.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3562.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3563.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3564.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3565.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3566.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3567.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3568.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3569.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3570.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3571.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3572.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3573.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3574.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3575.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3576.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3577.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3578.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3579.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3580.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3581.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3582.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3583.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3584.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3585.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3586.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3587.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3588.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3589.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3590.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3591.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3592.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3593.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3594.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3595.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3596.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3597.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3598.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3599.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3600.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3601.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3602.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3603.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3604.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3605.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3606.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3607.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3608.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3609.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3610.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3611.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3612.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3613.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3614.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3615.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3616.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3617.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3618.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3619.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3620.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3621.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3622.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3623.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3624.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3625.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3626.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3627.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3628.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3629.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3630.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3631.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3632.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3633.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3634.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3635.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3636.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3637.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3638.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3639.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3640.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3641.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3642.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3643.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3644.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3645.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3646.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3647.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3648.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3649.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3650.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3651.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3652.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3653.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3654.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3655.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3656.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3657.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3658.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3659.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3660.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3661.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3662.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3663.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3664.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3665.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3666.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3667.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3668.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3669.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3670.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3671.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3672.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3673.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3674.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3675.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3676.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3677.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3678.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3679.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3680.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3681.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3682.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3683.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3684.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3685.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3686.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3687.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3688.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3689.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3690.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3691.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3692.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3693.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3694.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3695.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3696.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3697.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3698.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3699.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3700.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3701.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3702.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3703.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3704.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3705.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3706.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3707.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3708.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3709.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3710.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3711.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3712.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3713.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3714.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3715.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3716.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3717.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3718.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3719.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3720.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3721.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3722.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3723.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3724.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3725.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3726.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3727.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3728.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3729.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3730.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3731.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3732.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3733.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3734.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3735.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3736.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3737.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3738.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3739.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3740.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3741.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3742.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3743.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3744.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3745.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3746.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3747.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3748.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3749.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3750.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3751.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3752.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3753.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3754.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3755.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3756.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3757.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3758.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3759.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3760.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3761.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3762.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3763.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3764.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3765.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3766.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3767.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3768.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3769.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3770.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3771.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3772.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3773.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3774.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3775.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3776.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3777.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3778.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3779.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3780.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3781.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3782.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3783.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3784.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3785.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3786.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3787.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3788.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3789.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3790.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3791.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3792.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3793.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3794.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3795.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3796.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3797.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3798.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3799.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3800.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3801.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3802.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3803.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3804.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3805.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3806.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3807.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3808.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3809.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3810.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3811.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3812.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3813.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3814.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3815.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3816.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3817.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3818.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3819.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3820.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3821.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3822.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3823.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3824.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3825.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3826.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3827.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3828.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3829.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3830.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3831.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3832.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3833.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3834.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3835.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3836.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3837.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3838.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3839.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3840.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3841.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3842.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3843.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3844.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3845.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3846.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3847.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3848.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3849.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3850.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3851.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3852.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3853.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3854.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3855.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3856.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3857.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3858.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3859.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3860.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3861.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3862.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3863.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3864.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3865.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3866.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3867.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3868.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3869.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3870.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3871.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3872.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3873.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3874.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3875.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3876.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3877.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3878.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3879.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3880.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3881.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3882.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3883.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3884.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3885.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3886.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3887.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3888.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3889.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3890.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3891.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3892.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3893.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3894.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3895.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3896.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3897.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3898.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3899.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3900.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3901.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3902.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3903.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3904.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3905.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3906.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3907.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3908.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3909.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3910.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3911.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3912.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3913.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3914.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3915.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3916.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3917.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3918.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3919.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3920.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3921.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3922.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3923.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3924.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3925.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3926.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3927.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3928.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3929.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3930.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3931.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3932.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3933.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3934.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3935.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3936.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3937.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3938.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3939.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3940.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3941.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3942.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3943.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3944.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3945.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3946.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3947.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3948.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3949.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3950.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3951.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3952.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3953.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3954.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3955.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3956.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3957.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3958.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3959.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3960.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3961.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3962.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3963.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3964.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3965.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3966.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3967.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3968.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3969.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3970.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3971.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3972.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3973.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3974.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3975.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3976.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3977.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3978.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3979.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3980.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3981.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3982.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3983.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3984.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3985.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3986.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3987.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3988.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3989.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3990.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3991.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3992.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3993.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3994.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3995.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3996.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3997.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3998.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_3999.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4000.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4001.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4002.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4003.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4004.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4005.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4006.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4007.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4008.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4009.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4010.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4011.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4012.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4013.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4014.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4015.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4016.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4017.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4018.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4019.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4020.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4021.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4022.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4023.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4024.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4025.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4026.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4027.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4028.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4029.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4030.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4031.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4032.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4033.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4034.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4035.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4036.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4037.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4038.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4039.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4040.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4041.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4042.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4043.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4044.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4045.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4046.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4047.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4048.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4049.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4050.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4051.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4052.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4053.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4054.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4055.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4056.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4057.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4058.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4059.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4060.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4061.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4062.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4063.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4064.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4065.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4066.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4067.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4068.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4069.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4070.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4071.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4072.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4073.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4074.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4075.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4076.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4077.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4078.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4079.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4080.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4081.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4082.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4083.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4084.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4085.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4086.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4087.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4088.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4089.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4090.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4091.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4092.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4093.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4094.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4095.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4096.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4097.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4098.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4099.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4100.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4101.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4102.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4103.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4104.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4105.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4106.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4107.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4108.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4109.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4110.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4111.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4112.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4113.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4114.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4115.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4116.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4117.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4118.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4119.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4120.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4121.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4122.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4123.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4124.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4125.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4126.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4127.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4128.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4129.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4130.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4131.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4132.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4133.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4134.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4135.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4136.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4137.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4138.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4139.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4140.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4141.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4142.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4143.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4144.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4145.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4146.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4147.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4148.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4149.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4150.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4151.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4152.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4153.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4154.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4155.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4156.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4157.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4158.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4159.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4160.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4161.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4162.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4163.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4164.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4165.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4166.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4167.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4168.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4169.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4170.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4171.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4172.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4173.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4174.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4175.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4176.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4177.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4178.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4179.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4180.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4181.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4182.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4183.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4184.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4185.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4186.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4187.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4188.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4189.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4190.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4191.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4192.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4193.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4194.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4195.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4196.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4197.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4198.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4199.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4200.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4201.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4202.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4203.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4204.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4205.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4206.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4207.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4208.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4209.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4210.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4211.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4212.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4213.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4214.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4215.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4216.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4217.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4218.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4219.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4220.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4221.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4222.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4223.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4224.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4225.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4226.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4227.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4228.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4229.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4230.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4231.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4232.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4233.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4234.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4235.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4236.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4237.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4238.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4239.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4240.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4241.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4242.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4243.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4244.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4245.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4246.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4247.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4248.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4249.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4250.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4251.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4252.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4253.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4254.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4255.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4256.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4257.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4258.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4259.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4260.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4261.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4262.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4263.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4264.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4265.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4266.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4267.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4268.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4269.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4270.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4271.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4272.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4273.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4274.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4275.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4276.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4277.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4278.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4279.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4280.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4281.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4282.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4283.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4284.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4285.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4286.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4287.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4288.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4289.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4290.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4291.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4292.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4293.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4294.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4295.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4296.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4297.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4298.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4299.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4300.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4301.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4302.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4303.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4304.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4305.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4306.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4307.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4308.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4309.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4310.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4311.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4312.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4313.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4314.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4315.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4316.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4317.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4318.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4319.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4320.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4321.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4322.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4323.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4324.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4325.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4326.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4327.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4328.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4329.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4330.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4331.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4332.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4333.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4334.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4335.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4336.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4337.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4338.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4339.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4340.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4341.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4342.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4343.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4344.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4345.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4346.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4347.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4348.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4349.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4350.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4351.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4352.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4353.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4354.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4355.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4356.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4357.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4358.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4359.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4360.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4361.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4362.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4363.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4364.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4365.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4366.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4367.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4368.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4369.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4370.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4371.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4372.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4373.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4374.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4375.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4376.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4377.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4378.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4379.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4380.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4381.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4382.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4383.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4384.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4385.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4386.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4387.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4388.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4389.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4390.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4391.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4392.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4393.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4394.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4395.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4396.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4397.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4398.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4399.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4400.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4401.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4402.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4403.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4404.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4405.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4406.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4407.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4408.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4409.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4410.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4411.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4412.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4413.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4414.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4415.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4416.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4417.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4418.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4419.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4420.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4421.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4422.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4423.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4424.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4425.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4426.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4427.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4428.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4429.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4430.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4431.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4432.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4433.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4434.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4435.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4436.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4437.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4438.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4439.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4440.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4441.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4442.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4443.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4444.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4445.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4446.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4447.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4448.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4449.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4450.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4451.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4452.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4453.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4454.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4455.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4456.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4457.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4458.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4459.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4460.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4461.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4462.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4463.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4464.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4465.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4466.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4467.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4468.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4469.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4470.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4471.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4472.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4473.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4474.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4475.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4476.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4477.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4478.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4479.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4480.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4481.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4482.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4483.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4484.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4485.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4486.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4487.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4488.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4489.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4490.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4491.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4492.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4493.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4494.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4495.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4496.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4497.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4498.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4499.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4500.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4501.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4502.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4503.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4504.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4505.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4506.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4507.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4508.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4509.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4510.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4511.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4512.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4513.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4514.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4515.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4516.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4517.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4518.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4519.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4520.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4521.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4522.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4523.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4524.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4525.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4526.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4527.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4528.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4529.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4530.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4531.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4532.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4533.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4534.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4535.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4536.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4537.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4538.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4539.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4540.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4541.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4542.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4543.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4544.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4545.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4546.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4547.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4548.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4549.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4550.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4551.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4552.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4553.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4554.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4555.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4556.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4557.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4558.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4559.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4560.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4561.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4562.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4563.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4564.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4565.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4566.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4567.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4568.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4569.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4570.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4571.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4572.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4573.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4574.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4575.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4576.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4577.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4578.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4579.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4580.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4581.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4582.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4583.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4584.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4585.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4586.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4587.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4588.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4589.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4590.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4591.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4592.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4593.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4594.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4595.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4596.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4597.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4598.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4599.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4600.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4601.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4602.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4603.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4604.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4605.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4606.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4607.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4608.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4609.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4610.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4611.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4612.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4613.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4614.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4615.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4616.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4617.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4618.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4619.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4620.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4621.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4622.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4623.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4624.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4625.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4626.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4627.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4628.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4629.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4630.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4631.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4632.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4633.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4634.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4635.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4636.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4637.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4638.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4639.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4640.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4641.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4642.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4643.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4644.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4645.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4646.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4647.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4648.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4649.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4650.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4651.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4652.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4653.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4654.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4655.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4656.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4657.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4658.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4659.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4660.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4661.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4662.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4663.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4664.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4665.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4666.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4667.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4668.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4669.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4670.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4671.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4672.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4673.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4674.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4675.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4676.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4677.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4678.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4679.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4680.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4681.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4682.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4683.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4684.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4685.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4686.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4687.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4688.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4689.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4690.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4691.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4692.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4693.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4694.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4695.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4696.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4697.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4698.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4699.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4700.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4701.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4702.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4703.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4704.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4705.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4706.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4707.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4708.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4709.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4710.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4711.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4712.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4713.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4714.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4715.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4716.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4717.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4718.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4719.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4720.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4721.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4722.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4723.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4724.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4725.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4726.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4727.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4728.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4729.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4730.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4731.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4732.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4733.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4734.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4735.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4736.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4737.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4738.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4739.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4740.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4741.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4742.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4743.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4744.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4745.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4746.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4747.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4748.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4749.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4750.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4751.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4752.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4753.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4754.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4755.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4756.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4757.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4758.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4759.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4760.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4761.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4762.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4763.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4764.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4765.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4766.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4767.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4768.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4769.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4770.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4771.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4772.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4773.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4774.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4775.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4776.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4777.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4778.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4779.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4780.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4781.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4782.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4783.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4784.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4785.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4786.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4787.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4788.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4789.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4790.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4791.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4792.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4793.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4794.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4795.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4796.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4797.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4798.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4799.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4800.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4801.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4802.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4803.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4804.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4805.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4806.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4807.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4808.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4809.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4810.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4811.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4812.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4813.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4814.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4815.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4816.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4817.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4818.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4819.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4820.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4821.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4822.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4823.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4824.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4825.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4826.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4827.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4828.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4829.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4830.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4831.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4832.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4833.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4834.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4835.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4836.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4837.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4838.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4839.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4840.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4841.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4842.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4843.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4844.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4845.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4846.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4847.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4848.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4849.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4850.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4851.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4852.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4853.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4854.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4855.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4856.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4857.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4858.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4859.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4860.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4861.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4862.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4863.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4864.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4865.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4866.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4867.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4868.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4869.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4870.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4871.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4872.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4873.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4874.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4875.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4876.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4877.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4878.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4879.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4880.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4881.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4882.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4883.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4884.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4885.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4886.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4887.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4888.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4889.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4890.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4891.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4892.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4893.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4894.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4895.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4896.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4897.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4898.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4899.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4900.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4901.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4902.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4903.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4904.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4905.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4906.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4907.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4908.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4909.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4910.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4911.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4912.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4913.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4914.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4915.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4916.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4917.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4918.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4919.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4920.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4921.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4922.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4923.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4924.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4925.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4926.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4927.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4928.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4929.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4930.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4931.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4932.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4933.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4934.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4935.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4936.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4937.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4938.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4939.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4940.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4941.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4942.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4943.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4944.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4945.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4946.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4947.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4948.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4949.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4950.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4951.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4952.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4953.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4954.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4955.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4956.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4957.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4958.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4959.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4960.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4961.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4962.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4963.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4964.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4965.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4966.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4967.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4968.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4969.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4970.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4971.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4972.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4973.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4974.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4975.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4976.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4977.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4978.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4979.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4980.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4981.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4982.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4983.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4984.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4985.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4986.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4987.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4988.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4989.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4990.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4991.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4992.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4993.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4994.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4995.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4996.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4997.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4998.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train/data_4999.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n }\n ],\n ""episode_metadata_val"": [\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0000.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0001.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0002.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0003.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0004.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0005.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0006.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0007.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0008.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0009.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0010.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0011.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0012.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0013.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0014.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0015.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0016.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0017.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0018.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0019.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0020.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0021.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0022.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0023.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0024.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0025.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0026.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0027.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0028.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0029.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0030.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0031.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0032.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0033.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0034.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0035.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0036.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0037.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0038.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0039.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0040.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0041.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0042.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0043.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0044.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0045.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0046.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0047.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0048.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0049.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0050.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0051.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0052.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0053.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0054.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0055.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0056.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0057.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0058.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0059.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0060.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0061.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0062.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0063.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0064.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0065.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0066.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0067.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0068.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0069.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0070.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0071.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0072.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0073.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0074.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0075.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0076.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0077.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0078.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0079.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0080.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0081.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0082.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0083.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0084.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0085.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0086.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0087.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0088.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0089.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0090.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0091.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0092.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0093.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0094.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0095.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0096.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0097.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0098.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0099.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0100.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0101.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0102.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0103.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0104.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0105.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0106.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0107.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0108.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0109.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0110.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0111.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0112.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0113.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0114.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0115.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0116.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0117.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0118.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0119.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0120.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0121.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0122.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0123.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0124.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0125.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0126.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0127.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0128.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0129.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0130.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0131.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0132.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0133.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0134.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0135.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0136.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0137.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0138.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0139.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0140.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0141.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0142.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0143.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0144.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0145.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0146.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0147.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0148.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0149.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0150.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0151.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0152.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0153.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0154.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0155.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0156.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0157.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0158.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0159.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0160.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0161.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0162.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0163.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0164.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0165.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0166.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0167.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0168.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0169.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0170.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0171.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0172.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0173.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0174.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0175.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0176.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0177.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0178.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0179.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0180.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0181.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0182.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0183.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0184.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0185.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0186.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0187.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0188.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0189.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0190.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0191.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0192.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0193.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0194.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0195.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0196.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0197.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0198.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0199.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0200.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0201.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0202.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0203.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0204.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0205.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0206.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0207.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0208.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0209.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0210.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0211.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0212.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0213.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0214.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0215.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0216.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0217.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0218.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0219.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0220.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0221.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0222.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0223.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0224.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0225.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0226.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0227.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0228.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0229.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0230.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0231.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0232.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0233.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0234.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0235.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0236.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0237.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0238.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0239.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0240.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0241.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0242.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0243.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0244.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0245.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0246.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0247.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0248.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val/data_0249.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n }\n ],\n ""episode_metadata_test"": [\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0000.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0001.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0002.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0003.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0004.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0005.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0006.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0007.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0008.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0009.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0010.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0011.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0012.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0013.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0014.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0015.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0016.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0017.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0018.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0019.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0020.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0021.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0022.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0023.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0024.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0025.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0026.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0027.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0028.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0029.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0030.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0031.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0032.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0033.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0034.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0035.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0036.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0037.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0038.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0039.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0040.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0041.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0042.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0043.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0044.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0045.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0046.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0047.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0048.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0049.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0050.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0051.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0052.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0053.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0054.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0055.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0056.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0057.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0058.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0059.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0060.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0061.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0062.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0063.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0064.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0065.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0066.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0067.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0068.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0069.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0070.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0071.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0072.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0073.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0074.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0075.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0076.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0077.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0078.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0079.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0080.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0081.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0082.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0083.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0084.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0085.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0086.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0087.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0088.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0089.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0090.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0091.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0092.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0093.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0094.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0095.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0096.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0097.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0098.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0099.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0100.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0101.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0102.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0103.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0104.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0105.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0106.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0107.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0108.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0109.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0110.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0111.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0112.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0113.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0114.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0115.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0116.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0117.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0118.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0119.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0120.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0121.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0122.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0123.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0124.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0125.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0126.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0127.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0128.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0129.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0130.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0131.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0132.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0133.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0134.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0135.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0136.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0137.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0138.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0139.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0140.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0141.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0142.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0143.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0144.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0145.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0146.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0147.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0148.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0149.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0150.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0151.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0152.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0153.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0154.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0155.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0156.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0157.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0158.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0159.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0160.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0161.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0162.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0163.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0164.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0165.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0166.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0167.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0168.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0169.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0170.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0171.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0172.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0173.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0174.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0175.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0176.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0177.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0178.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0179.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0180.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0181.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0182.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0183.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0184.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0185.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0186.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0187.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0188.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0189.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0190.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0191.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0192.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0193.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0194.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0195.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0196.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0197.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0198.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0199.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0200.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0201.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0202.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0203.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0204.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0205.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0206.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0207.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0208.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0209.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0210.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0211.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0212.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0213.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0214.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0215.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0216.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0217.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0218.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0219.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0220.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0221.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0222.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0223.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0224.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0225.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0226.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0227.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0228.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0229.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0230.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0231.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0232.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0233.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0234.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0235.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0236.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0237.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0238.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0239.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0240.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0241.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0242.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0243.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0244.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0245.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0246.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0247.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0248.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n },\n {\n ""path"": ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/test/data_0249.array_record"",\n ""num_chunks"": 100,\n ""avg_seq_len"": 16.0\n }\n ]\n}",json,content +54,77467,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/metadata.json",81,0,"",json,selection_mouse +55,77701,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/metadata.json",81,1,"5",json,selection_mouse +56,77747,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/metadata.json",81,2,"50",json,selection_mouse +57,77969,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/metadata.json",81,3,"500",json,selection_mouse +58,79493,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/metadata.json",84,0,"",json,selection_mouse +59,80445,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/metadata.json",85,0,"",json,selection_mouse +60,80680,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/metadata.json",85,1,"0",json,selection_mouse +61,81119,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/metadata.json",86,0,"",json,selection_mouse +62,93877,"TERMINAL",0,0,"python",,terminal_command +63,93957,"TERMINAL",0,0,"]633;CPython 3.9.18 (main, Sep 4 2025, 00:00:00) \r\n[GCC 11.4.1 20231218 (Red Hat 11.4.1-3)] on linux\r\nType ""help"", ""copyright"", ""credits"" or ""license"" for more information.\r\n>>> ",,terminal_output +64,95756,"TERMINAL",0,0,"5",,terminal_output +65,95942,"TERMINAL",0,0,"0",,terminal_output +66,96256,"TERMINAL",0,0,"0",,terminal_output +67,96742,"TERMINAL",0,0,"",,terminal_output +68,96850,"TERMINAL",0,0,"",,terminal_output +69,97635,"TERMINAL",0,0,"0",,terminal_output +70,97686,"TERMINAL",0,0,"0",,terminal_output +71,98051,"TERMINAL",0,0,"_",,terminal_output +72,98552,"TERMINAL",0,0,"0",,terminal_output +73,98781,"TERMINAL",0,0,"0",,terminal_output +74,98845,"TERMINAL",0,0,"0",,terminal_output +75,99581,"TERMINAL",0,0,"*",,terminal_output +76,101063,"TERMINAL",0,0,"1",,terminal_output +77,101503,"TERMINAL",0,0,"6",,terminal_output +78,102026,"TERMINAL",0,0,"\r\n8000000\r\n>>> ",,terminal_output +79,107101,"TERMINAL",0,0,"5",,terminal_output +80,107167,"TERMINAL",0,0,"*",,terminal_output +81,107652,"TERMINAL",0,0,"1",,terminal_output +82,107801,"TERMINAL",0,0,"1",,terminal_output +83,107852,"TERMINAL",0,0,"0",,terminal_output +84,108267,"TERMINAL",0,0,"*",,terminal_output +85,108753,"TERMINAL",0,0,"1",,terminal_output +86,109833,"TERMINAL",0,0,"6",,terminal_output +87,110259,"TERMINAL",0,0,"",,terminal_output +88,110606,"TERMINAL",0,0,"",,terminal_output +89,110671,"TERMINAL",0,0,"",,terminal_output +90,110885,"TERMINAL",0,0,"",,terminal_output +91,111309,"TERMINAL",0,0,"",,terminal_output +92,111433,"TERMINAL",0,0,"",,terminal_output +93,112196,"TERMINAL",0,0,"_*110*16",,terminal_output +94,112440,"TERMINAL",0,0,"0*110*16",,terminal_output +95,112506,"TERMINAL",0,0,"0*110*16",,terminal_output +96,112876,"TERMINAL",0,0,"0*110*16",,terminal_output +97,113721,"TERMINAL",0,0,"\r\n8800000\r\n>>> ",,terminal_output +98,124763,"TERMINAL",0,0,"\r\nKeyboardInterrupt\r\n>>> ",,terminal_output +99,125381,"TERMINAL",0,0,"\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect",,terminal_output +100,126466,"TERMINAL",0,0,"srun",,terminal_focus +101,133700,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=15:00:00\n#SBATCH --partition=large\n#SBATCH --cpus-per-task=16\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/preprocess/breakout/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/preprocess/breakout/%x_%j.log\n#SBATCH --job-name=preprocess_breakout_chunked\n\nsource .venv/bin/activate\n\npython input_pipeline/generate_breakout_dataset.py \\n --num_episodes_train 500000 \\n --min_episode_length 16 \\n --max_episode_length 16 \\n --chunk_size 16 \\n --output_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect",shellscript,tab +102,135416,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",540,0,"",shellscript,selection_command +103,138868,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",540,0,"0",shellscript,content +104,138870,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",541,0,"",shellscript,selection_keyboard +105,139380,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",540,0,"",shellscript,selection_command +106,140330,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",738,0,"",shellscript,selection_mouse +107,140332,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",737,0,"",shellscript,selection_command +108,140695,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",738,0,"",shellscript,selection_command +109,140940,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",738,0,"_",shellscript,content +110,140944,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",739,0,"",shellscript,selection_keyboard +111,141671,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",739,0,"b",shellscript,content +112,141673,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",740,0,"",shellscript,selection_keyboard +113,141766,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",740,0,"i",shellscript,content +114,141767,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",741,0,"",shellscript,selection_keyboard +115,141927,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",741,0,"g",shellscript,content +116,141928,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",742,0,"",shellscript,selection_keyboard +117,142340,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",741,0,"",shellscript,selection_command +118,144415,"TERMINAL",0,0,"r",,terminal_output +119,144480,"TERMINAL",0,0,"u",,terminal_output +120,144547,"TERMINAL",0,0,"n",,terminal_output +121,144867,"TERMINAL",0,0,"n",,terminal_output +122,145006,"TERMINAL",0,0,"e",,terminal_output +123,145072,"TERMINAL",0,0,"r",,terminal_output +124,145159,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0403:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine_jobs]$ ",,terminal_output +125,145625,"TERMINAL",0,0,"s",,terminal_output +126,145864,"TERMINAL",0,0,"yn",,terminal_output +127,146441,"TERMINAL",0,0,"c",,terminal_output +128,146567,"TERMINAL",0,0,"-",,terminal_output +129,146809,"TERMINAL",0,0,"r",,terminal_output +130,146885,"TERMINAL",0,0,"u",,terminal_output +131,146950,"TERMINAL",0,0,"n",,terminal_output +132,147117,"TERMINAL",0,0,"n",,terminal_output +133,147187,"TERMINAL",0,0,"e",,terminal_output +134,147244,"TERMINAL",0,0,"r",,terminal_output +135,147363,"TERMINAL",0,0,"\r\n[?2004l\rsending incremental file list\r\n",,terminal_output +136,151058,"TERMINAL",0,0,"genie.py\r\nsample.py\r\ntrain_lam.py\r\n",,terminal_output +137,152054,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/default_runs/train_lam_default.sh\r\nslurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh\r\n",,terminal_output +138,152165,"TERMINAL",0,0,"\r\nsent 93,815 bytes received 310 bytes 17,113.64 bytes/sec\r\ntotal size is 128,633,714 speedup is 1,366.63\r\n]0;tum_cte0515@hkn0403:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine_jobs]$ ",,terminal_output +139,156741,"TERMINAL",0,0,"s",,terminal_output +140,156809,"TERMINAL",0,0,"b",,terminal_output +141,156934,"TERMINAL",0,0,"a",,terminal_output +142,157040,"TERMINAL",0,0,"t",,terminal_output +143,157150,"TERMINAL",0,0,"c",,terminal_output +144,157214,"TERMINAL",0,0,"h",,terminal_output +145,157353,"TERMINAL",0,0," ",,terminal_output +146,157655,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh",,terminal_output +147,158916,"TERMINAL",0,0,"\rslurm/jobs/mihir/horeka/preprocessing/breakout_chunked.sh\r\n[?2004l\rSubmitted batch job 3512651\r\n]0;tum_cte0515@hkn0403:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine_jobs]$ ",,terminal_output +148,162848,"TERMINAL",0,0,"d",,terminal_output +149,162984,"TERMINAL",0,0,"e",,terminal_output +150,163092,"TERMINAL",0,0,"v",,terminal_output +151,163180,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0403:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine]$ ",,terminal_output +152,313586,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --job-name=train_dyn_default_breakout_longer\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_10m_gt_actions_split_small/train\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_10m_gt_actions_split_small/val\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3502552\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3502552\n\nenv | grep SLURM\n\nexport PYTHONUNBUFFERED=1\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --image_height=10 \\n --image_width=10 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=120 \\n --init_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=500 \\n --log_checkpoint_interval=500 \\n --dyna_type=maskgit \\n --log \\n --name=breakout-dyn-default-$slurm_job_id \\n --tags dyn breakout default \\n --entity instant-uv \\n --project jafar \\n --patch_size 4 \\n --lam_patch_size 4 \\n --warmup_steps 100 \\n --wsd_decay_steps 1000 \\n --num_steps 5000 \\n --data_dir $array_records_dir_train \\n --val_data_dir $array_records_dir_val \\n --tokenizer_checkpoint $tokenizer_checkpoint \\n --lam_checkpoint $lam_checkpoint \\n --val_interval 500 \\n --eval_full_frame \\n",shellscript,tab +153,316003,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1136,0,"",shellscript,selection_command +154,317102,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1236,0,"",shellscript,selection_mouse +155,317119,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1235,0,"",shellscript,selection_command +156,317474,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1236,0,"",shellscript,selection_command +157,317907,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1229,7,"",shellscript,content +158,318009,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1228,1,"",shellscript,content +159,318085,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1217,11,"",shellscript,content +160,318245,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1216,1,"",shellscript,content +161,318393,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1213,3,"",shellscript,content +162,318521,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1212,1,"",shellscript,content +163,318666,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1204,8,"",shellscript,content +164,318806,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1203,1,"",shellscript,content +165,318955,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1192,11,"",shellscript,content +166,319099,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1191,1,"",shellscript,content +167,319602,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1177,14,"",shellscript,content +168,319621,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1176,1,"",shellscript,content +169,319683,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1165,11,"",shellscript,content +170,319914,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1164,1,"",shellscript,content +171,320473,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1157,7,"",shellscript,content +172,320533,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1156,1,"",shellscript,content +173,320603,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1147,9,"",shellscript,content +174,320776,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1146,1,"",shellscript,content +175,320925,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1142,4,"",shellscript,content +176,321077,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1141,1,"",shellscript,content +177,321233,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1137,4,"",shellscript,content +178,321572,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1135,2,"",shellscript,content +179,321912,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1135,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502",shellscript,content +180,324064,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1135,106,"",shellscript,content +181,324498,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1135,0,"=/",shellscript,content +182,326279,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,0,"",shellscript,selection_command +183,327393,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1265,0,"",shellscript,selection_mouse +184,328387,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,0,"",shellscript,selection_mouse +185,329228,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,1,"",shellscript,content +186,329736,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,4,"",shellscript,content +187,329753,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,1,"",shellscript,content +188,329819,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,4,"",shellscript,content +189,329820,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,1,"",shellscript,content +190,329839,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,9,"",shellscript,content +191,329907,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,1,"",shellscript,content +192,329908,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,7,"",shellscript,content +193,329931,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,1,"",shellscript,content +194,329964,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,11,"",shellscript,content +195,330023,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,1,"",shellscript,content +196,330024,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,14,"",shellscript,content +197,330082,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,1,"",shellscript,content +198,330083,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,11,"",shellscript,content +199,330113,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,1,"",shellscript,content +200,330178,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,8,"",shellscript,content +201,330182,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,1,"",shellscript,content +202,330206,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,9,"",shellscript,content +203,330270,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,1,"",shellscript,content +204,330271,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,11,"",shellscript,content +205,330303,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,1,"",shellscript,content +206,330618,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,7,"",shellscript,content +207,331244,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1159,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502",shellscript,content +208,338434,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1119,0,"",shellscript,selection_mouse +209,339060,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1137,0,"",shellscript,selection_mouse +210,339406,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1136,1,"",shellscript,content +211,339670,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1136,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576",shellscript,content +212,343666,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1256,0,"",shellscript,selection_mouse +213,344219,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1255,0,"",shellscript,selection_command +214,346344,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\n#SBATCH --job-name=train_dyn_default_breakout_longer\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_10m_gt_actions_split/train\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_10m_gt_actions_split/val\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3502552\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3502552\n\nenv | grep SLURM\n\nexport PYTHONUNBUFFERED=1\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --image_height=10 \\n --image_width=10 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=120 \\n --init_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=250 \\n --log_checkpoint_interval=250 \\n --dyna_type=maskgit \\n --log \\n --name=breakout-dyn-default-gt-actions-$slurm_job_id \\n --tags dyn breakout default \\n --entity instant-uv \\n --project jafar \\n --patch_size 4 \\n --lam_patch_size 4 \\n --warmup_steps 100 \\n --wsd_decay_steps 1000 \\n --num_steps 5000 \\n --use_gt_actions \\n --data_dir $array_records_dir_train \\n --val_data_dir $array_records_dir_val \\n --tokenizer_checkpoint $tokenizer_checkpoint \\n --val_interval 250 \\n --eval_full_frame \\n",shellscript,tab +215,347355,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1244,0,"",shellscript,selection_mouse +216,348121,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1352,0,"\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502",shellscript,content +217,348123,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1353,0,"",shellscript,selection_command +218,348443,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1225,0,"",shellscript,selection_command +219,348779,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1225,128,"",shellscript,content +220,350677,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1134,0,"",shellscript,selection_mouse +221,351411,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1109,116,"",shellscript,content +222,356585,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",903,0,"",shellscript,selection_mouse +223,358461,"TERMINAL",0,0,"bash",,terminal_focus +224,359728,"TERMINAL",0,0,"ls",,terminal_command +225,361001,"TERMINAL",0,0,"pwd",,terminal_command +226,364934,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",767,0,"",shellscript,selection_mouse +227,366048,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",729,38,"",shellscript,content +228,366485,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",729,0,"breakout_episodes_perfect",shellscript,content +229,367808,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",891,0,"",shellscript,selection_mouse +230,368706,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",890,1,"",shellscript,content +231,369147,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",853,37,"",shellscript,content +232,369523,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",853,0,"breakout_episodes_perfect",shellscript,content +233,370900,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +234,373547,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",916,0,"",shellscript,selection_mouse +235,375205,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",872,44,"",shellscript,content +236,375430,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",872,0,"breakout_episodes_perfect",shellscript,content +237,376823,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",773,0,"",shellscript,selection_mouse +238,377569,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",729,44,"",shellscript,content +239,377820,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",729,0,"breakout_episodes_perfect",shellscript,content +240,378598,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1198,0,"",shellscript,selection_mouse +241,379128,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1326,0,"",shellscript,selection_mouse +242,491809,"TERMINAL",0,0,"srun",,terminal_focus +243,494184,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",0,0,"",shellscript,tab +244,547375,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +245,549482,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",864,0,"",shellscript,selection_mouse +246,549681,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",853,25,"breakout_episodes_perfect",shellscript,selection_mouse +247,550213,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",743,0,"",shellscript,selection_mouse +248,550364,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",729,25,"breakout_episodes_perfect",shellscript,selection_mouse +249,555261,"TERMINAL",0,0,"s",,terminal_output +250,555315,"TERMINAL",0,0,"h",,terminal_output +251,555414,"TERMINAL",0,0," ",,terminal_output +252,555627,"TERMINAL",0,0,"breakout_episodes_perfect",,terminal_output +253,556531,"TERMINAL",0,0,"",,terminal_output +254,558062,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",0,0,"",shellscript,tab +255,560706,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1409,0,"",shellscript,selection_mouse +256,563153,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",,terminal_output +257,563934,"TERMINAL",0,0,"\rslurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=48:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=250 \\r\n --log_checkpoint_interval=250 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-default-gt-actions-$slurm_job_id \\r\n --tags dyn breakout default \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 100 \\r\n --wsd_decay_steps 1000 \\r\n --num_steps 5000 \\r\n --use_gt_actions \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_interval 250 \\r\n --eval_full_frame \\r\n",,terminal_output +258,564043,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=4095309\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0403\r\nSLURM_JOB_START_TIME=1758621669\r\nSLURM_STEP_NODELIST=hkn0403\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1758625269\r\nSLURM_PMI2_SRUN_PORT=33711\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3512647\r\nSLURM_PTY_PORT=37671\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=45\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0403\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=246\r\nSLURM_NODELIST=hkn0403\r\nSLURM_SRUN_COMM_PORT=35797\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3512647\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0403\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=35797\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0403\r\n",,terminal_output +259,564174,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +260,565754,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3512647.0 task 0: running\r\n",,terminal_output +261,565941,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3512647.0\r\nsrun: forcing job termination\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/train_dynamics.py"", line 11, in \r\n from jax.sharding import Mesh, PartitionSpec, NamedSharding\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/jax/__init__.py"", line 82, in \r\n from jax._src.api import effects_barrier as effects_barrier\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/jax/_src/api.py"", line 43, in \r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\n from jax._src import stages\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/jax/_src/stages.py"", line 50, in \r\n from jax._src.interpreters import mlir\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/jax/_src/interpreters/mlir.py"", line 44, in \r\n from jax._src import path\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/jax/_src/path.py"", line 35, in \r\nslurmstepd: error: *** STEP 3512647.0 ON hkn0403 CANCELLED AT 2025-09-23T12:10:29 ***\r\n]0;tum_cte0515@hkn0403:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine]$ ",,terminal_output +262,566092,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0403:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine]$ ",,terminal_output +263,569693,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",0,0,"",shellscript,tab +264,573143,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1999,0,"",shellscript,selection_mouse +265,574081,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1998,1,"",shellscript,content +266,574174,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1997,1,"",shellscript,content +267,574530,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1996,1,"",shellscript,content +268,575579,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1996,0,"7",shellscript,content +269,575580,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1997,0,"",shellscript,selection_keyboard +270,575657,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1997,0,"5",shellscript,content +271,575659,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1998,0,"",shellscript,selection_keyboard +272,575757,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1998,0,"0",shellscript,content +273,575758,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1999,0,"",shellscript,selection_keyboard +274,578500,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",,terminal_output +275,580326,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=48:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=250 \\r\n --log_checkpoint_interval=250 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-default-gt-actions-$slurm_job_id \\r\n --tags dyn breakout default \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 100 \\r\n --wsd_decay_steps 1000 \\r\n --num_steps 5000 \\r\n --use_gt_actions \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_interval 750 \\r\n --eval_full_frame \\r\n",,terminal_output +276,580447,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=4095309\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0403\r\nSLURM_JOB_START_TIME=1758621669\r\nSLURM_STEP_NODELIST=hkn0403\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1758625269\r\nSLURM_PMI2_SRUN_PORT=33711\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3512647\r\nSLURM_PTY_PORT=37671\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=45\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0403\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=246\r\nSLURM_NODELIST=hkn0403\r\nSLURM_SRUN_COMM_PORT=35797\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3512647\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0403\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=35797\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0403\r\n",,terminal_output +277,580580,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +278,583709,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +279,586070,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +280,587293,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",2119,0,"",shellscript,selection_mouse +281,588037,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",2120,0,"",shellscript,selection_command +282,588416,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",2117,3,"",shellscript,content +283,589271,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",2117,0,"7",shellscript,content +284,589272,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",2118,0,"",shellscript,selection_keyboard +285,589759,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",2118,0,"5",shellscript,content +286,589760,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",2119,0,"",shellscript,selection_keyboard +287,589823,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",2119,0,"0",shellscript,content +288,589824,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",2120,0,"",shellscript,selection_keyboard +289,590312,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",2119,0,"",shellscript,selection_command +290,591528,"TERMINAL",0,0,"Counting all components: ['action_embed', 'dynamics', 'tokenizer']\r\n",,terminal_output +291,591878,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +292,592662,"TERMINAL",0,0,"wandb: creating run\r\n",,terminal_output +293,592803,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.21.3\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20250923_121055-3rw70eln\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run breakout-dyn-default-gt-actions-3512647\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/3rw70eln\r\n",,terminal_output +294,593160,"TERMINAL",0,0,"Parameter counts:\r\n{'action_embed': 192, 'dynamics': 26555904, 'tokenizer': 33750256, 'total': 60306352}\r\n",,terminal_output +295,600625,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +296,611551,"TERMINAL",0,0,"2025-09-23 12:11:15.537518: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:11:15.537951: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:11:15.537971: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:11:15.539777: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +297,640178,"TERMINAL",0,0,"Total memory size: 1.7 GB, Output size: 0.7 GB, Temp size: 1.0 GB, Argument size: 0.7 GB, Host temp size: 0.0 GB.\r\n",,terminal_output +298,640245,"TERMINAL",0,0,"FLOPs: 3.823e+10, Bytes: 3.007e+10 (28.0 GB), Intensity: 1.3 FLOPs/byte\r\nStarting training from step 0...\r\n",,terminal_output +299,640534,"TERMINAL",0,0,"\r\nMemstats: After params initialized:\r\n\tUsing (GB) 0.85 / 38.7 (2.196382%) on cuda:0\r\n",,terminal_output +300,692634,"TERMINAL",0,0,"Saved checkpoint at step 250\r\n",,terminal_output +301,721748,"TERMINAL",0,0,"Saved checkpoint at step 500\r\n",,terminal_output +302,750670,"TERMINAL",0,0,"Calculating validation metrics...\r\n",,terminal_output +303,757470,"TERMINAL",0,0,"2025-09-23 12:13:41.268865: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:13:41.268909: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:13:41.269402: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:13:41.269423: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +304,810605,"TERMINAL",0,0,"Step 750, validation loss: 1.0221530199050903\r\n",,terminal_output +305,811289,"TERMINAL",0,0,"Saved checkpoint at step 750\r\n",,terminal_output +306,834945,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",0,0,"",shellscript,tab +307,837933,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1442,0,"",shellscript,selection_mouse +308,839001,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1441,1,"",shellscript,content +309,839051,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1441,0,"1",shellscript,content +310,839054,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",1442,0,"",shellscript,selection_keyboard +311,863139,"TERMINAL",0,0,"Saved checkpoint at step 1000\r\n",,terminal_output +312,893893,"TERMINAL",0,0,"Saved checkpoint at step 1250\r\n",,terminal_output +313,923975,"TERMINAL",0,0,"Calculating validation metrics...\r\n",,terminal_output +314,946208,"TERMINAL",0,0,"Step 1500, validation loss: 0.8760607838630676\r\n",,terminal_output +315,946612,"TERMINAL",0,0,"Saved checkpoint at step 1500\r\n",,terminal_output +316,977313,"TERMINAL",0,0,"Saved checkpoint at step 1750\r\n",,terminal_output +317,1002922,"genie.py",0,0,"from typing import Dict\n\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\nimport orbax.checkpoint as ocp\n\nfrom models.dynamics import DynamicsMaskGIT, DynamicsCausal\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\n\nclass Genie(nnx.Module):\n """"""Genie model""""""\n\n def __init__(\n self,\n in_dim: int,\n tokenizer_dim: int,\n tokenizer_ffn_dim: int,\n latent_patch_dim: int,\n num_patch_latents: int,\n patch_size: int,\n tokenizer_num_blocks: int,\n tokenizer_num_heads: int,\n lam_dim: int,\n lam_ffn_dim: int,\n latent_action_dim: int,\n num_actions: int,\n lam_patch_size: int,\n lam_num_blocks: int,\n lam_num_heads: int,\n lam_co_train: bool,\n use_gt_actions: bool,\n dyna_type: str,\n dyna_dim: int,\n dyna_ffn_dim: int,\n dyna_num_blocks: int,\n dyna_num_heads: int,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n dropout: float = 0.0,\n mask_limit: float = 0.0,\n ):\n # --- Tokenizer ---\n self.in_dim = in_dim\n self.tokenizer_dim = tokenizer_dim\n self.tokenizer_ffn_dim = tokenizer_ffn_dim\n self.latent_patch_dim = latent_patch_dim\n self.num_patch_latents = num_patch_latents\n self.patch_size = patch_size\n self.tokenizer_num_blocks = tokenizer_num_blocks\n self.tokenizer_num_heads = tokenizer_num_heads\n # --- LAM ---\n self.lam_dim = lam_dim\n self.lam_ffn_dim = lam_ffn_dim\n self.latent_action_dim = latent_action_dim\n self.num_actions = num_actions\n self.lam_patch_size = lam_patch_size\n self.lam_num_blocks = lam_num_blocks\n self.lam_num_heads = lam_num_heads\n self.lam_co_train = lam_co_train\n self.use_gt_actions = use_gt_actions\n # --- Dynamics ---\n self.dyna_type = dyna_type\n self.dyna_dim = dyna_dim\n self.dyna_ffn_dim = dyna_ffn_dim\n self.dyna_num_blocks = dyna_num_blocks\n self.dyna_num_heads = dyna_num_heads\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.dropout = dropout\n self.mask_limit = mask_limit\n self.decode = decode\n\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n ffn_dim=self.tokenizer_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n if self.use_gt_actions:\n self.action_embed = nnx.Embed(\n self.num_actions, self.latent_action_dim, rngs=rngs\n )\n self.lam = None\n else:\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n ffn_dim=self.lam_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.action_embed = None\n if self.dyna_type == ""maskgit"":\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n elif self.dyna_type == ""causal"":\n self.dynamics = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n rngs=rngs,\n )\n else:\n raise ValueError(f""Invalid dynamics type: {self.dyna_type}"")\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n training: bool = True,\n ) -> Dict[str, jax.Array]:\n videos_BTHWC = batch[""videos""]\n tokenizer_outputs = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_indices_BTN = tokenizer_outputs[""indices""]\n latent_actions_BTm11L = None\n action_embeddings_BTm11L = None\n if self.use_gt_actions:\n assert self.action_embed is not None\n action_indices_E = None\n action_embeddings_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n action_embeddings_BTm11L = action_embeddings_BT1L[:, :-1]\n else:\n assert self.lam is not None\n lam_outputs = self.lam.vq_encode(videos_BTHWC, training=False)\n z_q_BTm11L = lam_outputs[""z_q""]\n action_indices_E = lam_outputs[""indices""]\n latent_actions_BTm11L = jax.lax.cond(\n self.lam_co_train,\n lambda: z_q_BTm11L,\n lambda: jax.lax.stop_gradient(z_q_BTm11L),\n )\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(token_indices_BTN),\n latent_actions=(\n action_embeddings_BTm11L\n if self.use_gt_actions\n else latent_actions_BTm11L\n ),\n )\n outputs[""mask_rng""] = batch[""rng""]\n dyna_logits_BTNV, dyna_mask = self.dynamics(outputs, training)\n outputs[""token_logits""] = dyna_logits_BTNV\n outputs[""mask""] = dyna_mask\n mle_indices_BTN = jnp.argmax(outputs[""token_logits""], axis=-1)\n H, W = batch[""videos""].shape[2:4]\n outputs[""recon""] = self.tokenizer.decode(mle_indices_BTN, (H, W))\n if action_indices_E is not None:\n outputs[""lam_indices""] = action_indices_E\n return outputs\n\n def sample(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n maskgit_steps: int = 25,\n ) -> tuple[jax.Array, jax.Array]:\n if self.dyna_type == ""maskgit"":\n return self.sample_maskgit(\n batch, seq_len, maskgit_steps, temperature, sample_argmax\n )\n elif self.dyna_type == ""causal"":\n return self.sample_causal(batch, seq_len, temperature, sample_argmax)\n else:\n raise ValueError(f""Dynamics model type unknown: {self.dyna_type}"")\n\n def sample_maskgit(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n steps: int = 25,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n P: S * N\n """"""\n assert isinstance(self.dynamics, DynamicsMaskGIT)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n init_logits_BSNV = jnp.zeros(\n shape=(*token_idxs_BSN.shape, self.num_patch_latents)\n )\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, :-1]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n # --- Extract submodule state ---\n dynamics_state = nnx.state(self.dynamics)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def maskgit_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, mask_BSN, action_tokens_EL = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_maskgit = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_maskgit, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_maskgit.patch_embed(token_idxs_BSN)\n mask_token_111M = dynamics_maskgit.mask_token.value\n mask_expanded_BSN1 = mask_BSN[..., None]\n vid_embed_BSNM = jnp.where(\n mask_expanded_BSN1, mask_token_111M, vid_embed_BSNM\n )\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_maskgit.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n vid_embed_BSNM += act_embed_BS1M\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (steps * 2))\n step_temp = temperature * (1.0 - unmasked_ratio)\n final_logits_BSNV = dynamics_maskgit.transformer(vid_embed_BSNM) / step_temp\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_BSN = jnp.argmax(final_logits_BSNV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_BSN = jax.random.categorical(_rng, final_logits_BSNV)\n gather_fn = jax.vmap(jax.vmap(jax.vmap(lambda x, y: x[y])))\n final_token_probs_BSN = gather_fn(\n jax.nn.softmax(final_logits_BSNV), sampled_token_idxs_BSN\n )\n final_token_probs_BSN += ~mask_BSN\n # Update masked tokens and logits only\n token_idxs_BSN = jnp.where(mask_BSN, sampled_token_idxs_BSN, token_idxs_BSN)\n logits_BSNV = jnp.where(\n jnp.expand_dims(mask_BSN, -1), final_logits_BSNV, logits_BSNV\n )\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n final_token_probs_flat_BP = einops.rearrange(\n final_token_probs_BSN, ""b s n -> b (s n)""\n )\n idx_mask_P = (\n jnp.arange(final_token_probs_flat_BP.shape[-1])\n <= N - num_unmasked_tokens\n )\n sorted_idxs_BP = jnp.argsort(final_token_probs_flat_BP, axis=-1)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask_P))\n mask_flat_BP = einops.rearrange(mask_BSN, ""b s n -> b (s n)"")\n new_mask_flat_BP = mask_update_fn(mask_flat_BP, sorted_idxs_BP)\n new_mask_BSN = einops.rearrange(new_mask_flat_BP, ""b (s n) -> b s n"", n=N)\n\n new_carry = (\n rng,\n token_idxs_BSN,\n logits_BSNV,\n new_mask_BSN,\n action_tokens_EL,\n )\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # Mask current frame (i.e., t == step_t)\n mask_S = jnp.arange(seq_len) == step_t\n mask_BSN = jnp.broadcast_to(mask_S[None, :, None], (B, seq_len, N)).astype(\n bool\n )\n masked_token_idxs_BSN = current_token_idxs_BSN * ~mask_BSN\n masked_logits_BSNV = current_logits_BSNV * jnp.expand_dims(~mask_BSN, -1)\n\n # --- Initialize and run MaskGIT loop ---\n init_carry_maskgit = (\n step_rng,\n masked_token_idxs_BSN,\n masked_logits_BSNV,\n mask_BSN,\n action_tokens_EL,\n )\n final_carry_maskgit = maskgit_step_fn(init_carry_maskgit, jnp.arange(steps))\n updated_token_idxs_BSN = final_carry_maskgit[1]\n updated_logits_BSNV = final_carry_maskgit[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, init_logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def sample_causal(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n assert isinstance(self.dynamics, DynamicsCausal)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n logits_BSNV = jnp.zeros((*token_idxs_BSN.shape, self.num_patch_latents))\n dynamics_state = nnx.state(self.dynamics)\n\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, 1:]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def causal_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step_n: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_causal = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=self.decode,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_causal, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_causal.patch_embed(token_idxs_BSN)\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_causal.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n vid_embed_BSNp1M = jnp.concatenate([act_embed_BS1M, vid_embed_BSNM], axis=2)\n final_logits_BTNp1V = (\n dynamics_causal.transformer(vid_embed_BSNp1M, (step_t, step_n))\n / temperature\n )\n final_logits_BV = final_logits_BTNp1V[:, step_t, step_n, :]\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_B = jnp.argmax(final_logits_BV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_B = jax.random.categorical(_rng, final_logits_BV)\n # Update next tokens only\n token_idxs_BSN = token_idxs_BSN.at[:, step_t, step_n].set(\n sampled_token_idxs_B\n )\n logits_BSNV = logits_BSNV.at[:, step_t, step_n].set(final_logits_BV)\n\n new_carry = (rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t)\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # --- Initialize and run causal loop ---\n init_carry_causal = (\n step_rng,\n current_token_idxs_BSN,\n current_logits_BSNV,\n action_tokens_EL,\n step_t,\n )\n final_carry_causal = causal_step_fn(init_carry_causal, jnp.arange(N))\n updated_token_idxs_BSN = final_carry_causal[1]\n updated_logits_BSNV = final_carry_causal[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def vq_encode(self, batch: Dict[str, jax.Array], training: bool) -> jax.Array:\n # --- Preprocess videos ---\n assert self.lam is not None\n video_BTHWC = batch[""videos""]\n lam_output: Dict[str, jax.Array] = self.lam.vq_encode(\n video_BTHWC, training=training\n )\n lam_indices_E = lam_output[""indices""]\n return lam_indices_E\n\n\n# FIXME (f.srambical): add conversion script for old checkpoints\ndef restore_genie_components(\n optimizer: nnx.Optimizer,\n sharding: jax.sharding.NamedSharding,\n rng: jax.Array,\n args,\n) -> nnx.Optimizer:\n """"""Restore pre-trained Genie components""""""\n rng_tokenizer, rng_lam = jax.random.split(rng)\n rngs_tokenizer = nnx.Rngs(rng_tokenizer)\n rngs_lam = nnx.Rngs(rng_lam)\n\n tx = optimizer.tx\n model = optimizer.model\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n tokenizer_checkpoint_manager = ocp.CheckpointManager(\n directory=args.tokenizer_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n ffn_dim=args.tokenizer_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_tokenizer,\n )\n dummy_tokenizer_optimizer = nnx.Optimizer(dummy_tokenizer, tx)\n dummy_tokenizer_optimizer_state = nnx.state(dummy_tokenizer_optimizer)\n abstract_sharded_tokenizer_optimizer_state = _create_abstract_sharded_pytree(\n dummy_tokenizer_optimizer_state, sharding\n )\n restored_tokenizer = tokenizer_checkpoint_manager.restore(\n step=tokenizer_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_tokenizer_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_tokenizer_optimizer.model, restored_tokenizer.model)\n model.tokenizer = dummy_tokenizer_optimizer.model\n tokenizer_checkpoint_manager.close()\n\n if args.lam_checkpoint:\n lam_checkpoint_manager = ocp.CheckpointManager(\n directory=args.lam_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n ffn_dim=args.lam_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_lam,\n )\n dummy_lam_optimizer = nnx.Optimizer(dummy_lam, tx)\n dummy_lam_optimizer_state = nnx.state(dummy_lam_optimizer)\n abstract_sharded_lam_optimizer_state = _create_abstract_sharded_pytree(\n dummy_lam_optimizer_state, sharding\n )\n restored_lam_optimizer = lam_checkpoint_manager.restore(\n step=lam_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_lam_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_lam_optimizer.model, restored_lam_optimizer.model)\n model.lam = dummy_lam_optimizer.model\n # Remove the LAM decoder to save memory and avoid unnecessary computation.\n del model.lam.decoder\n lam_checkpoint_manager.close()\n\n # Reinitialize the optimizer states\n optimizer = nnx.Optimizer(model, tx)\n return optimizer\n\n\ndef _create_abstract_sharded_pytree(\n pytree_template: nnx.GraphState, sharding_spec: jax.sharding.NamedSharding\n) -> jax.Array:\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)\n",python,tab +318,1002926,"genie.py",7231,6,"sample",python,selection_command +319,1007673,"TERMINAL",0,0,"Saved checkpoint at step 2000\r\n",,terminal_output +320,1037543,"TERMINAL",0,0,"Calculating validation metrics...\r\n",,terminal_output +321,1053096,"genie.py",18523,0,"",python,selection_mouse +322,1054420,"genie.py",18522,1,"",python,content +323,1054786,"genie.py",18523,0,"",python,selection_command +324,1055399,"genie.py",18523,0,"-",python,content +325,1055401,"genie.py",18524,0,"",python,selection_keyboard +326,1055471,"genie.py",18524,0,"1",python,content +327,1055472,"genie.py",18525,0,"",python,selection_keyboard +328,1057553,"genie.py",18524,0,"",python,selection_command +329,1059848,"TERMINAL",0,0,"Step 2250, validation loss: 0.7719724178314209\r\n",,terminal_output +330,1060290,"TERMINAL",0,0,"Saved checkpoint at step 2250\r\n",,terminal_output +331,1091137,"TERMINAL",0,0,"Saved checkpoint at step 2500\r\n",,terminal_output +332,1121557,"TERMINAL",0,0,"Saved checkpoint at step 2750\r\n",,terminal_output +333,1151473,"TERMINAL",0,0,"Calculating validation metrics...\r\n",,terminal_output +334,1173722,"TERMINAL",0,0,"Step 3000, validation loss: 0.7339697480201721\r\n",,terminal_output +335,1174087,"TERMINAL",0,0,"Saved checkpoint at step 3000\r\n",,terminal_output +336,1181731,"train_tokenizer.py",0,0,"import os\n\nos.environ.setdefault(""XLA_PYTHON_CLIENT_MEM_FRACTION"", ""0.98"")\n\nfrom dataclasses import dataclass, field\nfrom typing import cast, Optional\n\nimport einops\nimport itertools\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\nfrom utils.train_utils import (\n get_lr_schedule,\n count_parameters_by_component,\n print_mem_stats,\n print_compiled_memory_stats,\n print_compiled_cost_analysis,\n)\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n init_lr: float = 0.0\n max_lr: float = 3e-4\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 20000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n ffn_dim: int = 2048\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 4\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\ndef build_model(args: Args, rng: jax.Array) -> tuple[TokenizerVQVAE, jax.Array]:\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n return (\n TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n ffn_dim=args.ffn_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n ),\n rng,\n )\n\n\ndef build_optimizer(\n model: TokenizerVQVAE, args: Args\n) -> tuple[nnx.Optimizer, optax.Schedule]:\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.param_dtype, # moments in full precision\n )\n optimizer = nnx.Optimizer(model, tx)\n return optimizer, lr_schedule\n\n\ndef build_mesh_and_sharding(\n num_devices: int,\n) -> tuple[Mesh, NamedSharding, NamedSharding]:\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n return mesh, replicated_sharding, videos_sharding\n\n\ndef shard_optimizer_states(\n optimizer: nnx.Optimizer, replicated_sharding: NamedSharding\n) -> None:\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n\ndef build_dataloader(args: Args, data_dir: str) -> grain.DataLoaderIterator:\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(data_dir, x)\n for x in os.listdir(data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n return grain_iterator\n\n\ndef build_checkpoint_manager(args: Args) -> ocp.CheckpointManager:\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n return checkpoint_manager\n\n\ndef restore_checkpoint_if_needed(\n args: Args,\n checkpoint_manager: ocp.CheckpointManager,\n optimizer: nnx.Optimizer,\n train_iterator: grain.DataLoaderIterator,\n val_iterator: Optional[grain.DataLoaderIterator],\n restore_step: Optional[int] = None,\n) -> tuple[int, nnx.Optimizer, grain.DataLoaderIterator, grain.DataLoaderIterator]:\n step = 0\n if restore_step is None:\n restore_step = checkpoint_manager.latest_step()\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n if val_iterator:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n val_dataloader_state=grain.checkpoint.CheckpointRestore(val_iterator), # type: ignore\n )\n else:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n )\n restored = checkpoint_manager.restore(restore_step, args=restore_args)\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n train_iterator = restored[""train_dataloader_state""]\n if val_iterator:\n val_iterator = restored[""val_dataloader_state""]\n step = restore_step or 0\n print(f""Restored dataloader and model state from step {step}"")\n return step, optimizer, train_iterator, val_iterator\n\n\ndef main(args: Args) -> None:\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n tokenizer, rng = build_model(args, rng)\n\n _, params, _ = nnx.split(tokenizer, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n optimizer, lr_schedule = build_optimizer(tokenizer, args)\n del tokenizer\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n _, replicated_sharding, videos_sharding = build_mesh_and_sharding(num_devices)\n\n shard_optimizer_states(optimizer, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n checkpoint_manager = build_checkpoint_manager(args)\n\n # --- Create DataLoaderIterator from dataloader ---\n train_iterator = build_dataloader(args, args.data_dir)\n val_iterator = None\n if args.val_data_dir:\n val_iterator = build_dataloader(args, args.val_data_dir)\n\n # --- Restore checkpoint ---\n step, optimizer, train_iterator, val_iterator = restore_checkpoint_if_needed(\n args, checkpoint_manager, optimizer, train_iterator, val_iterator\n )\n\n # --- Define loss and train step (close over args) ---\n def tokenizer_loss_fn(\n model: TokenizerVQVAE, inputs: dict, training: bool = False\n ) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n mse = jnp.square(gt - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n gt_clipped = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt_clipped, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt_clipped, recon)).mean()\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n @nnx.jit(donate_argnums=0)\n def train_step(\n optimizer: nnx.Optimizer, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(model: TokenizerVQVAE) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n model.train()\n return tokenizer_loss_fn(model, inputs, training=True)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n optimizer.model\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return loss, recon, metrics\n\n @nnx.jit\n def val_step(\n tokenizer: TokenizerVQVAE, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n tokenizer.eval()\n (loss, (recon, metrics)) = tokenizer_loss_fn(tokenizer, inputs, training=False)\n return loss, recon, metrics\n\n def calculate_validation_metrics(val_dataloader, tokenizer):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n batch = None\n recon = None\n for batch in val_dataloader:\n loss, recon, metrics = val_step(tokenizer, batch)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(\n f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}""\n )\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([float(m[key]) for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n val_metrics[""val_loss""] = val_loss\n return val_metrics, batch, recon\n\n # --- TRAIN LOOP ---\n dataloader_train = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n }\n for elem in train_iterator\n )\n dataloader_val = None\n if val_iterator:\n dataloader_val = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n }\n for elem in val_iterator\n )\n if jax.process_index() == 0:\n first_batch = next(dataloader_train)\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader_train = itertools.chain([first_batch], dataloader_train)\n print(f""Starting training from step {step}..."")\n first_step = step\n while step < args.num_steps:\n for batch in dataloader_train:\n # --- Train step ---\n loss, recon, metrics = train_step(optimizer, batch)\n if step == first_step:\n print_mem_stats(""After params initialized"")\n # metrics[""lr""] = lr_schedule(step)\n # print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Validation loss ---\n val_results = {}\n if dataloader_val and step % args.val_interval == 0:\n print(""Calculating validation metrics..."")\n val_metrics, val_gt_batch, val_recon = calculate_validation_metrics(\n dataloader_val, optimizer.model\n )\n print(f""Step {step}, validation loss: {val_metrics['val_loss']}"")\n val_results = {\n ""metrics"": val_metrics,\n ""gt_batch"": val_gt_batch,\n ""recon"": val_recon,\n }\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n log_dict = {""loss"": loss, ""step"": step, **metrics}\n if val_results:\n log_dict.update(val_results[""metrics""])\n wandb.log(log_dict)\n if step % args.log_image_interval == 0:\n gt_seq = batch[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if val_results and step % args.val_interval == 0:\n val_results[""gt_seq_val""] = (\n val_results[""gt_batch""][""videos""][0].astype(jnp.float32)\n / 255.0\n )\n val_results[""recon_seq_val""] = val_results[""recon""][0].clip(\n 0, 1\n )\n val_results[""val_comparison_seq""] = jnp.concatenate(\n (val_results[""gt_seq_val""], val_results[""recon_seq_val""]),\n axis=1,\n )\n val_results[""val_comparison_seq""] = einops.rearrange(\n val_results[""val_comparison_seq""] * 255,\n ""t h w c -> h (t w) c"",\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n if val_results and step % args.val_interval == 0:\n log_images.update(\n dict(\n val_image=wandb.Image(\n np.asarray(val_results[""gt_seq_val""][0])\n ),\n val_recon=wandb.Image(\n np.asarray(val_results[""recon_seq_val""][0])\n ),\n val_true_vs_recon=wandb.Image(\n np.asarray(\n val_results[""val_comparison_seq""].astype(\n np.uint8\n )\n )\n ),\n )\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n if val_iterator:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n val_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n val_iterator # type: ignore\n ),\n )\n else:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n )\n checkpoint_manager.save(step, args=ckpt_manager_args)\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n main(args)\n",python,tab +337,1182806,"train_dynamics.py",0,0,"import os\n\n\nos.environ.setdefault(""XLA_PYTHON_CLIENT_MEM_FRACTION"", ""0.98"")\n\nfrom dataclasses import dataclass, field\nimport itertools\nfrom typing import cast, Optional\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\nfrom utils.train_utils import (\n get_lr_schedule,\n count_parameters_by_component,\n print_mem_stats,\n print_compiled_memory_stats,\n print_compiled_cost_analysis,\n)\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n # Tokenizer\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_type: str = ""maskgit"" # supported options: maskgit, causal\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n use_gt_actions: bool = False\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n eval_full_frame: bool = False\n val_maskgit_steps: int = 25\n val_temperature: float = 1\n val_sample_argmax: bool = False\n wandb_id: str = """"\n\n\ndef build_model(args: Args, rng: jax.Array) -> tuple[Genie, jax.Array]:\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_actions=args.num_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n use_gt_actions=args.use_gt_actions,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n decode=False,\n rngs=rngs,\n )\n if args.use_gt_actions:\n assert (\n not args.lam_checkpoint\n ), ""Cannot use LAM when using ground-truth actions.""\n else:\n assert genie.lam is not None\n del genie.lam.decoder\n return genie, rng\n\n\ndef build_optimizer(genie: Genie, args: Args) -> tuple[nnx.Optimizer, optax.Schedule]:\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.param_dtype, # moments in full precision\n )\n optimizer = nnx.Optimizer(genie, tx)\n return optimizer, lr_schedule\n\n\ndef build_mesh_and_sharding(\n num_devices: int,\n) -> tuple[Mesh, NamedSharding, NamedSharding, NamedSharding]:\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n actions_sharding = NamedSharding(mesh, PartitionSpec(""data"", None))\n return mesh, replicated_sharding, videos_sharding, actions_sharding\n\n\ndef shard_optimizer_states(\n optimizer: nnx.Optimizer, replicated_sharding: NamedSharding\n) -> None:\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n\ndef build_dataloader(args: Args, data_dir: str) -> grain.DataLoaderIterator:\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(data_dir, x)\n for x in os.listdir(data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n return grain_iterator\n\n\ndef build_checkpoint_manager(args: Args) -> ocp.CheckpointManager:\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n return checkpoint_manager\n\n\ndef restore_or_initialize_components(\n args: Args,\n checkpoint_manager: ocp.CheckpointManager,\n optimizer: nnx.Optimizer,\n train_iterator: grain.DataLoaderIterator,\n rng: jax.Array,\n replicated_sharding: NamedSharding,\n val_iterator: Optional[grain.DataLoaderIterator],\n restore_step: Optional[int] = None,\n) -> tuple[\n int, nnx.Optimizer, grain.DataLoaderIterator, grain.DataLoaderIterator, jax.Array\n]:\n step = 0\n if restore_step is None:\n restore_step = checkpoint_manager.latest_step()\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n if val_iterator:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n val_dataloader_state=grain.checkpoint.CheckpointRestore(val_iterator), # type: ignore\n )\n else:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n )\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(), args=restore_args\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n train_iterator = restored[""train_dataloader_state""]\n if val_iterator:\n val_iterator = restored[""val_dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n rng, _rng = jax.random.split(rng)\n optimizer = restore_genie_components(optimizer, replicated_sharding, _rng, args)\n # NOTE: We have to remove the (unused) tokenizer vq dropout due flax.nnx lazily initializing modules.\n # Specifically, the first dynamics model checkpoint will contain the vq dropout module,\n # but the first full restore will fail due to nnx not initializing the module when\n # dropout is set to 0.0.\n del optimizer.model.tokenizer.vq.drop\n return step, optimizer, train_iterator, val_iterator, rng\n\n\ndef _calculate_step_metrics(\n outputs: dict[str, jax.Array],\n gt: jax.Array,\n num_actions: int,\n num_patch_latents: int,\n) -> tuple[jax.Array, dict]:\n mask = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n gt_val = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt_val, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt_val, recon)).mean()\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]),\n size=num_patch_latents,\n fill_value=0,\n )\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n psnr=psnr,\n ssim=ssim,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n if ""lam_indices"" in outputs.keys():\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]),\n size=num_actions,\n fill_value=0,\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n metrics[""codebook_usage_lam""] = codebook_usage_lam\n return ce_loss, metrics\n\n\ndef main(args: Args) -> None:\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n genie, rng = build_model(args, rng)\n _, params, _ = nnx.split(genie, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n optimizer, lr_schedule = build_optimizer(genie, args)\n del genie\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n _, replicated_sharding, videos_sharding, actions_sharding = build_mesh_and_sharding(\n num_devices\n )\n\n shard_optimizer_states(optimizer, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n checkpoint_manager = build_checkpoint_manager(args)\n\n # --- Create DataLoaderIterator from dataloader ---\n train_iterator = build_dataloader(args, args.data_dir)\n val_iterator = None\n if args.val_data_dir:\n val_iterator = build_dataloader(args, args.val_data_dir)\n\n # --- Restore checkpoint ---\n step, optimizer, train_iterator, val_iterator, rng = (\n restore_or_initialize_components(\n args,\n checkpoint_manager,\n optimizer,\n train_iterator,\n rng,\n replicated_sharding,\n val_iterator,\n )\n )\n\n # --- Define loss and train step (close over args) ---\n def dynamics_loss_fn(\n model: Genie,\n inputs: dict,\n training: bool = False,\n ) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n ce_loss, metrics = _calculate_step_metrics(\n outputs, gt, args.num_actions, args.num_patch_latents\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n @nnx.jit(donate_argnums=0)\n def train_step(\n optimizer: nnx.Optimizer, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(model: Genie) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n model.train()\n return dynamics_loss_fn(model, inputs, training=True)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n optimizer.model\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return loss, recon, metrics\n\n @nnx.jit\n def val_step(genie: Genie, inputs: dict) -> dict:\n """"""Evaluate model and compute metrics""""""\n genie.eval()\n (loss, (recon, metrics)) = dynamics_loss_fn(genie, inputs, training=False)\n val_output = {""loss"": loss, ""recon"": recon, ""metrics"": metrics}\n\n # --- Evaluate full frame prediction (sampling) ---\n if args.eval_full_frame:\n tokenizer_outputs = genie.tokenizer.vq_encode(\n inputs[""videos""], training=False\n )\n tokens_full_frame = tokenizer_outputs[""indices""]\n lam_indices = None\n if not args.use_gt_actions:\n lam_indices = genie.vq_encode(inputs, training=False)\n inputs[""latent_actions""] = lam_indices\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt[:, :-1].astype(\n args.dtype\n ) # remove last frame for generation\n recon_full_frame, logits_full_frame = genie.sample(\n inputs,\n args.seq_len,\n args.val_temperature,\n args.val_sample_argmax,\n args.val_maskgit_steps,\n )\n step_outputs = {\n ""recon"": recon_full_frame,\n ""token_logits"": logits_full_frame,\n ""video_tokens"": tokens_full_frame,\n ""mask"": jnp.zeros_like(tokens_full_frame).at[:, -1].set(True),\n }\n if lam_indices is not None:\n step_outputs[""lam_indices""] = lam_indices\n\n loss_full_frame, metrics_full_frame = _calculate_step_metrics(\n step_outputs, gt, args.num_actions, args.num_patch_latents\n )\n val_output.update(\n {\n ""loss_full_frame"": loss_full_frame,\n ""recon_full_frame"": recon_full_frame,\n ""metrics_full_frame"": metrics_full_frame,\n }\n )\n return val_output\n\n def calculate_validation_metrics(val_dataloader, genie, rng):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n loss_full_frame_per_step = []\n metrics_full_frame_per_step = []\n batch = None\n recon = None\n recon_full_frame = None\n for batch in val_dataloader:\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n val_outputs = val_step(genie, batch)\n loss_per_step.append(val_outputs[""loss""])\n metrics_per_step.append(val_outputs[""metrics""])\n recon = val_outputs[""recon""]\n if args.eval_full_frame:\n loss_full_frame_per_step.append(val_outputs[""loss_full_frame""])\n metrics_full_frame_per_step.append(val_outputs[""metrics_full_frame""])\n recon_full_frame = val_outputs[""recon_full_frame""]\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(\n f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}""\n )\n\n val_metrics = {\n f""val_{key}"": np.mean([float(m[key]) for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n val_metrics[""val_loss""] = np.mean(loss_per_step)\n if args.eval_full_frame:\n val_metrics_full_frame = {\n f""val_full_frame_{key}"": np.mean(\n [float(m[key]) for m in metrics_full_frame_per_step]\n )\n for key in metrics_full_frame_per_step[0].keys()\n }\n val_metrics.update(val_metrics_full_frame)\n val_metrics[""val_loss_full_frame""] = np.mean(loss_full_frame_per_step)\n return val_metrics, batch, recon, recon_full_frame\n\n # --- TRAIN LOOP ---\n dataloader_train = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in train_iterator\n )\n dataloader_val = None\n if val_iterator:\n dataloader_val = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in val_iterator\n )\n if jax.process_index() == 0:\n first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader_train = itertools.chain([first_batch], dataloader_train)\n print(f""Starting training from step {step}..."")\n first_step = step\n while step < args.num_steps:\n for batch in dataloader_train:\n # --- Train step ---\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n loss, recon, metrics = train_step(optimizer, batch)\n if step == first_step:\n print_mem_stats(""After params initialized"")\n # metrics[""lr""] = lr_schedule(step)\n # print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Validation loss ---\n val_results = {}\n if dataloader_val and step % args.val_interval == 0:\n rng, _rng_mask_val = jax.random.split(rng, 2)\n print(""Calculating validation metrics..."")\n val_metrics, val_gt_batch, val_recon, val_recon_full_frame = (\n calculate_validation_metrics(\n dataloader_val, optimizer.model, _rng_mask_val\n )\n )\n print(f""Step {step}, validation loss: {val_metrics['val_loss']}"")\n val_results = {\n ""metrics"": val_metrics,\n ""gt_batch"": val_gt_batch,\n ""recon"": val_recon,\n ""full_frame"": val_recon_full_frame,\n }\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n log_dict = {""loss"": loss, ""step"": step, **metrics}\n if val_results:\n log_dict.update(val_results[""metrics""])\n wandb.log(log_dict)\n if step % args.log_image_interval == 0:\n gt_seq = batch[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if val_results:\n val_results[""gt_seq_val""] = (\n val_results[""gt_batch""][""videos""][0].astype(jnp.float32)\n / 255.0\n )\n val_results[""recon_seq_val""] = val_results[""recon""][0].clip(\n 0, 1\n )\n val_comparison_seq = jnp.concatenate(\n (val_results[""gt_seq_val""], val_results[""recon_seq_val""]),\n axis=1,\n )\n val_results[""val_comparison_seq""] = einops.rearrange(\n val_comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if args.eval_full_frame:\n val_results[""full_frame_seq_val""] = val_results[\n ""full_frame""\n ][0].clip(0, 1)\n val_results[""val_full_frame_comparison_seq""] = (\n jnp.concatenate(\n (\n val_results[""gt_seq_val""],\n val_results[""full_frame_seq_val""],\n ),\n axis=1,\n )\n )\n val_results[""val_full_frame_comparison_seq""] = (\n einops.rearrange(\n val_results[""val_full_frame_comparison_seq""] * 255,\n ""t h w c -> h (t w) c"",\n )\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n if val_results:\n log_images.update(\n dict(\n val_image=wandb.Image(\n np.asarray(\n val_results[""gt_seq_val""][args.seq_len - 1]\n )\n ),\n val_recon=wandb.Image(\n np.asarray(\n val_results[""recon_seq_val""][\n args.seq_len - 1\n ]\n )\n ),\n val_true_vs_recon=wandb.Image(\n np.asarray(\n val_results[""val_comparison_seq""].astype(\n np.uint8\n )\n )\n ),\n )\n )\n if args.eval_full_frame:\n log_images.update(\n dict(\n val_full_frame=wandb.Image(\n np.asarray(\n val_results[""full_frame_seq_val""][\n args.seq_len - 1\n ]\n )\n ),\n val_true_vs_full_frame=wandb.Image(\n np.asarray(\n val_results[\n ""val_full_frame_comparison_seq""\n ].astype(np.uint8)\n )\n ),\n )\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n if val_iterator:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n val_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n val_iterator # type: ignore\n ),\n )\n else:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n )\n checkpoint_manager.save(step, args=ckpt_manager_args)\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n main(args)\n",python,tab +338,1193094,"train_dynamics.py",22357,0,"",python,selection_mouse +339,1193118,"train_dynamics.py",22356,0,"",python,selection_command +340,1193228,"train_dynamics.py",22356,1,"1",python,selection_mouse +341,1193243,"train_dynamics.py",22357,0,"",python,selection_command +342,1193298,"train_dynamics.py",22325,32,": {loss}"")\n step += 1",python,selection_mouse +343,1193298,"train_dynamics.py",22321,36,"loss: {loss}"")\n step += 1",python,selection_mouse +344,1193312,"train_dynamics.py",22317,40,"p}, loss: {loss}"")\n step += 1",python,selection_mouse +345,1193335,"train_dynamics.py",22313,44,"{step}, loss: {loss}"")\n step += 1",python,selection_mouse +346,1193350,"train_dynamics.py",22308,49,"Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +347,1193365,"train_dynamics.py",22305,52,"(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +348,1193418,"train_dynamics.py",22303,54,"nt(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +349,1193419,"train_dynamics.py",22299,58," print(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +350,1193467,"train_dynamics.py",22298,59,"# print(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +351,1193468,"train_dynamics.py",22297,60," # print(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +352,1193523,"train_dynamics.py",22296,61," # print(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +353,1193580,"train_dynamics.py",22295,62," # print(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +354,1193581,"train_dynamics.py",22293,64," # print(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +355,1193593,"train_dynamics.py",22292,65," # print(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +356,1193616,"train_dynamics.py",22291,66," # print(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +357,1193630,"train_dynamics.py",22290,67," # print(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +358,1193657,"train_dynamics.py",22241,116," # metrics[""lr""] = lr_schedule(step)\n # print(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +359,1193673,"train_dynamics.py",22240,117," # metrics[""lr""] = lr_schedule(step)\n # print(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +360,1193699,"train_dynamics.py",22239,118," # metrics[""lr""] = lr_schedule(step)\n # print(f""Step {step}, loss: {loss}"")\n step += 1",python,selection_mouse +361,1194207,"train_dynamics.py",22239,0,"",python,selection_mouse +362,1194208,"train_dynamics.py",22238,12," ",python,selection_mouse +363,1194429,"train_dynamics.py",22238,50," # metrics[""lr""] = lr_schedule(step)\n ",python,selection_mouse +364,1194430,"train_dynamics.py",22238,51," # metrics[""lr""] = lr_schedule(step)\n ",python,selection_mouse +365,1194458,"train_dynamics.py",22238,52," # metrics[""lr""] = lr_schedule(step)\n ",python,selection_mouse +366,1194482,"train_dynamics.py",22238,53," # metrics[""lr""] = lr_schedule(step)\n ",python,selection_mouse +367,1194559,"train_dynamics.py",22238,104," # metrics[""lr""] = lr_schedule(step)\n # print(f""Step {step}, loss: {loss}"")\n ",python,selection_mouse +368,1196017,"train_dynamics.py",22342,0,"",python,selection_mouse +369,1204801,"TERMINAL",0,0,"Saved checkpoint at step 3250\r\n",,terminal_output +370,1235158,"TERMINAL",0,0,"Saved checkpoint at step 3500\r\n",,terminal_output +371,1235804,"train_dynamics.py",0,0,"",python,tab +372,1236053,"train_dynamics.py",16790,0,"",python,selection_command +373,1243457,"train_dynamics.py",0,0,"",python,tab +374,1265022,"TERMINAL",0,0,"Calculating validation metrics...\r\n",,terminal_output +375,1287282,"TERMINAL",0,0,"Step 3750, validation loss: 0.739980936050415\r\n",,terminal_output +376,1287584,"TERMINAL",0,0,"Saved checkpoint at step 3750\r\n",,terminal_output +377,1318043,"TERMINAL",0,0,"Saved checkpoint at step 4000\r\n",,terminal_output +378,1348952,"TERMINAL",0,0,"Saved checkpoint at step 4250\r\n",,terminal_output +379,1378627,"TERMINAL",0,0,"Calculating validation metrics...\r\n",,terminal_output +380,1400950,"TERMINAL",0,0,"Step 4500, validation loss: 0.6556053161621094\r\n",,terminal_output +381,1401277,"TERMINAL",0,0,"Saved checkpoint at step 4500\r\n",,terminal_output +382,1431987,"TERMINAL",0,0,"Saved checkpoint at step 4750\r\n",,terminal_output +383,1462615,"TERMINAL",0,0,"Saved checkpoint at step 5000\r\n",,terminal_output +384,1516181,"TERMINAL",0,0,"wandb: \r\nwandb: 🚀 View run breakout-dyn-default-gt-actions-3512647 at: https://wandb.ai/instant-uv/jafar/runs/3rw70eln\r\nwandb: Find logs at: ../../../../../hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20250923_121055-3rw70eln/logs\r\n",,terminal_output +385,1517976,"TERMINAL",0,0,"]0;tum_cte0515@hkn0403:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine]$ ",,terminal_output +386,1971016,"train_dynamics.py",22552,0,"",python,selection_mouse +387,1971017,"train_dynamics.py",22551,0,"",python,selection_command +388,1971721,"TERMINAL",0,0,"bash",,terminal_focus +389,1977444,"TERMINAL",0,0,"srun",,terminal_focus +390,1977918,"TERMINAL",0,0,"q",,terminal_output +391,1978101,"TERMINAL",0,0,"u",,terminal_output +392,1978171,"TERMINAL",0,0,"e",,terminal_output +393,1978235,"TERMINAL",0,0,"u",,terminal_output +394,1978460,"TERMINAL",0,0,"e",,terminal_output +395,1978597,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: squeue --mehkn0403.localdomain: Tue Sep 23 12:34:02 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3511158 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3511159 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)3512647 dev_accel interact tum_cte0 R32:53\t 1 hkn04033512651 large preproce tum_cte0 R30:17\t 1 hkn1901",,terminal_output +396,1979889,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0403:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine]$ ",,terminal_output +397,1984072,"TERMINAL",0,0,"queue",,terminal_output +398,1984752,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh",,terminal_output +399,1986018,"TERMINAL",0,0,"",,terminal_output +400,1986545,"TERMINAL",0,0,"",,terminal_output +401,1986915,"TERMINAL",0,0,"",,terminal_output +402,1987077,"TERMINAL",0,0,"",,terminal_output +403,1987262,"TERMINAL",0,0,"",,terminal_output +404,1987717,"TERMINAL",0,0,"gt_actions.sh ",,terminal_output +405,1988358,"TERMINAL",0,0,"",,terminal_output +406,1989126,"TERMINAL",0,0,"",,terminal_output +407,1989352,"TERMINAL",0,0,"",,terminal_output +408,1989534,"TERMINAL",0,0,"",,terminal_output +409,1989687,"TERMINAL",0,0,"",,terminal_output +410,1989842,"TERMINAL",0,0,"",,terminal_output +411,1990020,"TERMINAL",0,0,"",,terminal_output +412,1990351,"TERMINAL",0,0,".",,terminal_output +413,1990514,"TERMINAL",0,0,"sh ",,terminal_output +414,1996762,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +415,1999730,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1591,0,"",shellscript,selection_mouse +416,1999859,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1589,3,"500",shellscript,selection_mouse +417,2003391,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",1558,0,"",shellscript,selection_mouse +418,2006150,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=48:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=500 \\r\n --log_checkpoint_interval=500 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-default-$slurm_job_id \\r\n --tags dyn breakout default \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 100 \\r\n --wsd_decay_steps 1000 \\r\n --num_steps 5000 \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --lam_checkpoint $lam_checkpoint \\r\n --val_interval 750 \\r\n --eval_full_frame \\r\n",,terminal_output +419,2006270,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=4095309\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0403\r\nSLURM_JOB_START_TIME=1758621669\r\nSLURM_STEP_NODELIST=hkn0403\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1758625269\r\nSLURM_PMI2_SRUN_PORT=33711\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3512647\r\nSLURM_PTY_PORT=37671\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=45\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0403\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=246\r\nSLURM_NODELIST=hkn0403\r\nSLURM_SRUN_COMM_PORT=35797\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3512647\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0403\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=35797\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0403\r\n",,terminal_output +420,2006412,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +421,2008647,"TERMINAL",0,0,"bash",,terminal_focus +422,2009113,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +423,2011207,"TERMINAL",0,0,"dev",,terminal_command +424,2012664,"TERMINAL",0,0,"git status",,terminal_command +425,2012715,"TERMINAL",0,0,"]633;C",,terminal_output +426,2012961,"TERMINAL",0,0,"On branch generate-minatar-breakout-dataset\r\n",,terminal_output +427,2013027,"TERMINAL",0,0,"Your branch is ahead of 'origin/generate-minatar-breakout-dataset' by 10 commits.\r\n (use ""git push"" to publish your local commits)\r\n\r\n",,terminal_output +428,2013082,"TERMINAL",0,0,"Last commands done (2 commands done):\r\n pick ba37453 feat: generate coinrun dataset with val split\r\n pick faadd10 feat: implemented validation loss for all three models\r\nNext commands to do (26 remaining commands):\r\n pick 9a17dbb fix: pass val data path to dataloader\r\n pick 6e69cdb fix typo in image logging\r\n (use ""git rebase --edit-todo"" to view and edit)\r\nYou are currently editing a commit while rebasing branch 'gt-actions' on 'c7522f2'.\r\n (use ""git commit --amend"" to amend the current commit)\r\n (use ""git rebase --continue"" once you are satisfied with your changes)\r\n\r\nChanges not staged for commit:\r\n (use ""git add ..."" to update what will be committed)\r\n (use ""git restore ..."" to discard changes in working directory)\r\n\tmodified: genie.py\r\n\tmodified: input_pipeline/generate_breakout_dataset.py\r\n\tmodified: sample.py\r\n\tmodified: train_dynamics.py\r\n\tmodified: train_lam.py\r\n\tmodified: train_tokenizer.py\r\n\r\nUntracked files:\r\n (use ""git add ..."" to include in what will be committed)\r\n\tdiff.diff\r\n\tinput_pipeline/generate_breakout_dataset_agent.py\r\n\tkiller.sh\r\n\tkiller_partition.sh\r\n\tlog.log\r\n\toverfit_dir.zip\r\n\trequirements-franz.txt\r\n\tsamples/\r\n\tscripts_cremers/\r\n\tslurm/\r\n\ttest.py\r\n\tutils/visualizer.py\r\n\r\nno changes added to commit (use ""git add"" and/or ""git commit -a"")\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +429,2015303,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +430,2015657,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +431,2016491,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.21.3\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20250923_123439-jhpw2lqa\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run breakout-dyn-default-3512647\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/jhpw2lqa\r\n",,terminal_output +432,2016600,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 16900976, 'tokenizer': 33750256, 'total': 77207136}\r\n",,terminal_output +433,2018776,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +434,2018792,"TERMINAL",0,0,"git diff",,terminal_command +435,2018890,"TERMINAL",0,0,"]633;C[?1h=\rdiff --git a/genie.py b/genie.py\r\nindex bcb23e6..7fa3f40 100644\r\n--- a/genie.py\r\n+++ b/genie.py\r\n@@ -263,7 +263,7 @@ class Genie(nnx.Module):\r\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\r\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\r\n )\r\n- latent_actions_BTm11L = latent_actions_BT1L[:, 1:]\r\n+ latent_actions_BTm11L = latent_actions_BT1L[:, :-1]\r\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\r\n else:\r\n assert self.lam is not None\r\n@@ -452,7 +452,7 @@ class Genie(nnx.Module):\r\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\r\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\r\n )\r\n:",,terminal_output +436,2019442,"TERMINAL",0,0,"\r- latent_actions_BTm11L = latent_actions_BT1L[:, 1:]\r\n:",,terminal_output +437,2019688,"TERMINAL",0,0,"\r+ latent_actions_BTm11L = latent_actions_BT1L[:, :-1]\r\n:",,terminal_output +438,2019872,"TERMINAL",0,0,"\r action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\r\n:",,terminal_output +439,2020044,"TERMINAL",0,0,"\r else:\r\n:",,terminal_output +440,2020191,"TERMINAL",0,0,"\r assert self.lam is not None\r\n:",,terminal_output +441,2020303,"TERMINAL",0,0,"\rdiff --git a/input_pipeline/generate_breakout_dataset.py b/input_pipeline/generate_breakout_dataset.py\r\n:",,terminal_output +442,2020474,"TERMINAL",0,0,"\rindex 88928b9..4aa5586 100644\r\n:",,terminal_output +443,2020646,"TERMINAL",0,0,"\r--- a/input_pipeline/generate_breakout_dataset.py\r\n:",,terminal_output +444,2020705,"TERMINAL",0,0,"\r+++ b/input_pipeline/generate_breakout_dataset.py\r\n:",,terminal_output +445,2020862,"TERMINAL",0,0,"\r@@ -117,8 +117,8 @@ def generate_episodes(num_episodes: int, split: str):\r\n:",,terminal_output +446,2021701,"TERMINAL",0,0,"\r obs_chunks.extend(obs_chunks_data)\r\n:\r act_chunks.extend(act_chunks_data)\r\n:\r \r\n:\r- ep_metadata, obs_chunks, file_idx, act_chunks = save_chunks(\r\n:\r- obs_chunks, file_idx, args.chunks_per_file, output_dir_split, act_chunks\r\n:\r+ ep_metadata, file_idx, obs_chunks, act_chunks = save_chunks(\r\n:\r+ file_idx, args.chunks_per_file, output_dir_split, obs_chunks, act_chunks\r\n:\r )\r\n:\r episode_metadata.extend(ep_metadata)\r\n:\r \r\n:\rdiff --git a/sample.py b/sample.py\r\n:\rindex b5d5a22..6cf11b4 100644\r\n:",,terminal_output +447,2021868,"TERMINAL",0,0,"\r--- a/sample.py\r\n:",,terminal_output +448,2022308,"TERMINAL",0,0,"\r+++ b/sample.py\r\n:",,terminal_output +449,2022528,"TERMINAL",0,0,"\r@@ -237,7 +237,7 @@ if __name__ == ""__main__"":\r\n:",,terminal_output +450,2022691,"TERMINAL",0,0,"\r if action_batch_E is not None:\r\n:",,terminal_output +451,2022831,"TERMINAL",0,0,"\r action_batch_BSm11 = jnp.reshape(action_batch_E, (B, S - 1, 1))\r\n:",,terminal_output +452,2023001,"TERMINAL",0,0,"\r else:\r\n:",,terminal_output +453,2023175,"TERMINAL",0,0,"\r- action_batch_BSm11 = jnp.expand_dims(batch[""actions""][:, 1:], -1)\r\n:",,terminal_output +454,2023298,"TERMINAL",0,0,"\r+ action_batch_BSm11 = jnp.expand_dims(batch[""actions""][:, :-1], -1)\r\n:",,terminal_output +455,2023917,"TERMINAL",0,0,"\r for t, img in enumerate(imgs[1:]):\r\n:\r d = ImageDraw.Draw(img)\r\n:\r for row in range(B):\r\n:\rdiff --git a/train_dynamics.py b/train_dynamics.py\r\n:\rindex b34ba98..ad89aae 100644\r\n:",,terminal_output +456,2024008,"TERMINAL",0,0,"\r--- a/train_dynamics.py\r\n:\r+++ b/train_dynamics.py\r\n:",,terminal_output +457,2024315,"TERMINAL",0,0,"\r@@ -482,12 +482,14 @@ def main(args: Args) -> None:\r\n:\r \r\n:\r # --- Evaluate full frame prediction (sampling) ---\r\n:\r if args.eval_full_frame:\r\n:\r- lam_indices = genie.vq_encode(inputs, training=False)\r\n:\r tokenizer_outputs = genie.tokenizer.vq_encode(\r\n:\r inputs[""videos""], training=False\r\n:\r )\r\n:\r tokens_full_frame = tokenizer_outputs[""indices""]\r\n:\r- inputs[""latent_actions""] = lam_indices\r\n:\r+ lam_indices = None\r\n:",,terminal_output +458,2024629,"TERMINAL",0,0,"\rM for t, img in enumerate(imgs[1:]):\r\n\r:",,terminal_output +459,2025521,"TERMINAL",0,0,"\rM+ action_batch_BSm11 = jnp.expand_dims(batch[""actions""][:, :-1], -1)\r\n\r:\rM- action_batch_BSm11 = jnp.expand_dims(batch[""actions""][:, 1:], -1)\r\n\r:\rM else:\r\n\r:\rM action_batch_BSm11 = jnp.reshape(action_batch_E, (B, S - 1, 1))\r\n\r:\rM if action_batch_E is not None:\r\n\r:\rM@@ -237,7 +237,7 @@ if __name__ == ""__main__"":\r\n\r:\rM+++ b/sample.py\r\n\r:\rM--- a/sample.py\r\n\r:\rMindex b5d5a22..6cf11b4 100644\r\n\r:\rMdiff --git a/sample.py b/sample.py\r\n\r:\rM \r\n\r:\rM episode_metadata.extend(ep_metadata)\r\n\r:\rM )\r\n\r:\rM+ file_idx, args.chunks_per_file, output_dir_split, obs_chunks, act_chunks\r\n\r:\rM+ ep_metadata, file_idx, obs_chunks, act_chunks = save_chunks(\r\n\r:",,terminal_output +460,2025938,"TERMINAL",0,0,"\r for row in range(B):\r\n:",,terminal_output +461,2026178,"TERMINAL",0,0,"\rdiff --git a/train_dynamics.py b/train_dynamics.py\r\n:",,terminal_output +462,2026304,"TERMINAL",0,0,"\rindex b34ba98..ad89aae 100644\r\n:",,terminal_output +463,2026539,"TERMINAL",0,0,"\r--- a/train_dynamics.py\r\n:",,terminal_output +464,2027439,"TERMINAL",0,0,"\r+++ b/train_dynamics.py\r\n:",,terminal_output +465,2027805,"TERMINAL",0,0,"\r@@ -482,12 +482,14 @@ def main(args: Args) -> None:\r\n:",,terminal_output +466,2029606,"TERMINAL",0,0,"\rMdiff --git a/sample.py b/sample.py\r\n\r:",,terminal_output +467,2030509,"TERMINAL",0,0,"\rM \r\n\r:\rM episode_metadata.extend(ep_metadata)\r\n\r:\rM )\r\n\r:\rM+ file_idx, args.chunks_per_file, output_dir_split, obs_chunks, act_chunks\r\n\r:\rM+ ep_metadata, file_idx, obs_chunks, act_chunks = save_chunks(\r\n\r:\rM- obs_chunks, file_idx, args.chunks_per_file, output_dir_split, act_chunks\r\n\r:\rM- ep_metadata, obs_chunks, file_idx, act_chunks = save_chunks(\r\n\r:\rM \r\n\r:\rM act_chunks.extend(act_chunks_data)\r\n\r:\rM obs_chunks.extend(obs_chunks_data)\r\n\r:\rM@@ -117,8 +117,8 @@ def generate_episodes(num_episodes: int, split: str):\r\n\r:\rM+++ b/input_pipeline/generate_breakout_dataset.py\r\n\r:\rM--- a/input_pipeline/generate_breakout_dataset.py\r\n\r:\rMindex 88928b9..4aa5586 100644\r\n\r:\rMdiff --git a/input_pipeline/generate_breakout_dataset.py b/input_pipeline/generate_breakout_dataset.py\r\n\r:",,terminal_output +468,2030587,"TERMINAL",0,0,"\rM assert self.lam is not None\r\n\r:\rM else:\r\n\r:\rM action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\r\n\r:",,terminal_output +469,2031685,"TERMINAL",0,0,"\r \r\n:",,terminal_output +470,2032206,"TERMINAL",0,0,"\rdiff --git a/sample.py b/sample.py\r\n:",,terminal_output +471,2032380,"TERMINAL",0,0,"\rindex b5d5a22..6cf11b4 100644\r\n:\r--- a/sample.py\r\n:\r+++ b/sample.py\r\n:\r@@ -237,7 +237,7 @@ if __name__ == ""__main__"":\r\n:\r if action_batch_E is not None:\r\n:",,terminal_output +472,2032453,"TERMINAL",0,0,"\r action_batch_BSm11 = jnp.reshape(action_batch_E, (B, S - 1, 1))\r\n:\r else:\r\n:\r- action_batch_BSm11 = jnp.expand_dims(batch[""actions""][:, 1:], -1)\r\n:\r+ action_batch_BSm11 = jnp.expand_dims(batch[""actions""][:, :-1], -1)\r\n:",,terminal_output +473,2032534,"TERMINAL",0,0,"\r for t, img in enumerate(imgs[1:]):\r\n:\r d = ImageDraw.Draw(img)\r\n:",,terminal_output +474,2032821,"TERMINAL",0,0,"\r for row in range(B):\r\n:\rdiff --git a/train_dynamics.py b/train_dynamics.py\r\n:\rindex b34ba98..ad89aae 100644\r\n:\r--- a/train_dynamics.py\r\n:\r+++ b/train_dynamics.py\r\n:\r@@ -482,12 +482,14 @@ def main(args: Args) -> None:\r\n:\r \r\n:\r # --- Evaluate full frame prediction (sampling) ---\r\n:\r if args.eval_full_frame:\r\n:",,terminal_output +475,2033027,"TERMINAL",0,0,"\r- lam_indices = genie.vq_encode(inputs, training=False)\r\n:\r tokenizer_outputs = genie.tokenizer.vq_encode(\r\n:\r inputs[""videos""], training=False\r\n:\r )\r\n:\r tokens_full_frame = tokenizer_outputs[""indices""]\r\n:\r- inputs[""latent_actions""] = lam_indices\r\n:\r+ lam_indices = None\r\n:",,terminal_output +476,2033250,"TERMINAL",0,0,"\r+ if not args.use_gt_actions:\r\n:\r+ lam_indices = genie.vq_encode(inputs, training=False)\r\n:\r+ inputs[""latent_actions""] = lam_indices\r\n:\r gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\r\n:\r inputs[""videos""] = gt[:, :-1].astype(\r\n:",,terminal_output +477,2033305,"TERMINAL",0,0,"\r args.dtype\r\n:",,terminal_output +478,2033372,"TERMINAL",0,0,"\r@@ -504,8 +506,10 @@ def main(args: Args) -> None:\r\n:\r ""token_logits"": logits_full_frame,\r\n:\r ""video_tokens"": tokens_full_frame,\r\n:\r ""mask"": jnp.zeros_like(tokens_full_frame).at[:, -1].set(True),\r\n:\r- ""lam_indices"": lam_indices,\r\n:\r }\r\n:\r+ if lam_indices is not None:\r\n:",,terminal_output +479,2033674,"TERMINAL",0,0,"2025-09-23 12:34:57.659398: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:34:57.659836: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:34:57.659863: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:34:57.660278: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:34:57.662044: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:34:57.663597: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +480,2039304,"TERMINAL",0,0,"srun",,terminal_focus +481,2040016,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3512647.2 task 0: running\r\n",,terminal_output +482,2040617,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3512647.2\r\nsrun: forcing job termination\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-3:\r\n2025-09-23 12:35:04.553034: F external/xla/xla/service/gpu/autotuning/gemm_fusion_autotuner.cc:1136] Non-OK-status: executable.status()\r\nStatus: INTERNAL: ptxas exited with non-zero error code 2, output: - Failure occured when compiling fusion gemm_fusion_dot.84 with config '{block_m:128,block_n:256,block_k:32,split_k:1,num_stages:4,num_warps:8,num_ctas:1}'\r\nFused HLO computation:\r\n%gemm_fusion_dot.84_computation (parameter_0.86: bf16[1800,32], parameter_1.86: bf16[32,512]) -> bf16[1800,512] {\r\n %parameter_0.86 = bf16[1800,32]{1,0} parameter(0)\r\n %parameter_1.86 = bf16[32,512]{1,0} parameter(1)\r\n ROOT %dot.625 = bf16[1800,512]{1,0} dot(%parameter_0.86, %parameter_1.86), lhs_contracting_dims={1}, rhs_contracting_dims={0}, metadata={op_name=""jit(train_step)/jit(main)/dot_general"" source_file=""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/flax/nnx/nn/linear.py"" source_line=387}\r\n}\r\nProcess SpawnProcess-2:\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3512647.2 ON hkn0403 CANCELLED AT 2025-09-23T12:35:04 ***\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\nKeyboardInterrupt\r\nKeyboardInterrupt\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\nTraceback (most recent call last):\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\nTraceback (most recent call last):\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\nKeyboardInterrupt\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\nKeyboardInterrupt\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\nKeyboardInterrupt\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\nKeyboardInterrupt\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\nKeyboardInterrupt\r\nTraceback (most recent call last):\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\nKeyboardInterrupt\r\n",,terminal_output +483,2040846,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3512647.2\r\nsrun: job abort in progress\r\n",,terminal_output +484,2040912,"TERMINAL",0,0,"]0;tum_cte0515@hkn0403:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine]$ ",,terminal_output +485,2041054,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0403:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine]$ ",,terminal_output +486,2042231,"TERMINAL",0,0,"r",,terminal_output +487,2042310,"TERMINAL",0,0,"un",,terminal_output +488,2042438,"TERMINAL",0,0,"n",,terminal_output +489,2042581,"TERMINAL",0,0,"e",,terminal_output +490,2042634,"TERMINAL",0,0,"r",,terminal_output +491,2043937,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0403:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine_jobs]$ ",,terminal_output +492,2044214,"TERMINAL",0,0,"s",,terminal_output +493,2044382,"TERMINAL",0,0,"y",,terminal_output +494,2044455,"TERMINAL",0,0,"n",,terminal_output +495,2044526,"TERMINAL",0,0,"c",,terminal_output +496,2044717,"TERMINAL",0,0,"-",,terminal_output +497,2044947,"TERMINAL",0,0,"r",,terminal_output +498,2045083,"TERMINAL",0,0,"un",,terminal_output +499,2045293,"TERMINAL",0,0,"ne",,terminal_output +500,2045345,"TERMINAL",0,0,"r",,terminal_output +501,2045434,"TERMINAL",0,0,"\r\n[?2004l\rsending incremental file list\r\ngenie.py\r\nslurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh\r\nslurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu_gt_actions.sh\r\n",,terminal_output +502,2045499,"TERMINAL",0,0,"\r\nsent 65,055 bytes received 272 bytes 130,654.00 bytes/sec\r\ntotal size is 128,633,535 speedup is 1,969.07\r\n]0;tum_cte0515@hkn0403:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine_jobs]$ ",,terminal_output +503,2045852,"TERMINAL",0,0,"sync-runner",,terminal_output +504,2046015,"TERMINAL",0,0,"runner",,terminal_output +505,2046408,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh ",,terminal_output +506,2047157,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=48:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=500 \\r\n --log_checkpoint_interval=500 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-default-$slurm_job_id \\r\n --tags dyn breakout default \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 100 \\r\n --wsd_decay_steps 1000 \\r\n --num_steps 5000 \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --lam_checkpoint $lam_checkpoint \\r\n --val_interval 750 \\r\n --eval_full_frame \\r\n",,terminal_output +507,2047294,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh: line 20: .venv/bin/activate: No such file or directory\r\n]0;tum_cte0515@hkn0403:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine_jobs]$ ",,terminal_output +508,2050365,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh ",,terminal_output +509,2053236,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +510,2053237,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",608,0,"",shellscript,selection_command +511,2055314,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",608,0,"#",shellscript,content +512,2055316,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",609,0,"",shellscript,selection_keyboard +513,2055410,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",609,0," ",shellscript,content +514,2055411,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",610,0,"",shellscript,selection_keyboard +515,2055717,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",609,0,"",shellscript,selection_command +516,2057281,"TERMINAL",0,0,"\rync-runner",,terminal_output +517,2057648,"TERMINAL",0,0,"h slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh ",,terminal_output +518,2058112,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=48:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=500 \\r\n --log_checkpoint_interval=500 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-default-$slurm_job_id \\r\n --tags dyn breakout default \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 100 \\r\n --wsd_decay_steps 1000 \\r\n --num_steps 5000 \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --lam_checkpoint $lam_checkpoint \\r\n --val_interval 750 \\r\n --eval_full_frame \\r\n",,terminal_output +519,2058225,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh: line 20: .venv/bin/activate: No such file or directory\r\n]0;tum_cte0515@hkn0403:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine_jobs]$ ",,terminal_output +520,2059624,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh ",,terminal_output +521,2061668,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +522,2063919,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=48:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=500 \\r\n --log_checkpoint_interval=500 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-default-$slurm_job_id \\r\n --tags dyn breakout default \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 100 \\r\n --wsd_decay_steps 1000 \\r\n --num_steps 5000 \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --lam_checkpoint $lam_checkpoint \\r\n --val_interval 750 \\r\n --eval_full_frame \\r\n",,terminal_output +523,2064057,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh: line 20: .venv/bin/activate: No such file or directory\r\n]0;tum_cte0515@hkn0403:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine_jobs]$ ",,terminal_output +524,2065430,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh ",,terminal_output +525,2069828,"TERMINAL",0,0,"\r",,terminal_output +526,2070243,"TERMINAL",0,0,"s",,terminal_output +527,2070555,"TERMINAL",0,0,"y",,terminal_output +528,2070620,"TERMINAL",0,0,"n",,terminal_output +529,2070756,"TERMINAL",0,0,"c",,terminal_output +530,2070822,"TERMINAL",0,0,"-",,terminal_output +531,2071012,"TERMINAL",0,0,"r",,terminal_output +532,2071173,"TERMINAL",0,0,"un",,terminal_output +533,2071307,"TERMINAL",0,0,"n",,terminal_output +534,2071418,"TERMINAL",0,0,"e",,terminal_output +535,2071466,"TERMINAL",0,0,"r",,terminal_output +536,2071598,"TERMINAL",0,0,"\r\n[?2004l\rsending incremental file list\r\nslurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh\r\n\r\nsent 34,809 bytes received 226 bytes 70,070.00 bytes/sec\r\ntotal size is 128,633,537 speedup is 3,671.57\r\n]0;tum_cte0515@hkn0403:~/Projects/jasmine_jobs[?2004h(jasmine) [tum_cte0515@hkn0403 jasmine_jobs]$ ",,terminal_output +537,2071994,"TERMINAL",0,0,"sync-runner",,terminal_output +538,2072111,"TERMINAL",0,0,"h slurm/jobs/mihir/horeka/breakout/default_runs/train_dyn_single_gpu.sh ",,terminal_output +539,2072479,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=48:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/breakout/dyn/%x_%j.log\r\n#SBATCH --job-name=train_dyn_default_breakout_longer\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\n# source .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/train\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_breakout/breakout_episodes_perfect/val\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nlam_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/lam/interactive/3512576\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/breakout/tokenizer/interactive/3512502\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=10 \\r\n --image_width=10 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --max_lr=3e-5 \\r\n --log_image_interval=500 \\r\n --log_checkpoint_interval=500 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=breakout-dyn-default-$slurm_job_id \\r\n --tags dyn breakout default \\r\n --entity instant-uv \\r\n --project jafar \\r\n --patch_size 4 \\r\n --lam_patch_size 4 \\r\n --warmup_steps 100 \\r\n --wsd_decay_steps 1000 \\r\n --num_steps 5000 \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --lam_checkpoint $lam_checkpoint \\r\n --val_interval 750 \\r\n --eval_full_frame \\r\n",,terminal_output +540,2072627,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=4095309\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0403\r\nSLURM_JOB_START_TIME=1758621669\r\nSLURM_STEP_NODELIST=hkn0403\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1758625269\r\nSLURM_PMI2_SRUN_PORT=33711\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3512647\r\nSLURM_PTY_PORT=37671\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=45\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0403\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=246\r\nSLURM_NODELIST=hkn0403\r\nSLURM_SRUN_COMM_PORT=35797\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3512647\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0403\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=35797\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0403\r\n",,terminal_output +541,2072754,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +542,2073305,"TERMINAL",0,0,"git",,terminal_focus +543,2074861,"TERMINAL",0,0,"\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +544,2075477,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +545,2076808,"TERMINAL",0,0,"git stash",,terminal_command +546,2076846,"TERMINAL",0,0,"]633;C",,terminal_output +547,2077298,"TERMINAL",0,0,"Saved working directory and index state WIP on generate-minatar-breakout-dataset: 0d69ea3 Merge branch 'gt-actions' into generate-minatar-breakout-dataset\r\n",,terminal_output +548,2077411,"TERMINAL",0,0,"]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +549,2081737,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +550,2082112,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +551,2083025,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.21.3\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs/wandb/run-20250923_123546-bp9lkamg\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run breakout-dyn-default-3512647\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/bp9lkamg\r\n",,terminal_output +552,2083192,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 16900976, 'tokenizer': 33750256, 'total': 77207136}\r\n",,terminal_output +553,2085332,"TERMINAL",0,0,"git checkout gt-actions",,terminal_command +554,2085391,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +555,2085422,"",0,0,"Switched from branch 'generate-minatar-breakout-dataset' to 'gt-actions'",,git_branch_checkout +556,2093082,"genie.py",0,0,"from typing import Dict\n\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\nimport orbax.checkpoint as ocp\n\nfrom models.dynamics import DynamicsMaskGIT, DynamicsCausal\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\n\nclass Genie(nnx.Module):\n """"""Genie model""""""\n\n def __init__(\n self,\n in_dim: int,\n tokenizer_dim: int,\n tokenizer_ffn_dim: int,\n latent_patch_dim: int,\n num_patch_latents: int,\n patch_size: int,\n tokenizer_num_blocks: int,\n tokenizer_num_heads: int,\n lam_dim: int,\n lam_ffn_dim: int,\n latent_action_dim: int,\n num_actions: int,\n lam_patch_size: int,\n lam_num_blocks: int,\n lam_num_heads: int,\n lam_co_train: bool,\n use_gt_actions: bool,\n dyna_type: str,\n dyna_dim: int,\n dyna_ffn_dim: int,\n dyna_num_blocks: int,\n dyna_num_heads: int,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n dropout: float = 0.0,\n mask_limit: float = 0.0,\n ):\n # --- Tokenizer ---\n self.in_dim = in_dim\n self.tokenizer_dim = tokenizer_dim\n self.tokenizer_ffn_dim = tokenizer_ffn_dim\n self.latent_patch_dim = latent_patch_dim\n self.num_patch_latents = num_patch_latents\n self.patch_size = patch_size\n self.tokenizer_num_blocks = tokenizer_num_blocks\n self.tokenizer_num_heads = tokenizer_num_heads\n # --- LAM ---\n self.lam_dim = lam_dim\n self.lam_ffn_dim = lam_ffn_dim\n self.latent_action_dim = latent_action_dim\n self.num_actions = num_actions\n self.lam_patch_size = lam_patch_size\n self.lam_num_blocks = lam_num_blocks\n self.lam_num_heads = lam_num_heads\n self.lam_co_train = lam_co_train\n self.use_gt_actions = use_gt_actions\n # --- Dynamics ---\n self.dyna_type = dyna_type\n self.dyna_dim = dyna_dim\n self.dyna_ffn_dim = dyna_ffn_dim\n self.dyna_num_blocks = dyna_num_blocks\n self.dyna_num_heads = dyna_num_heads\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.dropout = dropout\n self.mask_limit = mask_limit\n self.decode = decode\n\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n ffn_dim=self.tokenizer_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n if self.use_gt_actions:\n self.action_embed = nnx.Embed(\n self.num_actions, self.latent_action_dim, rngs=rngs\n )\n self.lam = None\n else:\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n ffn_dim=self.lam_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.action_embed = None\n if self.dyna_type == ""maskgit"":\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n elif self.dyna_type == ""causal"":\n self.dynamics = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n rngs=rngs,\n )\n else:\n raise ValueError(f""Invalid dynamics type: {self.dyna_type}"")\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n training: bool = True,\n ) -> Dict[str, jax.Array]:\n videos_BTHWC = batch[""videos""]\n tokenizer_outputs = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_indices_BTN = tokenizer_outputs[""indices""]\n latent_actions_BTm11L = None\n action_embeddings_BTm11L = None\n if self.use_gt_actions:\n assert self.action_embed is not None\n action_indices_E = None\n action_embeddings_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n action_embeddings_BTm11L = action_embeddings_BT1L[:, :-1]\n else:\n assert self.lam is not None\n lam_outputs = self.lam.vq_encode(videos_BTHWC, training=False)\n z_q_BTm11L = lam_outputs[""z_q""]\n action_indices_E = lam_outputs[""indices""]\n latent_actions_BTm11L = jax.lax.cond(\n self.lam_co_train,\n lambda: z_q_BTm11L,\n lambda: jax.lax.stop_gradient(z_q_BTm11L),\n )\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(token_indices_BTN),\n latent_actions=(\n action_embeddings_BTm11L\n if self.use_gt_actions\n else latent_actions_BTm11L\n ),\n )\n outputs[""mask_rng""] = batch[""rng""]\n dyna_logits_BTNV, dyna_mask = self.dynamics(outputs, training)\n outputs[""token_logits""] = dyna_logits_BTNV\n outputs[""mask""] = dyna_mask\n mle_indices_BTN = jnp.argmax(outputs[""token_logits""], axis=-1)\n H, W = batch[""videos""].shape[2:4]\n outputs[""recon""] = self.tokenizer.decode(mle_indices_BTN, (H, W))\n if action_indices_E is not None:\n outputs[""lam_indices""] = action_indices_E\n return outputs\n\n def sample(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n maskgit_steps: int = 25,\n ) -> tuple[jax.Array, jax.Array]:\n if self.dyna_type == ""maskgit"":\n return self.sample_maskgit(\n batch, seq_len, maskgit_steps, temperature, sample_argmax\n )\n elif self.dyna_type == ""causal"":\n return self.sample_causal(batch, seq_len, temperature, sample_argmax)\n else:\n raise ValueError(f""Dynamics model type unknown: {self.dyna_type}"")\n\n def sample_maskgit(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n steps: int = 25,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n P: S * N\n """"""\n assert isinstance(self.dynamics, DynamicsMaskGIT)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n init_logits_BSNV = jnp.zeros(\n shape=(*token_idxs_BSN.shape, self.num_patch_latents)\n )\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, 1:]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n # --- Extract submodule state ---\n dynamics_state = nnx.state(self.dynamics)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def maskgit_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, mask_BSN, action_tokens_EL = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_maskgit = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_maskgit, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_maskgit.patch_embed(token_idxs_BSN)\n mask_token_111M = dynamics_maskgit.mask_token.value\n mask_expanded_BSN1 = mask_BSN[..., None]\n vid_embed_BSNM = jnp.where(\n mask_expanded_BSN1, mask_token_111M, vid_embed_BSNM\n )\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_maskgit.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n vid_embed_BSNM += act_embed_BS1M\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (steps * 2))\n step_temp = temperature * (1.0 - unmasked_ratio)\n final_logits_BSNV = dynamics_maskgit.transformer(vid_embed_BSNM) / step_temp\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_BSN = jnp.argmax(final_logits_BSNV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_BSN = jax.random.categorical(_rng, final_logits_BSNV)\n gather_fn = jax.vmap(jax.vmap(jax.vmap(lambda x, y: x[y])))\n final_token_probs_BSN = gather_fn(\n jax.nn.softmax(final_logits_BSNV), sampled_token_idxs_BSN\n )\n final_token_probs_BSN += ~mask_BSN\n # Update masked tokens and logits only\n token_idxs_BSN = jnp.where(mask_BSN, sampled_token_idxs_BSN, token_idxs_BSN)\n logits_BSNV = jnp.where(\n jnp.expand_dims(mask_BSN, -1), final_logits_BSNV, logits_BSNV\n )\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n final_token_probs_flat_BP = einops.rearrange(\n final_token_probs_BSN, ""b s n -> b (s n)""\n )\n idx_mask_P = (\n jnp.arange(final_token_probs_flat_BP.shape[-1])\n <= N - num_unmasked_tokens\n )\n sorted_idxs_BP = jnp.argsort(final_token_probs_flat_BP, axis=-1)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask_P))\n mask_flat_BP = einops.rearrange(mask_BSN, ""b s n -> b (s n)"")\n new_mask_flat_BP = mask_update_fn(mask_flat_BP, sorted_idxs_BP)\n new_mask_BSN = einops.rearrange(new_mask_flat_BP, ""b (s n) -> b s n"", n=N)\n\n new_carry = (\n rng,\n token_idxs_BSN,\n logits_BSNV,\n new_mask_BSN,\n action_tokens_EL,\n )\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # Mask current frame (i.e., t == step_t)\n mask_S = jnp.arange(seq_len) == step_t\n mask_BSN = jnp.broadcast_to(mask_S[None, :, None], (B, seq_len, N)).astype(\n bool\n )\n masked_token_idxs_BSN = current_token_idxs_BSN * ~mask_BSN\n masked_logits_BSNV = current_logits_BSNV * jnp.expand_dims(~mask_BSN, -1)\n\n # --- Initialize and run MaskGIT loop ---\n init_carry_maskgit = (\n step_rng,\n masked_token_idxs_BSN,\n masked_logits_BSNV,\n mask_BSN,\n action_tokens_EL,\n )\n final_carry_maskgit = maskgit_step_fn(init_carry_maskgit, jnp.arange(steps))\n updated_token_idxs_BSN = final_carry_maskgit[1]\n updated_logits_BSNV = final_carry_maskgit[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, init_logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def sample_causal(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n assert isinstance(self.dynamics, DynamicsCausal)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n logits_BSNV = jnp.zeros((*token_idxs_BSN.shape, self.num_patch_latents))\n dynamics_state = nnx.state(self.dynamics)\n\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, 1:]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def causal_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step_n: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_causal = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=self.decode,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_causal, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_causal.patch_embed(token_idxs_BSN)\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_causal.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n vid_embed_BSNp1M = jnp.concatenate([act_embed_BS1M, vid_embed_BSNM], axis=2)\n final_logits_BTNp1V = (\n dynamics_causal.transformer(vid_embed_BSNp1M, (step_t, step_n))\n / temperature\n )\n final_logits_BV = final_logits_BTNp1V[:, step_t, step_n, :]\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_B = jnp.argmax(final_logits_BV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_B = jax.random.categorical(_rng, final_logits_BV)\n # Update next tokens only\n token_idxs_BSN = token_idxs_BSN.at[:, step_t, step_n].set(\n sampled_token_idxs_B\n )\n logits_BSNV = logits_BSNV.at[:, step_t, step_n].set(final_logits_BV)\n\n new_carry = (rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t)\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # --- Initialize and run causal loop ---\n init_carry_causal = (\n step_rng,\n current_token_idxs_BSN,\n current_logits_BSNV,\n action_tokens_EL,\n step_t,\n )\n final_carry_causal = causal_step_fn(init_carry_causal, jnp.arange(N))\n updated_token_idxs_BSN = final_carry_causal[1]\n updated_logits_BSNV = final_carry_causal[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def vq_encode(self, batch: Dict[str, jax.Array], training: bool) -> jax.Array:\n # --- Preprocess videos ---\n assert self.lam is not None\n video_BTHWC = batch[""videos""]\n lam_output: Dict[str, jax.Array] = self.lam.vq_encode(\n video_BTHWC, training=training\n )\n lam_indices_E = lam_output[""indices""]\n return lam_indices_E\n\n\n# FIXME (f.srambical): add conversion script for old checkpoints\ndef restore_genie_components(\n optimizer: nnx.Optimizer,\n sharding: jax.sharding.NamedSharding,\n rng: jax.Array,\n args,\n) -> nnx.Optimizer:\n """"""Restore pre-trained Genie components""""""\n rng_tokenizer, rng_lam = jax.random.split(rng)\n rngs_tokenizer = nnx.Rngs(rng_tokenizer)\n rngs_lam = nnx.Rngs(rng_lam)\n\n tx = optimizer.tx\n model = optimizer.model\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n tokenizer_checkpoint_manager = ocp.CheckpointManager(\n directory=args.tokenizer_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n ffn_dim=args.tokenizer_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_tokenizer,\n )\n dummy_tokenizer_optimizer = nnx.Optimizer(dummy_tokenizer, tx)\n dummy_tokenizer_optimizer_state = nnx.state(dummy_tokenizer_optimizer)\n abstract_sharded_tokenizer_optimizer_state = _create_abstract_sharded_pytree(\n dummy_tokenizer_optimizer_state, sharding\n )\n restored_tokenizer = tokenizer_checkpoint_manager.restore(\n step=tokenizer_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_tokenizer_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_tokenizer_optimizer.model, restored_tokenizer.model)\n model.tokenizer = dummy_tokenizer_optimizer.model\n tokenizer_checkpoint_manager.close()\n\n if args.lam_checkpoint:\n lam_checkpoint_manager = ocp.CheckpointManager(\n directory=args.lam_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n ffn_dim=args.lam_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_lam,\n )\n dummy_lam_optimizer = nnx.Optimizer(dummy_lam, tx)\n dummy_lam_optimizer_state = nnx.state(dummy_lam_optimizer)\n abstract_sharded_lam_optimizer_state = _create_abstract_sharded_pytree(\n dummy_lam_optimizer_state, sharding\n )\n restored_lam_optimizer = lam_checkpoint_manager.restore(\n step=lam_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_lam_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_lam_optimizer.model, restored_lam_optimizer.model)\n model.lam = dummy_lam_optimizer.model\n # Remove the LAM decoder to save memory and avoid unnecessary computation.\n del model.lam.decoder\n lam_checkpoint_manager.close()\n\n # Reinitialize the optimizer states\n optimizer = nnx.Optimizer(model, tx)\n return optimizer\n\n\ndef _create_abstract_sharded_pytree(\n pytree_template: nnx.GraphState, sharding_spec: jax.sharding.NamedSharding\n) -> jax.Array:\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)\n",python,tab +557,2093084,"genie.py",9989,2,"1:",python,selection_command +558,2096142,"genie.py",9990,0,"",python,selection_mouse +559,2097088,"genie.py",9991,0,"",python,selection_command +560,2097488,"genie.py",9990,1,"",python,content +561,2097599,"genie.py",9989,1,"",python,content +562,2098248,"genie.py",9989,0,":",python,content +563,2098250,"genie.py",9990,0,"",python,selection_keyboard +564,2098430,"genie.py",9990,0,"-",python,content +565,2098431,"genie.py",9991,0,"",python,selection_keyboard +566,2098489,"genie.py",9991,0,"1",python,content +567,2098490,"genie.py",9992,0,"",python,selection_keyboard +568,2100699,"TERMINAL",0,0,"2025-09-23 12:36:04.685011: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:36:04.685469: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:36:04.685499: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:36:04.685918: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:36:04.687689: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:36:04.689259: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +569,2101508,"genie.py",10754,2,"1:",python,selection_command +570,2110187,"genie.py",18522,2,"1:",python,selection_command +571,2112017,"genie.py",19192,2,"1:",python,selection_command +572,2114706,"sample.py",0,0,"from dataclasses import dataclass\nimport time\nimport os\nimport optax\n\nimport dm_pix as pix\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nimport numpy as np\nimport orbax.checkpoint as ocp\nfrom PIL import Image, ImageDraw\nimport tyro\nfrom flax import nnx\n\nfrom genie import Genie\nfrom utils.dataloader import get_dataloader\n\n\n@dataclass\nclass Args:\n # Experiment\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data/coinrun_episodes""\n checkpoint: str = """"\n # Sampling\n batch_size: int = 1\n maskgit_steps: int = 25\n temperature: float = 1.0\n sample_argmax: bool = True\n start_frame: int = 1\n # Tokenizer checkpoint\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n # LAM checkpoint\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n use_gt_actions: bool = False\n # Dynamics checkpoint\n dyna_type: str = ""maskgit""\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\nif __name__ == ""__main__"":\n """"""\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n jax.distributed.initialize()\n\n rng = jax.random.key(args.seed)\n\n # --- Load Genie checkpoint ---\n rngs = nnx.Rngs(rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_actions=args.num_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=False,\n use_gt_actions=args.use_gt_actions,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n # FIXME (f.srambical): implement spatiotemporal KV caching and set decode=True\n decode=False,\n rngs=rngs,\n )\n\n del genie.tokenizer.vq.drop\n # Need to delete lam decoder for checkpoint loading\n if not args.use_gt_actions:\n assert genie.lam is not None\n del genie.lam.decoder\n\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n dummy_tx = optax.adamw(\n learning_rate=optax.linear_schedule(0.0001, 0.0001, 10000),\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n dummy_optimizer = nnx.Optimizer(genie, dummy_tx)\n\n abstract_optimizer = nnx.eval_shape(lambda: dummy_optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(dummy_optimizer, restored_optimizer_state)\n\n # --- Define sampling function ---\n def _sampling_fn(model: Genie, batch: dict) -> jax.Array:\n """"""Runs Genie.sample with pre-defined generation hyper-parameters.""""""\n assert args.dyna_type in [\n ""maskgit"",\n ""causal"",\n ], f""Invalid dynamics type: {args.dyna_type}""\n frames, _ = model.sample(\n batch,\n args.seq_len,\n args.temperature,\n args.sample_argmax,\n args.maskgit_steps,\n )\n return frames\n\n # --- Define autoregressive sampling loop ---\n def _autoreg_sample(genie, rng, batch):\n batch[""videos""] = batch[""videos""][:, : args.start_frame]\n batch[""rng""] = rng\n generated_vid_BSHWC = _sampling_fn(genie, batch)\n return generated_vid_BSHWC\n\n # --- Get video + latent actions ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n args.batch_size,\n args.image_height,\n args.image_width,\n args.image_channels,\n # We don't use workers in order to avoid grain shutdown issues (https://github.com/google/grain/issues/398)\n num_workers=0,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n dataloader = iter(dataloader)\n batch = next(dataloader)\n gt_video = jnp.asarray(batch[""videos""], dtype=jnp.float32) / 255.0\n batch[""videos""] = gt_video.astype(args.dtype)\n # Get latent actions for all videos in the batch\n action_batch_E = None\n if not args.use_gt_actions:\n action_batch_E = genie.vq_encode(batch, training=False)\n batch[""latent_actions""] = action_batch_E\n\n # --- Sample + evaluate video ---\n recon_video_BSHWC = _autoreg_sample(genie, rng, batch)\n recon_video_BSHWC = recon_video_BSHWC.astype(jnp.float32)\n gt = (\n gt_video[:, : recon_video_BSHWC.shape[1]]\n .clip(0, 1)\n .reshape(-1, *gt_video.shape[2:])\n )\n recon = recon_video_BSHWC.clip(0, 1).reshape(-1, *recon_video_BSHWC.shape[2:])\n ssim = jnp.asarray(\n pix.ssim(gt[:, args.start_frame :], recon[:, args.start_frame :])\n ).mean()\n print(f""SSIM: {ssim}"")\n\n # --- Construct video ---\n true_videos = (gt_video * 255).astype(np.uint8)\n pred_videos = (recon_video_BSHWC * 255).astype(np.uint8)\n video_comparison = np.zeros((2, *recon_video_BSHWC.shape), dtype=np.uint8)\n video_comparison[0] = true_videos[:, : args.seq_len]\n video_comparison[1] = pred_videos\n frames = einops.rearrange(video_comparison, ""n b t h w c -> t (b h) (n w) c"")\n\n # --- Save video ---\n imgs = [Image.fromarray(img) for img in frames]\n # Write actions on each frame, on each row (i.e., for each video in the batch, on the GT row)\n B, S, _, _, _ = batch[""videos""].shape\n if action_batch_E is not None:\n action_batch_BSm11 = jnp.reshape(action_batch_E, (B, S - 1, 1))\n else:\n action_batch_BSm11 = jnp.expand_dims(batch[""actions""][:, 1:], -1)\n for t, img in enumerate(imgs[1:]):\n d = ImageDraw.Draw(img)\n for row in range(B):\n action = action_batch_BSm11[row, t, 0]\n y_offset = row * batch[""videos""].shape[2] + 2\n d.text((2, y_offset), f""{action}"", fill=255)\n imgs[0].save(\n f""generation_{time.time()}.gif"",\n save_all=True,\n append_images=imgs[1:],\n duration=250,\n loop=0,\n )\n",python,tab +573,2114708,"sample.py",7706,2,"1:",python,selection_command +574,2129968,"sample.py",7708,0,"",python,selection_mouse +575,2131220,"sample.py",7707,1,"",python,content +576,2131343,"sample.py",7706,1,"",python,content +577,2131735,"sample.py",7706,0,":",python,content +578,2131736,"sample.py",7707,0,"",python,selection_keyboard +579,2132661,"sample.py",7707,0,"-",python,content +580,2132662,"sample.py",7708,0,"",python,selection_keyboard +581,2132739,"sample.py",7708,0,"1",python,content +582,2132740,"sample.py",7709,0,"",python,selection_keyboard +583,2146498,"TERMINAL",0,0,"Total memory size: 2.5 GB, Output size: 0.9 GB, Temp size: 1.7 GB, Argument size: 0.9 GB, Host temp size: 0.0 GB.\r\n",,terminal_output +584,2146582,"TERMINAL",0,0,"FLOPs: 1.214e+11, Bytes: 4.622e+10 (43.0 GB), Intensity: 2.6 FLOPs/byte\r\nStarting training from step 0...\r\n",,terminal_output +585,2147095,"TERMINAL",0,0,"\r\nMemstats: After params initialized:\r\n\tUsing (GB) 1.1 / 38.7 (2.842377%) on cuda:0\r\n",,terminal_output +586,2148853,"genie.py",0,0,"",python,tab +587,2148854,"genie.py",18522,2,"1:",python,selection_command +588,2150990,"genie.py",18524,0,"",python,selection_mouse +589,2151663,"genie.py",18523,1,"",python,content +590,2151837,"genie.py",18522,1,"",python,content +591,2152165,"genie.py",18522,0,":",python,content +592,2152166,"genie.py",18523,0,"",python,selection_keyboard +593,2152498,"genie.py",18523,0,".",python,content +594,2152500,"genie.py",18524,0,"",python,selection_keyboard +595,2152569,"genie.py",18524,0,"1",python,content +596,2152570,"genie.py",18525,0,"",python,selection_keyboard +597,2153211,"genie.py",18524,1,"",python,content +598,2153329,"genie.py",18523,1,"",python,content +599,2153595,"genie.py",18523,0,"-",python,content +600,2153596,"genie.py",18524,0,"",python,selection_keyboard +601,2153696,"genie.py",18524,0,"1",python,content +602,2153696,"genie.py",18525,0,"",python,selection_keyboard +603,2157359,"sample.py",0,0,"",python,tab +604,2157360,"sample.py",7749,2,"1:",python,selection_command +605,2163629,"sample.py",8091,2,"1:",python,selection_command +606,2165980,"train_lam.py",0,0,"import os\n\nos.environ.setdefault(""XLA_PYTHON_CLIENT_MEM_FRACTION"", ""0.98"")\n\nfrom dataclasses import dataclass, field\nimport itertools\nfrom typing import cast, Optional\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom models.lam import LatentActionModel\nfrom utils.dataloader import get_dataloader\nfrom utils.train_utils import (\n get_lr_schedule,\n count_parameters_by_component,\n print_mem_stats,\n print_compiled_memory_stats,\n print_compiled_cost_analysis,\n)\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n vq_beta: float = 0.25\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n vq_reset_thresh: int = 50\n # LAM\n model_dim: int = 512\n ffn_dim: int = 2048\n latent_dim: int = 32\n num_latents: int = 6\n patch_size: int = 16\n num_blocks: int = 4\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.0\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_lam""\n tags: list[str] = field(default_factory=lambda: [""lam""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\ndef build_model(args: Args, rng: jax.Array) -> tuple[LatentActionModel, jax.Array]:\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n return (\n LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n ffn_dim=args.ffn_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n ),\n rng,\n )\n\n\ndef build_optimizer(\n model: LatentActionModel, args: Args\n) -> tuple[nnx.Optimizer, optax.Schedule]:\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.param_dtype, # moments in full precision\n )\n optimizer = nnx.Optimizer(model, tx)\n return optimizer, lr_schedule\n\n\ndef build_mesh_and_sharding(\n num_devices: int,\n) -> tuple[Mesh, NamedSharding, NamedSharding]:\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n return mesh, replicated_sharding, videos_sharding\n\n\ndef shard_optimizer_states(\n optimizer: nnx.Optimizer, replicated_sharding: NamedSharding\n) -> None:\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n\ndef build_dataloader(args: Args, data_dir: str) -> grain.DataLoaderIterator:\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(data_dir, x)\n for x in os.listdir(data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n return grain_iterator\n\n\ndef build_checkpoint_manager(args: Args) -> ocp.CheckpointManager:\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n return checkpoint_manager\n\n\ndef restore_checkpoint_if_needed(\n args: Args,\n checkpoint_manager: ocp.CheckpointManager,\n optimizer: nnx.Optimizer,\n train_iterator: grain.DataLoaderIterator,\n val_iterator: Optional[grain.DataLoaderIterator],\n restore_step: Optional[int] = None,\n) -> tuple[int, nnx.Optimizer, grain.DataLoaderIterator, grain.DataLoaderIterator]:\n step = 0\n if restore_step is None:\n restore_step = checkpoint_manager.latest_step()\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n if val_iterator:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n val_dataloader_state=grain.checkpoint.CheckpointRestore(val_iterator), # type: ignore\n )\n else:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n )\n restored = checkpoint_manager.restore(restore_step, args=restore_args)\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n train_iterator = restored[""train_dataloader_state""]\n if val_iterator:\n val_iterator = restored[""val_dataloader_state""]\n step = restore_step or 0\n print(f""Restored dataloader and model state from step {step}"")\n return step, optimizer, train_iterator, val_iterator\n\n\ndef enable_sowing(lam: LatentActionModel) -> None:\n for model in [lam.encoder, lam.decoder]:\n setattr(model, ""sow_logits"", True)\n for blk in getattr(model, ""blocks"", []):\n setattr(blk, ""sow_weights"", True)\n setattr(blk, ""sow_activations"", True)\n\n\ndef main(args: Args) -> None:\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n lam, rng = build_model(args, rng)\n\n # Count parameters\n _, params, _ = nnx.split(lam, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n optimizer, lr_schedule = build_optimizer(lam, args)\n del lam\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n _, replicated_sharding, videos_sharding = build_mesh_and_sharding(num_devices)\n\n shard_optimizer_states(optimizer, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n checkpoint_manager = build_checkpoint_manager(args)\n\n # --- Create DataLoaderIterator from dataloader ---\n train_iterator = build_dataloader(args, args.data_dir)\n val_iterator = None\n if args.val_data_dir:\n val_iterator = build_dataloader(args, args.val_data_dir)\n\n # --- Restore checkpoint ---\n step, optimizer, train_iterator, val_iterator = restore_checkpoint_if_needed(\n args, checkpoint_manager, optimizer, train_iterator, val_iterator\n )\n\n # --- Define loss and train step (close over args) ---\n def lam_loss_fn(\n model: LatentActionModel, inputs: dict, training: bool = True\n ) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n gt_val = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt_val, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt_val, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=(index_counts != 0).mean(),\n )\n return loss, (outputs[""recon""], index_counts, metrics)\n\n @nnx.jit(donate_argnums=0)\n def train_step(\n optimizer: nnx.Optimizer,\n inputs: dict,\n action_last_active: jax.Array,\n rng: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, dict]:\n def loss_fn(\n model: LatentActionModel,\n ) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n model.train()\n return lam_loss_fn(model, inputs, training=True)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(optimizer.model)\n optimizer.update(grads)\n\n # --- Reset inactive latent actions ---\n codebook = optimizer.model.vq.codebook\n num_codes = len(codebook)\n active_codes = idx_counts != 0.0\n action_last_active = jnp.where(active_codes, 0, action_last_active + 1)\n p_code = active_codes / active_codes.sum()\n reset_idxs = jax.random.choice(rng, num_codes, shape=(num_codes,), p=p_code)\n do_reset = action_last_active >= args.vq_reset_thresh\n new_codebook = jnp.where(\n jnp.expand_dims(do_reset, -1), codebook[reset_idxs], codebook.value\n )\n optimizer.model.vq.codebook.value = new_codebook\n action_last_active = jnp.where(do_reset, 0, action_last_active)\n return loss, recon, action_last_active, metrics\n\n @nnx.jit\n def val_step(\n lam: LatentActionModel, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\n def calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n batch = None\n recon = None\n for batch in val_dataloader:\n loss, recon, metrics = val_step(lam, batch)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(\n f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}""\n )\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([float(m[key]) for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n val_metrics[""val_loss""] = val_loss\n return val_metrics, batch, recon\n\n # --- TRAIN LOOP ---\n dataloader_train = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n }\n for elem in train_iterator\n )\n dataloader_val = None\n if val_iterator:\n dataloader_val = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n }\n for elem in val_iterator\n )\n action_last_active = jnp.zeros(args.num_latents, dtype=jnp.int32)\n if jax.process_index() == 0:\n first_batch = next(dataloader_train)\n compiled = train_step.lower(\n optimizer, first_batch, action_last_active, rng\n ).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader_train = itertools.chain([first_batch], dataloader_train)\n print(f""Starting training from step {step}..."")\n first_step = step\n while step < args.num_steps:\n for batch in dataloader_train:\n # --- Train step ---\n rng, _rng = jax.random.split(rng, 2)\n loss, recon, action_last_active, metrics = train_step(\n optimizer, batch, action_last_active, _rng\n )\n if step == first_step:\n print_mem_stats(""After params initialized"")\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Validation loss ---\n val_results = {}\n if dataloader_val and step % args.val_interval == 0:\n print(""Calculating validation metrics..."")\n val_metrics, val_gt_batch, val_recon = calculate_validation_metrics(\n dataloader_val, optimizer.model\n )\n print(f""Step {step}, validation loss: {val_metrics['val_loss']}"")\n val_results = {\n ""metrics"": val_metrics,\n ""gt_batch"": val_gt_batch,\n ""recon"": val_recon,\n }\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n log_dict = {""loss"": loss, ""step"": step, **metrics}\n if val_results:\n log_dict.update(val_results[""metrics""])\n wandb.log(log_dict)\n if step % args.log_image_interval == 0:\n gt_seq = batch[""videos""][0, 1:].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if val_results:\n val_results[""gt_seq_val""] = (\n val_results[""gt_batch""][""videos""][0, 1:].astype(jnp.float32)\n / 255.0\n )\n val_results[""recon_seq_val""] = val_results[""recon""][0].clip(\n 0, 1\n )\n val_results[""val_comparison_seq""] = jnp.concatenate(\n (val_results[""gt_seq_val""], val_results[""recon_seq_val""]),\n axis=1,\n )\n val_results[""val_comparison_seq""] = einops.rearrange(\n val_results[""val_comparison_seq""] * 255,\n ""t h w c -> h (t w) c"",\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n if val_results:\n log_images.update(\n {\n ""val_image"": wandb.Image(\n np.asarray(val_results[""gt_seq_val""][0])\n ),\n ""val_recon"": wandb.Image(\n np.asarray(val_results[""recon_seq_val""][0])\n ),\n ""val_true_vs_recon"": wandb.Image(\n np.asarray(\n val_results[""val_comparison_seq""].astype(\n np.uint8\n )\n )\n ),\n }\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n if val_iterator:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n val_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n val_iterator # type: ignore\n ),\n )\n else:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n )\n checkpoint_manager.save(step, args=ckpt_manager_args)\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n main(args)\n",python,tab +607,2165982,"train_lam.py",11648,2,"1:",python,selection_command +608,2168723,"train_lam.py",17992,2,"1:",python,selection_command +609,2171378,"train_lam.py",18459,2,"1:",python,selection_command +610,2177990,"models/dynamics.py",0,0,"from typing import Dict\n\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\n\nfrom utils.nn import STTransformer, Transformer\n\n\nclass DynamicsMaskGIT(nnx.Module):\n """"""\n MaskGIT dynamics model\n\n Dimension keys:\n B: batch size\n T: sequence length\n N: number of patches per frame\n L: latent dimension\n V: vocabulary size (number of latents)\n """"""\n\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_latents: int,\n latent_action_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n mask_limit: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_latents = num_latents\n self.latent_action_dim = latent_action_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.mask_limit = mask_limit\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.transformer = STTransformer(\n self.model_dim,\n self.model_dim,\n self.ffn_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.patch_embed = nnx.Embed(self.num_latents, self.model_dim, rngs=rngs)\n self.mask_token = nnx.Param(\n nnx.initializers.lecun_uniform()(rngs.params(), (1, 1, 1, self.model_dim))\n )\n self.action_up = nnx.Linear(\n self.latent_action_dim,\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n training: bool = True,\n ) -> tuple[jax.Array, jax.Array]:\n # --- Mask videos ---\n video_tokens_BTN = batch[""video_tokens""]\n latent_actions_BTm11L = batch[""latent_actions""]\n vid_embed_BTNM = self.patch_embed(video_tokens_BTN)\n if training:\n batch_size = vid_embed_BTNM.shape[0]\n _rng_prob, *_rngs_mask = jax.random.split(batch[""mask_rng""], batch_size + 1)\n mask_prob = jax.random.uniform(\n _rng_prob, shape=(batch_size,), minval=self.mask_limit\n )\n per_sample_shape = vid_embed_BTNM.shape[1:-1]\n mask = jax.vmap(\n lambda rng, prob: jax.random.bernoulli(rng, prob, per_sample_shape),\n in_axes=(0, 0),\n )(jnp.asarray(_rngs_mask), mask_prob)\n mask = mask.at[:, 0].set(False)\n vid_embed_BTNM = jnp.where(\n jnp.expand_dims(mask, -1), self.mask_token.value, vid_embed_BTNM\n )\n else:\n mask = jnp.ones_like(video_tokens_BTN)\n\n # --- Predict transition ---\n act_embed_BTm11M = self.action_up(latent_actions_BTm11L)\n padded_act_embed_BT1M = jnp.pad(\n act_embed_BTm11M, ((0, 0), (1, 0), (0, 0), (0, 0))\n )\n padded_act_embed_BTNM = jnp.broadcast_to(\n padded_act_embed_BT1M, vid_embed_BTNM.shape\n )\n vid_embed_BTNM += padded_act_embed_BTNM\n logits_BTNV = self.transformer(vid_embed_BTNM)\n return logits_BTNV, mask\n\n\nclass DynamicsCausal(nnx.Module):\n """"""Causal dynamics model""""""\n\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_latents: int,\n latent_action_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_latents = num_latents\n self.latent_action_dim = latent_action_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.decode = decode\n\n self.transformer = Transformer(\n self.model_dim,\n self.model_dim,\n self.ffn_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=self.decode,\n rngs=rngs,\n )\n self.patch_embed = nnx.Embed(self.num_latents, self.model_dim, rngs=rngs)\n self.action_up = nnx.Linear(\n self.latent_action_dim,\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n training: bool = True,\n ) -> tuple[jax.Array, jax.Array]:\n video_tokens_BTN = batch[""video_tokens""]\n latent_actions_BTm11L = batch[""latent_actions""]\n vid_embed_BTNM = self.patch_embed(video_tokens_BTN)\n act_embed_BTm11M = self.action_up(latent_actions_BTm11L)\n padded_act_embed_BT1M = jnp.pad(\n act_embed_BTm11M, ((0, 0), (1, 0), (0, 0), (0, 0))\n )\n vid_embed_BTNp1M = jnp.concatenate(\n [padded_act_embed_BT1M, vid_embed_BTNM], axis=2\n )\n logits_BTNp1V = self.transformer(vid_embed_BTNp1M)\n logits_BTNV = logits_BTNp1V[:, :, :-1]\n return logits_BTNV, jnp.ones_like(video_tokens_BTN)\n",python,tab +611,2177991,"models/dynamics.py",2653,2,"1:",python,selection_command +612,2184000,"models/lam.py",0,0,"from typing import Dict\n\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\n\nfrom utils.preprocess import patchify, unpatchify\nfrom utils.nn import STTransformer, VectorQuantizer\n\n\nclass LatentActionModel(nnx.Module):\n """"""Latent Action ST-ViVit VQ-VAE\n\n Dimension keys:\n B: batch size\n T: sequence length\n N: number of patches per frame\n M: model dimension\n L: latent dimension\n E: B * (T - 1)\n H: height\n W: width\n C: number of channels (n_dim)\n P: patch token dimension (patch_size^2 * C)\n\n Tm1: T - 1\n Np1: N + 1\n """"""\n\n def __init__(\n self,\n in_dim: int,\n model_dim: int,\n ffn_dim: int,\n latent_dim: int,\n num_latents: int,\n patch_size: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n codebook_dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n ):\n self.in_dim = in_dim\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.latent_dim = latent_dim\n self.num_latents = num_latents\n self.patch_size = patch_size\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.codebook_dropout = codebook_dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.patch_token_dim = self.in_dim * self.patch_size**2\n self.encoder = STTransformer(\n self.patch_token_dim,\n self.model_dim,\n self.ffn_dim,\n self.latent_dim,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.action_in = nnx.Param(\n nnx.initializers.lecun_uniform()(\n rngs.params(), (1, 1, 1, self.patch_token_dim)\n )\n )\n self.vq = VectorQuantizer(\n self.latent_dim,\n self.num_latents,\n self.codebook_dropout,\n self.dtype,\n rngs=rngs,\n )\n self.patch_up = nnx.Linear(\n self.patch_token_dim,\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.action_up = nnx.Linear(\n self.latent_dim,\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.decoder = STTransformer(\n self.model_dim,\n self.model_dim,\n self.ffn_dim,\n self.patch_token_dim,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n\n def __call__(\n self, batch: Dict[str, jax.Array], training: bool = True\n ) -> Dict[str, jax.Array]:\n # --- Encode + VQ ---\n H, W = batch[""videos""].shape[2:4]\n videos_BTHWC = batch[""videos""]\n outputs = self.vq_encode(videos_BTHWC, training)\n patch_BTNP = outputs[""patches""]\n z_q_BTm11L = outputs[""z_q""]\n action_BTm11M = self.action_up(z_q_BTm11L)\n patch_BTm1NM = self.patch_up(patch_BTNP[:, :-1])\n action_BTm1NM = jnp.broadcast_to(action_BTm11M, patch_BTm1NM.shape)\n video_action_patches_BTm1NM = action_BTm1NM + patch_BTm1NM\n del outputs[""patches""], patch_BTNP, patch_BTm1NM\n\n # --- Decode ---\n video_recon_BTm1P = self.decoder(video_action_patches_BTm1NM)\n video_recon_BTm1P = video_recon_BTm1P.astype(jnp.float32)\n video_recon_BTm1P = nnx.sigmoid(video_recon_BTm1P)\n video_recon_BTm1P = video_recon_BTm1P.astype(self.dtype)\n video_recon_BTm1HWC = unpatchify(video_recon_BTm1P, self.patch_size, H, W)\n outputs[""recon""] = video_recon_BTm1HWC\n return outputs\n\n def vq_encode(\n self, videos_BTHWC: jax.Array, training: bool = True\n ) -> Dict[str, jax.Array]:\n # --- Preprocess videos ---\n B, T = videos_BTHWC.shape[:2]\n patch_BTNP = patchify(videos_BTHWC, self.patch_size)\n action_pad_BT1P = jnp.broadcast_to(\n self.action_in.value, (B, T, 1, self.patch_token_dim)\n )\n padded_patch_BTNp1P = jnp.concatenate((action_pad_BT1P, patch_BTNP), axis=2)\n\n # --- Encode ---\n z_BTNp1L = self.encoder(padded_patch_BTNp1P)\n # Get latent action for all future frames\n z_BTm1L = z_BTNp1L[:, 1:, 0]\n\n # --- Vector quantize ---\n z_EL = z_BTm1L.reshape(B * (T - 1), self.latent_dim)\n z_q_EL, z_EL, emb_EL, indices_E = self.vq(z_EL, training)\n z_q_BTm11L = z_q_EL.reshape(B, T - 1, 1, self.latent_dim)\n return dict(\n patches=patch_BTNP, z_q=z_q_BTm11L, z=z_EL, emb=emb_EL, indices=indices_E\n )\n",python,tab +613,2184001,"models/lam.py",4812,2,"1:",python,selection_command +614,2192956,"genie.py",0,0,"",python,tab +615,2192957,"genie.py",9930,0,"",python,selection_command +616,2228186,"sample.py",0,0,"",python,tab +617,2228187,"sample.py",7641,0,"",python,selection_command +618,2248251,"TERMINAL",0,0,"git status",,terminal_command +619,2248344,"TERMINAL",0,0,"]633;COn branch gt-actions\r\nYour branch is up to date with 'origin/gt-actions'.\r\n\r\n",,terminal_output +620,2248373,"TERMINAL",0,0,"Last commands done (2 commands done):\r\n pick ba37453 feat: generate coinrun dataset with val split\r\n pick faadd10 feat: implemented validation loss for all three models\r\nNext commands to do (26 remaining commands):\r\n pick 9a17dbb fix: pass val data path to dataloader\r\n pick 6e69cdb fix typo in image logging\r\n (use ""git rebase --edit-todo"" to view and edit)\r\nYou are currently editing a commit while rebasing branch 'gt-actions' on 'c7522f2'.\r\n (use ""git commit --amend"" to amend the current commit)\r\n (use ""git rebase --continue"" once you are satisfied with your changes)\r\n\r\nChanges to be committed:\r\n (use ""git restore --staged ..."" to unstage)\r\n\tmodified: genie.py\r\n\tmodified: sample.py\r\n\r\nUntracked files:\r\n (use ""git add ..."" to include in what will be committed)\r\n\tdiff.diff\r\n\tinput_pipeline/generate_breakout_dataset_agent.py\r\n\tkiller.sh\r\n\tkiller_partition.sh\r\n\tlog.log\r\n\toverfit_dir.zip\r\n\trequirements-franz.txt\r\n\tsamples/\r\n\tscripts_cremers/\r\n\tslurm/\r\n\ttest.py\r\n\tutils/visualizer.py\r\n\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +621,2259714,"TERMINAL",0,0,"git pull",,terminal_command +622,2259765,"TERMINAL",0,0,"]633;C",,terminal_output +623,2261594,"TERMINAL",0,0,"remote: Enumerating objects: 7, done.\r\nremote: Counting objects: 14% (1/7)\rremote: Counting objects: 28% (2/7)\rremote: Counting objects: 42% (3/7)\rremote: Counting objects: 57% (4/7)\rremote: Counting objects: 71% (5/7)\rremote: Counting objects: 85% (6/7)\rremote: Counting objects: 100% (7/7)\rremote: Counting objects: 100% (7/7), done.\r\nremote: Compressing objects: 25% (1/4)\rremote: Compressing objects: 50% (2/4)\rremote: Compressing objects: 75% (3/4)\rremote: Compressing objects: 100% (4/4)\rremote: Compressing objects: 100% (4/4), done.\r\nremote: Total 7 (delta 3), reused 3 (delta 3), pack-reused 0 (from 0)\r\nUnpacking objects: 14% (1/7)\rUnpacking objects: 28% (2/7)\rUnpacking objects: 42% (3/7)\rUnpacking objects: 57% (4/7)\rUnpacking objects: 71% (5/7)\rUnpacking objects: 85% (6/7)\r",,terminal_output +624,2261658,"TERMINAL",0,0,"Unpacking objects: 100% (7/7)\rUnpacking objects: 100% (7/7), 8.34 KiB | 70.00 KiB/s, done.\r\n",,terminal_output +625,2261939,"TERMINAL",0,0,"From github.com:p-doom/jasmine\r\n a5a539c..8fbd8eb gt-actions -> origin/gt-actions\r\n 35e26ae..3b83a0b generate-atari-dataset -> origin/generate-atari-dataset\r\n",,terminal_output +626,2262009,"TERMINAL",0,0,"Saved checkpoint at step 500\r\n",,terminal_output +627,2262020,"TERMINAL",0,0,"Updating a5a539c..8fbd8eb\r\nFast-forward\r\n",,terminal_output +628,2262055,"TERMINAL",0,0," train_dynamics.py | 4 ++--\r\n 1 file changed, 2 insertions(+), 2 deletions(-)\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +629,2273930,"TERMINAL",0,0,"git commit -m ""modify slicing towards new data format""",,terminal_command +630,2273973,"TERMINAL",0,0,"]633;C",,terminal_output +631,2274144,"TERMINAL",0,0,"g",,terminal_output +632,2274210,"TERMINAL",0,0,"i",,terminal_output +633,2274423,"TERMINAL",0,0,"t",,terminal_output +634,2274603,"TERMINAL",0,0," ",,terminal_output +635,2275204,"TERMINAL",0,0,"p",,terminal_output +636,2275314,"TERMINAL",0,0,"u",,terminal_output +637,2275447,"TERMINAL",0,0,"s",,terminal_output +638,2275512,"TERMINAL",0,0,"h",,terminal_output +639,2276617,"TERMINAL",0,0,"black....................................................................",,terminal_output +640,2279658,"TERMINAL",0,0,"Passed\r\n",,terminal_output +641,2279683,"TERMINAL",0,0,"[gt-actions 33b1895] modify slicing towards new data format\r\n 2 files changed, 3 insertions(+), 3 deletions(-)\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +642,2280644,"TERMINAL",0,0,"git push",,terminal_command +643,2280716,"TERMINAL",0,0,"]633;C",,terminal_output +644,2282076,"TERMINAL",0,0,"Enumerating objects: 7, done.\r\nCounting objects: 14% (1/7)\rCounting objects: 28% (2/7)\rCounting objects: 42% (3/7)\rCounting objects: 57% (4/7)\rCounting objects: 71% (5/7)\rCounting objects: 85% (6/7)\rCounting objects: 100% (7/7)\rCounting objects: 100% (7/7), done.\r\nDelta compression using up to 152 threads\r\nCompressing objects: 25% (1/4)\rCompressing objects: 50% (2/4)\rCompressing objects: 75% (3/4)\rCompressing objects: 100% (4/4)\rCompressing objects: 100% (4/4), done.\r\nWriting objects: 25% (1/4)\rWriting objects: 50% (2/4)\rWriting objects: 75% (3/4)\rWriting objects: 100% (4/4)\rWriting objects: 100% (4/4), 407 bytes | 407.00 KiB/s, done.\r\nTotal 4 (delta 3), reused 0 (delta 0), pack-reused 0\r\n",,terminal_output +645,2282162,"TERMINAL",0,0,"remote: Resolving deltas: 0% (0/3)\rremote: Resolving deltas: 33% (1/3)\rremote: Resolving deltas: 66% (2/3)\rremote: Resolving deltas: 100% (3/3)\rremote: Resolving deltas: 100% (3/3), completed with 3 local objects.\r\n",,terminal_output +646,2282403,"TERMINAL",0,0,"To github.com:p-doom/jasmine.git\r\n 8fbd8eb..33b1895 gt-actions -> gt-actions\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +647,2290671,"TERMINAL",0,0,"git branch",,terminal_command +648,2290715,"TERMINAL",0,0,"]633;C[?1h=\r action-mapper\r\n add-wandb-name-and-tags\r\n before-nnx\r\n causal-mem-reduce\r\n causal-spatiotemporal-kv-cache\r\n causal-st-transformer\r\n causal-transformer-dynamics-model\r\n causal-transformer-nnx-no-kv-cache\r\n coinrun-gt-actions\r\n convert-to-jax-array-in-iter\r\n correct-batched-sampling\r\n dev\r\n dont-let-tf-see-gpu\r\n feat/darkness-filter\r\n feat/explicit-image-dims\r\n fix-action-padding-lam-future-information-access\r\n fix-sampling\r\n:",,terminal_output +649,2292586,"TERMINAL",0,0,"\r fix-transformer-forwardpass\r\n:",,terminal_output +650,2293462,"TERMINAL",0,0,"\r fix/spatiotemporal-pe-once-in-STTransformer\r\n:\r generate-minatar-breakout-dataset\r\n:\r grad-norm-log-and-clip\r\n:\r grain-dataloader\r\n:\r* gt-actions\r\n:\r input_pipeline/add-npy2array_record\r\n:\r logging-variants\r\n:\r lr-schedules\r\n:\r main\r\n:\r maskgit-different-maskprob-per-sample\r\n:\r maskgit-sampling-iterative-unmasking-fix\r\n:\r metrics-logging-for-dynamics-model\r\n:\r monkey-patch\r\n:\r new-arch-sampling\r\n:",,terminal_output +651,2293604,"TERMINAL",0,0,"\r preprocess_video\r\n:",,terminal_output +652,2300937,"TERMINAL",0,0,"\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +653,2303431,"TERMINAL",0,0,"git checkout generate-minatar-breakout-dataset",,terminal_command +654,2303441,"TERMINAL",0,0,"]633;CSwitched to branch 'generate-minatar-breakout-dataset'\r\nYour branch is ahead of 'origin/generate-minatar-breakout-dataset' by 10 commits.\r\n (use ""git push"" to publish your local commits)\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +655,2303667,"sample.py",7641,75," action_batch_BSm11 = jnp.expand_dims(batch[""actions""][:, 1:], -1)\n",python,content +656,2303711,"TERMINAL",0,0,"Calculating validation metrics...\r\n",,terminal_output +657,2305430,"sample.py",0,0,"Switched from branch 'gt-actions' to 'generate-minatar-breakout-dataset'",python,git_branch_checkout +658,2310942,"TERMINAL",0,0,"2025-09-23 12:39:34.894923: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:39:34.895522: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-09-23 12:39:34.895542: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +659,2314022,"TERMINAL",0,0,"git stash pop",,terminal_command +660,2314096,"TERMINAL",0,0,"]633;COn branch generate-minatar-breakout-dataset\r\nYour branch is ahead of 'origin/generate-minatar-breakout-dataset' by 10 commits.\r\n (use ""git push"" to publish your local commits)\r\n\r\nLast commands done (2 commands done):\r\n pick ba37453 feat: generate coinrun dataset with val split\r\n pick faadd10 feat: implemented validation loss for all three models\r\nNext commands to do (26 remaining commands):\r\n pick 9a17dbb fix: pass val data path to dataloader\r\n pick 6e69cdb fix typo in image logging\r\n (use ""git rebase --edit-todo"" to view and edit)\r\nYou are currently editing a commit while rebasing branch 'gt-actions' on 'c7522f2'.\r\n (use ""git commit --amend"" to amend the current commit)\r\n (use ""git rebase --continue"" once you are satisfied with your changes)\r\n\r\nChanges not staged for commit:\r\n (use ""git add ..."" to update what will be committed)\r\n (use ""git restore ..."" to discard changes in working directory)\r\n\tmodified: genie.py\r\n\tmodified: input_pipeline/generate_breakout_dataset.py\r\n\tmodified: sample.py\r\n\tmodified: train_dynamics.py\r\n\tmodified: train_lam.py\r\n\tmodified: train_tokenizer.py\r\n\r\nUntracked files:\r\n (use ""git add ..."" to include in what will be committed)\r\n\tdiff.diff\r\n\tinput_pipeline/generate_breakout_dataset_agent.py\r\n\tkiller.sh\r\n\tkiller_partition.sh\r\n\tlog.log\r\n\toverfit_dir.zip\r\n\trequirements-franz.txt\r\n\tsamples/\r\n\tscripts_cremers/\r\n\tslurm/\r\n\ttest.py\r\n\tutils/visualizer.py\r\n\r\nno changes added to commit (use ""git add"" and/or ""git commit -a"")\r\n",,terminal_output +661,2314115,"TERMINAL",0,0,"Dropped refs/stash@{0} (7e8d393bd225e761edba528abd2f41161cf8a0c7)\r\n]0;tum_cte0515@hkn1993:~/Projects/jasmine",,terminal_output +662,2314358,"sample.py",7641,74," action_batch_BSm11 = jnp.expand_dims(batch[""actions""][:, :-1], -1)\n",python,content +663,2318027,"train_dynamics.py",0,0,"import os\n\n\nos.environ.setdefault(""XLA_PYTHON_CLIENT_MEM_FRACTION"", ""0.98"")\n\nfrom dataclasses import dataclass, field\nimport itertools\nfrom typing import cast, Optional\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\nfrom utils.train_utils import (\n get_lr_schedule,\n count_parameters_by_component,\n print_mem_stats,\n print_compiled_memory_stats,\n print_compiled_cost_analysis,\n)\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n # Tokenizer\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_type: str = ""maskgit"" # supported options: maskgit, causal\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n use_gt_actions: bool = False\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n eval_full_frame: bool = False\n val_maskgit_steps: int = 25\n val_temperature: float = 1\n val_sample_argmax: bool = False\n wandb_id: str = """"\n\n\ndef build_model(args: Args, rng: jax.Array) -> tuple[Genie, jax.Array]:\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_actions=args.num_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n use_gt_actions=args.use_gt_actions,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n decode=False,\n rngs=rngs,\n )\n if args.use_gt_actions:\n assert (\n not args.lam_checkpoint\n ), ""Cannot use LAM when using ground-truth actions.""\n else:\n assert genie.lam is not None\n del genie.lam.decoder\n return genie, rng\n\n\ndef build_optimizer(genie: Genie, args: Args) -> tuple[nnx.Optimizer, optax.Schedule]:\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.param_dtype, # moments in full precision\n )\n optimizer = nnx.Optimizer(genie, tx)\n return optimizer, lr_schedule\n\n\ndef build_mesh_and_sharding(\n num_devices: int,\n) -> tuple[Mesh, NamedSharding, NamedSharding, NamedSharding]:\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n actions_sharding = NamedSharding(mesh, PartitionSpec(""data"", None))\n return mesh, replicated_sharding, videos_sharding, actions_sharding\n\n\ndef shard_optimizer_states(\n optimizer: nnx.Optimizer, replicated_sharding: NamedSharding\n) -> None:\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n\ndef build_dataloader(args: Args, data_dir: str) -> grain.DataLoaderIterator:\n image_shape = (args.image_height, args.image_width, args.image_channels)\n array_record_files = [\n os.path.join(data_dir, x)\n for x in os.listdir(data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n return grain_iterator\n\n\ndef build_checkpoint_manager(args: Args) -> ocp.CheckpointManager:\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n return checkpoint_manager\n\n\ndef restore_or_initialize_components(\n args: Args,\n checkpoint_manager: ocp.CheckpointManager,\n optimizer: nnx.Optimizer,\n train_iterator: grain.DataLoaderIterator,\n rng: jax.Array,\n replicated_sharding: NamedSharding,\n val_iterator: Optional[grain.DataLoaderIterator],\n restore_step: Optional[int] = None,\n) -> tuple[\n int, nnx.Optimizer, grain.DataLoaderIterator, grain.DataLoaderIterator, jax.Array\n]:\n step = 0\n if restore_step is None:\n restore_step = checkpoint_manager.latest_step()\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n if val_iterator:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n val_dataloader_state=grain.checkpoint.CheckpointRestore(val_iterator), # type: ignore\n )\n else:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n )\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(), args=restore_args\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n train_iterator = restored[""train_dataloader_state""]\n if val_iterator:\n val_iterator = restored[""val_dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n rng, _rng = jax.random.split(rng)\n optimizer = restore_genie_components(optimizer, replicated_sharding, _rng, args)\n # NOTE: We have to remove the (unused) tokenizer vq dropout due flax.nnx lazily initializing modules.\n # Specifically, the first dynamics model checkpoint will contain the vq dropout module,\n # but the first full restore will fail due to nnx not initializing the module when\n # dropout is set to 0.0.\n del optimizer.model.tokenizer.vq.drop\n return step, optimizer, train_iterator, val_iterator, rng\n\n\ndef _calculate_step_metrics(\n outputs: dict[str, jax.Array],\n gt: jax.Array,\n num_actions: int,\n num_patch_latents: int,\n) -> tuple[jax.Array, dict]:\n mask = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n gt_val = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt_val, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt_val, recon)).mean()\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]),\n size=num_patch_latents,\n fill_value=0,\n )\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n psnr=psnr,\n ssim=ssim,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n if ""lam_indices"" in outputs.keys():\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]),\n size=num_actions,\n fill_value=0,\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n metrics[""codebook_usage_lam""] = codebook_usage_lam\n return ce_loss, metrics\n\n\ndef main(args: Args) -> None:\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n genie, rng = build_model(args, rng)\n _, params, _ = nnx.split(genie, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n optimizer, lr_schedule = build_optimizer(genie, args)\n del genie\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n _, replicated_sharding, videos_sharding, actions_sharding = build_mesh_and_sharding(\n num_devices\n )\n\n shard_optimizer_states(optimizer, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n checkpoint_manager = build_checkpoint_manager(args)\n\n # --- Create DataLoaderIterator from dataloader ---\n train_iterator = build_dataloader(args, args.data_dir)\n val_iterator = None\n if args.val_data_dir:\n val_iterator = build_dataloader(args, args.val_data_dir)\n\n # --- Restore checkpoint ---\n step, optimizer, train_iterator, val_iterator, rng = (\n restore_or_initialize_components(\n args,\n checkpoint_manager,\n optimizer,\n train_iterator,\n rng,\n replicated_sharding,\n val_iterator,\n )\n )\n\n # --- Define loss and train step (close over args) ---\n def dynamics_loss_fn(\n model: Genie,\n inputs: dict,\n training: bool = False,\n ) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n ce_loss, metrics = _calculate_step_metrics(\n outputs, gt, args.num_actions, args.num_patch_latents\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n @nnx.jit(donate_argnums=0)\n def train_step(\n optimizer: nnx.Optimizer, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(model: Genie) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n model.train()\n return dynamics_loss_fn(model, inputs, training=True)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n optimizer.model\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return loss, recon, metrics\n\n @nnx.jit\n def val_step(genie: Genie, inputs: dict) -> dict:\n """"""Evaluate model and compute metrics""""""\n genie.eval()\n (loss, (recon, metrics)) = dynamics_loss_fn(genie, inputs, training=False)\n val_output = {""loss"": loss, ""recon"": recon, ""metrics"": metrics}\n\n # --- Evaluate full frame prediction (sampling) ---\n if args.eval_full_frame:\n tokenizer_outputs = genie.tokenizer.vq_encode(\n inputs[""videos""], training=False\n )\n tokens_full_frame = tokenizer_outputs[""indices""]\n lam_indices = None\n if not args.use_gt_actions:\n lam_indices = genie.vq_encode(inputs, training=False)\n inputs[""latent_actions""] = lam_indices\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt[:, :-1].astype(\n args.dtype\n ) # remove last frame for generation\n recon_full_frame, logits_full_frame = genie.sample(\n inputs,\n args.seq_len,\n args.val_temperature,\n args.val_sample_argmax,\n args.val_maskgit_steps,\n )\n step_outputs = {\n ""recon"": recon_full_frame,\n ""token_logits"": logits_full_frame,\n ""video_tokens"": tokens_full_frame,\n ""mask"": jnp.zeros_like(tokens_full_frame).at[:, -1].set(True),\n }\n if lam_indices is not None:\n step_outputs[""lam_indices""] = lam_indices\n\n loss_full_frame, metrics_full_frame = _calculate_step_metrics(\n step_outputs, gt, args.num_actions, args.num_patch_latents\n )\n val_output.update(\n {\n ""loss_full_frame"": loss_full_frame,\n ""recon_full_frame"": recon_full_frame,\n ""metrics_full_frame"": metrics_full_frame,\n }\n )\n return val_output\n\n def calculate_validation_metrics(val_dataloader, genie, rng):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n loss_full_frame_per_step = []\n metrics_full_frame_per_step = []\n batch = None\n recon = None\n recon_full_frame = None\n for batch in val_dataloader:\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n val_outputs = val_step(genie, batch)\n loss_per_step.append(val_outputs[""loss""])\n metrics_per_step.append(val_outputs[""metrics""])\n recon = val_outputs[""recon""]\n if args.eval_full_frame:\n loss_full_frame_per_step.append(val_outputs[""loss_full_frame""])\n metrics_full_frame_per_step.append(val_outputs[""metrics_full_frame""])\n recon_full_frame = val_outputs[""recon_full_frame""]\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(\n f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}""\n )\n\n val_metrics = {\n f""val_{key}"": np.mean([float(m[key]) for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n val_metrics[""val_loss""] = np.mean(loss_per_step)\n if args.eval_full_frame:\n val_metrics_full_frame = {\n f""val_full_frame_{key}"": np.mean(\n [float(m[key]) for m in metrics_full_frame_per_step]\n )\n for key in metrics_full_frame_per_step[0].keys()\n }\n val_metrics.update(val_metrics_full_frame)\n val_metrics[""val_loss_full_frame""] = np.mean(loss_full_frame_per_step)\n return val_metrics, batch, recon, recon_full_frame\n\n # --- TRAIN LOOP ---\n dataloader_train = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in train_iterator\n )\n dataloader_val = None\n if val_iterator:\n dataloader_val = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in val_iterator\n )\n if jax.process_index() == 0:\n first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader_train = itertools.chain([first_batch], dataloader_train)\n print(f""Starting training from step {step}..."")\n first_step = step\n while step < args.num_steps:\n for batch in dataloader_train:\n # --- Train step ---\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n loss, recon, metrics = train_step(optimizer, batch)\n if step == first_step:\n print_mem_stats(""After params initialized"")\n # metrics[""lr""] = lr_schedule(step)\n # print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Validation loss ---\n val_results = {}\n if dataloader_val and step % args.val_interval == 0:\n rng, _rng_mask_val = jax.random.split(rng, 2)\n print(""Calculating validation metrics..."")\n val_metrics, val_gt_batch, val_recon, val_recon_full_frame = (\n calculate_validation_metrics(\n dataloader_val, optimizer.model, _rng_mask_val\n )\n )\n print(f""Step {step}, validation loss: {val_metrics['val_loss']}"")\n val_results = {\n ""metrics"": val_metrics,\n ""gt_batch"": val_gt_batch,\n ""recon"": val_recon,\n ""full_frame"": val_recon_full_frame,\n }\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n log_dict = {""loss"": loss, ""step"": step, **metrics}\n if val_results:\n log_dict.update(val_results[""metrics""])\n wandb.log(log_dict)\n if step % args.log_image_interval == 0:\n gt_seq = batch[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if val_results:\n val_results[""gt_seq_val""] = (\n val_results[""gt_batch""][""videos""][0].astype(jnp.float32)\n / 255.0\n )\n val_results[""recon_seq_val""] = val_results[""recon""][0].clip(\n 0, 1\n )\n val_comparison_seq = jnp.concatenate(\n (val_results[""gt_seq_val""], val_results[""recon_seq_val""]),\n axis=1,\n )\n val_results[""val_comparison_seq""] = einops.rearrange(\n val_comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if args.eval_full_frame:\n val_results[""full_frame_seq_val""] = val_results[\n ""full_frame""\n ][0].clip(0, 1)\n val_results[""val_full_frame_comparison_seq""] = (\n jnp.concatenate(\n (\n val_results[""gt_seq_val""],\n val_results[""full_frame_seq_val""],\n ),\n axis=1,\n )\n )\n val_results[""val_full_frame_comparison_seq""] = (\n einops.rearrange(\n val_results[""val_full_frame_comparison_seq""] * 255,\n ""t h w c -> h (t w) c"",\n )\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n if val_results:\n log_images.update(\n dict(\n val_image=wandb.Image(\n np.asarray(\n val_results[""gt_seq_val""][args.seq_len - 1]\n )\n ),\n val_recon=wandb.Image(\n np.asarray(\n val_results[""recon_seq_val""][\n args.seq_len - 1\n ]\n )\n ),\n val_true_vs_recon=wandb.Image(\n np.asarray(\n val_results[""val_comparison_seq""].astype(\n np.uint8\n )\n )\n ),\n )\n )\n if args.eval_full_frame:\n log_images.update(\n dict(\n val_full_frame=wandb.Image(\n np.asarray(\n val_results[""full_frame_seq_val""][\n args.seq_len - 1\n ]\n )\n ),\n val_true_vs_full_frame=wandb.Image(\n np.asarray(\n val_results[\n ""val_full_frame_comparison_seq""\n ].astype(np.uint8)\n )\n ),\n )\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n if val_iterator:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n val_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n val_iterator # type: ignore\n ),\n )\n else:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n )\n checkpoint_manager.save(step, args=ckpt_manager_args)\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n main(args)\n",python,tab +664,2328445,"genie.py",0,0,"from typing import Dict\n\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\nimport orbax.checkpoint as ocp\n\nfrom models.dynamics import DynamicsMaskGIT, DynamicsCausal\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\n\nclass Genie(nnx.Module):\n """"""Genie model""""""\n\n def __init__(\n self,\n in_dim: int,\n tokenizer_dim: int,\n tokenizer_ffn_dim: int,\n latent_patch_dim: int,\n num_patch_latents: int,\n patch_size: int,\n tokenizer_num_blocks: int,\n tokenizer_num_heads: int,\n lam_dim: int,\n lam_ffn_dim: int,\n latent_action_dim: int,\n num_actions: int,\n lam_patch_size: int,\n lam_num_blocks: int,\n lam_num_heads: int,\n lam_co_train: bool,\n use_gt_actions: bool,\n dyna_type: str,\n dyna_dim: int,\n dyna_ffn_dim: int,\n dyna_num_blocks: int,\n dyna_num_heads: int,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n dropout: float = 0.0,\n mask_limit: float = 0.0,\n ):\n # --- Tokenizer ---\n self.in_dim = in_dim\n self.tokenizer_dim = tokenizer_dim\n self.tokenizer_ffn_dim = tokenizer_ffn_dim\n self.latent_patch_dim = latent_patch_dim\n self.num_patch_latents = num_patch_latents\n self.patch_size = patch_size\n self.tokenizer_num_blocks = tokenizer_num_blocks\n self.tokenizer_num_heads = tokenizer_num_heads\n # --- LAM ---\n self.lam_dim = lam_dim\n self.lam_ffn_dim = lam_ffn_dim\n self.latent_action_dim = latent_action_dim\n self.num_actions = num_actions\n self.lam_patch_size = lam_patch_size\n self.lam_num_blocks = lam_num_blocks\n self.lam_num_heads = lam_num_heads\n self.lam_co_train = lam_co_train\n self.use_gt_actions = use_gt_actions\n # --- Dynamics ---\n self.dyna_type = dyna_type\n self.dyna_dim = dyna_dim\n self.dyna_ffn_dim = dyna_ffn_dim\n self.dyna_num_blocks = dyna_num_blocks\n self.dyna_num_heads = dyna_num_heads\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.dropout = dropout\n self.mask_limit = mask_limit\n self.decode = decode\n\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n ffn_dim=self.tokenizer_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n if self.use_gt_actions:\n self.action_embed = nnx.Embed(\n self.num_actions, self.latent_action_dim, rngs=rngs\n )\n self.lam = None\n else:\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n ffn_dim=self.lam_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.action_embed = None\n if self.dyna_type == ""maskgit"":\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n elif self.dyna_type == ""causal"":\n self.dynamics = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n rngs=rngs,\n )\n else:\n raise ValueError(f""Invalid dynamics type: {self.dyna_type}"")\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n training: bool = True,\n ) -> Dict[str, jax.Array]:\n videos_BTHWC = batch[""videos""]\n tokenizer_outputs = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_indices_BTN = tokenizer_outputs[""indices""]\n latent_actions_BTm11L = None\n action_embeddings_BTm11L = None\n if self.use_gt_actions:\n assert self.action_embed is not None\n action_indices_E = None\n action_embeddings_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n action_embeddings_BTm11L = action_embeddings_BT1L[:, :-1]\n else:\n assert self.lam is not None\n lam_outputs = self.lam.vq_encode(videos_BTHWC, training=False)\n z_q_BTm11L = lam_outputs[""z_q""]\n action_indices_E = lam_outputs[""indices""]\n latent_actions_BTm11L = jax.lax.cond(\n self.lam_co_train,\n lambda: z_q_BTm11L,\n lambda: jax.lax.stop_gradient(z_q_BTm11L),\n )\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(token_indices_BTN),\n latent_actions=(\n action_embeddings_BTm11L\n if self.use_gt_actions\n else latent_actions_BTm11L\n ),\n )\n outputs[""mask_rng""] = batch[""rng""]\n dyna_logits_BTNV, dyna_mask = self.dynamics(outputs, training)\n outputs[""token_logits""] = dyna_logits_BTNV\n outputs[""mask""] = dyna_mask\n mle_indices_BTN = jnp.argmax(outputs[""token_logits""], axis=-1)\n H, W = batch[""videos""].shape[2:4]\n outputs[""recon""] = self.tokenizer.decode(mle_indices_BTN, (H, W))\n if action_indices_E is not None:\n outputs[""lam_indices""] = action_indices_E\n return outputs\n\n def sample(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n maskgit_steps: int = 25,\n ) -> tuple[jax.Array, jax.Array]:\n if self.dyna_type == ""maskgit"":\n return self.sample_maskgit(\n batch, seq_len, maskgit_steps, temperature, sample_argmax\n )\n elif self.dyna_type == ""causal"":\n return self.sample_causal(batch, seq_len, temperature, sample_argmax)\n else:\n raise ValueError(f""Dynamics model type unknown: {self.dyna_type}"")\n\n def sample_maskgit(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n steps: int = 25,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n P: S * N\n """"""\n assert isinstance(self.dynamics, DynamicsMaskGIT)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n init_logits_BSNV = jnp.zeros(\n shape=(*token_idxs_BSN.shape, self.num_patch_latents)\n )\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, :-1]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n # --- Extract submodule state ---\n dynamics_state = nnx.state(self.dynamics)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def maskgit_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, mask_BSN, action_tokens_EL = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_maskgit = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_maskgit, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_maskgit.patch_embed(token_idxs_BSN)\n mask_token_111M = dynamics_maskgit.mask_token.value\n mask_expanded_BSN1 = mask_BSN[..., None]\n vid_embed_BSNM = jnp.where(\n mask_expanded_BSN1, mask_token_111M, vid_embed_BSNM\n )\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_maskgit.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n vid_embed_BSNM += act_embed_BS1M\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (steps * 2))\n step_temp = temperature * (1.0 - unmasked_ratio)\n final_logits_BSNV = dynamics_maskgit.transformer(vid_embed_BSNM) / step_temp\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_BSN = jnp.argmax(final_logits_BSNV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_BSN = jax.random.categorical(_rng, final_logits_BSNV)\n gather_fn = jax.vmap(jax.vmap(jax.vmap(lambda x, y: x[y])))\n final_token_probs_BSN = gather_fn(\n jax.nn.softmax(final_logits_BSNV), sampled_token_idxs_BSN\n )\n final_token_probs_BSN += ~mask_BSN\n # Update masked tokens and logits only\n token_idxs_BSN = jnp.where(mask_BSN, sampled_token_idxs_BSN, token_idxs_BSN)\n logits_BSNV = jnp.where(\n jnp.expand_dims(mask_BSN, -1), final_logits_BSNV, logits_BSNV\n )\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n final_token_probs_flat_BP = einops.rearrange(\n final_token_probs_BSN, ""b s n -> b (s n)""\n )\n idx_mask_P = (\n jnp.arange(final_token_probs_flat_BP.shape[-1])\n <= N - num_unmasked_tokens\n )\n sorted_idxs_BP = jnp.argsort(final_token_probs_flat_BP, axis=-1)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask_P))\n mask_flat_BP = einops.rearrange(mask_BSN, ""b s n -> b (s n)"")\n new_mask_flat_BP = mask_update_fn(mask_flat_BP, sorted_idxs_BP)\n new_mask_BSN = einops.rearrange(new_mask_flat_BP, ""b (s n) -> b s n"", n=N)\n\n new_carry = (\n rng,\n token_idxs_BSN,\n logits_BSNV,\n new_mask_BSN,\n action_tokens_EL,\n )\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # Mask current frame (i.e., t == step_t)\n mask_S = jnp.arange(seq_len) == step_t\n mask_BSN = jnp.broadcast_to(mask_S[None, :, None], (B, seq_len, N)).astype(\n bool\n )\n masked_token_idxs_BSN = current_token_idxs_BSN * ~mask_BSN\n masked_logits_BSNV = current_logits_BSNV * jnp.expand_dims(~mask_BSN, -1)\n\n # --- Initialize and run MaskGIT loop ---\n init_carry_maskgit = (\n step_rng,\n masked_token_idxs_BSN,\n masked_logits_BSNV,\n mask_BSN,\n action_tokens_EL,\n )\n final_carry_maskgit = maskgit_step_fn(init_carry_maskgit, jnp.arange(steps))\n updated_token_idxs_BSN = final_carry_maskgit[1]\n updated_logits_BSNV = final_carry_maskgit[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, init_logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def sample_causal(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n assert isinstance(self.dynamics, DynamicsCausal)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n logits_BSNV = jnp.zeros((*token_idxs_BSN.shape, self.num_patch_latents))\n dynamics_state = nnx.state(self.dynamics)\n\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, :-1]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def causal_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step_n: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_causal = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=self.decode,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_causal, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_causal.patch_embed(token_idxs_BSN)\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_causal.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n vid_embed_BSNp1M = jnp.concatenate([act_embed_BS1M, vid_embed_BSNM], axis=2)\n final_logits_BTNp1V = (\n dynamics_causal.transformer(vid_embed_BSNp1M, (step_t, step_n))\n / temperature\n )\n final_logits_BV = final_logits_BTNp1V[:, step_t, step_n, :]\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_B = jnp.argmax(final_logits_BV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_B = jax.random.categorical(_rng, final_logits_BV)\n # Update next tokens only\n token_idxs_BSN = token_idxs_BSN.at[:, step_t, step_n].set(\n sampled_token_idxs_B\n )\n logits_BSNV = logits_BSNV.at[:, step_t, step_n].set(final_logits_BV)\n\n new_carry = (rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t)\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # --- Initialize and run causal loop ---\n init_carry_causal = (\n step_rng,\n current_token_idxs_BSN,\n current_logits_BSNV,\n action_tokens_EL,\n step_t,\n )\n final_carry_causal = causal_step_fn(init_carry_causal, jnp.arange(N))\n updated_token_idxs_BSN = final_carry_causal[1]\n updated_logits_BSNV = final_carry_causal[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def vq_encode(self, batch: Dict[str, jax.Array], training: bool) -> jax.Array:\n # --- Preprocess videos ---\n assert self.lam is not None\n video_BTHWC = batch[""videos""]\n lam_output: Dict[str, jax.Array] = self.lam.vq_encode(\n video_BTHWC, training=training\n )\n lam_indices_E = lam_output[""indices""]\n return lam_indices_E\n\n\n# FIXME (f.srambical): add conversion script for old checkpoints\ndef restore_genie_components(\n optimizer: nnx.Optimizer,\n sharding: jax.sharding.NamedSharding,\n rng: jax.Array,\n args,\n) -> nnx.Optimizer:\n """"""Restore pre-trained Genie components""""""\n rng_tokenizer, rng_lam = jax.random.split(rng)\n rngs_tokenizer = nnx.Rngs(rng_tokenizer)\n rngs_lam = nnx.Rngs(rng_lam)\n\n tx = optimizer.tx\n model = optimizer.model\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n tokenizer_checkpoint_manager = ocp.CheckpointManager(\n directory=args.tokenizer_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n ffn_dim=args.tokenizer_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_tokenizer,\n )\n dummy_tokenizer_optimizer = nnx.Optimizer(dummy_tokenizer, tx)\n dummy_tokenizer_optimizer_state = nnx.state(dummy_tokenizer_optimizer)\n abstract_sharded_tokenizer_optimizer_state = _create_abstract_sharded_pytree(\n dummy_tokenizer_optimizer_state, sharding\n )\n restored_tokenizer = tokenizer_checkpoint_manager.restore(\n step=tokenizer_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_tokenizer_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_tokenizer_optimizer.model, restored_tokenizer.model)\n model.tokenizer = dummy_tokenizer_optimizer.model\n tokenizer_checkpoint_manager.close()\n\n if args.lam_checkpoint:\n lam_checkpoint_manager = ocp.CheckpointManager(\n directory=args.lam_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n ffn_dim=args.lam_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_lam,\n )\n dummy_lam_optimizer = nnx.Optimizer(dummy_lam, tx)\n dummy_lam_optimizer_state = nnx.state(dummy_lam_optimizer)\n abstract_sharded_lam_optimizer_state = _create_abstract_sharded_pytree(\n dummy_lam_optimizer_state, sharding\n )\n restored_lam_optimizer = lam_checkpoint_manager.restore(\n step=lam_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_lam_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_lam_optimizer.model, restored_lam_optimizer.model)\n model.lam = dummy_lam_optimizer.model\n # Remove the LAM decoder to save memory and avoid unnecessary computation.\n del model.lam.decoder\n lam_checkpoint_manager.close()\n\n # Reinitialize the optimizer states\n optimizer = nnx.Optimizer(model, tx)\n return optimizer\n\n\ndef _create_abstract_sharded_pytree(\n pytree_template: nnx.GraphState, sharding_spec: jax.sharding.NamedSharding\n) -> jax.Array:\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)\n",python,tab +665,2328448,"genie.py",9930,0,"",python,selection_command +666,2369109,"TERMINAL",0,0,"Step 750, validation loss: 0.6088705062866211\r\n",,terminal_output +667,2428368,"genie.py",0,0,"from typing import Dict\n\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\nimport orbax.checkpoint as ocp\n\nfrom models.dynamics import DynamicsMaskGIT, DynamicsCausal\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\n\nclass Genie(nnx.Module):\n """"""Genie model""""""\n\n def __init__(\n self,\n in_dim: int,\n tokenizer_dim: int,\n tokenizer_ffn_dim: int,\n latent_patch_dim: int,\n num_patch_latents: int,\n patch_size: int,\n tokenizer_num_blocks: int,\n tokenizer_num_heads: int,\n lam_dim: int,\n lam_ffn_dim: int,\n latent_action_dim: int,\n num_actions: int,\n lam_patch_size: int,\n lam_num_blocks: int,\n lam_num_heads: int,\n lam_co_train: bool,\n use_gt_actions: bool,\n dyna_type: str,\n dyna_dim: int,\n dyna_ffn_dim: int,\n dyna_num_blocks: int,\n dyna_num_heads: int,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n dropout: float = 0.0,\n mask_limit: float = 0.0,\n ):\n # --- Tokenizer ---\n self.in_dim = in_dim\n self.tokenizer_dim = tokenizer_dim\n self.tokenizer_ffn_dim = tokenizer_ffn_dim\n self.latent_patch_dim = latent_patch_dim\n self.num_patch_latents = num_patch_latents\n self.patch_size = patch_size\n self.tokenizer_num_blocks = tokenizer_num_blocks\n self.tokenizer_num_heads = tokenizer_num_heads\n # --- LAM ---\n self.lam_dim = lam_dim\n self.lam_ffn_dim = lam_ffn_dim\n self.latent_action_dim = latent_action_dim\n self.num_actions = num_actions\n self.lam_patch_size = lam_patch_size\n self.lam_num_blocks = lam_num_blocks\n self.lam_num_heads = lam_num_heads\n self.lam_co_train = lam_co_train\n self.use_gt_actions = use_gt_actions\n # --- Dynamics ---\n self.dyna_type = dyna_type\n self.dyna_dim = dyna_dim\n self.dyna_ffn_dim = dyna_ffn_dim\n self.dyna_num_blocks = dyna_num_blocks\n self.dyna_num_heads = dyna_num_heads\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.dropout = dropout\n self.mask_limit = mask_limit\n self.decode = decode\n\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n ffn_dim=self.tokenizer_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n if self.use_gt_actions:\n self.action_embed = nnx.Embed(\n self.num_actions, self.latent_action_dim, rngs=rngs\n )\n self.lam = None\n else:\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n ffn_dim=self.lam_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.action_embed = None\n if self.dyna_type == ""maskgit"":\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n elif self.dyna_type == ""causal"":\n self.dynamics = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n rngs=rngs,\n )\n else:\n raise ValueError(f""Invalid dynamics type: {self.dyna_type}"")\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n training: bool = True,\n ) -> Dict[str, jax.Array]:\n videos_BTHWC = batch[""videos""]\n tokenizer_outputs = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_indices_BTN = tokenizer_outputs[""indices""]\n latent_actions_BTm11L = None\n action_embeddings_BTm11L = None\n if self.use_gt_actions:\n assert self.action_embed is not None\n action_indices_E = None\n action_embeddings_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n action_embeddings_BTm11L = action_embeddings_BT1L[:, :-1]\n else:\n assert self.lam is not None\n lam_outputs = self.lam.vq_encode(videos_BTHWC, training=False)\n z_q_BTm11L = lam_outputs[""z_q""]\n action_indices_E = lam_outputs[""indices""]\n latent_actions_BTm11L = jax.lax.cond(\n self.lam_co_train,\n lambda: z_q_BTm11L,\n lambda: jax.lax.stop_gradient(z_q_BTm11L),\n )\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(token_indices_BTN),\n latent_actions=(\n action_embeddings_BTm11L\n if self.use_gt_actions\n else latent_actions_BTm11L\n ),\n )\n outputs[""mask_rng""] = batch[""rng""]\n dyna_logits_BTNV, dyna_mask = self.dynamics(outputs, training)\n outputs[""token_logits""] = dyna_logits_BTNV\n outputs[""mask""] = dyna_mask\n mle_indices_BTN = jnp.argmax(outputs[""token_logits""], axis=-1)\n H, W = batch[""videos""].shape[2:4]\n outputs[""recon""] = self.tokenizer.decode(mle_indices_BTN, (H, W))\n if action_indices_E is not None:\n outputs[""lam_indices""] = action_indices_E\n return outputs\n\n def sample(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n maskgit_steps: int = 25,\n ) -> tuple[jax.Array, jax.Array]:\n if self.dyna_type == ""maskgit"":\n return self.sample_maskgit(\n batch, seq_len, maskgit_steps, temperature, sample_argmax\n )\n elif self.dyna_type == ""causal"":\n return self.sample_causal(batch, seq_len, temperature, sample_argmax)\n else:\n raise ValueError(f""Dynamics model type unknown: {self.dyna_type}"")\n\n def sample_maskgit(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n steps: int = 25,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n P: S * N\n """"""\n assert isinstance(self.dynamics, DynamicsMaskGIT)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n init_logits_BSNV = jnp.zeros(\n shape=(*token_idxs_BSN.shape, self.num_patch_latents)\n )\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, 1:]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n # --- Extract submodule state ---\n dynamics_state = nnx.state(self.dynamics)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def maskgit_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, mask_BSN, action_tokens_EL = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_maskgit = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_maskgit, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_maskgit.patch_embed(token_idxs_BSN)\n mask_token_111M = dynamics_maskgit.mask_token.value\n mask_expanded_BSN1 = mask_BSN[..., None]\n vid_embed_BSNM = jnp.where(\n mask_expanded_BSN1, mask_token_111M, vid_embed_BSNM\n )\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_maskgit.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n vid_embed_BSNM += act_embed_BS1M\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (steps * 2))\n step_temp = temperature * (1.0 - unmasked_ratio)\n final_logits_BSNV = dynamics_maskgit.transformer(vid_embed_BSNM) / step_temp\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_BSN = jnp.argmax(final_logits_BSNV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_BSN = jax.random.categorical(_rng, final_logits_BSNV)\n gather_fn = jax.vmap(jax.vmap(jax.vmap(lambda x, y: x[y])))\n final_token_probs_BSN = gather_fn(\n jax.nn.softmax(final_logits_BSNV), sampled_token_idxs_BSN\n )\n final_token_probs_BSN += ~mask_BSN\n # Update masked tokens and logits only\n token_idxs_BSN = jnp.where(mask_BSN, sampled_token_idxs_BSN, token_idxs_BSN)\n logits_BSNV = jnp.where(\n jnp.expand_dims(mask_BSN, -1), final_logits_BSNV, logits_BSNV\n )\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n final_token_probs_flat_BP = einops.rearrange(\n final_token_probs_BSN, ""b s n -> b (s n)""\n )\n idx_mask_P = (\n jnp.arange(final_token_probs_flat_BP.shape[-1])\n <= N - num_unmasked_tokens\n )\n sorted_idxs_BP = jnp.argsort(final_token_probs_flat_BP, axis=-1)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask_P))\n mask_flat_BP = einops.rearrange(mask_BSN, ""b s n -> b (s n)"")\n new_mask_flat_BP = mask_update_fn(mask_flat_BP, sorted_idxs_BP)\n new_mask_BSN = einops.rearrange(new_mask_flat_BP, ""b (s n) -> b s n"", n=N)\n\n new_carry = (\n rng,\n token_idxs_BSN,\n logits_BSNV,\n new_mask_BSN,\n action_tokens_EL,\n )\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # Mask current frame (i.e., t == step_t)\n mask_S = jnp.arange(seq_len) == step_t\n mask_BSN = jnp.broadcast_to(mask_S[None, :, None], (B, seq_len, N)).astype(\n bool\n )\n masked_token_idxs_BSN = current_token_idxs_BSN * ~mask_BSN\n masked_logits_BSNV = current_logits_BSNV * jnp.expand_dims(~mask_BSN, -1)\n\n # --- Initialize and run MaskGIT loop ---\n init_carry_maskgit = (\n step_rng,\n masked_token_idxs_BSN,\n masked_logits_BSNV,\n mask_BSN,\n action_tokens_EL,\n )\n final_carry_maskgit = maskgit_step_fn(init_carry_maskgit, jnp.arange(steps))\n updated_token_idxs_BSN = final_carry_maskgit[1]\n updated_logits_BSNV = final_carry_maskgit[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, init_logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def sample_causal(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n assert isinstance(self.dynamics, DynamicsCausal)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n logits_BSNV = jnp.zeros((*token_idxs_BSN.shape, self.num_patch_latents))\n dynamics_state = nnx.state(self.dynamics)\n\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, 1:]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def causal_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step_n: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_causal = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=self.decode,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_causal, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_causal.patch_embed(token_idxs_BSN)\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_causal.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n vid_embed_BSNp1M = jnp.concatenate([act_embed_BS1M, vid_embed_BSNM], axis=2)\n final_logits_BTNp1V = (\n dynamics_causal.transformer(vid_embed_BSNp1M, (step_t, step_n))\n / temperature\n )\n final_logits_BV = final_logits_BTNp1V[:, step_t, step_n, :]\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_B = jnp.argmax(final_logits_BV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_B = jax.random.categorical(_rng, final_logits_BV)\n # Update next tokens only\n token_idxs_BSN = token_idxs_BSN.at[:, step_t, step_n].set(\n sampled_token_idxs_B\n )\n logits_BSNV = logits_BSNV.at[:, step_t, step_n].set(final_logits_BV)\n\n new_carry = (rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t)\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # --- Initialize and run causal loop ---\n init_carry_causal = (\n step_rng,\n current_token_idxs_BSN,\n current_logits_BSNV,\n action_tokens_EL,\n step_t,\n )\n final_carry_causal = causal_step_fn(init_carry_causal, jnp.arange(N))\n updated_token_idxs_BSN = final_carry_causal[1]\n updated_logits_BSNV = final_carry_causal[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def vq_encode(self, batch: Dict[str, jax.Array], training: bool) -> jax.Array:\n # --- Preprocess videos ---\n assert self.lam is not None\n video_BTHWC = batch[""videos""]\n lam_output: Dict[str, jax.Array] = self.lam.vq_encode(\n video_BTHWC, training=training\n )\n lam_indices_E = lam_output[""indices""]\n return lam_indices_E\n\n\n# FIXME (f.srambical): add conversion script for old checkpoints\ndef restore_genie_components(\n optimizer: nnx.Optimizer,\n sharding: jax.sharding.NamedSharding,\n rng: jax.Array,\n args,\n) -> nnx.Optimizer:\n """"""Restore pre-trained Genie components""""""\n rng_tokenizer, rng_lam = jax.random.split(rng)\n rngs_tokenizer = nnx.Rngs(rng_tokenizer)\n rngs_lam = nnx.Rngs(rng_lam)\n\n tx = optimizer.tx\n model = optimizer.model\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n tokenizer_checkpoint_manager = ocp.CheckpointManager(\n directory=args.tokenizer_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n ffn_dim=args.tokenizer_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_tokenizer,\n )\n dummy_tokenizer_optimizer = nnx.Optimizer(dummy_tokenizer, tx)\n dummy_tokenizer_optimizer_state = nnx.state(dummy_tokenizer_optimizer)\n abstract_sharded_tokenizer_optimizer_state = _create_abstract_sharded_pytree(\n dummy_tokenizer_optimizer_state, sharding\n )\n restored_tokenizer = tokenizer_checkpoint_manager.restore(\n step=tokenizer_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_tokenizer_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_tokenizer_optimizer.model, restored_tokenizer.model)\n model.tokenizer = dummy_tokenizer_optimizer.model\n tokenizer_checkpoint_manager.close()\n\n if args.lam_checkpoint:\n lam_checkpoint_manager = ocp.CheckpointManager(\n directory=args.lam_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n ffn_dim=args.lam_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_lam,\n )\n dummy_lam_optimizer = nnx.Optimizer(dummy_lam, tx)\n dummy_lam_optimizer_state = nnx.state(dummy_lam_optimizer)\n abstract_sharded_lam_optimizer_state = _create_abstract_sharded_pytree(\n dummy_lam_optimizer_state, sharding\n )\n restored_lam_optimizer = lam_checkpoint_manager.restore(\n step=lam_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_lam_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_lam_optimizer.model, restored_lam_optimizer.model)\n model.lam = dummy_lam_optimizer.model\n # Remove the LAM decoder to save memory and avoid unnecessary computation.\n del model.lam.decoder\n lam_checkpoint_manager.close()\n\n # Reinitialize the optimizer states\n optimizer = nnx.Optimizer(model, tx)\n return optimizer\n\n\ndef _create_abstract_sharded_pytree(\n pytree_template: nnx.GraphState, sharding_spec: jax.sharding.NamedSharding\n) -> jax.Array:\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)\n",python,tab +668,2428371,"genie.py",9814,0,"",python,selection_mouse +669,2428483,"genie.py",9805,12,"action_embed",python,selection_mouse +670,2429370,"genie.py",9889,0,"",python,selection_mouse +671,2429993,"genie.py",9811,0,"",python,selection_mouse +672,2430121,"genie.py",9805,12,"action_embed",python,selection_mouse +673,2431462,"genie.py",9929,0,"",python,selection_mouse +674,2431463,"genie.py",9928,0,"",python,selection_command +675,2431552,"genie.py",9929,0,"",python,selection_mouse +676,2431553,"genie.py",9928,0,"",python,selection_command +677,2432109,"genie.py",9976,0,"",python,selection_mouse +678,2432244,"genie.py",9966,19,"latent_actions_BT1L",python,selection_mouse +679,2432831,"genie.py",9813,0,"",python,selection_mouse +680,2432990,"genie.py",9805,12,"action_embed",python,selection_mouse +681,2441678,"genie.py",0,0,"",python,tab +682,2441682,"train_dynamics.py",0,0,"",python,tab +683,2441731,"TERMINAL",0,0,"Saved checkpoint at step 1000\r\n",,terminal_output +684,2447133,"sample.py",0,0,"from dataclasses import dataclass\nimport time\nimport os\nimport optax\n\nimport dm_pix as pix\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nimport numpy as np\nimport orbax.checkpoint as ocp\nfrom PIL import Image, ImageDraw\nimport tyro\nfrom flax import nnx\n\nfrom genie import Genie\nfrom utils.dataloader import get_dataloader\n\n\n@dataclass\nclass Args:\n # Experiment\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data/coinrun_episodes""\n checkpoint: str = """"\n # Sampling\n batch_size: int = 1\n maskgit_steps: int = 25\n temperature: float = 1.0\n sample_argmax: bool = True\n start_frame: int = 1\n # Tokenizer checkpoint\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n # LAM checkpoint\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n use_gt_actions: bool = False\n # Dynamics checkpoint\n dyna_type: str = ""maskgit""\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\nif __name__ == ""__main__"":\n """"""\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n jax.distributed.initialize()\n\n rng = jax.random.key(args.seed)\n\n # --- Load Genie checkpoint ---\n rngs = nnx.Rngs(rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_actions=args.num_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=False,\n use_gt_actions=args.use_gt_actions,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n # FIXME (f.srambical): implement spatiotemporal KV caching and set decode=True\n decode=False,\n rngs=rngs,\n )\n\n del genie.tokenizer.vq.drop\n # Need to delete lam decoder for checkpoint loading\n if not args.use_gt_actions:\n assert genie.lam is not None\n del genie.lam.decoder\n\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n dummy_tx = optax.adamw(\n learning_rate=optax.linear_schedule(0.0001, 0.0001, 10000),\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n dummy_optimizer = nnx.Optimizer(genie, dummy_tx)\n\n abstract_optimizer = nnx.eval_shape(lambda: dummy_optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n ),\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(dummy_optimizer, restored_optimizer_state)\n\n # --- Define sampling function ---\n def _sampling_fn(model: Genie, batch: dict) -> jax.Array:\n """"""Runs Genie.sample with pre-defined generation hyper-parameters.""""""\n assert args.dyna_type in [\n ""maskgit"",\n ""causal"",\n ], f""Invalid dynamics type: {args.dyna_type}""\n frames, _ = model.sample(\n batch,\n args.seq_len,\n args.temperature,\n args.sample_argmax,\n args.maskgit_steps,\n )\n return frames\n\n # --- Define autoregressive sampling loop ---\n def _autoreg_sample(genie, rng, batch):\n batch[""videos""] = batch[""videos""][:, : args.start_frame]\n batch[""rng""] = rng\n generated_vid_BSHWC = _sampling_fn(genie, batch)\n return generated_vid_BSHWC\n\n # --- Get video + latent actions ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n args.batch_size,\n args.image_height,\n args.image_width,\n args.image_channels,\n # We don't use workers in order to avoid grain shutdown issues (https://github.com/google/grain/issues/398)\n num_workers=0,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n dataloader = iter(dataloader)\n batch = next(dataloader)\n gt_video = jnp.asarray(batch[""videos""], dtype=jnp.float32) / 255.0\n batch[""videos""] = gt_video.astype(args.dtype)\n # Get latent actions for all videos in the batch\n action_batch_E = None\n if not args.use_gt_actions:\n action_batch_E = genie.vq_encode(batch, training=False)\n batch[""latent_actions""] = action_batch_E\n\n # --- Sample + evaluate video ---\n recon_video_BSHWC = _autoreg_sample(genie, rng, batch)\n recon_video_BSHWC = recon_video_BSHWC.astype(jnp.float32)\n gt = (\n gt_video[:, : recon_video_BSHWC.shape[1]]\n .clip(0, 1)\n .reshape(-1, *gt_video.shape[2:])\n )\n recon = recon_video_BSHWC.clip(0, 1).reshape(-1, *recon_video_BSHWC.shape[2:])\n ssim = jnp.asarray(\n pix.ssim(gt[:, args.start_frame :], recon[:, args.start_frame :])\n ).mean()\n print(f""SSIM: {ssim}"")\n\n # --- Construct video ---\n true_videos = (gt_video * 255).astype(np.uint8)\n pred_videos = (recon_video_BSHWC * 255).astype(np.uint8)\n video_comparison = np.zeros((2, *recon_video_BSHWC.shape), dtype=np.uint8)\n video_comparison[0] = true_videos[:, : args.seq_len]\n video_comparison[1] = pred_videos\n frames = einops.rearrange(video_comparison, ""n b t h w c -> t (b h) (n w) c"")\n\n # --- Save video ---\n imgs = [Image.fromarray(img) for img in frames]\n # Write actions on each frame, on each row (i.e., for each video in the batch, on the GT row)\n B, S, _, _, _ = batch[""videos""].shape\n if action_batch_E is not None:\n action_batch_BSm11 = jnp.reshape(action_batch_E, (B, S - 1, 1))\n else:\n action_batch_BSm11 = jnp.expand_dims(batch[""actions""][:, :-1], -1)\n for t, img in enumerate(imgs[1:]):\n d = ImageDraw.Draw(img)\n for row in range(B):\n action = action_batch_BSm11[row, t, 0]\n y_offset = row * batch[""videos""].shape[2] + 2\n d.text((2, y_offset), f""{action}"", fill=255)\n imgs[0].save(\n f""generation_{time.time()}.gif"",\n save_all=True,\n append_images=imgs[1:],\n duration=250,\n loop=0,\n )\n",python,tab +685,2450689,"sample.py",6295,0,"",python,selection_mouse +686,2450779,"sample.py",6291,14,"action_batch_E",python,selection_mouse +687,2451389,"sample.py",6268,0,"",python,selection_mouse +688,2452106,"sample.py",6282,0,"",python,selection_mouse +689,2463247,"sample.py",6367,0,"",python,selection_mouse +690,2463387,"sample.py",6362,14,"latent_actions",python,selection_mouse +691,2473440,"sample.py",6362,0,"",python,selection_mouse +692,2527522,"TERMINAL",0,0,"Calculating validation metrics...\r\n",,terminal_output +693,2541906,"sample.py",7577,0,"",python,selection_mouse +694,2542061,"sample.py",7567,18,"action_batch_BSm11",python,selection_mouse +695,2542697,"sample.py",7662,0,"",python,selection_mouse +696,2542853,"sample.py",7649,18,"action_batch_BSm11",python,selection_mouse +697,2543435,"sample.py",7583,0,"",python,selection_mouse +698,2543647,"sample.py",7567,18,"action_batch_BSm11",python,selection_mouse +699,2546861,"sample.py",6301,0,"",python,selection_mouse +700,2547435,"sample.py",6360,0,"",python,selection_mouse +701,2548036,"sample.py",6361,0,"",python,selection_mouse +702,2548584,"sample.py",6268,0,"",python,selection_mouse +703,2549191,"sample.py",6300,0,"",python,selection_mouse +704,2550300,"TERMINAL",0,0,"Step 1500, validation loss: 0.36969342827796936\r\n",,terminal_output +705,2550387,"TERMINAL",0,0,"Saved checkpoint at step 1500\r\n",,terminal_output +706,2634243,"TERMINAL",0,0,"Saved checkpoint at step 2000\r\n",,terminal_output +707,2677381,"TERMINAL",0,0,"Calculating validation metrics...\r\n",,terminal_output +708,2700307,"TERMINAL",0,0,"Step 2250, validation loss: 0.3902597725391388\r\n",,terminal_output +709,3405594,"sample.py",0,0,"Switched from branch 'generate-minatar-breakout-dataset' to 'gt-actions'",python,git_branch_checkout +710,3835637,"sample.py",0,0,"Switched from branch 'gt-actions' to 'generate-minatar-breakout-dataset'",python,git_branch_checkout +711,4035635,"sample.py",0,0,"Switched from branch 'generate-minatar-breakout-dataset' to 'gt-actions'",python,git_branch_checkout +712,4250664,"sample.py",0,0,"Switched from branch 'gt-actions' to 'generate-minatar-breakout-dataset'",python,git_branch_checkout +713,7570922,"sample.py",0,0,"Switched from branch 'generate-minatar-breakout-dataset' to 'main'",python,git_branch_checkout +714,7655947,"sample.py",0,0,"Switched from branch 'main' to 'hotfix/sampling-shapes-error'",python,git_branch_checkout diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-7f667af1-e8dc-4534-8563-a4ef7be9de461754942097068-2025_08_11-21.55.24.740/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-7f667af1-e8dc-4534-8563-a4ef7be9de461754942097068-2025_08_11-21.55.24.740/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..3507c0739ae9c80f4c4f83ff4181151a2224f81b --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-7f667af1-e8dc-4534-8563-a4ef7be9de461754942097068-2025_08_11-21.55.24.740/source.csv @@ -0,0 +1,22 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +2,461,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"9:55:24 PM [info] Activating crowd-code\n9:55:24 PM [info] Recording started\n9:55:24 PM [info] Initializing git provider using file system watchers...\n9:55:24 PM [info] Git repository found\n9:55:24 PM [info] Git provider initialized successfully\n",Log,tab +3,483,"TERMINAL",0,0,"git branch",,terminal_command +4,537,"TERMINAL",0,0,"]633;E;2025-08-11 21:55:24 git branch;686e78ec-8807-43f4-8f60-628f6dc2af90]633;C[?1h=\r",,terminal_output +5,732,"TERMINAL",0,0,"* (HEAD detached at ab43d16)\r\n add-wandb-name-and-tags\r\n before-nnx\r\n causal-mem-reduce\r\n causal-spatiotemporal-kv-cache\r\n causal-st-transformer\r\n causal-transformer-dynamics-model\r\n causal-transformer-nnx-no-kv-cache\r\n convert-to-jax-array-in-iter\r\n correct-batched-sampling\r\n dev\r\n dont-let-tf-see-gpu\r\n feat/explicit-image-dims\r\n fix-action-padding-lam-future-information-access\r\n fix-sampling\r\n fix-transformer-forwardpass\r\n fix/spatiotemporal-pe-once-in-STTransformer\r\n grad-norm-log-and-clip\r\n grain-dataloader\r\n input_pipeline/add-npy2array_record\r\n logging-variants\r\n lr-schedules\r\n main\r\n maskgit-different-maskprob-per-sample\r\n:",,terminal_output +6,2249,"TERMINAL",0,0,"\r maskgit-sampling-iterative-unmasking-fix\r\n:",,terminal_output +7,2427,"TERMINAL",0,0,"\r metrics-logging-for-dynamics-model\r\n:",,terminal_output +8,2573,"TERMINAL",0,0,"\r monkey-patch\r\n:",,terminal_output +9,2811,"TERMINAL",0,0,"\r new-arch-sampling\r\n:",,terminal_output +10,2907,"TERMINAL",0,0,"\r preprocess_video\r\n:",,terminal_output +11,2986,"TERMINAL",0,0,"\r refactor-tmp\r\n:",,terminal_output +12,3470,"TERMINAL",0,0,"\rM causal-st-transformer\r\n\r:",,terminal_output +13,4286,"TERMINAL",0,0,"\rM causal-spatiotemporal-kv-cache\r\n\r:\rM causal-mem-reduce\r\n\r:\rM before-nnx\r\n\r:\rM add-wandb-name-and-tags\r\n\r:\rM* (HEAD detached at ab43d16)\r\n\r:\r\r:\r\r:\r\r:\r\r:\r\r:\r\r:\r\r:",,terminal_output +14,4417,"TERMINAL",0,0,"\r\r:\r\r:\r\r:\r\r:\r\r:",,terminal_output +15,4470,"TERMINAL",0,0,"\r\r:\r\r:",,terminal_output +16,4524,"TERMINAL",0,0,"\r\r:",,terminal_output +17,4848,"TERMINAL",0,0,"\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +18,7630,"TERMINAL",0,0,"git checkout main",,terminal_command +19,7730,"TERMINAL",0,0,"]633;E;2025-08-11 21:55:32 git checkout main;686e78ec-8807-43f4-8f60-628f6dc2af90]633;Cerror: Your local changes to the following files would be overwritten by checkout:\r\n\ttrain_dynamics.py\r\nPlease commit your changes or stash them before you switch branches.\r\nAborting\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;1",,terminal_output +20,11378,"TERMINAL",0,0,"git stash",,terminal_command +21,11436,"TERMINAL",0,0,"]633;E;2025-08-11 21:55:36 git stash;686e78ec-8807-43f4-8f60-628f6dc2af90]633;C",,terminal_output +22,11610,"TERMINAL",0,0,"Saved working directory and index state WIP on (no branch): ab43d16 fix: missing reshape and shape suffixes in `sample.py` (#134)\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-89289833-78d5-4138-b766-a3025aedfd811759399195678-2025_10_02-12.00.33.280/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-89289833-78d5-4138-b766-a3025aedfd811759399195678-2025_10_02-12.00.33.280/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..3ebd9c335e968f4ae4aa3aa4b84516b388f35a1e --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-89289833-78d5-4138-b766-a3025aedfd811759399195678-2025_10_02-12.00.33.280/source.csv @@ -0,0 +1,4696 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +2,365,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:00:33 PM [info] Activating crowd-code\n12:00:33 PM [info] Recording started\n12:00:33 PM [info] Initializing git provider using file system watchers...\n",Log,tab +3,479,"extension-output-pdoom-org.crowd-code-#1-crowd-code",153,0,"12:00:33 PM [info] Git repository found\n12:00:33 PM [info] Git provider initialized successfully\n12:00:33 PM [info] Initial git state: [object Object]\n",Log,content +4,1707,"TERMINAL",0,0,"git branc^C",,terminal_command +5,13388,"TERMINAL",0,0,"cat ~/.bashrc",,terminal_command +6,13398,"TERMINAL",0,0,"]633;C# .bashrc\r\n\r\n# Source global definitions\r\nif [ -f /etc/bashrc ]; then\r\n\t. /etc/bashrc\r\nfi\r\n\r\n# User specific environment\r\nif ! [[ ""$PATH"" =~ ""$HOME/.local/bin:$HOME/bin:"" ]]\r\nthen\r\n PATH=""$HOME/.local/bin:$HOME/bin:$PATH""\r\nfi\r\nexport PATH\r\n\r\nalias idle='sinfo_t_idle'\r\nexport ws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\nalias salloc_node='salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8'\r\nalias salloc_cpu='salloc --time=01:00:00 --partition=dev_cpuonly --nodes=1 --cpus-per-task=128'\r\nalias sync-runner=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs/""\r\nalias sync-jafar=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jafar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""\r\nalias sync-runner-2=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_2/""\r\nalias sync-runner-3=""sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jasmine /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""\r\nalias runner=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs/""\r\nalias runner-2=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_2/""\r\nalias runner-3=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine_jobs_3/""\r\nalias jafar=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar/""\r\nalias jafar-runner=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/""\r\nalias dev=""cd /home/hk-project-p0023960/tum_cte0515/Projects/jasmine/""\r\nalias smi=""watch -n1 nvidia-smi""\r\nalias idling=""watch -n1 sinfo_t_idle""\r\nalias queue='watch -n1 squeue --me'\r\nalias fqueue='watch -n 1 ""squeue -o \""%.10i %.16P %.30j %.8u %.8T %.10M %.9l %.6D %R\""""'\r\nalias fsacct_week='sacct --format=""JobID%15,JobName%30,Partition%16,AllocCPUS%3,State%12,Elapsed%10,Timelimit%10"" --starttime $(date -d ""last week"" +%Y-%m-%d) | grep -vE ""*.batch|*.extern|*.inter|bash|python|CANCELLED|echo""'\r\nalias logs=""cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir""\r\n\r\nalias sbatch_dir=""sh /home/hk-project-p0023960/tum_cte0515/sbatch_dir.sh $@""\r\nalias branch=""git rev-parse --abbrev-ref HEAD""\r\n\r\n\r\noverlapper() {\r\n if [ -z ""$1"" ]; then\r\n echo ""Usage: overlap ""\r\n else\r\n srun --jobid=""$1"" --pty /usr/bin/bash\r\n fi\r\n}\r\n# Uncomment the following line if you don't like systemctl's auto-paging feature:\r\n# export SYSTEMD_PAGER=\r\n\r\n# User specific aliases and functions\r\nexport PATH=""$HOME/ffmpeg:$PATH""\r\n\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +7,249939,"TERMINAL",0,0,"bash",,terminal_focus +8,383493,"TERMINAL",0,0,"bash",,terminal_focus +9,468500,"TERMINAL",0,0,"branch",,terminal_command +10,468533,"TERMINAL",0,0,"]633;Cdynamics_coinrun_500m_dataset_29519\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +11,480525,"TERMINAL",0,0,"git branch | pytorch",,terminal_command +12,480596,"TERMINAL",0,0,"]633;Cbash: pytorch: command not found...\r\n",,terminal_output +13,481877,"TERMINAL",0,0,"]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +14,485678,"TERMINAL",0,0,"git branch | grep pytorch",,terminal_command +15,490887,"TERMINAL",0,0,"git checkout ablation/use-pytorch-dataloader",,terminal_command +16,490924,"TERMINAL",0,0,"]633;C",,terminal_output +17,491441,"TERMINAL",0,0,"Switched to branch 'ablation/use-pytorch-dataloader'\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +18,493756,"TERMINAL",0,0,"git pull",,terminal_command +19,493810,"TERMINAL",0,0,"]633;C",,terminal_output +20,495309,"",0,0,"Switched from branch 'dynamics_coinrun_500m_dataset_29519' to 'ablation/use-pytorch-dataloader'",,git_branch_checkout +21,495685,"TERMINAL",0,0,"There is no tracking information for the current branch.\r\nPlease specify which branch you want to merge with.\r\nSee git-pull(1) for details.\r\n\r\n git pull \r\n\r\nIf you wish to set tracking information for this branch you can do so with:\r\n\r\n git branch --set-upstream-to=origin/ ablation/use-pytorch-dataloader\r\n\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +22,505387,"jasmine/train_dynamics.py",0,0,"import os\n\n\nos.environ.setdefault(""XLA_PYTHON_CLIENT_MEM_FRACTION"", ""0.98"")\n\nfrom dataclasses import dataclass, field\nimport itertools\nfrom typing import cast, Optional\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport flax.nnx as nnx\nfrom torch.utils.data import DataLoader\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader_torch import get_dataloader\nfrom utils.train_utils import (\n get_lr_schedule,\n count_parameters_by_component,\n print_mem_stats,\n print_compiled_memory_stats,\n print_compiled_cost_analysis,\n)\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n # Tokenizer\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_type: str = ""maskgit"" # supported options: maskgit, causal\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n z_loss_weight: float = 0.0\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n use_gt_actions: bool = False\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n eval_full_frame: bool = False\n val_maskgit_steps: int = 25\n val_temperature: float = 1\n val_sample_argmax: bool = False\n wandb_id: str = """"\n\n\ndef build_model(args: Args, rng: jax.Array) -> tuple[Genie, jax.Array]:\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_actions=args.num_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n use_gt_actions=args.use_gt_actions,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n decode=False,\n rngs=rngs,\n )\n if args.use_gt_actions:\n assert (\n not args.lam_checkpoint\n ), ""Cannot use LAM when using ground-truth actions.""\n else:\n assert genie.lam is not None\n del genie.lam.decoder\n return genie, rng\n\n\ndef build_optimizer(genie: Genie, args: Args) -> nnx.ModelAndOptimizer:\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.param_dtype, # moments in full precision\n )\n optimizer = nnx.ModelAndOptimizer(genie, tx)\n return optimizer\n\n\ndef build_mesh_and_sharding(\n num_devices: int,\n) -> tuple[Mesh, NamedSharding, NamedSharding, NamedSharding]:\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n actions_sharding = NamedSharding(mesh, PartitionSpec(""data"", None))\n return mesh, replicated_sharding, videos_sharding, actions_sharding\n\n\ndef shard_optimizer_states(\n optimizer: nnx.ModelAndOptimizer, replicated_sharding: NamedSharding\n) -> None:\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n\ndef build_dataloader(args: Args, data_dir: str) -> DataLoader:\n return get_dataloader(data_dir, args.seq_len, args.batch_size)\n\n\ndef build_checkpoint_manager(args: Args) -> Optional[ocp.CheckpointManager]:\n if args.restore_ckpt or args.save_ckpt:\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n return checkpoint_manager\n else:\n return None\n\n\ndef restore_or_initialize_components(\n args: Args,\n checkpoint_manager: Optional[ocp.CheckpointManager],\n optimizer: nnx.ModelAndOptimizer,\n rng: jax.Array,\n replicated_sharding: NamedSharding,\n restore_step: Optional[int] = None,\n) -> tuple[\n int,\n nnx.ModelAndOptimizer,\n jax.Array,\n]:\n step = 0\n if checkpoint_manager and restore_step is None:\n restore_step = checkpoint_manager.latest_step()\n if args.restore_ckpt:\n assert checkpoint_manager is not None\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n )\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(), args=restore_args\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n rng, _rng = jax.random.split(rng)\n optimizer = restore_genie_components(optimizer, replicated_sharding, _rng, args)\n return step, optimizer, rng\n\n\ndef _calculate_top_k_accuracy(\n token_logits_BTNV: jax.Array,\n video_tokens_BTN: jax.Array,\n mask_BTN: jax.Array,\n k: int,\n) -> jax.Array:\n _, topk_indices_BTNK = jax.lax.top_k(token_logits_BTNV, k)\n topk_correct = jnp.any(\n topk_indices_BTNK == video_tokens_BTN[..., jnp.newaxis], axis=-1\n )\n topk_acc = (mask_BTN * topk_correct).sum() / mask_BTN.sum()\n return topk_acc\n\n\ndef _calculate_step_metrics(\n outputs: dict[str, jax.Array],\n gt: jax.Array,\n num_actions: int,\n num_patch_latents: int,\n) -> tuple[jax.Array, dict]:\n mask_BTN = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask_BTN * ce_loss).sum() / mask_BTN.sum()\n z_val = jax.nn.logsumexp(outputs[""token_logits""], axis=-1)\n z_loss_metric = (mask_BTN * (z_val**2)).sum() / mask_BTN.sum()\n\n masked_token_top_1_acc = _calculate_top_k_accuracy(\n outputs[""token_logits""], outputs[""video_tokens""], mask_BTN, 1\n )\n masked_token_top_2_acc = _calculate_top_k_accuracy(\n outputs[""token_logits""], outputs[""video_tokens""], mask_BTN, 2\n )\n masked_token_top_5_acc = _calculate_top_k_accuracy(\n outputs[""token_logits""], outputs[""video_tokens""], mask_BTN, 5\n )\n masked_token_top_16_acc = _calculate_top_k_accuracy(\n outputs[""token_logits""], outputs[""video_tokens""], mask_BTN, 16\n )\n\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n gt_val = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt_val, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt_val, recon)).mean()\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]),\n size=num_patch_latents,\n fill_value=0,\n )\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_top1_accuracy=masked_token_top_1_acc,\n masked_token_top2_accuracy=masked_token_top_2_acc,\n masked_token_top5_accuracy=masked_token_top_5_acc,\n masked_token_top16_accuracy=masked_token_top_16_acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n z_loss=z_loss_metric,\n psnr=psnr,\n ssim=ssim,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n if ""lam_indices"" in outputs.keys():\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]),\n size=num_actions,\n fill_value=0,\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n metrics[""codebook_usage_lam""] = codebook_usage_lam\n return ce_loss, metrics\n\n\ndef main(args: Args) -> None:\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n genie, rng = build_model(args, rng)\n _, params, _ = nnx.split(genie, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n optimizer = build_optimizer(genie, args)\n del genie\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n _, replicated_sharding, videos_sharding, actions_sharding = build_mesh_and_sharding(\n num_devices\n )\n\n shard_optimizer_states(optimizer, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n checkpoint_manager = build_checkpoint_manager(args)\n\n # --- Create DataLoaderIterator from dataloader ---\n print(""============\n\n"", args.data_dir, ""\n\n============"")\n train_iterator = build_dataloader(args, args.data_dir)\n val_iterator = None\n if args.val_data_dir:\n val_iterator = build_dataloader(args, args.val_data_dir)\n\n # --- Restore checkpoint ---\n step, optimizer, rng = restore_or_initialize_components(\n args,\n checkpoint_manager,\n optimizer,\n rng,\n replicated_sharding,\n )\n\n # --- Define loss and train step (close over args) ---\n def dynamics_loss_fn(\n model: Genie,\n inputs: dict,\n ) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs)\n ce_loss, metrics = _calculate_step_metrics(\n outputs, gt, args.num_actions, args.num_patch_latents\n )\n z_loss = metrics[""z_loss""]\n total_loss = ce_loss + args.z_loss_weight * z_loss\n metrics[""total_loss""] = total_loss\n return total_loss, (outputs[""recon""], metrics)\n\n @nnx.jit(donate_argnums=0)\n def train_step(\n optimizer: nnx.ModelAndOptimizer, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(model: Genie) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n model.train()\n return dynamics_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n optimizer.model\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return loss, recon, metrics\n\n @nnx.jit\n def val_step(genie: Genie, inputs: dict) -> dict:\n """"""Evaluate model and compute metrics""""""\n genie.eval()\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n (loss, (recon, metrics)) = dynamics_loss_fn(genie, inputs)\n val_output = {""loss"": loss, ""recon"": recon, ""metrics"": metrics}\n\n # --- Evaluate full frame prediction (sampling) ---\n if args.eval_full_frame:\n inputs[""videos""] = gt.astype(args.dtype)\n tokenizer_outputs = genie.tokenizer.vq_encode(\n inputs[""videos""], training=False\n )\n tokens_full_frame = tokenizer_outputs[""indices""]\n lam_indices_E = None\n if not args.use_gt_actions:\n lam_indices_E = genie.vq_encode(inputs, training=False)\n inputs[""latent_actions""] = lam_indices_E\n inputs[""videos""] = inputs[""videos""][\n :, :-1\n ] # remove last frame for generation\n recon_full_frame, logits_full_frame = genie.sample(\n inputs,\n args.seq_len,\n args.val_temperature,\n args.val_sample_argmax,\n args.val_maskgit_steps,\n )\n # Calculate metrics for the last frame only\n step_outputs = {\n ""recon"": recon_full_frame[:, -1],\n ""token_logits"": logits_full_frame[:, -1],\n ""video_tokens"": tokens_full_frame[:, -1],\n ""mask"": jnp.ones_like(tokens_full_frame[:, -1]),\n }\n if lam_indices_E is not None:\n lam_indices_B = lam_indices_E.reshape((-1, args.seq_len - 1))[:, -1]\n step_outputs[""lam_indices""] = lam_indices_B\n\n loss_full_frame, metrics_full_frame = _calculate_step_metrics(\n step_outputs, gt[:, -1], args.num_actions, args.num_patch_latents\n )\n val_output.update(\n {\n ""loss_full_frame"": loss_full_frame,\n ""recon_full_frame"": recon_full_frame,\n ""metrics_full_frame"": metrics_full_frame,\n }\n )\n return val_output\n\n def calculate_validation_metrics(val_dataloader, genie, rng):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n loss_full_frame_per_step = []\n metrics_full_frame_per_step = []\n batch = None\n recon = None\n recon_full_frame = None\n for batch in val_dataloader:\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n val_outputs = val_step(genie, batch)\n loss_per_step.append(val_outputs[""loss""])\n metrics_per_step.append(val_outputs[""metrics""])\n recon = val_outputs[""recon""]\n if args.eval_full_frame:\n loss_full_frame_per_step.append(val_outputs[""loss_full_frame""])\n metrics_full_frame_per_step.append(val_outputs[""metrics_full_frame""])\n recon_full_frame = val_outputs[""recon_full_frame""]\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(\n f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}""\n )\n\n val_metrics = {\n f""val_{key}"": np.mean([float(m[key]) for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n val_metrics[""val_loss""] = np.mean(loss_per_step)\n if args.eval_full_frame:\n val_metrics_full_frame = {\n f""val_full_frame_{key}"": np.mean(\n [float(m[key]) for m in metrics_full_frame_per_step]\n )\n for key in metrics_full_frame_per_step[0].keys()\n }\n val_metrics.update(val_metrics_full_frame)\n val_metrics[""val_full_frame_loss""] = np.mean(loss_full_frame_per_step)\n return val_metrics, batch, recon, recon_full_frame\n\n # --- TRAIN LOOP ---\n dataloader_train = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in train_iterator\n )\n dataloader_val = None\n if val_iterator:\n dataloader_val = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in val_iterator\n )\n if jax.process_index() == 0:\n first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader_train = itertools.chain([first_batch], dataloader_train)\n print(f""Starting training from step {step}..."")\n first_step = step\n while step < args.num_steps:\n for batch in dataloader_train:\n # --- Train step ---\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n loss, recon, metrics = train_step(optimizer, batch)\n if step == first_step:\n print_mem_stats(""After params initialized"")\n step += 1\n\n # --- Validation loss ---\n val_results = {}\n if dataloader_val and step % args.val_interval == 0:\n rng, _rng_mask_val = jax.random.split(rng, 2)\n print(""Calculating validation metrics..."")\n val_metrics, val_gt_batch, val_recon, val_recon_full_frame = (\n calculate_validation_metrics(\n dataloader_val, optimizer.model, _rng_mask_val\n )\n )\n print(f""Step {step}, validation loss: {val_metrics['val_loss']}"")\n val_results = {\n ""metrics"": val_metrics,\n ""gt_batch"": val_gt_batch,\n ""recon"": val_recon,\n ""full_frame"": val_recon_full_frame,\n }\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n log_dict = {""loss"": loss, ""step"": step, **metrics}\n if val_results:\n log_dict.update(val_results[""metrics""])\n wandb.log(log_dict)\n if step % args.log_image_interval == 0:\n gt_seq = batch[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if val_results:\n val_results[""gt_seq_val""] = (\n val_results[""gt_batch""][""videos""][0].astype(jnp.float32)\n / 255.0\n )\n val_results[""recon_seq_val""] = val_results[""recon""][0].clip(\n 0, 1\n )\n val_comparison_seq = jnp.concatenate(\n (val_results[""gt_seq_val""], val_results[""recon_seq_val""]),\n axis=1,\n )\n val_results[""val_comparison_seq""] = einops.rearrange(\n val_comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if args.eval_full_frame:\n val_results[""full_frame_seq_val""] = val_results[\n ""full_frame""\n ][0].clip(0, 1)\n val_results[""val_full_frame_comparison_seq""] = (\n jnp.concatenate(\n (\n val_results[""gt_seq_val""],\n val_results[""full_frame_seq_val""],\n ),\n axis=1,\n )\n )\n val_results[""val_full_frame_comparison_seq""] = (\n einops.rearrange(\n val_results[""val_full_frame_comparison_seq""] * 255,\n ""t h w c -> h (t w) c"",\n )\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n if val_results:\n log_images.update(\n dict(\n val_image=wandb.Image(\n np.asarray(\n val_results[""gt_seq_val""][args.seq_len - 1]\n )\n ),\n val_recon=wandb.Image(\n np.asarray(\n val_results[""recon_seq_val""][\n args.seq_len - 1\n ]\n )\n ),\n val_true_vs_recon=wandb.Image(\n np.asarray(\n val_results[""val_comparison_seq""].astype(\n np.uint8\n )\n )\n ),\n )\n )\n if args.eval_full_frame:\n log_images.update(\n dict(\n val_full_frame=wandb.Image(\n np.asarray(\n val_results[""full_frame_seq_val""][\n args.seq_len - 1\n ]\n )\n ),\n val_true_vs_full_frame=wandb.Image(\n np.asarray(\n val_results[\n ""val_full_frame_comparison_seq""\n ].astype(np.uint8)\n )\n ),\n )\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n assert checkpoint_manager is not None\n optimizer_state = nnx.state(optimizer)\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n )\n checkpoint_manager.save(step, args=ckpt_manager_args)\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n if checkpoint_manager:\n checkpoint_manager.close()\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n main(args)\n",python,tab +23,539750,"jasmine/train_dynamics.py",6179,0,"",python,selection_command +24,540933,"jasmine/train_dynamics.py",6220,0,"",python,selection_command +25,541349,"jasmine/train_dynamics.py",6247,0,"",python,selection_command +26,541559,"jasmine/train_dynamics.py",13394,0,"",python,selection_command +27,541730,"jasmine/train_dynamics.py",13418,0,"",python,selection_command +28,543722,"jasmine/train_dynamics.py",13474,0,"",python,selection_command +29,544632,"jasmine/train_dynamics.py",13418,0,"",python,selection_command +30,544838,"jasmine/train_dynamics.py",13474,0,"",python,selection_command +31,587544,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=2-00:00:00\n#SBATCH --cpus-per-task=8\n#SBATCH --gres=gpu:1\n#SBATCH --partition=accelerated\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics/%x_%j.log\n#SBATCH --job-name=train_dynamics_coinrun_og_reproduction\n\n# Log the sbatch script\ncat $0\n\nsource .venv/bin/activate\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\ntags=""coinrun_og dynanmics 10m_dataset repro_mihir""\n\nnpy_records_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m""\n\n# TODO mihir: update the tokenizer and lam checkpoints\ntokenizer_ckpt_dir=""""\nlam_ckpt_dir=""""\n\nCHECKPOINT_DIR=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/dynamics/${job_name}/${slurm_job_id}""\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python jasmine/train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --tokenizer_checkpoint=""${tokenizer_ckpt_dir}"" \\n --lam_checkpoint=""${lam_ckpt_dir}"" \\n --log_image_interval=1000 \\n --log \\n --name=""${job_name}_${slurm_job_id}"" \\n --tags ${tags} \\n --entity instant-uv \\n --project jafar \\n --data_dir $npy_records_dir \\n --wandb_id $slurm_job_id\n",shellscript,tab +32,601514,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",760,0,"",shellscript,selection_mouse +33,601684,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",741,20,"coinrun_episodes_10m",shellscript,selection_mouse +34,601856,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",741,22,"coinrun_episodes_10m""\n",shellscript,selection_mouse +35,602431,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",683,78,"workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m",shellscript,selection_mouse +36,602513,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",682,79,"/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m",shellscript,selection_mouse +37,602540,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",678,83,"work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m",shellscript,selection_mouse +38,602703,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",677,84,"/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m",shellscript,selection_mouse +39,602731,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",673,88,"hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m",shellscript,selection_mouse +40,603481,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",672,89,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m",shellscript,selection_mouse +41,604408,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",668,0,"",shellscript,selection_mouse +42,604559,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",655,15,"npy_records_dir",shellscript,selection_mouse +43,604787,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",655,108,"npy_records_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m""\n",shellscript,selection_mouse +44,609780,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=00:20:00\n#SBATCH --partition=dev_accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\n#SBATCH --job-name=train_dyn_single_gpu\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_test\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_split/val\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955\n\nenv | grep SLURM\n\nexport PYTHONUNBUFFERED=1\n\nsrun python jasmine/train_dynamics.py \\n --save_ckpt \\n --image_height=64 \\n --image_width=64 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=110 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=50 \\n --log_checkpoint_interval=2 \\n --dyna_type=maskgit \\n --log \\n --name=coinrun-dyn-dev-$slurm_job_id \\n --tags dyn coinrun dev \\n --entity instant-uv \\n --project jafar \\n --warmup_steps 0 \\n --wsd_decay_steps 0 \\n --num_steps 10 \\n --data_dir $array_records_dir_train \\n --tokenizer_checkpoint $tokenizer_checkpoint \\n --val_data_dir $array_records_dir_val \\n --val_interval 2 \\n --eval_full_frame \\n --val_steps 5\n",shellscript,tab +45,617289,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1114,0,"",shellscript,selection_mouse +46,617416,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1092,26,"coinrun_episodes_10m_split",shellscript,selection_mouse +47,619609,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,0,"",shellscript,selection_mouse +48,619728,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,2,"/h",shellscript,selection_mouse +49,619796,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,3,"/hk",shellscript,selection_mouse +50,619827,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,4,"/hkf",shellscript,selection_mouse +51,619864,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,5,"/hkfs",shellscript,selection_mouse +52,619894,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,6,"/hkfs/",shellscript,selection_mouse +53,619925,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,7,"/hkfs/w",shellscript,selection_mouse +54,619926,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,8,"/hkfs/wo",shellscript,selection_mouse +55,619959,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,9,"/hkfs/wor",shellscript,selection_mouse +56,619995,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,11,"/hkfs/work/",shellscript,selection_mouse +57,619996,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,12,"/hkfs/work/w",shellscript,selection_mouse +58,620024,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,13,"/hkfs/work/wo",shellscript,selection_mouse +59,620069,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",912,111,"k/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +60,620069,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",913,110,"/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +61,620104,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",914,109,"workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +62,620104,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",916,107,"rkspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +63,620144,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",918,105,"space/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +64,620145,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",920,103,"ace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +65,620145,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",923,100,"/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +66,620178,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",925,98,"cratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +67,620179,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",802,221,"h/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +68,620209,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",806,217,"m_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +69,620210,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",809,214,"nd3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +70,620242,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",812,211,"695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +71,620243,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",816,207,"jafa_ws_shared/data_coinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +72,620274,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",819,204,"a_ws_shared/data_coinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +73,620275,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",821,202,"ws_shared/data_coinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +74,620302,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",823,200,"_shared/data_coinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +75,620336,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",826,197,"ared/data_coinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +76,620337,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",829,194,"d/data_coinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +77,620368,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",833,190,"ta_coinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +78,620369,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",837,186,"oinrun/coinrun_episodes_10m_split/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +79,620397,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",966,57,"inrun/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +80,620398,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",968,55,"run/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +81,620429,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",970,53,"n/coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +82,620430,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",972,51,"coinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +83,620463,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",973,50,"oinrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +84,620571,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",974,49,"inrun_episodes_10m_chunked\narray_records_dir_val=",shellscript,selection_mouse +85,620572,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,75,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinru",shellscript,selection_mouse +86,620572,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,77,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_",shellscript,selection_mouse +87,620588,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,82,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episo",shellscript,selection_mouse +88,620602,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,84,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episode",shellscript,selection_mouse +89,620621,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,100,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_split/val\n",shellscript,selection_mouse +90,620980,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,93,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_spl",shellscript,selection_mouse +91,621014,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,94,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_spli",shellscript,selection_mouse +92,621193,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1023,95,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_split",shellscript,selection_mouse +93,623330,"TERMINAL",0,0,"bash",,terminal_focus +94,625081,"TERMINAL",0,0,"cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_split",,terminal_command +95,625656,"TERMINAL",0,0,"ls",,terminal_command +96,625688,"TERMINAL",0,0,"]633;Cmetadata.json test train val\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_split",,terminal_output +97,631060,"TERMINAL",0,0,"cd val/",,terminal_command +98,636527,"TERMINAL",0,0,"ls -l | head",,terminal_command +99,636601,"TERMINAL",0,0,"]633;C",,terminal_output +100,637135,"TERMINAL",0,0,"total 371520\r\n-rw-rw----+ 1 tum_cte0515 hk-project-p0023960 655360 Sep 4 19:48 episode_0.array_record\r\n-rw-rw----+ 1 tum_cte0515 hk-project-p0023960 851968 Sep 4 19:48 episode_100.array_record\r\n-rw-rw----+ 1 tum_cte0515 hk-project-p0023960 983040 Sep 4 19:48 episode_101.array_record\r\n-rw-rw----+ 1 tum_cte0515 hk-project-p0023960 786432 Sep 4 19:48 episode_102.array_record\r\n-rw-rw----+ 1 tum_cte0515 hk-project-p0023960 524288 Sep 4 19:48 episode_103.array_record\r\n-rw-rw----+ 1 tum_cte0515 hk-project-p0023960 983040 Sep 4 19:48 episode_104.array_record\r\n-rw-rw----+ 1 tum_cte0515 hk-project-p0023960 524288 Sep 4 19:48 episode_105.array_record\r\n-rw-rw----+ 1 tum_cte0515 hk-project-p0023960 720896 Sep 4 19:48 episode_106.array_record\r\n-rw-rw----+ 1 tum_cte0515 hk-project-p0023960 720896 Sep 4 19:48 episode_107.array_record\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_split/val",,terminal_output +101,639536,"TERMINAL",0,0,"cd ..",,terminal_command +102,639740,"TERMINAL",0,0,"ls",,terminal_command +103,640417,"TERMINAL",0,0,"cd ..",,terminal_command +104,640720,"TERMINAL",0,0,"ls",,terminal_command +105,640791,"TERMINAL",0,0,"]633;C",,terminal_output +106,640889,"TERMINAL",0,0,"array_records coinrun_episodes_10m_chunked coinrun_episodes_10m_gt_actions_split coinrun_episodes_500m_gt_actions_split\r\ncoinrun_episodes coinrun_episodes_10m_gt_actions coinrun_episodes_10m_gt_actions_split_test coinrun_episodes_test\r\ncoinrun_episodes_10m coinrun_episodes_10m_gt_actions_distinct_seed coinrun_episodes_10m_split\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun",,terminal_output +107,643515,"TERMINAL",0,0,"cd coinrun_episodes",,terminal_command +108,646310,"TERMINAL",0,0,"ls",,terminal_command +109,646382,"TERMINAL",0,0,"]633;C",,terminal_output +110,647578,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes",,terminal_output +111,651003,"TERMINAL",0,0,"ls -l | head",,terminal_command +112,651067,"TERMINAL",0,0,"]633;C",,terminal_output +113,651449,"TERMINAL",0,0,"total 73932112\r\n-rw-r--r-- 1 tum_cte0515 hk-project-p0023960 3244160 Jun 10 16:09 episode_0.npy\r\n-rw-r--r-- 1 tum_cte0515 hk-project-p0023960 4067456 Jun 10 16:12 episode_1000.npy\r\n-rw-r--r-- 1 tum_cte0515 hk-project-p0023960 8810624 Jun 10 16:12 episode_1001.npy\r\n-rw-r--r-- 1 tum_cte0515 hk-project-p0023960 3244160 Jun 10 16:12 episode_1002.npy\r\n-rw-r--r-- 1 tum_cte0515 hk-project-p0023960 6094976 Jun 10 16:12 episode_1003.npy\r\n-rw-r--r-- 1 tum_cte0515 hk-project-p0023960 8736896 Jun 10 16:12 episode_1004.npy\r\n-rw-r--r-- 1 tum_cte0515 hk-project-p0023960 1953920 Jun 10 16:12 episode_1005.npy\r\n-rw-r--r-- 1 tum_cte0515 hk-project-p0023960 9990272 Jun 10 16:12 episode_1006.npy\r\n-rw-r--r-- 1 tum_cte0515 hk-project-p0023960 5456000 Jun 10 16:12 episode_1007.npy\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes",,terminal_output +114,656744,"TERMINAL",0,0,"pwd",,terminal_command +115,659880,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",876,0,"",shellscript,selection_mouse +116,659884,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",875,0,"",shellscript,selection_command +117,659920,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",875,1,"n",shellscript,selection_mouse +118,659956,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",876,0,"",shellscript,selection_command +119,660989,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",871,5,"",shellscript,content +120,661346,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",870,1,"",shellscript,content +121,661447,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",844,26,"",shellscript,content +122,661582,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",843,1,"",shellscript,content +123,661746,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",831,12,"",shellscript,content +124,661901,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",830,1,"",shellscript,content +125,662038,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",816,14,"",shellscript,content +126,662207,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",815,1,"",shellscript,content +127,662335,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",804,11,"",shellscript,content +128,662476,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",803,1,"",shellscript,content +129,662634,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",796,7,"",shellscript,content +130,662779,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",795,1,"",shellscript,content +131,663054,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",786,9,"",shellscript,content +132,663327,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",785,1,"",shellscript,content +133,663637,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",781,4,"",shellscript,content +134,663911,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",780,1,"",shellscript,content +135,664170,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",776,4,"",shellscript,content +136,664620,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",775,1,"",shellscript,content +137,664933,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",775,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes",shellscript,content +138,665572,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",970,0,"",shellscript,selection_command +139,665735,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1094,0,"",shellscript,selection_command +140,666973,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",985,0,"",shellscript,selection_command +141,668971,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",986,0,"",shellscript,selection_command +142,669476,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",987,0,"",shellscript,selection_command +143,669490,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",988,0,"",shellscript,selection_command +144,669525,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",989,0,"",shellscript,selection_command +145,669558,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",990,0,"",shellscript,selection_command +146,669591,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",991,0,"",shellscript,selection_command +147,669622,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",992,0,"",shellscript,selection_command +148,669655,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",993,0,"",shellscript,selection_command +149,669690,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",994,0,"",shellscript,selection_command +150,669723,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",995,0,"",shellscript,selection_command +151,669755,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",996,0,"",shellscript,selection_command +152,669788,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",997,0,"",shellscript,selection_command +153,669823,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",998,0,"",shellscript,selection_command +154,669854,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",999,0,"",shellscript,selection_command +155,669857,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1000,0,"",shellscript,selection_command +156,669933,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1001,0,"",shellscript,selection_command +157,669933,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1002,0,"",shellscript,selection_command +158,669947,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1003,0,"",shellscript,selection_command +159,670012,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1004,0,"",shellscript,selection_command +160,670013,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1005,0,"",shellscript,selection_command +161,670299,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1006,0,"",shellscript,selection_command +162,670452,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1007,0,"",shellscript,selection_command +163,671452,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1007,99,"",shellscript,content +164,671454,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1006,0,"",shellscript,selection_command +165,671849,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1007,0,"",shellscript,selection_command +166,672216,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1007,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes",shellscript,content +167,672577,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1091,0,"",shellscript,selection_command +168,674562,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",940,0,"",shellscript,selection_mouse +169,675616,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",861,124,"",shellscript,content +170,675705,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",751,0,"",shellscript,selection_command +171,675849,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",634,0,"",shellscript,selection_command +172,676171,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",634,117,"",shellscript,content +173,676267,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",744,0,"",shellscript,selection_command +174,676458,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",852,0,"",shellscript,selection_command +175,676635,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",853,0,"",shellscript,selection_command +176,677134,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",878,0,"",shellscript,selection_command +177,677156,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",905,0,"",shellscript,selection_command +178,677203,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",906,0,"",shellscript,selection_command +179,677262,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1026,0,"",shellscript,selection_command +180,677294,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1051,0,"",shellscript,selection_command +181,677323,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1052,0,"",shellscript,selection_command +182,677353,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1192,0,"",shellscript,selection_command +183,677381,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1193,0,"",shellscript,selection_command +184,677416,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1210,0,"",shellscript,selection_command +185,677447,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1211,0,"",shellscript,selection_command +186,677474,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1237,0,"",shellscript,selection_command +187,677501,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1238,0,"",shellscript,selection_command +188,677528,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1278,0,"",shellscript,selection_command +189,677585,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1296,0,"",shellscript,selection_command +190,677585,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1320,0,"",shellscript,selection_command +191,677613,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1343,0,"",shellscript,selection_command +192,677640,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1376,0,"",shellscript,selection_command +193,677667,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1399,0,"",shellscript,selection_command +194,677697,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1417,0,"",shellscript,selection_command +195,677755,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1437,0,"",shellscript,selection_command +196,677784,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1467,0,"",shellscript,selection_command +197,677813,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1501,0,"",shellscript,selection_command +198,677840,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1527,0,"",shellscript,selection_command +199,677867,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1539,0,"",shellscript,selection_command +200,677897,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1582,0,"",shellscript,selection_command +201,677925,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1611,0,"",shellscript,selection_command +202,677955,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1637,0,"",shellscript,selection_command +203,678013,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1659,0,"",shellscript,selection_command +204,678014,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1682,0,"",shellscript,selection_command +205,678048,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1708,0,"",shellscript,selection_command +206,678075,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1729,0,"",shellscript,selection_command +207,678106,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1771,0,"",shellscript,selection_command +208,678135,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1822,0,"",shellscript,selection_command +209,678162,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1866,0,"",shellscript,selection_command +210,678211,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1889,0,"",shellscript,selection_command +211,678231,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1913,0,"",shellscript,selection_command +212,679506,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1889,0,"",shellscript,selection_command +213,680004,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1866,0,"",shellscript,selection_command +214,680031,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1822,0,"",shellscript,selection_command +215,680065,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1771,0,"",shellscript,selection_command +216,680097,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1729,0,"",shellscript,selection_command +217,680141,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1708,0,"",shellscript,selection_command +218,680171,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1682,0,"",shellscript,selection_command +219,680202,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1659,0,"",shellscript,selection_command +220,680234,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1637,0,"",shellscript,selection_command +221,680264,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1611,0,"",shellscript,selection_command +222,680300,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1582,0,"",shellscript,selection_command +223,680301,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1539,0,"",shellscript,selection_command +224,680355,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1527,0,"",shellscript,selection_command +225,680356,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1501,0,"",shellscript,selection_command +226,680414,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1467,0,"",shellscript,selection_command +227,680573,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1437,0,"",shellscript,selection_command +228,680842,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1467,0,"",shellscript,selection_command +229,680999,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1501,0,"",shellscript,selection_command +230,681150,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1527,0,"",shellscript,selection_command +231,685583,"TERMINAL",0,0,"bash",,terminal_focus +232,688470,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +233,688523,"TERMINAL",0,0,"]633;Csalloc: Granted job allocation 3536670\r\n",,terminal_output +234,688645,"TERMINAL",0,0,"salloc: Waiting for resource configuration\r\n",,terminal_output +235,696896,"TERMINAL",0,0,"s",,terminal_output +236,697064,"TERMINAL",0,0,"o",,terminal_output +237,697159,"TERMINAL",0,0,"ur",,terminal_output +238,697323,"TERMINAL",0,0,"c",,terminal_output +239,697480,"TERMINAL",0,0,"e",,terminal_output +240,697574,"TERMINAL",0,0," ",,terminal_output +241,698359,"TERMINAL",0,0,".",,terminal_output +242,698417,"TERMINAL",0,0,"v",,terminal_output +243,698564,"TERMINAL",0,0,"e",,terminal_output +244,698730,"TERMINAL",0,0,"n",,terminal_output +245,699167,"TERMINAL",0,0,"\t",,terminal_output +246,699699,"TERMINAL",0,0,"b",,terminal_output +247,699872,"TERMINAL",0,0,"i",,terminal_output +248,699934,"TERMINAL",0,0,"n",,terminal_output +249,700125,"TERMINAL",0,0,"\t",,terminal_output +250,700443,"TERMINAL",0,0,"a",,terminal_output +251,700602,"TERMINAL",0,0,"c",,terminal_output +252,700763,"TERMINAL",0,0,"t",,terminal_output +253,700822,"TERMINAL",0,0,"i",,terminal_output +254,700980,"TERMINAL",0,0,"\t",,terminal_output +255,701358,"TERMINAL",0,0,"\r\n",,terminal_output +256,715709,"TERMINAL",0,0,"salloc: Nodes hkn0401 are ready for job\r\n",,terminal_output +257,715842,"TERMINAL",0,0,"source .ven\tbin\tacti\t\r\n",,terminal_output +258,716513,"TERMINAL",0,0,"]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h[tum_cte0515@hkn0401 jasmine]$ source .venv/bin/activate\r\n[?2004l\r]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +259,719703,"TERMINAL",0,0,"s",,terminal_output +260,719758,"TERMINAL",0,0,"h",,terminal_output +261,719988,"TERMINAL",0,0," ",,terminal_output +262,720329,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +263,720838,"TERMINAL",0,0,"\rslurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --val_interval 2 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +264,721026,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +265,721127,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +266,734794,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +267,741412,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +268,741827,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +269,742768,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_121254-u5gwj4e9\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/u5gwj4e9\r\n",,terminal_output +270,743098,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +271,744540,"TERMINAL",0,0,"============\r\n\r\n /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes \r\n\r\n============\r\n",,terminal_output +272,745384,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +273,749809,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 745, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 587, in main\r\n first_batch = next(dataloader_train)\r\n ^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 554, in \r\n dataloader_train = (\r\n ^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/torch/utils/data/dataloader.py"", line 734, in __next__\r\n data = self._next_data()\r\n ^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/torch/utils/data/dataloader.py"", line 790, in _next_data\r\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/torch/utils/data/_utils/fetch.py"", line 52, in fetch\r\n data = [self.dataset[idx] for idx in possibly_batched_index]\r\n ~~~~~~~~~~~~^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py"", line 19, in __getitem__\r\n episode = np.load(self.metadata[idx][""path""])\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/numpy/lib/_npyio_impl.py"", line 454, in load\r\n fid = stack.enter_context(open(os.fspath(file), ""rb""))\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nFileNotFoundError: [Errno 2] No such file or directory: 'data/coinrun_episodes/episode_15.npy'\r\n",,terminal_output +274,750831,"TERMINAL",0,0,"wandb: \r\nwandb: 🚀 View run coinrun-dyn-dev-3536670 at: https://wandb.ai/instant-uv/jafar/runs/u5gwj4e9\r\nwandb: Find logs at: ../../../../../hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_121254-u5gwj4e9/logs\r\n",,terminal_output +275,756158,"TERMINAL",0,0,"W1002 12:13:09.144347 1469662 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugonly job_name: ""jax_worker"": UNAVAILABLE: Cancelling all calls\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_status:14, grpc_message:""Cancelling all calls""}\r\n",,terminal_output +276,756690,"TERMINAL",0,0,"srun: error: hkn0401: task 0: Exited with exit code 1\r\n]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +277,758111,"jasmine/train_dynamics.py",0,0,"",python,tab +278,759246,"jasmine/train_dynamics.py",13414,0,"",python,selection_mouse +279,759661,"jasmine/train_dynamics.py",13469,0,"",python,selection_mouse +280,759829,"jasmine/train_dynamics.py",13468,8,"data_dir",python,selection_mouse +281,760144,"jasmine/train_dynamics.py",13466,0,"",python,selection_mouse +282,760325,"jasmine/train_dynamics.py",13463,4,"args",python,selection_mouse +283,760627,"jasmine/train_dynamics.py",13472,0,"",python,selection_mouse +284,760825,"jasmine/train_dynamics.py",13468,8,"data_dir",python,selection_mouse +285,761166,"jasmine/train_dynamics.py",13551,0,"",python,selection_mouse +286,761295,"jasmine/train_dynamics.py",13547,8,"data_dir",python,selection_mouse +287,761683,"jasmine/train_dynamics.py",13544,0,"",python,selection_mouse +288,761841,"jasmine/train_dynamics.py",13542,4,"args",python,selection_mouse +289,762103,"jasmine/train_dynamics.py",13552,0,"",python,selection_mouse +290,762261,"jasmine/train_dynamics.py",13547,8,"data_dir",python,selection_mouse +291,762831,"jasmine/train_dynamics.py",13526,0,"",python,selection_mouse +292,763835,"jasmine/train_dynamics.py",6173,0,"",python,selection_command +293,765107,"jasmine/train_dynamics.py",6261,0,"",python,selection_mouse +294,765375,"jasmine/train_dynamics.py",6258,8,"data_dir",python,selection_mouse +295,765724,"jasmine/train_dynamics.py",6205,0,"",python,selection_mouse +296,765902,"jasmine/train_dynamics.py",6202,8,"data_dir",python,selection_mouse +297,766186,"jasmine/train_dynamics.py",6270,0,"",python,selection_mouse +298,766321,"jasmine/train_dynamics.py",6268,4,"args",python,selection_mouse +299,766542,"jasmine/train_dynamics.py",6275,0,"",python,selection_mouse +300,766731,"jasmine/train_dynamics.py",6273,7,"seq_len",python,selection_mouse +301,767102,"jasmine/train_dynamics.py",6272,0,"",python,selection_mouse +302,767254,"jasmine/train_dynamics.py",6268,4,"args",python,selection_mouse +303,767629,"jasmine/train_dynamics.py",6206,0,"",python,selection_mouse +304,767780,"jasmine/train_dynamics.py",6202,8,"data_dir",python,selection_mouse +305,768138,"jasmine/train_dynamics.py",6299,0,"",python,selection_mouse +306,768603,"jasmine/train_dynamics.py",6266,0,"",python,selection_mouse +307,768734,"jasmine/train_dynamics.py",6258,8,"data_dir",python,selection_mouse +308,769030,"jasmine/train_dynamics.py",6207,0,"",python,selection_mouse +309,769191,"jasmine/train_dynamics.py",6202,8,"data_dir",python,selection_mouse +310,769539,"jasmine/train_dynamics.py",6266,0,"",python,selection_mouse +311,769895,"jasmine/train_dynamics.py",6265,0,"",python,selection_mouse +312,770079,"jasmine/train_dynamics.py",6264,0,"",python,selection_mouse +313,770562,"jasmine/train_dynamics.py",6263,0,"",python,selection_mouse +314,770710,"jasmine/train_dynamics.py",6258,8,"data_dir",python,selection_mouse +315,772568,"jasmine/train_dynamics.py",6269,0,"",python,selection_mouse +316,772713,"jasmine/train_dynamics.py",6268,4,"args",python,selection_mouse +317,773115,"jasmine/train_dynamics.py",6274,0,"",python,selection_mouse +318,773274,"jasmine/train_dynamics.py",6273,7,"seq_len",python,selection_mouse +319,774237,"jasmine/train_dynamics.py",6252,0,"",python,selection_mouse +320,774606,"jasmine/utils/dataloader_torch.py",0,0,"# file copied from https://raw.githubusercontent.com/FLAIROx/jafar/refs/heads/main/utils/dataloader.py\nfrom pathlib import Path\n\nimport jax.numpy as jnp\nimport numpy as np\nfrom torch.utils.data import Dataset, DataLoader\n\n\nclass VideoDataset(Dataset):\n def __init__(self, data_dir, seq_len):\n self.data_dir = Path(data_dir)\n self.seq_len = seq_len\n self.metadata = np.load(self.data_dir / ""metadata.npy"", allow_pickle=True)\n\n def __len__(self):\n return len(self.metadata)\n\n def __getitem__(self, idx):\n episode = np.load(self.metadata[idx][""path""])\n start_idx = np.random.randint(0, len(episode) - self.seq_len + 1)\n seq = episode[start_idx : start_idx + self.seq_len]\n return seq.astype(np.float32) / 255.0\n\n\ndef collate_fn(batch):\n """"""Convert batch of numpy arrays to JAX array""""""\n return jnp.array(np.stack(batch))\n\n\ndef get_dataloader(data_dir, seq_len, batch_size):\n dataset = VideoDataset(data_dir, seq_len)\n return DataLoader(\n dataset,\n batch_size=batch_size,\n shuffle=True,\n collate_fn=collate_fn,\n )\n",python,tab +321,774608,"jasmine/utils/dataloader_torch.py",895,0,"",python,selection_command +322,775674,"jasmine/utils/dataloader_torch.py",913,0,"",python,selection_mouse +323,775815,"jasmine/utils/dataloader_torch.py",910,8,"data_dir",python,selection_mouse +324,786703,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",0,0,"# file copied from https://raw.githubusercontent.com/FLAIROx/jafar/refs/heads/main/utils/dataloader.py\nfrom pathlib import Path\n\nimport jax.numpy as jnp\nimport numpy as np\nfrom torch.utils.data import Dataset, DataLoader\n\n\nclass VideoDataset(Dataset):\n def __init__(self, data_dir, seq_len):\n self.data_dir = Path(data_dir)\n self.seq_len = seq_len\n self.metadata = np.load(self.data_dir / ""metadata.npy"", allow_pickle=True)\n\n def __len__(self):\n return len(self.metadata)\n\n def __getitem__(self, idx):\n episode = np.load(self.metadata[idx][""path""])\n start_idx = np.random.randint(0, len(episode) - self.seq_len + 1)\n seq = episode[start_idx : start_idx + self.seq_len]\n return seq.astype(np.float32) / 255.0\n\n\ndef collate_fn(batch):\n """"""Convert batch of numpy arrays to JAX array""""""\n return jnp.array(np.stack(batch))\n\n\ndef get_dataloader(data_dir, seq_len, batch_size):\n dataset = VideoDataset(data_dir, seq_len)\n return DataLoader(\n dataset,\n batch_size=batch_size,\n shuffle=True,\n collate_fn=collate_fn,\n )\n",python,tab +325,786704,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",539,0,"",python,selection_command +326,789098,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",577,0,"",python,selection_mouse +327,789228,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",570,8,"metadata",python,selection_mouse +328,793039,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",404,0,"",python,selection_mouse +329,793186,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",402,8,"data_dir",python,selection_mouse +330,793833,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",397,0,"",python,selection_mouse +331,793972,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",397,4,"self",python,selection_mouse +332,794351,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",364,0,"",python,selection_mouse +333,794351,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",363,0,"",python,selection_command +334,794507,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",357,7,"seq_len",python,selection_mouse +335,794520,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",358,6,"eq_len",python,selection_command +336,794827,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",404,0,"",python,selection_mouse +337,794974,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",402,8,"data_dir",python,selection_mouse +338,795676,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",400,0,"",python,selection_mouse +339,795835,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",397,4,"self",python,selection_mouse +340,796122,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",404,0,"",python,selection_mouse +341,796255,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",402,8,"data_dir",python,selection_mouse +342,796988,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",399,0,"",python,selection_mouse +343,797127,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",397,4,"self",python,selection_mouse +344,797476,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",403,0,"",python,selection_mouse +345,797639,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",402,8,"data_dir",python,selection_mouse +346,803838,"TERMINAL",0,0,"bash",,terminal_focus +347,806194,"TERMINAL",0,0,"dev",,terminal_command +348,808502,"TERMINAL",0,0,"cd ..",,terminal_command +349,809392,"TERMINAL",0,0,"ls",,terminal_command +350,809426,"TERMINAL",0,0,"]633;C",,terminal_output +351,809529,"TERMINAL",0,0,"checkpoints jafar jasmine jasmine_jobs jasmine_jobs_2 jasmine_jobs_3 tmp\r\n]0;tum_cte0515@hkn1991:~/Projects",,terminal_output +352,827952,"TERMINAL",0,0,"ls jafar/generate_dataset.py",,terminal_command +353,830184,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",0,0,"""""""\nGenerates a dataset of random-action CoinRun episodes.\nEpisodes are saved individually as memory-mapped files for efficient loading.\n""""""\n\nfrom dataclasses import dataclass\nfrom pathlib import Path\n\nfrom gym3 import types_np\nimport numpy as np\nfrom procgen import ProcgenGym3Env\nimport tyro\n\n\n@dataclass\nclass Args:\n num_episodes: int = 10000\n output_dir: str = ""data/coinrun_episodes""\n min_episode_length: int = 50\n\n\nargs = tyro.cli(Args)\noutput_dir = Path(args.output_dir)\noutput_dir.mkdir(parents=True, exist_ok=True)\n\n# --- Generate episodes ---\ni = 0\nmetadata = []\nwhile i < args.num_episodes:\n seed = np.random.randint(0, 10000)\n env = ProcgenGym3Env(num=1, env_name=""coinrun"", start_level=seed)\n dataseq = []\n\n # --- Run episode ---\n for j in range(1000):\n env.act(types_np.sample(env.ac_space, bshape=(env.num,)))\n rew, obs, first = env.observe()\n dataseq.append(obs[""rgb""])\n if first:\n break\n\n # --- Save episode ---\n if len(dataseq) >= args.min_episode_length:\n episode_data = np.concatenate(dataseq, axis=0)\n episode_path = output_dir / f""episode_{i}.npy""\n np.save(episode_path, episode_data.astype(np.uint8))\n metadata.append({""path"": str(episode_path), ""length"": len(dataseq)})\n print(f""Episode {i} completed, length: {len(dataseq)}"")\n i += 1\n else:\n print(f""Episode too short ({len(dataseq)}), resampling..."")\n\n# --- Save metadata ---\nnp.save(output_dir / ""metadata.npy"", metadata)\nprint(f""Dataset generated with {len(metadata)} valid episodes"")\n",python,tab +354,830930,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",227,0,"",python,selection_mouse +355,830963,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",226,0,"",python,selection_command +356,847996,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/dataloader_torch.py",0,0,"",python,tab +357,849573,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",0,0,"",python,tab +358,850781,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +359,853508,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,0,"",shellscript,selection_mouse +360,853624,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,2,"/h",shellscript,selection_mouse +361,853636,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,4,"/hkf",shellscript,selection_mouse +362,853650,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,6,"/hkfs/",shellscript,selection_mouse +363,853677,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,8,"/hkfs/wo",shellscript,selection_mouse +364,853706,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,10,"/hkfs/work",shellscript,selection_mouse +365,853738,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,15,"/hkfs/work/work",shellscript,selection_mouse +366,853739,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,17,"/hkfs/work/worksp",shellscript,selection_mouse +367,853770,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,21,"/hkfs/work/workspace/",shellscript,selection_mouse +368,853803,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,129,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR",shellscript,selection_mouse +369,853835,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,32,"/hkfs/work/workspace/scratch/tum",shellscript,selection_mouse +370,853865,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,37,"/hkfs/work/workspace/scratch/tum_ind3",shellscript,selection_mouse +371,853866,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,39,"/hkfs/work/workspace/scratch/tum_ind369",shellscript,selection_mouse +372,853897,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,42,"/hkfs/work/workspace/scratch/tum_ind3695-j",shellscript,selection_mouse +373,853898,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,45,"/hkfs/work/workspace/scratch/tum_ind3695-jafa",shellscript,selection_mouse +374,853929,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,48,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws",shellscript,selection_mouse +375,853930,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,53,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shar",shellscript,selection_mouse +376,853964,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",905,16,"\nCHECKPOINT_DIR=",shellscript,selection_mouse +377,854035,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",904,17,"\n\nCHECKPOINT_DIR=",shellscript,selection_mouse +378,854038,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",877,44,"\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=",shellscript,selection_mouse +379,854272,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",904,17,"\n\nCHECKPOINT_DIR=",shellscript,selection_mouse +380,854360,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",905,16,"\nCHECKPOINT_DIR=",shellscript,selection_mouse +381,854620,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,96,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slur",shellscript,selection_mouse +382,854634,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,95,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slu",shellscript,selection_mouse +383,854659,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,94,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$sl",shellscript,selection_mouse +384,854667,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,93,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$s",shellscript,selection_mouse +385,854692,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,91,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/",shellscript,selection_mouse +386,854701,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,90,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name",shellscript,selection_mouse +387,854724,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,89,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_nam",shellscript,selection_mouse +388,854733,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,87,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_n",shellscript,selection_mouse +389,854755,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",921,86,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_",shellscript,selection_mouse +390,855175,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1138,0,"",shellscript,selection_mouse +391,938594,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1209,0,"",shellscript,selection_mouse +392,938596,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1208,0,"",shellscript,selection_command +393,938699,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1208,1,"M",shellscript,selection_mouse +394,938700,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1209,0,"",shellscript,selection_command +395,939134,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1192,0,"",shellscript,selection_mouse +396,939472,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1210,0,"",shellscript,selection_mouse +397,940105,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1237,0,"",shellscript,selection_mouse +398,940589,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1236,0,"",shellscript,selection_mouse +399,940596,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1235,0,"",shellscript,selection_command +400,940628,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1235,1,"1",shellscript,selection_mouse +401,940634,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1236,0,"",shellscript,selection_command +402,941248,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1436,0,"",shellscript,selection_mouse +403,941249,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1435,0,"",shellscript,selection_command +404,942889,"jasmine/utils/dataloader_torch.py",0,0,"",python,tab +405,950010,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +406,953575,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,0,"",shellscript,selection_mouse +407,953696,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,2,"/h",shellscript,selection_mouse +408,953708,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,8,"/hkfs/wo",shellscript,selection_mouse +409,953723,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,12,"/hkfs/work/w",shellscript,selection_mouse +410,953743,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,17,"/hkfs/work/worksp",shellscript,selection_mouse +411,953774,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,25,"/hkfs/work/workspace/scra",shellscript,selection_mouse +412,953784,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,30,"/hkfs/work/workspace/scratch/t",shellscript,selection_mouse +413,953814,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,34,"/hkfs/work/workspace/scratch/tum_i",shellscript,selection_mouse +414,953821,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,38,"/hkfs/work/workspace/scratch/tum_ind36",shellscript,selection_mouse +415,953856,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,41,"/hkfs/work/workspace/scratch/tum_ind3695-",shellscript,selection_mouse +416,953856,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,44,"/hkfs/work/workspace/scratch/tum_ind3695-jaf",shellscript,selection_mouse +417,953891,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,49,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_",shellscript,selection_mouse +418,953892,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",633,25,"\narray_records_dir_train=",shellscript,selection_mouse +419,953965,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",632,26,"\n\narray_records_dir_train=",shellscript,selection_mouse +420,954203,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",633,25,"\narray_records_dir_train=",shellscript,selection_mouse +421,954232,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,67,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinru",shellscript,selection_mouse +422,954271,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,66,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinr",shellscript,selection_mouse +423,954305,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,67,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinru",shellscript,selection_mouse +424,954372,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,68,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun",shellscript,selection_mouse +425,954419,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,69,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/",shellscript,selection_mouse +426,954421,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,71,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/co",shellscript,selection_mouse +427,954421,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,73,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coin",shellscript,selection_mouse +428,954461,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,76,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun",shellscript,selection_mouse +429,954462,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,77,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_",shellscript,selection_mouse +430,954462,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,79,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_ep",shellscript,selection_mouse +431,954480,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,80,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_epi",shellscript,selection_mouse +432,954584,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,81,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_epis",shellscript,selection_mouse +433,954720,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,82,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episo",shellscript,selection_mouse +434,954800,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,83,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episod",shellscript,selection_mouse +435,954899,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,84,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episode",shellscript,selection_mouse +436,955019,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,85,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes",shellscript,selection_mouse +437,955686,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",846,0,"",shellscript,selection_command +438,956603,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",846,0,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",shellscript,content +439,956604,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",901,0,"",shellscript,selection_keyboard +440,957893,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",846,55,"",shellscript,content +441,959616,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,0,"",shellscript,selection_mouse +442,959719,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,1,"=",shellscript,selection_mouse +443,959757,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,3,"=/h",shellscript,selection_mouse +444,959797,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,5,"=/hkf",shellscript,selection_mouse +445,959798,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,6,"=/hkfs",shellscript,selection_mouse +446,959834,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,8,"=/hkfs/w",shellscript,selection_mouse +447,959835,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,12,"=/hkfs/work/",shellscript,selection_mouse +448,959875,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,14,"=/hkfs/work/wo",shellscript,selection_mouse +449,959876,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,18,"=/hkfs/work/worksp",shellscript,selection_mouse +450,959877,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,22,"=/hkfs/work/workspace/",shellscript,selection_mouse +451,959908,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",633,24,"\narray_records_dir_train",shellscript,selection_mouse +452,959992,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",632,25,"\n\narray_records_dir_train",shellscript,selection_mouse +453,960028,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",633,24,"\narray_records_dir_train",shellscript,selection_mouse +454,960163,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,82,"=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_epis",shellscript,selection_mouse +455,960175,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,83,"=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episo",shellscript,selection_mouse +456,960212,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,84,"=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episod",shellscript,selection_mouse +457,960308,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,85,"=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episode",shellscript,selection_mouse +458,960390,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,86,"=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes",shellscript,selection_mouse +459,966169,"TERMINAL",0,0,"cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes",,terminal_command +460,967542,"TERMINAL",0,0,"ls",,terminal_command +461,967592,"TERMINAL",0,0,"]633;C",,terminal_output +462,967842,"TERMINAL",0,0,"episode_0.npy episode_2000.npy episode_3001.npy episode_4002.npy episode_5003.npy episode_6004.npy episode_7005.npy episode_8006.npy episode_9007.npy\r\nepisode_1000.npy episode_2001.npy episode_3002.npy episode_4003.npy episode_5004.npy episode_6005.npy episode_7006.npy episode_8007.npy episode_9008.npy\r\nepisode_1001.npy episode_2002.npy episode_3003.npy episode_4004.npy episode_5005.npy episode_6006.npy episode_7007.npy episode_8008.npy episode_9009.npy\r\nepisode_1002.npy episode_2003.npy episode_3004.npy episode_4005.npy episode_5006.npy episode_6007.npy episode_7008.npy episode_8009.npy episode_900.npy\r\nepisode_1003.npy episode_2004.npy episode_3005.npy episode_4006.npy episode_5007.npy episode_6008.npy episode_7009.npy episode_800.npy episode_9010.npy\r\nepisode_1004.npy episode_2005.npy episode_3006.npy episode_4007.npy episode_5008.npy episode_6009.npy episode_700.npy episode_8010.npy episode_9011.npy\r\nepisode_1005.npy episode_2006.npy episode_3007.npy episode_4008.npy episode_5009.npy episode_600.npy episode_7010.npy episode_8011.npy episode_9012.npy\r\nepisode_1006.npy episode_2007.npy episode_3008.npy episode_4009.npy episode_500.npy episode_6010.npy episode_7011.npy episode_8012.npy episode_9013.npy\r\nepisode_1007.npy episode_2008.npy episode_3009.npy episode_400.npy episode_5010.npy episode_6011.npy episode_7012.npy episode_8013.npy episode_9014.npy\r\nepisode_1008.npy episode_2009.npy episode_300.npy episode_4010.npy episode_5011.npy episode_6012.npy episode_7013.npy episode_8014.npy episode_9015.npy\r\nepisode_1009.npy episode_200.npy episode_3010.npy episode_4011.npy episode_5012.npy episode_6013.npy episode_7014.npy episode_8015.npy episode_9016.npy\r\nepisode_100.npy episode_2010.npy episode_3011.npy episode_4012.npy episode_5013.npy episode_6014.npy episode_7015.npy episode_8016.npy episode_9017.npy\r\nepisode_1010.npy episode_2011.npy episode_3012.npy episode_4013.npy episode_5014.npy episode_6015.npy episode_7016.npy episode_8017.npy episode_9018.npy\r\nepisode_1011.npy episode_2012.npy episode_3013.npy episode_4014.npy episode_5015.npy episode_6016.npy episode_7017.npy episode_8018.npy episode_9019.npy\r\nepisode_1012.npy episode_2013.npy episode_3014.npy episode_4015.npy episode_5016.npy episode_6017.npy episode_7018.npy episode_8019.npy episode_901.npy\r\nepisode_1013.npy episode_2014.npy episode_3015.npy episode_4016.npy episode_5017.npy episode_6018.npy episode_7019.npy episode_801.npy episode_9020.npy\r\nepisode_1014.npy episode_2015.npy episode_3016.npy episode_4017.npy episode_5018.npy episode_6019.npy episode_701.npy episode_8020.npy episode_9021.npy\r\nepisode_1015.npy episode_2016.npy episode_3017.npy episode_4018.npy episode_5019.npy episode_601.npy episode_7020.npy episode_8021.npy episode_9022.npy\r\nepisode_1016.npy episode_2017.npy episode_3018.npy episode_4019.npy episode_501.npy episode_6020.npy episode_7021.npy episode_8022.npy episode_9023.npy\r\nepisode_1017.npy episode_2018.npy episode_3019.npy episode_401.npy episode_5020.npy episode_6021.npy episode_7022.npy episode_8023.npy episode_9024.npy\r\nepisode_1018.npy episode_2019.npy episode_301.npy episode_4020.npy episode_5021.npy episode_6022.npy episode_7023.npy episode_8024.npy episode_9025.npy\r\nepisode_1019.npy episode_201.npy episode_3020.npy episode_4021.npy episode_5022.npy episode_6023.npy episode_7024.npy episode_8025.npy episode_9026.npy\r\nepisode_101.npy episode_2020.npy episode_3021.npy episode_4022.npy episode_5023.npy episode_6024.npy episode_7025.npy episode_8026.npy episode_9027.npy\r\nepisode_1020.npy episode_2021.npy episode_3022.npy episode_4023.npy episode_5024.npy episode_6025.npy episode_7026.npy episode_8027.npy episode_9028.npy\r\nepisode_1021.npy episode_2022.npy episode_3023.npy episode_4024.npy episode_5025.npy episode_6026.npy episode_7027.npy episode_8028.npy episode_9029.npy\r\nepisode_1022.npy episode_2023.npy episode_3024.npy episode_4025.npy episode_5026.npy episode_6027.npy episode_7028.npy episode_8029.npy episode_902.npy\r\nepisode_1023.npy episode_2024.npy episode_3025.npy episode_4026.npy episode_5027.npy episode_6028.npy episode_7029.npy episode_802.npy episode_9030.npy\r\nepisode_1024.npy episode_2025.npy episode_3026.npy episode_4027.npy episode_5028.npy episode_6029.npy episode_702.npy episode_8030.npy episode_9031.npy\r\nepisode_1025.npy episode_2026.npy episode_3027.npy episode_4028.npy episode_5029.npy episode_602.npy episode_7030.npy episode_8031.npy episode_9032.npy\r\nepisode_1026.npy episode_2027.npy episode_3028.npy episode_4029.npy episode_502.npy episode_6030.npy episode_7031.npy episode_8032.npy episode_9033.npy\r\nepisode_1027.npy episode_2028.npy episode_3029.npy episode_402.npy episode_5030.npy episode_6031.npy episode_7032.npy episode_8033.npy episode_9034.npy\r\nepisode_1028.npy episode_2029.npy episode_302.npy episode_4030.npy episode_5031.npy episode_6032.npy episode_7033.npy episode_8034.npy episode_9035.npy\r\nepisode_1029.npy episode_202.npy episode_3030.npy episode_4031.npy episode_5032.npy episode_6033.npy episode_7034.npy episode_8035.npy episode_9036.npy\r\nepisode_102.npy episode_2030.npy episode_3031.npy episode_4032.npy episode_5033.npy episode_6034.npy episode_7035.npy episode_8036.npy episode_9037.npy\r\nepisode_1030.npy episode_2031.npy episode_3032.npy episode_4033.npy episode_5034.npy episode_6035.npy episode_7036.npy episode_8037.npy episode_9038.npy\r\nepisode_1031.npy episode_2032.npy episode_3033.npy episode_4034.npy episode_5035.npy episode_6036.npy episode_7037.npy episode_8038.npy episode_9039.npy\r\nepisode_1032.npy episode_2033.npy episode_3034.npy episode_4035.npy episode_5036.npy episode_6037.npy episode_7038.npy episode_8039.npy episode_903.npy\r\nepisode_1033.npy episode_2034.npy episode_3035.npy episode_4036.npy episode_5037.npy episode_6038.npy episode_7039.npy episode_803.npy episode_9040.npy\r\nepisode_1034.npy episode_2035.npy episode_3036.npy episode_4037.npy episode_5038.npy episode_6039.npy episode_703.npy episode_8040.npy episode_9041.npy\r\nepisode_1035.npy episode_2036.npy episode_3037.npy episode_4038.npy episode_5039.npy episode_603.npy episode_7040.npy episode_8041.npy episode_9042.npy\r\nepisode_1036.npy episode_2037.npy episode_3038.npy episode_4039.npy episode_503.npy episode_6040.npy episode_7041.npy episode_8042.npy episode_9043.npy\r\nepisode_1037.npy episode_2038.npy episode_3039.npy episode_403.npy episode_5040.npy episode_6041.npy episode_7042.npy episode_8043.npy episode_9044.npy\r\nepisode_1038.npy episode_2039.npy episode_303.npy episode_4040.npy episode_5041.npy episode_6042.npy episode_7043.npy episode_8044.npy episode_9045.npy\r\nepisode_1039.npy episode_203.npy episode_3040.npy episode_4041.npy episode_5042.npy episode_6043.npy episode_7044.npy episode_8045.npy episode_9046.npy\r\nepisode_103.npy episode_2040.npy episode_3041.npy episode_4042.npy episode_5043.npy episode_6044.npy episode_7045.npy episode_8046.npy episode_9047.npy\r\nepisode_1040.npy episode_2041.npy episode_3042.npy episode_4043.npy episode_5044.npy episode_6045.npy episode_7046.npy episode_8047.npy episode_9048.npy\r\nepisode_1041.npy episode_2042.npy episode_3043.npy episode_4044.npy episode_5045.npy episode_6046.npy episode_7047.npy episode_8048.npy episode_9049.npy\r\nepisode_1042.npy episode_2043.npy episode_3044.npy episode_4045.npy episode_5046.npy episode_6047.npy episode_7048.npy episode_8049.npy episode_904.npy\r\nepisode_1043.npy episode_2044.npy episode_3045.npy episode_4046.npy episode_5047.npy episode_6048.npy episode_7049.npy episode_804.npy episode_9050.npy\r\nepisode_1044.npy episode_2045.npy episode_3046.npy episode_4047.npy episode_5048.npy episode_6049.npy episode_704.npy episode_8050.npy episode_9051.npy\r\nepisode_1045.npy episode_2046.npy episode_3047.npy episode_4048.npy episode_5049.npy episode_604.npy episode_7050.npy episode_8051.npy episode_9052.npy\r\nepisode_1046.npy episode_2047.npy episode_3048.npy episode_4049.npy episode_504.npy episode_6050.npy episode_7051.npy episode_8052.npy episode_9053.npy\r\nepisode_1047.npy episode_2048.npy episode_3049.npy episode_404.npy episode_5050.npy episode_6051.npy episode_7052.npy episode_8053.npy episode_9054.npy\r\nepisode_1048.npy episode_2049.npy episode_304.npy episode_4050.npy episode_5051.npy episode_6052.npy episode_7053.npy episode_8054.npy episode_9055.npy\r\nepisode_1049.npy episode_204.npy episode_3050.npy episode_4051.npy episode_5052.npy episode_6053.npy episode_7054.npy episode_8055.npy episode_9056.npy\r\nepisode_104.npy episode_2050.npy episode_3051.npy episode_4052.npy episode_5053.npy episode_6054.npy episode_7055.npy episode_8056.npy episode_9057.npy\r\nepisode_1050.npy episode_2051.npy episode_3052.npy episode_4053.npy episode_5054.npy episode_6055.npy episode_7056.npy episode_8057.npy episode_9058.npy\r\nepisode_1051.npy episode_2052.npy episode_3053.npy episode_4054.npy episode_5055.npy episode_6056.npy episode_7057.npy episode_8058.npy episode_9059.npy\r\nepisode_1052.npy episode_2053.npy episode_3054.npy episode_4055.npy episode_5056.npy episode_6057.npy episode_7058.npy episode_8059.npy episode_905.npy\r\nepisode_1053.npy episode_2054.npy episode_3055.npy episode_4056.npy episode_5057.npy episode_6058.npy episode_7059.npy episode_805.npy episode_9060.npy\r\nepisode_1054.npy episode_2055.npy episode_3056.npy episode_4057.npy episode_5058.npy episode_6059.npy episode_705.npy episode_8060.npy episode_9061.npy\r\nepisode_1055.npy episode_2056.npy episode_3057.npy episode_4058.npy episode_5059.npy episode_605.npy episode_7060.npy episode_8061.npy episode_9062.npy\r\nepisode_1056.npy episode_2057.npy episode_3058.npy episode_4059.npy episode_505.npy episode_6060.npy episode_7061.npy episode_8062.npy episode_9063.npy\r\nepisode_1057.npy episode_2058.npy episode_3059.npy episode_405.npy episode_5060.npy episode_6061.npy episode_7062.npy episode_8063.npy episode_9064.npy\r\nepisode_1058.npy episode_2059.npy episode_305.npy episode_4060.npy episode_5061.npy episode_6062.npy episode_7063.npy episode_8064.npy episode_9065.npy\r\nepisode_1059.npy episode_205.npy episode_3060.npy episode_4061.npy episode_5062.npy episode_6063.npy episode_7064.npy episode_8065.npy episode_9066.npy\r\nepisode_105.npy episode_2060.npy episode_3061.npy episode_4062.npy episode_5063.npy episode_6064.npy episode_7065.npy episode_8066.npy episode_9067.npy\r\nepisode_1060.npy episode_2061.npy episode_3062.npy episode_4063.npy episode_5064.npy episode_6065.npy episode_7066.npy episode_8067.npy episode_9068.npy\r\nepisode_1061.npy episode_2062.npy episode_3063.npy episode_4064.npy episode_5065.npy episode_6066.npy episode_7067.npy episode_8068.npy episode_9069.npy\r\nepisode_1062.npy episode_2063.npy episode_3064.npy episode_4065.npy episode_5066.npy episode_6067.npy episode_7068.npy episode_8069.npy episode_906.npy\r\nepisode_1063.npy episode_2064.npy episode_3065.npy episode_4066.npy episode_5067.npy episode_6068.npy episode_7069.npy episode_806.npy episode_9070.npy\r\nepisode_1064.npy episode_2065.npy episode_3066.npy episode_4067.npy episode_5068.npy episode_6069.npy episode_706.npy episode_8070.npy episode_9071.npy\r\nepisode_1065.npy episode_2066.npy episode_3067.npy episode_4068.npy episode_5069.npy episode_606.npy episode_7070.npy episode_8071.npy episode_9072.npy\r\nepisode_1066.npy episode_2067.npy episode_3068.npy episode_4069.npy episode_506.npy episode_6070.npy episode_7071.npy episode_8072.npy episode_9073.npy\r\nepisode_1067.npy episode_2068.npy episode_3069.npy episode_406.npy episode_5070.npy episode_6071.npy episode_7072.npy episode_8073.npy episode_9074.npy\r\nepisode_1068.npy episode_2069.npy episode_306.npy episode_4070.npy episode_5071.npy episode_6072.npy episode_7073.npy episode_8074.npy episode_9075.npy\r\nepisode_1069.npy episode_206.npy episode_3070.npy episode_4071.npy episode_5072.npy episode_6073.npy episode_7074.npy episode_8075.npy episode_9076.npy\r\nepisode_106.npy episode_2070.npy episode_3071.npy episode_4072.npy episode_5073.npy episode_6074.npy episode_7075.npy episode_8076.npy episode_9077.npy\r\nepisode_1070.npy episode_2071.npy episode_3072.npy episode_4073.npy episode_5074.npy episode_6075.npy episode_7076.npy episode_8077.npy episode_9078.npy\r\nepisode_1071.npy episode_2072.npy episode_3073.npy episode_4074.npy episode_5075.npy episode_6076.npy episode_7077.npy episode_8078.npy episode_9079.npy\r\nepisode_1072.npy episode_2073.npy episode_3074.npy episode_4075.npy episode_5076.npy episode_6077.npy episode_7078.npy episode_8079.npy episode_907.npy\r\nepisode_1073.npy episode_2074.npy episode_3075.npy episode_4076.npy episode_5077.npy episode_6078.npy episode_7079.npy episode_807.npy episode_9080.npy\r\nepisode_1074.npy episode_2075.npy episode_3076.npy episode_4077.npy episode_5078.npy episode_6079.npy episode_707.npy episode_8080.npy episode_9081.npy\r\nepisode_1075.npy episode_2076.npy episode_3077.npy episode_4078.npy episode_5079.npy episode_607.npy episode_7080.npy episode_8081.npy episode_9082.npy\r\nepisode_1076.npy episode_2077.npy episode_3078.npy episode_4079.npy episode_507.npy episode_6080.npy episode_7081.npy episode_8082.npy episode_9083.npy\r\nepisode_1077.npy episode_2078.npy episode_3079.npy episode_407.npy episode_5080.npy episode_6081.npy episode_7082.npy episode_8083.npy episode_9084.npy\r\nepisode_1078.npy episode_2079.npy episode_307.npy episode_4080.npy episode_5081.npy episode_6082.npy episode_7083.npy episode_8084.npy episode_9085.npy\r\nepisode_1079.npy episode_207.npy episode_3080.npy episode_4081.npy episode_5082.npy episode_6083.npy episode_7084.npy episode_8085.npy episode_9086.npy\r\nepisode_107.npy episode_2080.npy episode_3081.npy episode_4082.npy episode_5083.npy episode_6084.npy episode_7085.npy episode_8086.npy episode_9087.npy\r\nepisode_1080.npy episode_2081.npy episode_3082.npy episode_4083.npy episode_5084.npy episode_6085.npy episode_7086.npy episode_8087.npy episode_9088.npy\r\nepisode_1081.npy episode_2082.npy episode_3083.npy episode_4084.npy episode_5085.npy episode_6086.npy episode_7087.npy episode_8088.npy episode_9089.npy\r\nepisode_1082.npy episode_2083.npy episode_3084.npy episode_4085.npy episode_5086.npy episode_6087.npy episode_7088.npy episode_8089.npy episode_908.npy\r\nepisode_1083.npy episode_2084.npy episode_3085.npy episode_4086.npy episode_5087.npy episode_6088.npy episode_7089.npy episode_808.npy episode_9090.npy\r\nepisode_1084.npy episode_2085.npy episode_3086.npy episode_4087.npy episode_5088.npy episode_6089.npy episode_708.npy episode_8090.npy episode_9091.npy\r\nepisode_1085.npy episode_2086.npy episode_3087.npy episode_4088.npy episode_5089.npy episode_608.npy episode_7090.npy episode_8091.npy episode_9092.npy\r\nepisode_1086.npy episode_2087.npy episode_3088.npy episode_4089.npy episode_508.npy episode_6090.npy episode_7091.npy episode_8092.npy episode_9093.npy\r\nepisode_1087.npy episode_2088.npy episode_3089.npy episode_408.npy episode_5090.npy episode_6091.npy episode_7092.npy episode_8093.npy episode_9094.npy\r\nepisode_1088.npy episode_2089.npy episode_308.npy episode_4090.npy episode_5091.npy episode_6092.npy episode_7093.npy episode_8094.npy episode_9095.npy\r\nepisode_1089.npy episode_208.npy episode_3090.npy episode_4091.npy episode_5092.npy episode_6093.npy episode_7094.npy episode_8095.npy episode_9096.npy\r\nepisode_108.npy episode_2090.npy episode_3091.npy episode_4092.npy episode_5093.npy episode_6094.npy episode_7095.npy episode_8096.npy episode_9097.npy\r\nepisode_1090.npy episode_2091.npy episode_3092.npy episode_4093.npy episode_5094.npy episode_6095.npy episode_7096.npy episode_8097.npy episode_9098.npy\r\nepisode_1091.npy episode_2092.npy episode_3093.npy episode_4094.npy episode_5095.npy episode_6096.npy episode_7097.npy episode_8098.npy episode_9099.npy\r\nepisode_1092.npy episode_2093.npy episode_3094.npy episode_4095.npy episode_5096.npy episode_6097.npy episode_7098.npy episode_8099.npy episode_909.npy\r\nepisode_1093.npy episode_2094.npy episode_3095.npy episode_4096.npy episode_5097.npy episode_6098.npy episode_7099.npy episode_809.npy episode_90.npy\r\nepisode_1094.npy episode_2095.npy episode_3096.npy episode_4097.npy episode_5098.npy episode_6099.npy episode_709.npy episode_80.npy episode_9100.npy\r\nepisode_1095.npy episode_2096.npy episode_3097.npy episode_4098.npy episode_5099.npy episode_609.npy episode_70.npy episode_8100.npy episode_9101.npy\r\nepisode_1096.npy episode_2097.npy episode_3098.npy episode_4099.npy episode_509.npy episode_60.npy episode_7100.npy episode_8101.npy episode_9102.npy\r\nepisode_1097.npy episode_2098.npy episode_3099.npy episode_409.npy episode_50.npy episode_6100.npy episode_7101.npy episode_8102.npy episode_9103.npy\r\nepisode_1098.npy episode_2099.npy episode_309.npy episode_40.npy episode_5100.npy episode_6101.npy episode_7102.npy episode_8103.npy episode_9104.npy\r\nepisode_1099.npy episode_209.npy episode_30.npy episode_4100.npy episode_5101.npy episode_6102.npy episode_7103.npy episode_8104.npy episode_9105.npy\r\nepisode_109.npy episode_20.npy episode_3100.npy episode_4101.npy episode_5102.npy episode_6103.npy episode_7104.npy episode_8105.npy episode_9106.npy\r\nepisode_10.npy episode_2100.npy episode_3101.npy episode_4102.npy episode_5103.npy episode_6104.npy episode_7105.npy episode_8106.npy episode_9107.npy\r\nepisode_1100.npy episode_2101.npy episode_3102.npy episode_4103.npy episode_5104.npy episode_6105.npy episode_7106.npy episode_8107.npy episode_9108.npy\r\nepisode_1101.npy episode_2102.npy episode_3103.npy episode_4104.npy episode_5105.npy episode_6106.npy episode_7107.npy episode_8108.npy episode_9109.npy\r\nepisode_1102.npy episode_2103.npy episode_3104.npy episode_4105.npy episode_5106.npy episode_6107.npy episode_7108.npy episode_8109.npy episode_910.npy\r\nepisode_1103.npy episode_2104.npy episode_3105.npy episode_4106.npy episode_5107.npy episode_6108.npy episode_7109.npy episode_810.npy episode_9110.npy\r\nepisode_1104.npy episode_2105.npy episode_3106.npy episode_4107.npy episode_5108.npy episode_6109.npy episode_710.npy episode_8110.npy episode_9111.npy\r\nepisode_1105.npy episode_2106.npy episode_3107.npy episode_4108.npy episode_5109.npy episode_610.npy episode_7110.npy episode_8111.npy episode_9112.npy\r\nepisode_1106.npy episode_2107.npy episode_3108.npy episode_4109.npy episode_510.npy episode_6110.npy episode_7111.npy episode_8112.npy episode_9113.npy\r\nepisode_1107.npy episode_2108.npy episode_3109.npy episode_410.npy episode_5110.npy episode_6111.npy episode_7112.npy episode_8113.npy episode_9114.npy\r\nepisode_1108.npy episode_2109.npy episode_310.npy episode_4110.npy episode_5111.npy episode_6112.npy episode_7113.npy episode_8114.npy episode_9115.npy\r\nepisode_1109.npy episode_210.npy episode_3110.npy episode_4111.npy episode_5112.npy episode_6113.npy episode_7114.npy episode_8115.npy episode_9116.npy\r\nepisode_110.npy episode_2110.npy episode_3111.npy episode_4112.npy episode_5113.npy episode_6114.npy episode_7115.npy episode_8116.npy episode_9117.npy\r\nepisode_1110.npy episode_2111.npy episode_3112.npy episode_4113.npy episode_5114.npy episode_6115.npy episode_7116.npy episode_8117.npy episode_9118.npy\r\nepisode_1111.npy episode_2112.npy episode_3113.npy episode_4114.npy episode_5115.npy episode_6116.npy episode_7117.npy episode_8118.npy episode_9119.npy\r\nepisode_1112.npy episode_2113.npy episode_3114.npy episode_4115.npy episode_5116.npy episode_6117.npy episode_7118.npy episode_8119.npy episode_911.npy\r\nepisode_1113.npy episode_2114.npy episode_3115.npy episode_4116.npy episode_5117.npy episode_6118.npy episode_7119.npy episode_811.npy episode_9120.npy\r\nepisode_1114.npy episode_2115.npy episode_3116.npy episode_4117.npy episode_5118.npy episode_6119.npy episode_711.npy episode_8120.npy episode_9121.npy\r\nepisode_1115.npy episode_2116.npy episode_3117.npy episode_4118.npy episode_5119.npy episode_611.npy episode_7120.npy episode_8121.npy episode_9122.npy\r\nepisode_1116.npy episode_2117.npy episode_3118.npy episode_4119.npy episode_511.npy episode_6120.npy episode_7121.npy episode_8122.npy episode_9123.npy\r\nepisode_1117.npy episode_2118.npy episode_3119.npy episode_411.npy episode_5120.npy episode_6121.npy episode_7122.npy episode_8123.npy episode_9124.npy\r\nepisode_1118.npy episode_2119.npy episode_311.npy episode_4120.npy episode_5121.npy episode_6122.npy episode_7123.npy episode_8124.npy episode_9125.npy\r\nepisode_1119.npy episode_211.npy episode_3120.npy episode_4121.npy episode_5122.npy episode_6123.npy episode_7124.npy episode_8125.npy episode_9126.npy\r\nepisode_111.npy episode_2120.npy episode_3121.npy episode_4122.npy episode_5123.npy episode_6124.npy episode_7125.npy episode_8126.npy episode_9127.npy\r\nepisode_1120.npy episode_2121.npy episode_3122.npy episode_4123.npy episode_5124.npy episode_6125.npy episode_7126.npy episode_8127.npy episode_9128.npy\r\nepisode_1121.npy episode_2122.npy episode_3123.npy episode_4124.npy episode_5125.npy episode_6126.npy episode_7127.npy episode_8128.npy episode_9129.npy\r\nepisode_1122.npy episode_2123.npy episode_3124.npy episode_4125.npy episode_5126.npy episode_6127.npy episode_7128.npy episode_8129.npy episode_912.npy\r\nepisode_1123.npy episode_2124.npy episode_3125.npy episode_4126.npy episode_5127.npy episode_6128.npy episode_7129.npy episode_812.npy episode_9130.npy\r\nepisode_1124.npy episode_2125.npy episode_3126.npy episode_4127.npy episode_5128.npy episode_6129.npy episode_712.npy episode_8130.npy episode_9131.npy\r\nepisode_1125.npy episode_2126.npy episode_3127.npy episode_4128.npy episode_5129.npy episode_612.npy episode_7130.npy episode_8131.npy episode_9132.npy\r\nepisode_1126.npy episode_2127.npy episode_3128.npy episode_4129.npy episode_512.npy episode_6130.npy episode_7131.npy episode_8132.npy episode_9133.npy\r\nepisode_1127.npy episode_2128.npy episode_3129.npy episode_412.npy episode_5130.npy episode_6131.npy episode_7132.npy episode_8133.npy episode_9134.npy\r\nepisode_1128.npy episode_2129.npy episode_312.npy episode_4130.npy episode_5131.npy episode_6132.npy episode_7133.npy episode_8134.npy episode_9135.npy\r\nepisode_1129.npy episode_212.npy episode_3130.npy episode_4131.npy episode_5132.npy episode_6133.npy episode_7134.npy episode_8135.npy episode_9136.npy\r\nepisode_112.npy episode_2130.npy episode_3131.npy episode_4132.npy episode_5133.npy episode_6134.npy episode_7135.npy episode_8136.npy episode_9137.npy\r\nepisode_1130.npy episode_2131.npy episode_3132.npy episode_4133.npy episode_5134.npy episode_6135.npy episode_7136.npy episode_8137.npy episode_9138.npy\r\nepisode_1131.npy episode_2132.npy episode_3133.npy episode_4134.npy episode_5135.npy episode_6136.npy episode_7137.npy episode_8138.npy episode_9139.npy\r\nepisode_1132.npy episode_2133.npy episode_3134.npy episode_4135.npy episode_5136.npy episode_6137.npy episode_7138.npy episode_8139.npy episode_913.npy\r\nepisode_1133.npy episode_2134.npy episode_3135.npy episode_4136.npy episode_5137.npy episode_6138.npy episode_7139.npy episode_813.npy episode_9140.npy\r\nepisode_1134.npy episode_2135.npy episode_3136.npy episode_4137.npy episode_5138.npy episode_6139.npy episode_713.npy episode_8140.npy episode_9141.npy\r\nepisode_1135.npy episode_2136.npy episode_3137.npy episode_4138.npy episode_5139.npy episode_613.npy episode_7140.npy episode_8141.npy episode_9142.npy\r\nepisode_1136.npy episode_2137.npy episode_3138.npy episode_4139.npy episode_513.npy episode_6140.npy episode_7141.npy episode_8142.npy episode_9143.npy\r\nepisode_1137.npy episode_2138.npy episode_3139.npy episode_413.npy episode_5140.npy episode_6141.npy episode_7142.npy episode_8143.npy episode_9144.npy\r\nepisode_1138.npy episode_2139.npy episode_313.npy episode_4140.npy episode_5141.npy episode_6142.npy episode_7143.npy episode_8144.npy episode_9145.npy\r\nepisode_1139.npy episode_213.npy episode_3140.npy episode_4141.npy episode_5142.npy episode_6143.npy episode_7144.npy episode_8145.npy episode_9146.npy\r\nepisode_113.npy episode_2140.npy episode_3141.npy episode_4142.npy episode_5143.npy episode_6144.npy episode_7145.npy episode_8146.npy episode_9147.npy\r\nepisode_1140.npy episode_2141.npy episode_3142.npy episode_4143.npy episode_5144.npy episode_6145.npy episode_7146.npy episode_8147.npy episode_9148.npy\r\nepisode_1141.npy episode_2142.npy episode_3143.npy episode_4144.npy episode_5145.npy episode_6146.npy episode_7147.npy episode_8148.npy episode_9149.npy\r\nepisode_1142.npy episode_2143.npy episode_3144.npy episode_4145.npy episode_5146.npy episode_6147.npy episode_7148.npy episode_8149.npy episode_914.npy\r\nepisode_1143.npy episode_2144.npy episode_3145.npy episode_4146.npy episode_5147.npy episode_6148.npy episode_7149.npy episode_814.npy episode_9150.npy\r\nepisode_1144.npy episode_2145.npy episode_3146.npy episode_4147.npy episode_5148.npy episode_6149.npy episode_714.npy episode_8150.npy episode_9151.npy\r\nepisode_1145.npy episode_2146.npy episode_3147.npy episode_4148.npy episode_5149.npy episode_614.npy episode_7150.npy episode_8151.npy episode_9152.npy\r\nepisode_1146.npy episode_2147.npy episode_3148.npy episode_4149.npy episode_514.npy episode_6150.npy episode_7151.npy episode_8152.npy episode_9153.npy\r\nepisode_1147.npy episode_2148.npy episode_3149.npy episode_414.npy episode_5150.npy episode_6151.npy episode_7152.npy episode_8153.npy episode_9154.npy\r\nepisode_1148.npy episode_2149.npy episode_314.npy episode_4150.npy episode_5151.npy episode_6152.npy episode_7153.npy episode_8154.npy episode_9155.npy\r\nepisode_1149.npy episode_214.npy episode_3150.npy episode_4151.npy episode_5152.npy episode_6153.npy episode_7154.npy episode_8155.npy episode_9156.npy\r\nepisode_114.npy episode_2150.npy episode_3151.npy episode_4152.npy episode_5153.npy episode_6154.npy episode_7155.npy episode_8156.npy episode_9157.npy\r\nepisode_1150.npy episode_2151.npy episode_3152.npy episode_4153.npy episode_5154.npy episode_6155.npy episode_7156.npy episode_8157.npy episode_9158.npy\r\nepisode_1151.npy episode_2152.npy episode_3153.npy episode_4154.npy episode_5155.npy episode_6156.npy episode_7157.npy episode_8158.npy episode_9159.npy\r\nepisode_1152.npy episode_2153.npy episode_3154.npy episode_4155.npy episode_5156.npy episode_6157.npy episode_7158.npy episode_8159.npy episode_915.npy\r\nepisode_1153.npy episode_2154.npy episode_3155.npy episode_4156.npy episode_5157.npy episode_6158.npy episode_7159.npy episode_815.npy episode_9160.npy\r\nepisode_1154.npy episode_2155.npy episode_3156.npy episode_4157.npy episode_5158.npy episode_6159.npy episode_715.npy episode_8160.npy episode_9161.npy\r\nepisode_1155.npy episode_2156.npy episode_3157.npy episode_4158.npy episode_5159.npy episode_615.npy episode_7160.npy episode_8161.npy episode_9162.npy\r\nepisode_1156.npy episode_2157.npy episode_3158.npy episode_4159.npy episode_515.npy episode_6160.npy episode_7161.npy episode_8162.npy episode_9163.npy\r\nepisode_1157.npy episode_2158.npy episode_3159.npy episode_415.npy episode_5160.npy episode_6161.npy episode_7162.npy episode_8163.npy episode_9164.npy\r\nepisode_1158.npy episode_2159.npy episode_315.npy episode_4160.npy episode_5161.npy episode_6162.npy episode_7163.npy episode_8164.npy episode_9165.npy\r\nepisode_1159.npy episode_215.npy episode_3160.npy episode_4161.npy episode_5162.npy episode_6163.npy episode_7164.npy episode_8165.npy episode_9166.npy\r\nepisode_115.npy episode_2160.npy episode_3161.npy episode_4162.npy episode_5163.npy episode_6164.npy episode_7165.npy episode_8166.npy episode_9167.npy\r\nepisode_1160.npy episode_2161.npy episode_3162.npy episode_4163.npy episode_5164.npy episode_6165.npy episode_7166.npy episode_8167.npy episode_9168.npy\r\nepisode_1161.npy episode_2162.npy episode_3163.npy episode_4164.npy episode_5165.npy episode_6166.npy episode_7167.npy episode_8168.npy episode_9169.npy\r\nepisode_1162.npy episode_2163.npy episode_3164.npy episode_4165.npy episode_5166.npy episode_6167.npy episode_7168.npy episode_8169.npy episode_916.npy\r\nepisode_1163.npy episode_2164.npy episode_3165.npy episode_4166.npy episode_5167.npy episode_6168.npy episode_7169.npy episode_816.npy episode_9170.npy\r\nepisode_1164.npy episode_2165.npy episode_3166.npy episode_4167.npy episode_5168.npy episode_6169.npy episode_716.npy episode_8170.npy episode_9171.npy\r\nepisode_1165.npy episode_2166.npy episode_3167.npy episode_4168.npy episode_5169.npy episode_616.npy episode_7170.npy episode_8171.npy episode_9172.npy\r\nepisode_1166.npy episode_2167.npy episode_3168.npy episode_4169.npy episode_516.npy episode_6170.npy episode_7171.npy episode_8172.npy episode_9173.npy\r\nepisode_1167.npy episode_2168.npy episode_3169.npy episode_416.npy episode_5170.npy episode_6171.npy episode_7172.npy episode_8173.npy episode_9174.npy\r\nepisode_1168.npy episode_2169.npy episode_316.npy episode_4170.npy episode_5171.npy episode_6172.npy episode_7173.npy episode_8174.npy episode_9175.npy\r\nepisode_1169.npy episode_216.npy episode_3170.npy episode_4171.npy episode_5172.npy episode_6173.npy episode_7174.npy episode_8175.npy episode_9176.npy\r\nepisode_116.npy episode_2170.npy episode_3171.npy episode_4172.npy episode_5173.npy episode_6174.npy episode_7175.npy episode_8176.npy episode_9177.npy\r\nepisode_1170.npy episode_2171.npy episode_3172.npy episode_4173.npy episode_5174.npy episode_6175.npy episode_7176.npy episode_8177.npy episode_9178.npy\r\nepisode_1171.npy episode_2172.npy episode_3173.npy episode_4174.npy episode_5175.npy episode_6176.npy episode_7177.npy episode_8178.npy episode_9179.npy\r\nepisode_1172.npy episode_2173.npy episode_3174.npy episode_4175.npy episode_5176.npy episode_6177.npy episode_7178.npy episode_8179.npy episode_917.npy\r\nepisode_1173.npy episode_2174.npy episode_3175.npy episode_4176.npy episode_5177.npy episode_6178.npy episode_7179.npy episode_817.npy episode_9180.npy\r\nepisode_1174.npy episode_2175.npy episode_3176.npy episode_4177.npy episode_5178.npy episode_6179.npy episode_717.npy episode_8180.npy episode_9181.npy\r\nepisode_1175.npy episode_2176.npy episode_3177.npy episode_4178.npy episode_5179.npy episode_617.npy episode_7180.npy episode_8181.npy episode_9182.npy\r\nepisode_1176.npy episode_2177.npy episode_3178.npy episode_4179.npy episode_517.npy episode_6180.npy episode_7181.npy episode_8182.npy episode_9183.npy\r\nepisode_1177.npy episode_2178.npy episode_3179.npy episode_417.npy episode_5180.npy episode_6181.npy episode_7182.npy episode_8183.npy episode_9184.npy\r\nepisode_1178.npy episode_2179.npy episode_317.npy episode_4180.npy episode_5181.npy episode_6182.npy episode_7183.npy episode_8184.npy episode_9185.npy\r\nepisode_1179.npy episode_217.npy episode_3180.npy episode_4181.npy episode_5182.npy episode_6183.npy episode_7184.npy episode_8185.npy episode_9186.npy\r\nepisode_117.npy episode_2180.npy episode_3181.npy episode_4182.npy episode_5183.npy episode_6184.npy episode_7185.npy episode_8186.npy episode_9187.npy\r\nepisode_1180.npy episode_2181.npy episode_3182.npy episode_4183.npy episode_5184.npy episode_6185.npy episode_7186.npy episode_8187.npy episode_9188.npy\r\nepisode_1181.npy episode_2182.npy episode_3183.npy episode_4184.npy episode_5185.npy episode_6186.npy episode_7187.npy episode_8188.npy episode_9189.npy\r\nepisode_1182.npy episode_2183.npy episode_3184.npy episode_4185.npy episode_5186.npy episode_6187.npy episode_7188.npy episode_8189.npy episode_918.npy\r\nepisode_1183.npy episode_2184.npy episode_3185.npy episode_4186.npy episode_5187.npy episode_6188.npy episode_7189.npy episode_818.npy episode_9190.npy\r\nepisode_1184.npy episode_2185.npy episode_3186.npy episode_4187.npy episode_5188.npy episode_6189.npy episode_718.npy episode_8190.npy episode_9191.npy\r\nepisode_1185.npy episode_2186.npy episode_3187.npy episode_4188.npy episode_5189.npy episode_618.npy episode_7190.npy episode_8191.npy episode_9192.npy\r\nepisode_1186.npy episode_2187.npy episode_3188.npy episode_4189.npy episode_518.npy episode_6190.npy episode_7191.npy episode_8192.npy episode_9193.npy\r\nepisode_1187.npy episode_2188.npy episode_3189.npy episode_418.npy episode_5190.npy episode_6191.npy episode_7192.npy episode_8193.npy episode_9194.npy\r\nepisode_1188.npy episode_2189.npy episode_318.npy episode_4190.npy episode_5191.npy episode_6192.npy episode_7193.npy episode_8194.npy episode_9195.npy\r\nepisode_1189.npy episode_218.npy episode_3190.npy episode_4191.npy episode_5192.npy episode_6193.npy episode_7194.npy episode_8195.npy episode_9196.npy\r\nepisode_118.npy episode_2190.npy episode_3191.npy episode_4192.npy episode_5193.npy episode_6194.npy episode_7195.npy episode_8196.npy episode_9197.npy\r\nepisode_1190.npy episode_2191.npy episode_3192.npy episode_4193.npy episode_5194.npy episode_6195.npy episode_7196.npy episode_8197.npy episode_9198.npy\r\nepisode_1191.npy episode_2192.npy episode_3193.npy episode_4194.npy episode_5195.npy episode_6196.npy episode_7197.npy episode_8198.npy episode_9199.npy\r\nepisode_1192.npy episode_2193.npy episode_3194.npy episode_4195.npy episode_5196.npy episode_6197.npy episode_7198.npy episode_8199.npy episode_919.npy\r\nepisode_1193.npy episode_2194.npy episode_3195.npy episode_4196.npy episode_5197.npy episode_6198.npy episode_7199.npy episode_819.npy episode_91.npy\r\nepisode_1194.npy episode_2195.npy episode_3196.npy episode_4197.npy episode_5198.npy episode_6199.npy episode_719.npy episode_81.npy episode_9200.npy\r\nepisode_1195.npy episode_2196.npy episode_3197.npy episode_4198.npy episode_5199.npy episode_619.npy episode_71.npy episode_8200.npy episode_9201.npy\r\nepisode_1196.npy episode_2197.npy episode_3198.npy episode_4199.npy episode_519.npy episode_61.npy episode_7200.npy episode_8201.npy episode_9202.npy\r\nepisode_1197.npy episode_2198.npy episode_3199.npy episode_419.npy episode_51.npy episode_6200.npy episode_7201.npy episode_8202.npy episode_9203.npy\r\nepisode_1198.npy episode_2199.npy episode_319.npy episode_41.npy episode_5200.npy episode_6201.npy episode_7202.npy episode_8203.npy episode_9204.npy\r\nepisode_1199.npy episode_219.npy episode_31.npy episode_4200.npy episode_5201.npy episode_6202.npy episode_7203.npy episode_8204.npy episode_9205.npy\r\nepisode_119.npy episode_21.npy episode_3200.npy episode_4201.npy episode_5202.npy episode_6203.npy episode_7204.npy episode_8205.npy episode_9206.npy\r\nepisode_11.npy episode_2200.npy episode_3201.npy episode_4202.npy episode_5203.npy episode_6204.npy episode_7205.npy episode_8206.npy episode_9207.npy\r\nepisode_1200.npy episode_2201.npy episode_3202.npy episode_4203.npy episode_5204.npy episode_6205.npy episode_7206.npy episode_8207.npy episode_9208.npy\r\nepisode_1201.npy episode_2202.npy episode_3203.npy episode_4204.npy episode_5205.npy episode_6206.npy episode_7207.npy episode_8208.npy episode_9209.npy\r\nepisode_1202.npy episode_2203.npy episode_3204.npy episode_4205.npy episode_5206.npy episode_6207.npy episode_7208.npy episode_8209.npy episode_920.npy\r\nepisode_1203.npy episode_2204.npy episode_3205.npy episode_4206.npy episode_5207.npy episode_6208.npy episode_7209.npy episode_820.npy episode_9210.npy\r\nepisode_1204.npy episode_2205.npy episode_3206.npy episode_4207.npy episode_5208.npy episode_6209.npy episode_720.npy episode_8210.npy episode_9211.npy\r\nepisode_1205.npy episode_2206.npy episode_3207.npy episode_4208.npy episode_5209.npy episode_620.npy episode_7210.npy episode_8211.npy episode_9212.npy\r\nepisode_1206.npy episode_2207.npy episode_3208.npy episode_4209.npy episode_520.npy episode_6210.npy episode_7211.npy episode_8212.npy episode_9213.npy\r\nepisode_1207.npy episode_2208.npy episode_3209.npy episode_420.npy episode_5210.npy episode_6211.npy episode_7212.npy episode_8213.npy episode_9214.npy\r\nepisode_1208.npy episode_2209.npy episode_320.npy episode_4210.npy episode_5211.npy episode_6212.npy episode_7213.npy episode_8214.npy episode_9215.npy\r\nepisode_1209.npy episode_220.npy episode_3210.npy episode_4211.npy episode_5212.npy episode_6213.npy episode_7214.npy episode_8215.npy episode_9216.npy\r\nepisode_120.npy episode_2210.npy episode_3211.npy episode_4212.npy episode_5213.npy episode_6214.npy episode_7215.npy episode_8216.npy episode_9217.npy\r\nepisode_1210.npy episode_2211.npy episode_3212.npy episode_4213.npy episode_5214.npy episode_6215.npy episode_7216.npy episode_8217.npy episode_9218.npy\r\nepisode_1211.npy episode_2212.npy episode_3213.npy episode_4214.npy episode_5215.npy episode_6216.npy episode_7217.npy episode_8218.npy episode_9219.npy\r\nepisode_1212.npy episode_2213.npy episode_3214.npy episode_4215.npy episode_5216.npy episode_6217.npy episode_7218.npy episode_8219.npy episode_921.npy\r\nepisode_1213.npy episode_2214.npy episode_3215.npy episode_4216.npy episode_5217.npy episode_6218.npy episode_7219.npy episode_821.npy episode_9220.npy\r\nepisode_1214.npy episode_2215.npy episode_3216.npy episode_4217.npy episode_5218.npy episode_6219.npy episode_721.npy episode_8220.npy episode_9221.npy\r\nepisode_1215.npy episode_2216.npy episode_3217.npy episode_4218.npy episode_5219.npy episode_621.npy episode_7220.npy episode_8221.npy episode_9222.npy\r\nepisode_1216.npy episode_2217.npy episode_3218.npy episode_4219.npy episode_521.npy episode_6220.npy episode_7221.npy episode_8222.npy episode_9223.npy\r\nepisode_1217.npy episode_2218.npy episode_3219.npy episode_421.npy episode_5220.npy episode_6221.npy episode_7222.npy episode_8223.npy episode_9224.npy\r\nepisode_1218.npy episode_2219.npy episode_321.npy episode_4220.npy episode_5221.npy episode_6222.npy episode_7223.npy episode_8224.npy episode_9225.npy\r\nepisode_1219.npy episode_221.npy episode_3220.npy episode_4221.npy episode_5222.npy episode_6223.npy episode_7224.npy episode_8225.npy episode_9226.npy\r\nepisode_121.npy episode_2220.npy episode_3221.npy episode_4222.npy episode_5223.npy episode_6224.npy episode_7225.npy episode_8226.npy episode_9227.npy\r\nepisode_1220.npy episode_2221.npy episode_3222.npy episode_4223.npy episode_5224.npy episode_6225.npy episode_7226.npy episode_8227.npy episode_9228.npy\r\nepisode_1221.npy episode_2222.npy episode_3223.npy episode_4224.npy episode_5225.npy episode_6226.npy episode_7227.npy episode_8228.npy episode_9229.npy\r\nepisode_1222.npy episode_2223.npy episode_3224.npy episode_4225.npy episode_5226.npy episode_6227.npy episode_7228.npy episode_8229.npy episode_922.npy\r\nepisode_1223.npy episode_2224.npy episode_3225.npy episode_4226.npy episode_5227.npy episode_6228.npy episode_7229.npy episode_822.npy episode_9230.npy\r\nepisode_1224.npy episode_2225.npy episode_3226.npy episode_4227.npy episode_5228.npy episode_6229.npy episode_722.npy episode_8230.npy episode_9231.npy\r\nepisode_1225.npy episode_2226.npy episode_3227.npy episode_4228.npy episode_5229.npy episode_622.npy episode_7230.npy episode_8231.npy episode_9232.npy\r\nepisode_1226.npy episode_2227.npy episode_3228.npy episode_4229.npy episode_522.npy episode_6230.npy episode_7231.npy episode_8232.npy episode_9233.npy\r\nepisode_1227.npy episode_2228.npy episode_3229.npy episode_422.npy episode_5230.npy episode_6231.npy episode_7232.npy episode_8233.npy episode_9234.npy\r\nepisode_1228.npy episode_2229.npy episode_322.npy episode_4230.npy episode_5231.npy episode_6232.npy episode_7233.npy episode_8234.npy episode_9235.npy\r\nepisode_1229.npy episode_222.npy episode_3230.npy episode_4231.npy episode_5232.npy episode_6233.npy episode_7234.npy episode_8235.npy episode_9236.npy\r\nepisode_122.npy episode_2230.npy episode_3231.npy episode_4232.npy episode_5233.npy episode_6234.npy episode_7235.npy episode_8236.npy episode_9237.npy\r\nepisode_1230.npy episode_2231.npy episode_3232.npy episode_4233.npy episode_5234.npy episode_6235.npy episode_7236.npy episode_8237.npy episode_9238.npy\r\nepisode_1231.npy episode_2232.npy episode_3233.npy episode_4234.npy episode_5235.npy episode_6236.npy episode_7237.npy episode_8238.npy episode_9239.npy\r\nepisode_1232.npy episode_2233.npy episode_3234.npy episode_4235.npy episode_5236.npy episode_6237.npy episode_7238.npy episode_8239.npy episode_923.npy\r\nepisode_1233.npy episode_2234.npy episode_3235.npy episode_4236.npy episode_5237.npy episode_6238.npy episode_7239.npy episode_823.npy episode_9240.npy\r\nepisode_1234.npy episode_2235.npy episode_3236.npy episode_4237.npy episode_5238.npy episode_6239.npy episode_723.npy episode_8240.npy episode_9241.npy\r\nepisode_1235.npy episode_2236.npy episode_3237.npy episode_4238.npy episode_5239.npy episode_623.npy episode_7240.npy episode_8241.npy episode_9242.npy\r\nepisode_1236.npy episode_2237.npy episode_3238.npy episode_4239.npy episode_523.npy episode_6240.npy episode_7241.npy episode_8242.npy episode_9243.npy\r\nepisode_1237.npy episode_2238.npy episode_3239.npy episode_423.npy episode_5240.npy episode_6241.npy episode_7242.npy episode_8243.npy episode_9244.npy\r\nepisode_1238.npy episode_2239.npy episode_323.npy episode_4240.npy episode_5241.npy episode_6242.npy episode_7243.npy episode_8244.npy episode_9245.npy\r\nepisode_1239.npy episode_223.npy episode_3240.npy episode_4241.npy episode_5242.npy episode_6243.npy episode_7244.npy episode_8245.npy episode_9246.npy\r\nepisode_123.npy episode_2240.npy episode_3241.npy episode_4242.npy episode_5243.npy episode_6244.npy episode_7245.npy episode_8246.npy episode_9247.npy\r\nepisode_1240.npy episode_2241.npy episode_3242.npy episode_4243.npy episode_5244.npy episode_6245.npy episode_7246.npy episode_8247.npy episode_9248.npy\r\nepisode_1241.npy episode_2242.npy episode_3243.npy episode_4244.npy episode_5245.npy episode_6246.npy episode_7247.npy episode_8248.npy episode_9249.npy\r\nepisode_1242.npy episode_2243.npy episode_3244.npy episode_4245.npy episode_5246.npy episode_6247.npy episode_7248.npy episode_8249.npy episode_924.npy\r\nepisode_1243.npy episode_2244.npy episode_3245.npy episode_4246.npy episode_5247.npy episode_6248.npy episode_7249.npy episode_824.npy episode_9250.npy\r\nepisode_1244.npy episode_2245.npy episode_3246.npy episode_4247.npy episode_5248.npy episode_6249.npy episode_724.npy episode_8250.npy episode_9251.npy\r\nepisode_1245.npy episode_2246.npy episode_3247.npy episode_4248.npy episode_5249.npy episode_624.npy episode_7250.npy episode_8251.npy episode_9252.npy\r\nepisode_1246.npy episode_2247.npy episode_3248.npy episode_4249.npy episode_524.npy episode_6250.npy episode_7251.npy episode_8252.npy episode_9253.npy\r\nepisode_1247.npy episode_2248.npy episode_3249.npy episode_424.npy episode_5250.npy episode_6251.npy episode_7252.npy episode_8253.npy episode_9254.npy\r\nepisode_1248.npy episode_2249.npy episode_324.npy episode_4250.npy episode_5251.npy episode_6252.npy episode_7253.npy episode_8254.npy episode_9255.npy\r\nepisode_1249.npy episode_224.npy episode_3250.npy episode_4251.npy episode_5252.npy episode_6253.npy episode_7254.npy episode_8255.npy episode_9256.npy\r\nepisode_124.npy episode_2250.npy episode_3251.npy episode_4252.npy episode_5253.npy episode_6254.npy episode_7255.npy episode_8256.npy episode_9257.npy\r\nepisode_1250.npy episode_2251.npy episode_3252.npy episode_4253.npy episode_5254.npy episode_6255.npy episode_7256.npy episode_8257.npy episode_9258.npy\r\nepisode_1251.npy episode_2252.npy episode_3253.npy episode_4254.npy episode_5255.npy episode_6256.npy episode_7257.npy episode_8258.npy episode_9259.npy\r\nepisode_1252.npy episode_2253.npy episode_3254.npy episode_4255.npy episode_5256.npy episode_6257.npy episode_7258.npy episode_8259.npy episode_925.npy\r\nepisode_1253.npy episode_2254.npy episode_3255.npy episode_4256.npy episode_5257.npy episode_6258.npy episode_7259.npy episode_825.npy episode_9260.npy\r\nepisode_1254.npy episode_2255.npy episode_3256.npy episode_4257.npy episode_5258.npy episode_6259.npy episode_725.npy episode_8260.npy episode_9261.npy\r\nepisode_1255.npy episode_2256.npy episode_3257.npy episode_4258.npy episode_5259.npy episode_625.npy episode_7260.npy episode_8261.npy episode_9262.npy\r\nepisode_1256.npy episode_2257.npy episode_3258.npy episode_4259.npy episode_525.npy episode_6260.npy episode_7261.npy episode_8262.npy episode_9263.npy\r\nepisode_1257.npy episode_2258.npy episode_3259.npy episode_425.npy episode_5260.npy episode_6261.npy episode_7262.npy episode_8263.npy episode_9264.npy\r\nepisode_1258.npy episode_2259.npy episode_325.npy episode_4260.npy episode_5261.npy episode_6262.npy episode_7263.npy episode_8264.npy episode_9265.npy\r\nepisode_1259.npy episode_225.npy episode_3260.npy episode_4261.npy episode_5262.npy episode_6263.npy episode_7264.npy episode_8265.npy episode_9266.npy\r\nepisode_125.npy episode_2260.npy episode_3261.npy episode_4262.npy episode_5263.npy episode_6264.npy episode_7265.npy episode_8266.npy episode_9267.npy\r\nepisode_1260.npy episode_2261.npy episode_3262.npy episode_4263.npy episode_5264.npy episode_6265.npy episode_7266.npy episode_8267.npy episode_9268.npy\r\nepisode_1261.npy episode_2262.npy episode_3263.npy episode_4264.npy episode_5265.npy episode_6266.npy episode_7267.npy episode_8268.npy episode_9269.npy\r\nepisode_1262.npy episode_2263.npy episode_3264.npy episode_4265.npy episode_5266.npy episode_6267.npy episode_7268.npy episode_8269.npy episode_926.npy\r\nepisode_1263.npy episode_2264.npy episode_3265.npy episode_4266.npy episode_5267.npy episode_6268.npy episode_7269.npy episode_826.npy episode_9270.npy\r\nepisode_1264.npy episode_2265.npy episode_3266.npy episode_4267.npy episode_5268.npy episode_6269.npy episode_726.npy episode_8270.npy episode_9271.npy\r\nepisode_1265.npy episode_2266.npy episode_3267.npy episode_4268.npy episode_5269.npy episode_626.npy episode_7270.npy episode_8271.npy episode_9272.npy\r\nepisode_1266.npy episode_2267.npy episode_3268.npy episode_4269.npy episode_526.npy episode_6270.npy episode_7271.npy episode_8272.npy episode_9273.npy\r\nepisode_1267.npy episode_2268.npy episode_3269.npy episode_426.npy episode_5270.npy episode_6271.npy episode_7272.npy episode_8273.npy episode_9274.npy\r\nepisode_1268.npy episode_2269.npy episode_326.npy episode_4270.npy episode_5271.npy episode_6272.npy episode_7273.npy episode_8274.npy episode_9275.npy\r\nepisode_1269.npy episode_226.npy episode_3270.npy episode_4271.npy episode_5272.npy episode_6273.npy episode_7274.npy episode_8275.npy episode_9276.npy\r\nepisode_126.npy episode_2270.npy episode_3271.npy episode_4272.npy episode_5273.npy episode_6274.npy episode_7275.npy episode_8276.npy episode_9277.npy\r\nepisode_1270.npy episode_2271.npy episode_3272.npy episode_4273.npy episode_5274.npy episode_6275.npy episode_7276.npy episode_8277.npy episode_9278.npy\r\nepisode_1271.npy episode_2272.npy episode_3273.npy episode_4274.npy episode_5275.npy episode_6276.npy episode_7277.npy episode_8278.npy episode_9279.npy\r\nepisode_1272.npy episode_2273.npy episode_3274.npy episode_4275.npy episode_5276.npy episode_6277.npy episode_7278.npy episode_8279.npy episode_927.npy\r\nepisode_1273.npy episode_2274.npy episode_3275.npy episode_4276.npy episode_5277.npy episode_6278.npy episode_7279.npy episode_827.npy episode_9280.npy\r\nepisode_1274.npy episode_2275.npy episode_3276.npy episode_4277.npy episode_5278.npy episode_6279.npy episode_727.npy episode_8280.npy episode_9281.npy\r\nepisode_1275.npy episode_2276.npy episode_3277.npy episode_4278.npy episode_5279.npy episode_627.npy episode_7280.npy episode_8281.npy episode_9282.npy\r\nepisode_1276.npy episode_2277.npy episode_3278.npy episode_4279.npy episode_527.npy episode_6280.npy episode_7281.npy episode_8282.npy episode_9283.npy\r\nepisode_1277.npy episode_2278.npy episode_3279.npy episode_427.npy episode_5280.npy episode_6281.npy episode_7282.npy episode_8283.npy episode_9284.npy\r\nepisode_1278.npy episode_2279.npy episode_327.npy episode_4280.npy episode_5281.npy episode_6282.npy episode_7283.npy episode_8284.npy episode_9285.npy\r\nepisode_1279.npy episode_227.npy episode_3280.npy episode_4281.npy episode_5282.npy episode_6283.npy episode_7284.npy episode_8285.npy episode_9286.npy\r\nepisode_127.npy episode_2280.npy episode_3281.npy episode_4282.npy episode_5283.npy episode_6284.npy episode_7285.npy episode_8286.npy episode_9287.npy\r\nepisode_1280.npy episode_2281.npy episode_3282.npy episode_4283.npy episode_5284.npy episode_6285.npy episode_7286.npy episode_8287.npy episode_9288.npy\r\nepisode_1281.npy episode_2282.npy episode_3283.npy episode_4284.npy episode_5285.npy episode_6286.npy episode_7287.npy episode_8288.npy episode_9289.npy\r\nepisode_1282.npy episode_2283.npy episode_3284.npy episode_4285.npy episode_5286.npy episode_6287.npy episode_7288.npy episode_8289.npy episode_928.npy\r\nepisode_1283.npy episode_2284.npy episode_3285.npy episode_4286.npy episode_5287.npy episode_6288.npy episode_7289.npy episode_828.npy episode_9290.npy\r\nepisode_1284.npy episode_2285.npy episode_3286.npy episode_4287.npy episode_5288.npy episode_6289.npy episode_728.npy episode_8290.npy episode_9291.npy\r\nepisode_1285.npy episode_2286.npy episode_3287.npy episode_4288.npy episode_5289.npy episode_628.npy episode_7290.npy episode_8291.npy episode_9292.npy\r\nepisode_1286.npy episode_2287.npy episode_3288.npy episode_4289.npy episode_528.npy episode_6290.npy episode_7291.npy episode_8292.npy episode_9293.npy\r\nepisode_1287.npy episode_2288.npy episode_3289.npy episode_428.npy episode_5290.npy episode_6291.npy episode_7292.npy episode_8293.npy episode_9294.npy\r\nepisode_1288.npy episode_2289.npy episode_328.npy episode_4290.npy episode_5291.npy episode_6292.npy episode_7293.npy episode_8294.npy episode_9295.npy\r\nepisode_1289.npy episode_228.npy episode_3290.npy episode_4291.npy episode_5292.npy episode_6293.npy episode_7294.npy episode_8295.npy episode_9296.npy\r\nepisode_128.npy episode_2290.npy episode_3291.npy episode_4292.npy episode_5293.npy episode_6294.npy episode_7295.npy episode_8296.npy episode_9297.npy\r\nepisode_1290.npy episode_2291.npy episode_3292.npy episode_4293.npy episode_5294.npy episode_6295.npy episode_7296.npy episode_8297.npy episode_9298.npy\r\nepisode_1291.npy episode_2292.npy episode_3293.npy episode_4294.npy episode_5295.npy episode_6296.npy episode_7297.npy episode_8298.npy episode_9299.npy\r\nepisode_1292.npy episode_2293.npy episode_3294.npy episode_4295.npy episode_5296.npy episode_6297.npy episode_7298.npy episode_8299.npy episode_929.npy\r\nepisode_1293.npy episode_2294.npy episode_3295.npy episode_4296.npy episode_5297.npy episode_6298.npy episode_7299.npy episode_829.npy episode_92.npy\r\nepisode_1294.npy episode_2295.npy episode_3296.npy episode_4297.npy episode_5298.npy episode_6299.npy episode_729.npy episode_82.npy episode_9300.npy\r\nepisode_1295.npy episode_2296.npy episode_3297.npy episode_4298.npy episode_5299.npy episode_629.npy episode_72.npy episode_8300.npy episode_9301.npy\r\nepisode_1296.npy episode_2297.npy episode_3298.npy episode_4299.npy episode_529.npy episode_62.npy episode_7300.npy episode_8301.npy episode_9302.npy\r\nepisode_1297.npy episode_2298.npy episode_3299.npy episode_429.npy episode_52.npy episode_6300.npy episode_7301.npy episode_8302.npy episode_9303.npy\r\nepisode_1298.npy episode_2299.npy episode_329.npy episode_42.npy episode_5300.npy episode_6301.npy episode_7302.npy episode_8303.npy episode_9304.npy\r\nepisode_1299.npy episode_229.npy episode_32.npy episode_4300.npy episode_5301.npy episode_6302.npy episode_7303.npy episode_8304.npy episode_9305.npy\r\nepisode_129.npy episode_22.npy episode_3300.npy episode_4301.npy episode_5302.npy episode_6303.npy episode_7304.npy episode_8305.npy episode_9306.npy\r\nepisode_12.npy episode_2300.npy episode_3301.npy episode_4302.npy episode_5303.npy episode_6304.npy episode_7305.npy episode_8306.npy episode_9307.npy\r\nepisode_1300.npy episode_2301.npy episode_3302.npy episode_4303.npy episode_5304.npy episode_6305.npy episode_7306.npy episode_8307.npy episode_9308.npy\r\nepisode_1301.npy episode_2302.npy episode_3303.npy episode_4304.npy episode_5305.npy episode_6306.npy episode_7307.npy episode_8308.npy episode_9309.npy\r\nepisode_1302.npy episode_2303.npy episode_3304.npy episode_4305.npy episode_5306.npy episode_6307.npy episode_7308.npy episode_8309.npy episode_930.npy\r\nepisode_1303.npy episode_2304.npy episode_3305.npy episode_4306.npy episode_5307.npy episode_6308.npy episode_7309.npy episode_830.npy episode_9310.npy\r\nepisode_1304.npy episode_2305.npy episode_3306.npy episode_4307.npy episode_5308.npy episode_6309.npy episode_730.npy episode_8310.npy episode_9311.npy\r\nepisode_1305.npy episode_2306.npy episode_3307.npy episode_4308.npy episode_5309.npy episode_630.npy episode_7310.npy episode_8311.npy episode_9312.npy\r\nepisode_1306.npy episode_2307.npy episode_3308.npy episode_4309.npy episode_530.npy episode_6310.npy episode_7311.npy episode_8312.npy episode_9313.npy\r\nepisode_1307.npy episode_2308.npy episode_3309.npy episode_430.npy episode_5310.npy episode_6311.npy episode_7312.npy episode_8313.npy episode_9314.npy\r\nepisode_1308.npy episode_2309.npy episode_330.npy episode_4310.npy episode_5311.npy episode_6312.npy episode_7313.npy episode_8314.npy episode_9315.npy\r\nepisode_1309.npy episode_230.npy episode_3310.npy episode_4311.npy episode_5312.npy episode_6313.npy episode_7314.npy episode_8315.npy episode_9316.npy\r\nepisode_130.npy episode_2310.npy episode_3311.npy episode_4312.npy episode_5313.npy episode_6314.npy episode_7315.npy episode_8316.npy episode_9317.npy\r\nepisode_1310.npy episode_2311.npy episode_3312.npy episode_4313.npy episode_5314.npy episode_6315.npy episode_7316.npy episode_8317.npy episode_9318.npy\r\nepisode_1311.npy episode_2312.npy episode_3313.npy episode_4314.npy episode_5315.npy episode_6316.npy episode_7317.npy episode_8318.npy episode_9319.npy\r\nepisode_1312.npy episode_2313.npy episode_3314.npy episode_4315.npy episode_5316.npy episode_6317.npy episode_7318.npy episode_8319.npy episode_931.npy\r\nepisode_1313.npy episode_2314.npy episode_3315.npy episode_4316.npy episode_5317.npy episode_6318.npy episode_7319.npy episode_831.npy episode_9320.npy\r\nepisode_1314.npy episode_2315.npy episode_3316.npy episode_4317.npy episode_5318.npy episode_6319.npy episode_731.npy episode_8320.npy episode_9321.npy\r\nepisode_1315.npy episode_2316.npy episode_3317.npy episode_4318.npy episode_5319.npy episode_631.npy episode_7320.npy episode_8321.npy episode_9322.npy\r\nepisode_1316.npy episode_2317.npy episode_3318.npy episode_4319.npy episode_531.npy episode_6320.npy episode_7321.npy episode_8322.npy episode_9323.npy\r\nepisode_1317.npy episode_2318.npy episode_3319.npy episode_431.npy episode_5320.npy episode_6321.npy episode_7322.npy episode_8323.npy episode_9324.npy\r\nepisode_1318.npy episode_2319.npy episode_331.npy episode_4320.npy episode_5321.npy episode_6322.npy episode_7323.npy episode_8324.npy episode_9325.npy\r\nepisode_1319.npy episode_231.npy episode_3320.npy episode_4321.npy episode_5322.npy episode_6323.npy episode_7324.npy episode_8325.npy episode_9326.npy\r\nepisode_131.npy episode_2320.npy episode_3321.npy episode_4322.npy episode_5323.npy episode_6324.npy episode_7325.npy episode_8326.npy episode_9327.npy\r\nepisode_1320.npy episode_2321.npy episode_3322.npy episode_4323.npy episode_5324.npy episode_6325.npy episode_7326.npy episode_8327.npy episode_9328.npy\r\nepisode_1321.npy episode_2322.npy episode_3323.npy episode_4324.npy episode_5325.npy episode_6326.npy episode_7327.npy episode_8328.npy episode_9329.npy\r\nepisode_1322.npy episode_2323.npy episode_3324.npy episode_4325.npy episode_5326.npy episode_6327.npy episode_7328.npy episode_8329.npy episode_932.npy\r\nepisode_1323.npy episode_2324.npy episode_3325.npy episode_4326.npy episode_5327.npy episode_6328.npy episode_7329.npy episode_832.npy episode_9330.npy\r\nepisode_1324.npy episode_2325.npy episode_3326.npy episode_4327.npy episode_5328.npy episode_6329.npy episode_732.npy episode_8330.npy episode_9331.npy\r\nepisode_1325.npy episode_2326.npy episode_3327.npy episode_4328.npy episode_5329.npy episode_632.npy episode_7330.npy episode_8331.npy episode_9332.npy\r\nepisode_1326.npy episode_2327.npy episode_3328.npy episode_4329.npy episode_532.npy episode_6330.npy episode_7331.npy episode_8332.npy episode_9333.npy\r\nepisode_1327.npy episode_2328.npy episode_3329.npy episode_432.npy episode_5330.npy episode_6331.npy episode_7332.npy episode_8333.npy episode_9334.npy\r\nepisode_1328.npy episode_2329.npy episode_332.npy episode_4330.npy episode_5331.npy episode_6332.npy episode_7333.npy episode_8334.npy episode_9335.npy\r\nepisode_1329.npy episode_232.npy episode_3330.npy episode_4331.npy episode_5332.npy episode_6333.npy episode_7334.npy episode_8335.npy episode_9336.npy\r\nepisode_132.npy episode_2330.npy episode_3331.npy episode_4332.npy episode_5333.npy episode_6334.npy episode_7335.npy episode_8336.npy episode_9337.npy\r\nepisode_1330.npy episode_2331.npy episode_3332.npy episode_4333.npy episode_5334.npy episode_6335.npy episode_7336.npy episode_8337.npy episode_9338.npy\r\nepisode_1331.npy episode_2332.npy episode_3333.npy episode_4334.npy episode_5335.npy episode_6336.npy episode_7337.npy episode_8338.npy episode_9339.npy\r\nepisode_1332.npy episode_2333.npy episode_3334.npy episode_4335.npy episode_5336.npy episode_6337.npy episode_7338.npy episode_8339.npy episode_933.npy\r\nepisode_1333.npy episode_2334.npy episode_3335.npy episode_4336.npy episode_5337.npy episode_6338.npy episode_7339.npy episode_833.npy episode_9340.npy\r\nepisode_1334.npy episode_2335.npy episode_3336.npy episode_4337.npy episode_5338.npy episode_6339.npy episode_733.npy episode_8340.npy episode_9341.npy\r\nepisode_1335.npy episode_2336.npy episode_3337.npy episode_4338.npy episode_5339.npy episode_633.npy episode_7340.npy episode_8341.npy episode_9342.npy\r\nepisode_1336.npy episode_2337.npy episode_3338.npy episode_4339.npy episode_533.npy episode_6340.npy episode_7341.npy episode_8342.npy episode_9343.npy\r\nepisode_1337.npy episode_2338.npy episode_3339.npy episode_433.npy episode_5340.npy episode_6341.npy episode_7342.npy episode_8343.npy episode_9344.npy\r\nepisode_1338.npy episode_2339.npy episode_333.npy episode_4340.npy episode_5341.npy episode_6342.npy episode_7343.npy episode_8344.npy episode_9345.npy\r\nepisode_1339.npy episode_233.npy episode_3340.npy episode_4341.npy episode_5342.npy episode_6343.npy episode_7344.npy episode_8345.npy episode_9346.npy\r\nepisode_133.npy episode_2340.npy episode_3341.npy episode_4342.npy episode_5343.npy episode_6344.npy episode_7345.npy episode_8346.npy episode_9347.npy\r\nepisode_1340.npy episode_2341.npy episode_3342.npy episode_4343.npy episode_5344.npy episode_6345.npy episode_7346.npy episode_8347.npy episode_9348.npy\r\nepisode_1341.npy episode_2342.npy episode_3343.npy episode_4344.npy episode_5345.npy episode_6346.npy episode_7347.npy episode_8348.npy episode_9349.npy\r\nepisode_1342.npy episode_2343.npy episode_3344.npy episode_4345.npy episode_5346.npy episode_6347.npy episode_7348.npy episode_8349.npy episode_934.npy\r\nepisode_1343.npy episode_2344.npy episode_3345.npy episode_4346.npy episode_5347.npy episode_6348.npy episode_7349.npy episode_834.npy episode_9350.npy\r\nepisode_1344.npy episode_2345.npy episode_3346.npy episode_4347.npy episode_5348.npy episode_6349.npy episode_734.npy episode_8350.npy episode_9351.npy\r\nepisode_1345.npy episode_2346.npy episode_3347.npy episode_4348.npy episode_5349.npy episode_634.npy episode_7350.npy episode_8351.npy episode_9352.npy\r\nepisode_1346.npy episode_2347.npy episode_3348.npy episode_4349.npy episode_534.npy episode_6350.npy episode_7351.npy episode_8352.npy episode_9353.npy\r\nepisode_1347.npy episode_2348.npy episode_3349.npy episode_434.npy episode_5350.npy episode_6351.npy episode_7352.npy episode_8353.npy episode_9354.npy\r\nepisode_1348.npy episode_2349.npy episode_334.npy episode_4350.npy episode_5351.npy episode_6352.npy episode_7353.npy episode_8354.npy episode_9355.npy\r\nepisode_1349.npy episode_234.npy episode_3350.npy episode_4351.npy episode_5352.npy episode_6353.npy episode_7354.npy episode_8355.npy episode_9356.npy\r\nepisode_134.npy episode_2350.npy episode_3351.npy episode_4352.npy episode_5353.npy episode_6354.npy episode_7355.npy episode_8356.npy episode_9357.npy\r\nepisode_1350.npy episode_2351.npy episode_3352.npy episode_4353.npy episode_5354.npy episode_6355.npy episode_7356.npy episode_8357.npy episode_9358.npy\r\nepisode_1351.npy episode_2352.npy episode_3353.npy episode_4354.npy episode_5355.npy episode_6356.npy episode_7357.npy episode_8358.npy episode_9359.npy\r\nepisode_1352.npy episode_2353.npy episode_3354.npy episode_4355.npy episode_5356.npy episode_6357.npy episode_7358.npy episode_8359.npy episode_935.npy\r\nepisode_1353.npy episode_2354.npy episode_3355.npy episode_4356.npy episode_5357.npy episode_6358.npy episode_7359.npy episode_835.npy episode_9360.npy\r\nepisode_1354.npy episode_2355.npy episode_3356.npy episode_4357.npy episode_5358.npy episode_6359.npy episode_735.npy episode_8360.npy episode_9361.npy\r\nepisode_1355.npy episode_2356.npy episode_3357.npy episode_4358.npy episode_5359.npy episode_635.npy episode_7360.npy episode_8361.npy episode_9362.npy\r\nepisode_1356.npy episode_2357.npy episode_3358.npy episode_4359.npy episode_535.npy episode_6360.npy episode_7361.npy episode_8362.npy episode_9363.npy\r\nepisode_1357.npy episode_2358.npy episode_3359.npy episode_435.npy episode_5360.npy episode_6361.npy episode_7362.npy episode_8363.npy episode_9364.npy\r\nepisode_1358.npy episode_2359.npy episode_335.npy episode_4360.npy episode_5361.npy episode_6362.npy episode_7363.npy episode_8364.npy episode_9365.npy\r\nepisode_1359.npy episode_235.npy episode_3360.npy episode_4361.npy episode_5362.npy episode_6363.npy episode_7364.npy episode_8365.npy episode_9366.npy\r\nepisode_135.npy episode_2360.npy episode_3361.npy episode_4362.npy episode_5363.npy episode_6364.npy episode_7365.npy episode_8366.npy episode_9367.npy\r\nepisode_1360.npy episode_2361.npy episode_3362.npy episode_4363.npy episode_5364.npy episode_6365.npy episode_7366.npy episode_8367.npy episode_9368.npy\r\nepisode_1361.npy episode_2362.npy episode_3363.npy episode_4364.npy episode_5365.npy episode_6366.npy episode_7367.npy episode_8368.npy episode_9369.npy\r\nepisode_1362.npy episode_2363.npy episode_3364.npy episode_4365.npy episode_5366.npy episode_6367.npy episode_7368.npy episode_8369.npy episode_936.npy\r\nepisode_1363.npy episode_2364.npy episode_3365.npy episode_4366.npy episode_5367.npy episode_6368.npy episode_7369.npy episode_836.npy episode_9370.npy\r\nepisode_1364.npy episode_2365.npy episode_3366.npy episode_4367.npy episode_5368.npy episode_6369.npy episode_736.npy episode_8370.npy episode_9371.npy\r\nepisode_1365.npy episode_2366.npy episode_3367.npy episode_4368.npy episode_5369.npy episode_636.npy episode_7370.npy episode_8371.npy episode_9372.npy\r\nepisode_1366.npy episode_2367.npy episode_3368.npy episode_4369.npy episode_536.npy episode_6370.npy episode_7371.npy episode_8372.npy episode_9373.npy\r\nepisode_1367.npy episode_2368.npy episode_3369.npy episode_436.npy episode_5370.npy episode_6371.npy episode_7372.npy episode_8373.npy episode_9374.npy\r\nepisode_1368.npy episode_2369.npy episode_336.npy episode_4370.npy episode_5371.npy episode_6372.npy episode_7373.npy episode_8374.npy episode_9375.npy\r\nepisode_1369.npy episode_236.npy episode_3370.npy episode_4371.npy episode_5372.npy episode_6373.npy episode_7374.npy episode_8375.npy episode_9376.npy\r\nepisode_136.npy episode_2370.npy episode_3371.npy episode_4372.npy episode_5373.npy episode_6374.npy episode_7375.npy episode_8376.npy episode_9377.npy\r\nepisode_1370.npy episode_2371.npy episode_3372.npy episode_4373.npy episode_5374.npy episode_6375.npy episode_7376.npy episode_8377.npy episode_9378.npy\r\nepisode_1371.npy episode_2372.npy episode_3373.npy episode_4374.npy episode_5375.npy episode_6376.npy episode_7377.npy episode_8378.npy episode_9379.npy\r\nepisode_1372.npy episode_2373.npy episode_3374.npy episode_4375.npy episode_5376.npy episode_6377.npy episode_7378.npy episode_8379.npy episode_937.npy\r\nepisode_1373.npy episode_2374.npy episode_3375.npy episode_4376.npy episode_5377.npy episode_6378.npy episode_7379.npy episode_837.npy episode_9380.npy\r\nepisode_1374.npy episode_2375.npy episode_3376.npy episode_4377.npy episode_5378.npy episode_6379.npy episode_737.npy episode_8380.npy episode_9381.npy\r\nepisode_1375.npy episode_2376.npy episode_3377.npy episode_4378.npy episode_5379.npy episode_637.npy episode_7380.npy episode_8381.npy episode_9382.npy\r\nepisode_1376.npy episode_2377.npy episode_3378.npy episode_4379.npy episode_537.npy episode_6380.npy episode_7381.npy episode_8382.npy episode_9383.npy\r\nepisode_1377.npy episode_2378.npy episode_3379.npy episode_437.npy episode_5380.npy episode_6381.npy episode_7382.npy episode_8383.npy episode_9384.npy\r\nepisode_1378.npy episode_2379.npy episode_337.npy episode_4380.npy episode_5381.npy episode_6382.npy episode_7383.npy episode_8384.npy episode_9385.npy\r\nepisode_1379.npy episode_237.npy episode_3380.npy episode_4381.npy episode_5382.npy episode_6383.npy episode_7384.npy episode_8385.npy episode_9386.npy\r\nepisode_137.npy episode_2380.npy episode_3381.npy episode_4382.npy episode_5383.npy episode_6384.npy episode_7385.npy episode_8386.npy episode_9387.npy\r\nepisode_1380.npy episode_2381.npy episode_3382.npy episode_4383.npy episode_5384.npy episode_6385.npy episode_7386.npy episode_8387.npy episode_9388.npy\r\nepisode_1381.npy episode_2382.npy episode_3383.npy episode_4384.npy episode_5385.npy episode_6386.npy episode_7387.npy episode_8388.npy episode_9389.npy\r\nepisode_1382.npy episode_2383.npy episode_3384.npy episode_4385.npy episode_5386.npy episode_6387.npy episode_7388.npy episode_8389.npy episode_938.npy\r\nepisode_1383.npy episode_2384.npy episode_3385.npy episode_4386.npy episode_5387.npy episode_6388.npy episode_7389.npy episode_838.npy episode_9390.npy\r\nepisode_1384.npy episode_2385.npy episode_3386.npy episode_4387.npy episode_5388.npy episode_6389.npy episode_738.npy episode_8390.npy episode_9391.npy\r\nepisode_1385.npy episode_2386.npy episode_3387.npy episode_4388.npy episode_5389.npy episode_638.npy episode_7390.npy episode_8391.npy episode_9392.npy\r\nepisode_1386.npy episode_2387.npy episode_3388.npy episode_4389.npy episode_538.npy episode_6390.npy episode_7391.npy episode_8392.npy episode_9393.npy\r\nepisode_1387.npy episode_2388.npy episode_3389.npy episode_438.npy episode_5390.npy episode_6391.npy episode_7392.npy episode_8393.npy episode_9394.npy\r\nepisode_1388.npy episode_2389.npy episode_338.npy episode_4390.npy episode_5391.npy episode_6392.npy episode_7393.npy episode_8394.npy episode_9395.npy\r\nepisode_1389.npy episode_238.npy episode_3390.npy episode_4391.npy episode_5392.npy episode_6393.npy episode_7394.npy episode_8395.npy episode_9396.npy\r\nepisode_138.npy episode_2390.npy episode_3391.npy episode_4392.npy episode_5393.npy episode_6394.npy episode_7395.npy episode_8396.npy episode_9397.npy\r\nepisode_1390.npy episode_2391.npy episode_3392.npy episode_4393.npy episode_5394.npy episode_6395.npy episode_7396.npy episode_8397.npy episode_9398.npy\r\nepisode_1391.npy episode_2392.npy episode_3393.npy episode_4394.npy episode_5395.npy episode_6396.npy episode_7397.npy episode_8398.npy episode_9399.npy\r\nepisode_1392.npy episode_2393.npy episode_3394.npy episode_4395.npy episode_5396.npy episode_6397.npy episode_7398.npy episode_8399.npy episode_939.npy\r\nepisode_1393.npy episode_2394.npy episode_3395.npy episode_4396.npy episode_5397.npy episode_6398.npy episode_7399.npy episode_839.npy episode_93.npy\r\nepisode_1394.npy episode_2395.npy episode_3396.npy episode_4397.npy episode_5398.npy episode_6399.npy episode_739.npy episode_83.npy episode_9400.npy\r\nepisode_1395.npy episode_2396.npy episode_3397.npy episode_4398.npy episode_5399.npy episode_639.npy episode_73.npy episode_8400.npy episode_9401.npy\r\nepisode_1396.npy episode_2397.npy episode_3398.npy episode_4399.npy episode_539.npy episode_63.npy episode_7400.npy episode_8401.npy episode_9402.npy\r\nepisode_1397.npy episode_2398.npy episode_3399.npy episode_439.npy episode_53.npy episode_6400.npy episode_7401.npy episode_8402.npy episode_9403.npy\r\nepisode_1398.npy episode_2399.npy episode_339.npy episode_43.npy episode_5400.npy episode_6401.npy episode_7402.npy episode_8403.npy episode_9404.npy\r\nepisode_1399.npy episode_239.npy episode_33.npy episode_4400.npy episode_5401.npy episode_6402.npy episode_7403.npy episode_8404.npy episode_9405.npy\r\nepisode_139.npy episode_23.npy episode_3400.npy episode_4401.npy episode_5402.npy episode_6403.npy episode_7404.npy episode_8405.npy episode_9406.npy\r\nepisode_13.npy episode_2400.npy episode_3401.npy episode_4402.npy episode_5403.npy episode_6404.npy episode_7405.npy episode_8406.npy episode_9407.npy\r\nepisode_1400.npy episode_2401.npy episode_3402.npy episode_4403.npy episode_5404.npy episode_6405.npy episode_7406.npy episode_8407.npy episode_9408.npy\r\nepisode_1401.npy episode_2402.npy episode_3403.npy episode_4404.npy episode_5405.npy episode_6406.npy episode_7407.npy episode_8408.npy episode_9409.npy\r\nepisode_1402.npy episode_2403.npy episode_3404.npy episode_4405.npy episode_5406.npy episode_6407.npy episode_7408.npy episode_8409.npy episode_940.npy\r\nepisode_1403.npy episode_2404.npy episode_3405.npy episode_4406.npy episode_5407.npy episode_6408.npy episode_7409.npy episode_840.npy episode_9410.npy\r\nepisode_1404.npy episode_2405.npy episode_3406.npy episode_4407.npy episode_5408.npy episode_6409.npy episode_740.npy episode_8410.npy episode_9411.npy\r\nepisode_1405.npy episode_2406.npy episode_3407.npy episode_4408.npy episode_5409.npy episode_640.npy episode_7410.npy episode_8411.npy episode_9412.npy\r\nepisode_1406.npy episode_2407.npy episode_3408.npy episode_4409.npy episode_540.npy episode_6410.npy episode_7411.npy episode_8412.npy episode_9413.npy\r\nepisode_1407.npy episode_2408.npy episode_3409.npy episode_440.npy episode_5410.npy episode_6411.npy episode_7412.npy episode_8413.npy episode_9414.npy\r\nepisode_1408.npy episode_2409.npy episode_340.npy episode_4410.npy episode_5411.npy episode_6412.npy episode_7413.npy episode_8414.npy episode_9415.npy\r\nepisode_1409.npy episode_240.npy episode_3410.npy episode_4411.npy episode_5412.npy episode_6413.npy episode_7414.npy episode_8415.npy episode_9416.npy\r\nepisode_140.npy episode_2410.npy episode_3411.npy episode_4412.npy episode_5413.npy episode_6414.npy episode_7415.npy episode_8416.npy episode_9417.npy\r\nepisode_1410.npy episode_2411.npy episode_3412.npy episode_4413.npy episode_5414.npy episode_6415.npy episode_7416.npy episode_8417.npy episode_9418.npy\r\nepisode_1411.npy episode_2412.npy episode_3413.npy episode_4414.npy episode_5415.npy episode_6416.npy episode_7417.npy episode_8418.npy episode_9419.npy\r\nepisode_1412.npy episode_2413.npy episode_3414.npy episode_4415.npy episode_5416.npy episode_6417.npy episode_7418.npy episode_8419.npy episode_941.npy\r\nepisode_1413.npy episode_2414.npy episode_3415.npy episode_4416.npy episode_5417.npy episode_6418.npy episode_7419.npy episode_841.npy episode_9420.npy\r\nepisode_1414.npy episode_2415.npy episode_3416.npy episode_4417.npy episode_5418.npy episode_6419.npy episode_741.npy episode_8420.npy episode_9421.npy\r\nepisode_1415.npy episode_2416.npy episode_3417.npy episode_4418.npy episode_5419.npy episode_641.npy episode_7420.npy episode_8421.npy episode_9422.npy\r\nepisode_1416.npy episode_2417.npy episode_3418.npy episode_4419.npy episode_541.npy episode_6420.npy episode_7421.npy episode_8422.npy episode_9423.npy\r\nepisode_1417.npy episode_2418.npy episode_3419.npy episode_441.npy episode_5420.npy episode_6421.npy episode_7422.npy episode_8423.npy episode_9424.npy\r\nepisode_1418.npy episode_2419.npy episode_341.npy episode_4420.npy episode_5421.npy episode_6422.npy episode_7423.npy episode_8424.npy episode_9425.npy\r\nepisode_1419.npy episode_241.npy episode_3420.npy episode_4421.npy episode_5422.npy episode_6423.npy episode_7424.npy episode_8425.npy episode_9426.npy\r\nepisode_141.npy episode_2420.npy episode_3421.npy episode_4422.npy episode_5423.npy episode_6424.npy episode_7425.npy episode_8426.npy episode_9427.npy\r\nepisode_1420.npy episode_2421.npy episode_3422.npy episode_4423.npy episode_5424.npy episode_6425.npy episode_7426.npy episode_8427.npy episode_9428.npy\r\nepisode_1421.npy episode_2422.npy episode_3423.npy episode_4424.npy episode_5425.npy episode_6426.npy episode_7427.npy episode_8428.npy episode_9429.npy\r\nepisode_1422.npy episode_2423.npy episode_3424.npy episode_4425.npy episode_5426.npy episode_6427.npy episode_7428.npy episode_8429.npy episode_942.npy\r\nepisode_1423.npy episode_2424.npy episode_3425.npy episode_4426.npy episode_5427.npy episode_6428.npy episode_7429.npy episode_842.npy episode_9430.npy\r\nepisode_1424.npy episode_2425.npy episode_3426.npy episode_4427.npy episode_5428.npy episode_6429.npy episode_742.npy episode_8430.npy episode_9431.npy\r\nepisode_1425.npy episode_2426.npy episode_3427.npy episode_4428.npy episode_5429.npy episode_642.npy episode_7430.npy episode_8431.npy episode_9432.npy\r\nepisode_1426.npy episode_2427.npy episode_3428.npy episode_4429.npy episode_542.npy episode_6430.npy episode_7431.npy episode_8432.npy episode_9433.npy\r\nepisode_1427.npy episode_2428.npy episode_3429.npy episode_442.npy episode_5430.npy episode_6431.npy episode_7432.npy episode_8433.npy episode_9434.npy\r\nepisode_1428.npy episode_2429.npy episode_342.npy episode_4430.npy episode_5431.npy episode_6432.npy episode_7433.npy episode_8434.npy episode_9435.npy\r\nepisode_1429.npy episode_242.npy episode_3430.npy episode_4431.npy episode_5432.npy episode_6433.npy episode_7434.npy episode_8435.npy episode_9436.npy\r\nepisode_142.npy episode_2430.npy episode_3431.npy episode_4432.npy episode_5433.npy episode_6434.npy episode_7435.npy episode_8436.npy episode_9437.npy\r\nepisode_1430.npy episode_2431.npy episode_3432.npy episode_4433.npy episode_5434.npy episode_6435.npy episode_7436.npy episode_8437.npy episode_9438.npy\r\nepisode_1431.npy episode_2432.npy episode_3433.npy episode_4434.npy episode_5435.npy episode_6436.npy episode_7437.npy episode_8438.npy episode_9439.npy\r\nepisode_1432.npy episode_2433.npy episode_3434.npy episode_4435.npy episode_5436.npy episode_6437.npy episode_7438.npy episode_8439.npy episode_943.npy\r\nepisode_1433.npy episode_2434.npy episode_3435.npy episode_4436.npy episode_5437.npy episode_6438.npy episode_7439.npy episode_843.npy episode_9440.npy\r\nepisode_1434.npy episode_2435.npy episode_3436.npy episode_4437.npy episode_5438.npy episode_6439.npy episode_743.npy episode_8440.npy episode_9441.npy\r\nepisode_1435.npy episode_2436.npy episode_3437.npy episode_4438.npy episode_5439.npy episode_643.npy episode_7440.npy episode_8441.npy episode_9442.npy\r\nepisode_1436.npy episode_2437.npy episode_3438.npy episode_4439.npy episode_543.npy episode_6440.npy episode_7441.npy episode_8442.npy episode_9443.npy\r\nepisode_1437.npy episode_2438.npy episode_3439.npy episode_443.npy episode_5440.npy episode_6441.npy episode_7442.npy episode_8443.npy episode_9444.npy\r\nepisode_1438.npy episode_2439.npy episode_343.npy episode_4440.npy episode_5441.npy episode_6442.npy episode_7443.npy episode_8444.npy episode_9445.npy\r\nepisode_1439.npy episode_243.npy episode_3440.npy episode_4441.npy episode_5442.npy episode_6443.npy episode_7444.npy episode_8445.npy episode_9446.npy\r\nepisode_143.npy episode_2440.npy episode_3441.npy episode_4442.npy episode_5443.npy episode_6444.npy episode_7445.npy episode_8446.npy episode_9447.npy\r\nepisode_1440.npy episode_2441.npy episode_3442.npy episode_4443.npy episode_5444.npy episode_6445.npy episode_7446.npy episode_8447.npy episode_9448.npy\r\nepisode_1441.npy episode_2442.npy episode_3443.npy episode_4444.npy episode_5445.npy episode_6446.npy episode_7447.npy episode_8448.npy episode_9449.npy\r\nepisode_1442.npy episode_2443.npy episode_3444.npy episode_4445.npy episode_5446.npy episode_6447.npy episode_7448.npy episode_8449.npy episode_944.npy\r\nepisode_1443.npy episode_2444.npy episode_3445.npy episode_4446.npy episode_5447.npy episode_6448.npy episode_7449.npy episode_844.npy episode_9450.npy\r\nepisode_1444.npy episode_2445.npy episode_3446.npy episode_4447.npy episode_5448.npy episode_6449.npy episode_744.npy episode_8450.npy episode_9451.npy\r\nepisode_1445.npy episode_2446.npy episode_3447.npy episode_4448.npy episode_5449.npy episode_644.npy episode_7450.npy episode_8451.npy episode_9452.npy\r\nepisode_1446.npy episode_2447.npy episode_3448.npy episode_4449.npy episode_544.npy episode_6450.npy episode_7451.npy episode_8452.npy episode_9453.npy\r\nepisode_1447.npy episode_2448.npy episode_3449.npy episode_444.npy episode_5450.npy episode_6451.npy episode_7452.npy episode_8453.npy episode_9454.npy\r\nepisode_1448.npy episode_2449.npy episode_344.npy episode_4450.npy episode_5451.npy episode_6452.npy episode_7453.npy episode_8454.npy episode_9455.npy\r\nepisode_1449.npy episode_244.npy episode_3450.npy episode_4451.npy episode_5452.npy episode_6453.npy episode_7454.npy episode_8455.npy episode_9456.npy\r\nepisode_144.npy episode_2450.npy episode_3451.npy episode_4452.npy episode_5453.npy episode_6454.npy episode_7455.npy episode_8456.npy episode_9457.npy\r\nepisode_1450.npy episode_2451.npy episode_3452.npy episode_4453.npy episode_5454.npy episode_6455.npy episode_7456.npy episode_8457.npy episode_9458.npy\r\nepisode_1451.npy episode_2452.npy episode_3453.npy episode_4454.npy episode_5455.npy episode_6456.npy episode_7457.npy episode_8458.npy episode_9459.npy\r\nepisode_1452.npy episode_2453.npy episode_3454.npy episode_4455.npy episode_5456.npy episode_6457.npy episode_7458.npy episode_8459.npy episode_945.npy\r\nepisode_1453.npy episode_2454.npy episode_3455.npy episode_4456.npy episode_5457.npy episode_6458.npy episode_7459.npy episode_845.npy episode_9460.npy\r\nepisode_1454.npy episode_2455.npy episode_3456.npy episode_4457.npy episode_5458.npy episode_6459.npy episode_745.npy episode_8460.npy episode_9461.npy\r\nepisode_1455.npy episode_2456.npy episode_3457.npy episode_4458.npy episode_5459.npy episode_645.npy episode_7460.npy episode_8461.npy episode_9462.npy\r\nepisode_1456.npy episode_2457.npy episode_3458.npy episode_4459.npy episode_545.npy episode_6460.npy episode_7461.npy episode_8462.npy episode_9463.npy\r\nepisode_1457.npy episode_2458.npy episode_3459.npy episode_445.npy episode_5460.npy episode_6461.npy episode_7462.npy episode_8463.npy episode_9464.npy\r\nepisode_1458.npy episode_2459.npy episode_345.npy episode_4460.npy episode_5461.npy episode_6462.npy episode_7463.npy episode_8464.npy episode_9465.npy\r\nepisode_1459.npy episode_245.npy episode_3460.npy episode_4461.npy episode_5462.npy episode_6463.npy episode_7464.npy episode_8465.npy episode_9466.npy\r\nepisode_145.npy episode_2460.npy episode_3461.npy episode_4462.npy episode_5463.npy episode_6464.npy episode_7465.npy episode_8466.npy episode_9467.npy\r\nepisode_1460.npy episode_2461.npy episode_3462.npy episode_4463.npy episode_5464.npy episode_6465.npy episode_7466.npy episode_8467.npy episode_9468.npy\r\nepisode_1461.npy episode_2462.npy episode_3463.npy episode_4464.npy episode_5465.npy episode_6466.npy episode_7467.npy episode_8468.npy episode_9469.npy\r\nepisode_1462.npy episode_2463.npy episode_3464.npy episode_4465.npy episode_5466.npy episode_6467.npy episode_7468.npy episode_8469.npy episode_946.npy\r\nepisode_1463.npy episode_2464.npy episode_3465.npy episode_4466.npy episode_5467.npy episode_6468.npy episode_7469.npy episode_846.npy episode_9470.npy\r\nepisode_1464.npy episode_2465.npy episode_3466.npy episode_4467.npy episode_5468.npy episode_6469.npy episode_746.npy episode_8470.npy episode_9471.npy\r\nepisode_1465.npy episode_2466.npy episode_3467.npy episode_4468.npy episode_5469.npy episode_646.npy episode_7470.npy episode_8471.npy episode_9472.npy\r\nepisode_1466.npy episode_2467.npy episode_3468.npy episode_4469.npy episode_546.npy episode_6470.npy episode_7471.npy episode_8472.npy episode_9473.npy\r\nepisode_1467.npy episode_2468.npy episode_3469.npy episode_446.npy episode_5470.npy episode_6471.npy episode_7472.npy episode_8473.npy episode_9474.npy\r\nepisode_1468.npy episode_2469.npy episode_346.npy episode_4470.npy episode_5471.npy episode_6472.npy episode_7473.npy episode_8474.npy episode_9475.npy\r\nepisode_1469.npy episode_246.npy episode_3470.npy episode_4471.npy episode_5472.npy episode_6473.npy episode_7474.npy episode_8475.npy episode_9476.npy\r\nepisode_146.npy episode_2470.npy episode_3471.npy episode_4472.npy episode_5473.npy episode_6474.npy episode_7475.npy episode_8476.npy episode_9477.npy\r\nepisode_1470.npy episode_2471.npy episode_3472.npy episode_4473.npy episode_5474.npy episode_6475.npy episode_7476.npy episode_8477.npy episode_9478.npy\r\nepisode_1471.npy episode_2472.npy episode_3473.npy episode_4474.npy episode_5475.npy episode_6476.npy episode_7477.npy episode_8478.npy episode_9479.npy\r\nepisode_1472.npy episode_2473.npy episode_3474.npy episode_4475.npy episode_5476.npy episode_6477.npy episode_7478.npy episode_8479.npy episode_947.npy\r\nepisode_1473.npy episode_2474.npy episode_3475.npy episode_4476.npy episode_5477.npy episode_6478.npy episode_7479.npy episode_847.npy episode_9480.npy\r\nepisode_1474.npy episode_2475.npy episode_3476.npy episode_4477.npy episode_5478.npy episode_6479.npy episode_747.npy episode_8480.npy episode_9481.npy\r\nepisode_1475.npy episode_2476.npy episode_3477.npy episode_4478.npy episode_5479.npy episode_647.npy episode_7480.npy episode_8481.npy episode_9482.npy\r\nepisode_1476.npy episode_2477.npy episode_3478.npy episode_4479.npy episode_547.npy episode_6480.npy episode_7481.npy episode_8482.npy episode_9483.npy\r\nepisode_1477.npy episode_2478.npy episode_3479.npy episode_447.npy episode_5480.npy episode_6481.npy episode_7482.npy episode_8483.npy episode_9484.npy\r\nepisode_1478.npy episode_2479.npy episode_347.npy episode_4480.npy episode_5481.npy episode_6482.npy episode_7483.npy episode_8484.npy episode_9485.npy\r\nepisode_1479.npy episode_247.npy episode_3480.npy episode_4481.npy episode_5482.npy episode_6483.npy episode_7484.npy episode_8485.npy episode_9486.npy\r\nepisode_147.npy episode_2480.npy episode_3481.npy episode_4482.npy episode_5483.npy episode_6484.npy episode_7485.npy episode_8486.npy episode_9487.npy\r\nepisode_1480.npy episode_2481.npy episode_3482.npy episode_4483.npy episode_5484.npy episode_6485.npy episode_7486.npy episode_8487.npy episode_9488.npy\r\nepisode_1481.npy episode_2482.npy episode_3483.npy episode_4484.npy episode_5485.npy episode_6486.npy episode_7487.npy episode_8488.npy episode_9489.npy\r\nepisode_1482.npy episode_2483.npy episode_3484.npy episode_4485.npy episode_5486.npy episode_6487.npy episode_7488.npy episode_8489.npy episode_948.npy\r\nepisode_1483.npy episode_2484.npy episode_3485.npy episode_4486.npy episode_5487.npy episode_6488.npy episode_7489.npy episode_848.npy episode_9490.npy\r\nepisode_1484.npy episode_2485.npy episode_3486.npy episode_4487.npy episode_5488.npy episode_6489.npy episode_748.npy episode_8490.npy episode_9491.npy\r\nepisode_1485.npy episode_2486.npy episode_3487.npy episode_4488.npy episode_5489.npy episode_648.npy episode_7490.npy episode_8491.npy episode_9492.npy\r\nepisode_1486.npy episode_2487.npy episode_3488.npy episode_4489.npy episode_548.npy episode_6490.npy episode_7491.npy episode_8492.npy episode_9493.npy\r\nepisode_1487.npy episode_2488.npy episode_3489.npy episode_448.npy episode_5490.npy episode_6491.npy episode_7492.npy episode_8493.npy episode_9494.npy\r\nepisode_1488.npy episode_2489.npy episode_348.npy episode_4490.npy episode_5491.npy episode_6492.npy episode_7493.npy episode_8494.npy episode_9495.npy\r\nepisode_1489.npy episode_248.npy episode_3490.npy episode_4491.npy episode_5492.npy episode_6493.npy episode_7494.npy episode_8495.npy episode_9496.npy\r\nepisode_148.npy episode_2490.npy episode_3491.npy episode_4492.npy episode_5493.npy episode_6494.npy episode_7495.npy episode_8496.npy episode_9497.npy\r\nepisode_1490.npy episode_2491.npy episode_3492.npy episode_4493.npy episode_5494.npy episode_6495.npy episode_7496.npy episode_8497.npy episode_9498.npy\r\nepisode_1491.npy episode_2492.npy episode_3493.npy episode_4494.npy episode_5495.npy episode_6496.npy episode_7497.npy episode_8498.npy episode_9499.npy\r\nepisode_1492.npy episode_2493.npy episode_3494.npy episode_4495.npy episode_5496.npy episode_6497.npy episode_7498.npy episode_8499.npy episode_949.npy\r\nepisode_1493.npy episode_2494.npy episode_3495.npy episode_4496.npy episode_5497.npy episode_6498.npy episode_7499.npy episode_849.npy episode_94.npy\r\nepisode_1494.npy episode_2495.npy episode_3496.npy episode_4497.npy episode_5498.npy episode_6499.npy episode_749.npy episode_84.npy episode_9500.npy\r\nepisode_1495.npy episode_2496.npy episode_3497.npy episode_4498.npy episode_5499.npy episode_649.npy episode_74.npy episode_8500.npy episode_9501.npy\r\nepisode_1496.npy episode_2497.npy episode_3498.npy episode_4499.npy episode_549.npy episode_64.npy episode_7500.npy episode_8501.npy episode_9502.npy\r\nepisode_1497.npy episode_2498.npy episode_3499.npy episode_449.npy episode_54.npy episode_6500.npy episode_7501.npy episode_8502.npy episode_9503.npy\r\nepisode_1498.npy episode_2499.npy episode_349.npy episode_44.npy episode_5500.npy episode_6501.npy episode_7502.npy episode_8503.npy episode_9504.npy\r\nepisode_1499.npy episode_249.npy episode_34.npy episode_4500.npy episode_5501.npy episode_6502.npy episode_7503.npy episode_8504.npy episode_9505.npy\r\nepisode_149.npy episode_24.npy episode_3500.npy episode_4501.npy episode_5502.npy episode_6503.npy episode_7504.npy episode_8505.npy episode_9506.npy\r\nepisode_14.npy episode_2500.npy episode_3501.npy episode_4502.npy episode_5503.npy episode_6504.npy episode_7505.npy episode_8506.npy episode_9507.npy\r\nepisode_1500.npy episode_2501.npy episode_3502.npy episode_4503.npy episode_5504.npy episode_6505.npy episode_7506.npy episode_8507.npy episode_9508.npy\r\nepisode_1501.npy episode_2502.npy episode_3503.npy episode_4504.npy episode_5505.npy episode_6506.npy episode_7507.npy episode_8508.npy episode_9509.npy\r\nepisode_1502.npy episode_2503.npy episode_3504.npy episode_4505.npy episode_5506.npy episode_6507.npy episode_7508.npy episode_8509.npy episode_950.npy\r\nepisode_1503.npy episode_2504.npy episode_3505.npy episode_4506.npy episode_5507.npy episode_6508.npy episode_7509.npy episode_850.npy episode_9510.npy\r\nepisode_1504.npy episode_2505.npy episode_3506.npy episode_4507.npy episode_5508.npy episode_6509.npy episode_750.npy episode_8510.npy episode_9511.npy\r\nepisode_1505.npy episode_2506.npy episode_3507.npy episode_4508.npy episode_5509.npy episode_650.npy episode_7510.npy episode_8511.npy episode_9512.npy\r\nepisode_1506.npy episode_2507.npy episode_3508.npy episode_4509.npy episode_550.npy episode_6510.npy episode_7511.npy episode_8512.npy episode_9513.npy\r\nepisode_1507.npy episode_2508.npy episode_3509.npy episode_450.npy episode_5510.npy episode_6511.npy episode_7512.npy episode_8513.npy episode_9514.npy\r\nepisode_1508.npy episode_2509.npy episode_350.npy episode_4510.npy episode_5511.npy episode_6512.npy episode_7513.npy episode_8514.npy episode_9515.npy\r\nepisode_1509.npy episode_250.npy episode_3510.npy episode_4511.npy episode_5512.npy episode_6513.npy episode_7514.npy episode_8515.npy episode_9516.npy\r\nepisode_150.npy episode_2510.npy episode_3511.npy episode_4512.npy episode_5513.npy episode_6514.npy episode_7515.npy episode_8516.npy episode_9517.npy\r\nepisode_1510.npy episode_2511.npy episode_3512.npy episode_4513.npy episode_5514.npy episode_6515.npy episode_7516.npy episode_8517.npy episode_9518.npy\r\nepisode_1511.npy episode_2512.npy episode_3513.npy episode_4514.npy episode_5515.npy episode_6516.npy episode_7517.npy episode_8518.npy episode_9519.npy\r\nepisode_1512.npy episode_2513.npy episode_3514.npy episode_4515.npy episode_5516.npy episode_6517.npy episode_7518.npy episode_8519.npy episode_951.npy\r\nepisode_1513.npy episode_2514.npy episode_3515.npy episode_4516.npy episode_5517.npy episode_6518.npy episode_7519.npy episode_851.npy episode_9520.npy\r\nepisode_1514.npy episode_2515.npy episode_3516.npy episode_4517.npy episode_5518.npy episode_6519.npy episode_751.npy episode_8520.npy episode_9521.npy\r\nepisode_1515.npy episode_2516.npy episode_3517.npy episode_4518.npy episode_5519.npy episode_651.npy episode_7520.npy episode_8521.npy episode_9522.npy\r\nepisode_1516.npy episode_2517.npy episode_3518.npy episode_4519.npy episode_551.npy episode_6520.npy episode_7521.npy episode_8522.npy episode_9523.npy\r\nepisode_1517.npy episode_2518.npy episode_3519.npy episode_451.npy episode_5520.npy episode_6521.npy episode_7522.npy episode_8523.npy episode_9524.npy\r\nepisode_1518.npy episode_2519.npy episode_351.npy episode_4520.npy episode_5521.npy episode_6522.npy episode_7523.npy episode_8524.npy episode_9525.npy\r\nepisode_1519.npy episode_251.npy episode_3520.npy episode_4521.npy episode_5522.npy episode_6523.npy episode_7524.npy episode_8525.npy episode_9526.npy\r\nepisode_151.npy episode_2520.npy episode_3521.npy episode_4522.npy episode_5523.npy episode_6524.npy episode_7525.npy episode_8526.npy episode_9527.npy\r\nepisode_1520.npy episode_2521.npy episode_3522.npy episode_4523.npy episode_5524.npy episode_6525.npy episode_7526.npy episode_8527.npy episode_9528.npy\r\nepisode_1521.npy episode_2522.npy episode_3523.npy episode_4524.npy episode_5525.npy episode_6526.npy episode_7527.npy episode_8528.npy episode_9529.npy\r\nepisode_1522.npy episode_2523.npy episode_3524.npy episode_4525.npy episode_5526.npy episode_6527.npy episode_7528.npy episode_8529.npy episode_952.npy\r\nepisode_1523.npy episode_2524.npy episode_3525.npy episode_4526.npy episode_5527.npy episode_6528.npy episode_7529.npy episode_852.npy episode_9530.npy\r\nepisode_1524.npy episode_2525.npy episode_3526.npy episode_4527.npy episode_5528.npy episode_6529.npy episode_752.npy episode_8530.npy episode_9531.npy\r\nepisode_1525.npy episode_2526.npy episode_3527.npy episode_4528.npy episode_5529.npy episode_652.npy episode_7530.npy episode_8531.npy episode_9532.npy\r\nepisode_1526.npy episode_2527.npy episode_3528.npy episode_4529.npy episode_552.npy episode_6530.npy episode_7531.npy episode_8532.npy episode_9533.npy\r\nepisode_1527.npy episode_2528.npy episode_3529.npy episode_452.npy episode_5530.npy episode_6531.npy episode_7532.npy episode_8533.npy episode_9534.npy\r\nepisode_1528.npy episode_2529.npy episode_352.npy episode_4530.npy episode_5531.npy episode_6532.npy episode_7533.npy episode_8534.npy episode_9535.npy\r\nepisode_1529.npy episode_252.npy episode_3530.npy episode_4531.npy episode_5532.npy episode_6533.npy episode_7534.npy episode_8535.npy episode_9536.npy\r\nepisode_152.npy episode_2530.npy episode_3531.npy episode_4532.npy episode_5533.npy episode_6534.npy episode_7535.npy episode_8536.npy episode_9537.npy\r\nepisode_1530.npy episode_2531.npy episode_3532.npy episode_4533.npy episode_5534.npy episode_6535.npy episode_7536.npy episode_8537.npy episode_9538.npy\r\nepisode_1531.npy episode_2532.npy episode_3533.npy episode_4534.npy episode_5535.npy episode_6536.npy episode_7537.npy episode_8538.npy episode_9539.npy\r\nepisode_1532.npy episode_2533.npy episode_3534.npy episode_4535.npy episode_5536.npy episode_6537.npy episode_7538.npy episode_8539.npy episode_953.npy\r\nepisode_1533.npy episode_2534.npy episode_3535.npy episode_4536.npy episode_5537.npy episode_6538.npy episode_7539.npy episode_853.npy episode_9540.npy\r\nepisode_1534.npy episode_2535.npy episode_3536.npy episode_4537.npy episode_5538.npy episode_6539.npy episode_753.npy episode_8540.npy episode_9541.npy\r\nepisode_1535.npy episode_2536.npy episode_3537.npy episode_4538.npy episode_5539.npy episode_653.npy episode_7540.npy episode_8541.npy episode_9542.npy\r\nepisode_1536.npy episode_2537.npy episode_3538.npy episode_4539.npy episode_553.npy episode_6540.npy episode_7541.npy episode_8542.npy episode_9543.npy\r\nepisode_1537.npy episode_2538.npy episode_3539.npy episode_453.npy episode_5540.npy episode_6541.npy episode_7542.npy episode_8543.npy episode_9544.npy\r\nepisode_1538.npy episode_2539.npy episode_353.npy episode_4540.npy episode_5541.npy episode_6542.npy episode_7543.npy episode_8544.npy episode_9545.npy\r\nepisode_1539.npy episode_253.npy episode_3540.npy episode_4541.npy episode_5542.npy episode_6543.npy episode_7544.npy episode_8545.npy episode_9546.npy\r\nepisode_153.npy episode_2540.npy episode_3541.npy episode_4542.npy episode_5543.npy episode_6544.npy episode_7545.npy episode_8546.npy episode_9547.npy\r\nepisode_1540.npy episode_2541.npy episode_3542.npy episode_4543.npy episode_5544.npy episode_6545.npy episode_7546.npy episode_8547.npy episode_9548.npy\r\nepisode_1541.npy episode_2542.npy episode_3543.npy episode_4544.npy episode_5545.npy episode_6546.npy episode_7547.npy episode_8548.npy episode_9549.npy\r\nepisode_1542.npy episode_2543.npy episode_3544.npy episode_4545.npy episode_5546.npy episode_6547.npy episode_7548.npy episode_8549.npy episode_954.npy\r\nepisode_1543.npy episode_2544.npy episode_3545.npy episode_4546.npy episode_5547.npy episode_6548.npy episode_7549.npy episode_854.npy episode_9550.npy\r\nepisode_1544.npy episode_2545.npy episode_3546.npy episode_4547.npy episode_5548.npy episode_6549.npy episode_754.npy episode_8550.npy episode_9551.npy\r\nepisode_1545.npy episode_2546.npy episode_3547.npy episode_4548.npy episode_5549.npy episode_654.npy episode_7550.npy episode_8551.npy episode_9552.npy\r\nepisode_1546.npy episode_2547.npy episode_3548.npy episode_4549.npy episode_554.npy episode_6550.npy episode_7551.npy episode_8552.npy episode_9553.npy\r\nepisode_1547.npy episode_2548.npy episode_3549.npy episode_454.npy episode_5550.npy episode_6551.npy episode_7552.npy episode_8553.npy episode_9554.npy\r\nepisode_1548.npy episode_2549.npy episode_354.npy episode_4550.npy episode_5551.npy episode_6552.npy episode_7553.npy episode_8554.npy episode_9555.npy\r\nepisode_1549.npy episode_254.npy episode_3550.npy episode_4551.npy episode_5552.npy episode_6553.npy episode_7554.npy episode_8555.npy episode_9556.npy\r\nepisode_154.npy episode_2550.npy episode_3551.npy episode_4552.npy episode_5553.npy episode_6554.npy episode_7555.npy episode_8556.npy episode_9557.npy\r\nepisode_1550.npy episode_2551.npy episode_3552.npy episode_4553.npy episode_5554.npy episode_6555.npy episode_7556.npy episode_8557.npy episode_9558.npy\r\nepisode_1551.npy episode_2552.npy episode_3553.npy episode_4554.npy episode_5555.npy episode_6556.npy episode_7557.npy episode_8558.npy episode_9559.npy\r\nepisode_1552.npy episode_2553.npy episode_3554.npy episode_4555.npy episode_5556.npy episode_6557.npy episode_7558.npy episode_8559.npy episode_955.npy\r\nepisode_1553.npy episode_2554.npy episode_3555.npy episode_4556.npy episode_5557.npy episode_6558.npy episode_7559.npy episode_855.npy episode_9560.npy\r\nepisode_1554.npy episode_2555.npy episode_3556.npy episode_4557.npy episode_5558.npy episode_6559.npy episode_755.npy episode_8560.npy episode_9561.npy\r\nepisode_1555.npy episode_2556.npy episode_3557.npy episode_4558.npy episode_5559.npy episode_655.npy episode_7560.npy episode_8561.npy episode_9562.npy\r\nepisode_1556.npy episode_2557.npy episode_3558.npy episode_4559.npy episode_555.npy episode_6560.npy episode_7561.npy episode_8562.npy episode_9563.npy\r\nepisode_1557.npy episode_2558.npy episode_3559.npy episode_455.npy episode_5560.npy episode_6561.npy episode_7562.npy episode_8563.npy episode_9564.npy\r\nepisode_1558.npy episode_2559.npy episode_355.npy episode_4560.npy episode_5561.npy episode_6562.npy episode_7563.npy episode_8564.npy episode_9565.npy\r\nepisode_1559.npy episode_255.npy episode_3560.npy episode_4561.npy episode_5562.npy episode_6563.npy episode_7564.npy episode_8565.npy episode_9566.npy\r\nepisode_155.npy episode_2560.npy episode_3561.npy episode_4562.npy episode_5563.npy episode_6564.npy episode_7565.npy episode_8566.npy episode_9567.npy\r\nepisode_1560.npy episode_2561.npy episode_3562.npy episode_4563.npy episode_5564.npy episode_6565.npy episode_7566.npy episode_8567.npy episode_9568.npy\r\nepisode_1561.npy episode_2562.npy episode_3563.npy episode_4564.npy episode_5565.npy episode_6566.npy episode_7567.npy episode_8568.npy episode_9569.npy\r\nepisode_1562.npy episode_2563.npy episode_3564.npy episode_4565.npy episode_5566.npy episode_6567.npy episode_7568.np",,terminal_output +463,968017,"TERMINAL",0,0,"y episode_8569.npy episode_956.npy\r\nepisode_1563.npy episode_2564.npy episode_3565.npy episode_4566.npy episode_5567.npy episode_6568.npy episode_7569.npy episode_856.npy episode_9570.npy\r\nepisode_1564.npy episode_2565.npy episode_3566.npy episode_4567.npy episode_5568.npy episode_6569.npy episode_756.npy episode_8570.npy episode_9571.npy\r\nepisode_1565.npy episode_2566.npy episode_3567.npy episode_4568.npy episode_5569.npy episode_656.npy episode_7570.npy episode_8571.npy episode_9572.npy\r\nepisode_1566.npy episode_2567.npy episode_3568.npy episode_4569.npy episode_556.npy episode_6570.npy episode_7571.npy episode_8572.npy episode_9573.npy\r\nepisode_1567.npy episode_2568.npy episode_3569.npy episode_456.npy episode_5570.npy episode_6571.npy episode_7572.npy episode_8573.npy episode_9574.npy\r\nepisode_1568.npy episode_2569.npy episode_356.npy episode_4570.npy episode_5571.npy episode_6572.npy episode_7573.npy episode_8574.npy episode_9575.npy\r\nepisode_1569.npy episode_256.npy episode_3570.npy episode_4571.npy episode_5572.npy episode_6573.npy episode_7574.npy episode_8575.npy episode_9576.npy\r\nepisode_156.npy episode_2570.npy episode_3571.npy episode_4572.npy episode_5573.npy episode_6574.npy episode_7575.npy episode_8576.npy episode_9577.npy\r\nepisode_1570.npy episode_2571.npy episode_3572.npy episode_4573.npy episode_5574.npy episode_6575.npy episode_7576.npy episode_8577.npy episode_9578.npy\r\nepisode_1571.npy episode_2572.npy episode_3573.npy episode_4574.npy episode_5575.npy episode_6576.npy episode_7577.npy episode_8578.npy episode_9579.npy\r\nepisode_1572.npy episode_2573.npy episode_3574.npy episode_4575.npy episode_5576.npy episode_6577.npy episode_7578.npy episode_8579.npy episode_957.npy\r\nepisode_1573.npy episode_2574.npy episode_3575.npy episode_4576.npy episode_5577.npy episode_6578.npy episode_7579.npy episode_857.npy episode_9580.npy\r\nepisode_1574.npy episode_2575.npy episode_3576.npy episode_4577.npy episode_5578.npy episode_6579.npy episode_757.npy episode_8580.npy episode_9581.npy\r\nepisode_1575.npy episode_2576.npy episode_3577.npy episode_4578.npy episode_5579.npy episode_657.npy episode_7580.npy episode_8581.npy episode_9582.npy\r\nepisode_1576.npy episode_2577.npy episode_3578.npy episode_4579.npy episode_557.npy episode_6580.npy episode_7581.npy episode_8582.npy episode_9583.npy\r\nepisode_1577.npy episode_2578.npy episode_3579.npy episode_457.npy episode_5580.npy episode_6581.npy episode_7582.npy episode_8583.npy episode_9584.npy\r\nepisode_1578.npy episode_2579.npy episode_357.npy episode_4580.npy episode_5581.npy episode_6582.npy episode_7583.npy episode_8584.npy episode_9585.npy\r\nepisode_1579.npy episode_257.npy episode_3580.npy episode_4581.npy episode_5582.npy episode_6583.npy episode_7584.npy episode_8585.npy episode_9586.npy\r\nepisode_157.npy episode_2580.npy episode_3581.npy episode_4582.npy episode_5583.npy episode_6584.npy episode_7585.npy episode_8586.npy episode_9587.npy\r\nepisode_1580.npy episode_2581.npy episode_3582.npy episode_4583.npy episode_5584.npy episode_6585.npy episode_7586.npy episode_8587.npy episode_9588.npy\r\nepisode_1581.npy episode_2582.npy episode_3583.npy episode_4584.npy episode_5585.npy episode_6586.npy episode_7587.npy episode_8588.npy episode_9589.npy\r\nepisode_1582.npy episode_2583.npy episode_3584.npy episode_4585.npy episode_5586.npy episode_6587.npy episode_7588.npy episode_8589.npy episode_958.npy\r\nepisode_1583.npy episode_2584.npy episode_3585.npy episode_4586.npy episode_5587.npy episode_6588.npy episode_7589.npy episode_858.npy episode_9590.npy\r\nepisode_1584.npy episode_2585.npy episode_3586.npy episode_4587.npy episode_5588.npy episode_6589.npy episode_758.npy episode_8590.npy episode_9591.npy\r\nepisode_1585.npy episode_2586.npy episode_3587.npy episode_4588.npy episode_5589.npy episode_658.npy episode_7590.npy episode_8591.npy episode_9592.npy\r\nepisode_1586.npy episode_2587.npy episode_3588.npy episode_4589.npy episode_558.npy episode_6590.npy episode_7591.npy episode_8592.npy episode_9593.npy\r\nepisode_1587.npy episode_2588.npy episode_3589.npy episode_458.npy episode_5590.npy episode_6591.npy episode_7592.npy episode_8593.npy episode_9594.npy\r\nepisode_1588.npy episode_2589.npy episode_358.npy episode_4590.npy episode_5591.npy episode_6592.npy episode_7593.npy episode_8594.npy episode_9595.npy\r\nepisode_1589.npy episode_258.npy episode_3590.npy episode_4591.npy episode_5592.npy episode_6593.npy episode_7594.npy episode_8595.npy episode_9596.npy\r\nepisode_158.npy episode_2590.npy episode_3591.npy episode_4592.npy episode_5593.npy episode_6594.npy episode_7595.npy episode_8596.npy episode_9597.npy\r\nepisode_1590.npy episode_2591.npy episode_3592.npy episode_4593.npy episode_5594.npy episode_6595.npy episode_7596.npy episode_8597.npy episode_9598.npy\r\nepisode_1591.npy episode_2592.npy episode_3593.npy episode_4594.npy episode_5595.npy episode_6596.npy episode_7597.npy episode_8598.npy episode_9599.npy\r\nepisode_1592.npy episode_2593.npy episode_3594.npy episode_4595.npy episode_5596.npy episode_6597.npy episode_7598.npy episode_8599.npy episode_959.npy\r\nepisode_1593.npy episode_2594.npy episode_3595.npy episode_4596.npy episode_5597.npy episode_6598.npy episode_7599.npy episode_859.npy episode_95.npy\r\nepisode_1594.npy episode_2595.npy episode_3596.npy episode_4597.npy episode_5598.npy episode_6599.npy episode_759.npy episode_85.npy episode_9600.npy\r\nepisode_1595.npy episode_2596.npy episode_3597.npy episode_4598.npy episode_5599.npy episode_659.npy episode_75.npy episode_8600.npy episode_9601.npy\r\nepisode_1596.npy episode_2597.npy episode_3598.npy episode_4599.npy episode_559.npy episode_65.npy episode_7600.npy episode_8601.npy episode_9602.npy\r\nepisode_1597.npy episode_2598.npy episode_3599.npy episode_459.npy episode_55.npy episode_6600.npy episode_7601.npy episode_8602.npy episode_9603.npy\r\nepisode_1598.npy episode_2599.npy episode_359.npy episode_45.npy episode_5600.npy episode_6601.npy episode_7602.npy episode_8603.npy episode_9604.npy\r\nepisode_1599.npy episode_259.npy episode_35.npy episode_4600.npy episode_5601.npy episode_6602.npy episode_7603.npy episode_8604.npy episode_9605.npy\r\nepisode_159.npy episode_25.npy episode_3600.npy episode_4601.npy episode_5602.npy episode_6603.npy episode_7604.npy episode_8605.npy episode_9606.npy\r\nepisode_15.npy episode_2600.npy episode_3601.npy episode_4602.npy episode_5603.npy episode_6604.npy episode_7605.npy episode_8606.npy episode_9607.npy\r\nepisode_1600.npy episode_2601.npy episode_3602.npy episode_4603.npy episode_5604.npy episode_6605.npy episode_7606.npy episode_8607.npy episode_9608.npy\r\nepisode_1601.npy episode_2602.npy episode_3603.npy episode_4604.npy episode_5605.npy episode_6606.npy episode_7607.npy episode_8608.npy episode_9609.npy\r\nepisode_1602.npy episode_2603.npy episode_3604.npy episode_4605.npy episode_5606.npy episode_6607.npy episode_7608.npy episode_8609.npy episode_960.npy\r\nepisode_1603.npy episode_2604.npy episode_3605.npy episode_4606.npy episode_5607.npy episode_6608.npy episode_7609.npy episode_860.npy episode_9610.npy\r\nepisode_1604.npy episode_2605.npy episode_3606.npy episode_4607.npy episode_5608.npy episode_6609.npy episode_760.npy episode_8610.npy episode_9611.npy\r\nepisode_1605.npy episode_2606.npy episode_3607.npy episode_4608.npy episode_5609.npy episode_660.npy episode_7610.npy episode_8611.npy episode_9612.npy\r\nepisode_1606.npy episode_2607.npy episode_3608.npy episode_4609.npy episode_560.npy episode_6610.npy episode_7611.npy episode_8612.npy episode_9613.npy\r\nepisode_1607.npy episode_2608.npy episode_3609.npy episode_460.npy episode_5610.npy episode_6611.npy episode_7612.npy episode_8613.npy episode_9614.npy\r\nepisode_1608.npy episode_2609.npy episode_360.npy episode_4610.npy episode_5611.npy episode_6612.npy episode_7613.npy episode_8614.npy episode_9615.npy\r\nepisode_1609.npy episode_260.npy episode_3610.npy episode_4611.npy episode_5612.npy episode_6613.npy episode_7614.npy episode_8615.npy episode_9616.npy\r\nepisode_160.npy episode_2610.npy episode_3611.npy episode_4612.npy episode_5613.npy episode_6614.npy episode_7615.npy episode_8616.npy episode_9617.npy\r\nepisode_1610.npy episode_2611.npy episode_3612.npy episode_4613.npy episode_5614.npy episode_6615.npy episode_7616.npy episode_8617.npy episode_9618.npy\r\nepisode_1611.npy episode_2612.npy episode_3613.npy episode_4614.npy episode_5615.npy episode_6616.npy episode_7617.npy episode_8618.npy episode_9619.npy\r\nepisode_1612.npy episode_2613.npy episode_3614.npy episode_4615.npy episode_5616.npy episode_6617.npy episode_7618.npy episode_8619.npy episode_961.npy\r\nepisode_1613.npy episode_2614.npy episode_3615.npy episode_4616.npy episode_5617.npy episode_6618.npy episode_7619.npy episode_861.npy episode_9620.npy\r\nepisode_1614.npy episode_2615.npy episode_3616.npy episode_4617.npy episode_5618.npy episode_6619.npy episode_761.npy episode_8620.npy episode_9621.npy\r\nepisode_1615.npy episode_2616.npy episode_3617.npy episode_4618.npy episode_5619.npy episode_661.npy episode_7620.npy episode_8621.npy episode_9622.npy\r\nepisode_1616.npy episode_2617.npy episode_3618.npy episode_4619.npy episode_561.npy episode_6620.npy episode_7621.npy episode_8622.npy episode_9623.npy\r\nepisode_1617.npy episode_2618.npy episode_3619.npy episode_461.npy episode_5620.npy episode_6621.npy episode_7622.npy episode_8623.npy episode_9624.npy\r\nepisode_1618.npy episode_2619.npy episode_361.npy episode_4620.npy episode_5621.npy episode_6622.npy episode_7623.npy episode_8624.npy episode_9625.npy\r\nepisode_1619.npy episode_261.npy episode_3620.npy episode_4621.npy episode_5622.npy episode_6623.npy episode_7624.npy episode_8625.npy episode_9626.npy\r\nepisode_161.npy episode_2620.npy episode_3621.npy episode_4622.npy episode_5623.npy episode_6624.npy episode_7625.npy episode_8626.npy episode_9627.npy\r\nepisode_1620.npy episode_2621.npy episode_3622.npy episode_4623.npy episode_5624.npy episode_6625.npy episode_7626.npy episode_8627.npy episode_9628.npy\r\nepisode_1621.npy episode_2622.npy episode_3623.npy episode_4624.npy episode_5625.npy episode_6626.npy episode_7627.npy episode_8628.npy episode_9629.npy\r\nepisode_1622.npy episode_2623.npy episode_3624.npy episode_4625.npy episode_5626.npy episode_6627.npy episode_7628.npy episode_8629.npy episode_962.npy\r\nepisode_1623.npy episode_2624.npy episode_3625.npy episode_4626.npy episode_5627.npy episode_6628.npy episode_7629.npy episode_862.npy episode_9630.npy\r\nepisode_1624.npy episode_2625.npy episode_3626.npy episode_4627.npy episode_5628.npy episode_6629.npy episode_762.npy episode_8630.npy episode_9631.npy\r\nepisode_1625.npy episode_2626.npy episode_3627.npy episode_4628.npy episode_5629.npy episode_662.npy episode_7630.npy episode_8631.npy episode_9632.npy\r\nepisode_1626.npy episode_2627.npy episode_3628.npy episode_4629.npy episode_562.npy episode_6630.npy episode_7631.npy episode_8632.npy episode_9633.npy\r\nepisode_1627.npy episode_2628.npy episode_3629.npy episode_462.npy episode_5630.npy episode_6631.npy episode_7632.npy episode_8633.npy episode_9634.npy\r\nepisode_1628.npy episode_2629.npy episode_362.npy episode_4630.npy episode_5631.npy episode_6632.npy episode_7633.npy episode_8634.npy episode_9635.npy\r\nepisode_1629.npy episode_262.npy episode_3630.npy episode_4631.npy episode_5632.npy episode_6633.npy episode_7634.npy episode_8635.npy episode_9636.npy\r\nepisode_162.npy episode_2630.npy episode_3631.npy episode_4632.npy episode_5633.npy episode_6634.npy episode_7635.npy episode_8636.npy episode_9637.npy\r\nepisode_1630.npy episode_2631.npy episode_3632.npy episode_4633.npy episode_5634.npy episode_6635.npy episode_7636.npy episode_8637.npy episode_9638.npy\r\nepisode_1631.npy episode_2632.npy episode_3633.npy episode_4634.npy episode_5635.npy episode_6636.npy episode_7637.npy episode_8638.npy episode_9639.npy\r\nepisode_1632.npy episode_2633.npy episode_3634.npy episode_4635.npy episode_5636.npy episode_6637.npy episode_7638.npy episode_8639.npy episode_963.npy\r\nepisode_1633.npy episode_2634.npy episode_3635.npy episode_4636.npy episode_5637.npy episode_6638.npy episode_7639.npy episode_863.npy episode_9640.npy\r\nepisode_1634.npy episode_2635.npy episode_3636.npy episode_4637.npy episode_5638.npy episode_6639.npy episode_763.npy episode_8640.npy episode_9641.npy\r\nepisode_1635.npy episode_2636.npy episode_3637.npy episode_4638.npy episode_5639.npy episode_663.npy episode_7640.npy episode_8641.npy episode_9642.npy\r\nepisode_1636.npy episode_2637.npy episode_3638.npy episode_4639.npy episode_563.npy episode_6640.npy episode_7641.npy episode_8642.npy episode_9643.npy\r\nepisode_1637.npy episode_2638.npy episode_3639.npy episode_463.npy episode_5640.npy episode_6641.npy episode_7642.npy episode_8643.npy episode_9644.npy\r\nepisode_1638.npy episode_2639.npy episode_363.npy episode_4640.npy episode_5641.npy episode_6642.npy episode_7643.npy episode_8644.npy episode_9645.npy\r\nepisode_1639.npy episode_263.npy episode_3640.npy episode_4641.npy episode_5642.npy episode_6643.npy episode_7644.npy episode_8645.npy episode_9646.npy\r\nepisode_163.npy episode_2640.npy episode_3641.npy episode_4642.npy episode_5643.npy episode_6644.npy episode_7645.npy episode_8646.npy episode_9647.npy\r\nepisode_1640.npy episode_2641.npy episode_3642.npy episode_4643.npy episode_5644.npy episode_6645.npy episode_7646.npy episode_8647.npy episode_9648.npy\r\nepisode_1641.npy episode_2642.npy episode_3643.npy episode_4644.npy episode_5645.npy episode_6646.npy episode_7647.npy episode_8648.npy episode_9649.npy\r\nepisode_1642.npy episode_2643.npy episode_3644.npy episode_4645.npy episode_5646.npy episode_6647.npy episode_7648.npy episode_8649.npy episode_964.npy\r\nepisode_1643.npy episode_2644.npy episode_3645.npy episode_4646.npy episode_5647.npy episode_6648.npy episode_7649.npy episode_864.npy episode_9650.npy\r\nepisode_1644.npy episode_2645.npy episode_3646.npy episode_4647.npy episode_5648.npy episode_6649.npy episode_764.npy episode_8650.npy episode_9651.npy\r\nepisode_1645.npy episode_2646.npy episode_3647.npy episode_4648.npy episode_5649.npy episode_664.npy episode_7650.npy episode_8651.npy episode_9652.npy\r\nepisode_1646.npy episode_2647.npy episode_3648.npy episode_4649.npy episode_564.npy episode_6650.npy episode_7651.npy episode_8652.npy episode_9653.npy\r\nepisode_1647.npy episode_2648.npy episode_3649.npy episode_464.npy episode_5650.npy episode_6651.npy episode_7652.npy episode_8653.npy episode_9654.npy\r\nepisode_1648.npy episode_2649.npy episode_364.npy episode_4650.npy episode_5651.npy episode_6652.npy episode_7653.npy episode_8654.npy episode_9655.npy\r\nepisode_1649.npy episode_264.npy episode_3650.npy episode_4651.npy episode_5652.npy episode_6653.npy episode_7654.npy episode_8655.npy episode_9656.npy\r\nepisode_164.npy episode_2650.npy episode_3651.npy episode_4652.npy episode_5653.npy episode_6654.npy episode_7655.npy episode_8656.npy episode_9657.npy\r\nepisode_1650.npy episode_2651.npy episode_3652.npy episode_4653.npy episode_5654.npy episode_6655.npy episode_7656.npy episode_8657.npy episode_9658.npy\r\nepisode_1651.npy episode_2652.npy episode_3653.npy episode_4654.npy episode_5655.npy episode_6656.npy episode_7657.npy episode_8658.npy episode_9659.npy\r\nepisode_1652.npy episode_2653.npy episode_3654.npy episode_4655.npy episode_5656.npy episode_6657.npy episode_7658.npy episode_8659.npy episode_965.npy\r\nepisode_1653.npy episode_2654.npy episode_3655.npy episode_4656.npy episode_5657.npy episode_6658.npy episode_7659.npy episode_865.npy episode_9660.npy\r\nepisode_1654.npy episode_2655.npy episode_3656.npy episode_4657.npy episode_5658.npy episode_6659.npy episode_765.npy episode_8660.npy episode_9661.npy\r\nepisode_1655.npy episode_2656.npy episode_3657.npy episode_4658.npy episode_5659.npy episode_665.npy episode_7660.npy episode_8661.npy episode_9662.npy\r\nepisode_1656.npy episode_2657.npy episode_3658.npy episode_4659.npy episode_565.npy episode_6660.npy episode_7661.npy episode_8662.npy episode_9663.npy\r\nepisode_1657.npy episode_2658.npy episode_3659.npy episode_465.npy episode_5660.npy episode_6661.npy episode_7662.npy episode_8663.npy episode_9664.npy\r\nepisode_1658.npy episode_2659.npy episode_365.npy episode_4660.npy episode_5661.npy episode_6662.npy episode_7663.npy episode_8664.npy episode_9665.npy\r\nepisode_1659.npy episode_265.npy episode_3660.npy episode_4661.npy episode_5662.npy episode_6663.npy episode_7664.npy episode_8665.npy episode_9666.npy\r\nepisode_165.npy episode_2660.npy episode_3661.npy episode_4662.npy episode_5663.npy episode_6664.npy episode_7665.npy episode_8666.npy episode_9667.npy\r\nepisode_1660.npy episode_2661.npy episode_3662.npy episode_4663.npy episode_5664.npy episode_6665.npy episode_7666.npy episode_8667.npy episode_9668.npy\r\nepisode_1661.npy episode_2662.npy episode_3663.npy episode_4664.npy episode_5665.npy episode_6666.npy episode_7667.npy episode_8668.npy episode_9669.npy\r\nepisode_1662.npy episode_2663.npy episode_3664.npy episode_4665.npy episode_5666.npy episode_6667.npy episode_7668.npy episode_8669.npy episode_966.npy\r\nepisode_1663.npy episode_2664.npy episode_3665.npy episode_4666.npy episode_5667.npy episode_6668.npy episode_7669.npy episode_866.npy episode_9670.npy\r\nepisode_1664.npy episode_2665.npy episode_3666.npy episode_4667.npy episode_5668.npy episode_6669.npy episode_766.npy episode_8670.npy episode_9671.npy\r\nepisode_1665.npy episode_2666.npy episode_3667.npy episode_4668.npy episode_5669.npy episode_666.npy episode_7670.npy episode_8671.npy episode_9672.npy\r\nepisode_1666.npy episode_2667.npy episode_3668.npy episode_4669.npy episode_566.npy episode_6670.npy episode_7671.npy episode_8672.npy episode_9673.npy\r\nepisode_1667.npy episode_2668.npy episode_3669.npy episode_466.npy episode_5670.npy episode_6671.npy episode_7672.npy episode_8673.npy episode_9674.npy\r\nepisode_1668.npy episode_2669.npy episode_366.npy episode_4670.npy episode_5671.npy episode_6672.npy episode_7673.npy episode_8674.npy episode_9675.npy\r\nepisode_1669.npy episode_266.npy episode_3670.npy episode_4671.npy episode_5672.npy episode_6673.npy episode_7674.npy episode_8675.npy episode_9676.npy\r\nepisode_166.npy episode_2670.npy episode_3671.npy episode_4672.npy episode_5673.npy episode_6674.npy episode_7675.npy episode_8676.npy episode_9677.npy\r\nepisode_1670.npy episode_2671.npy episode_3672.npy episode_4673.npy episode_5674.npy episode_6675.npy episode_7676.npy episode_8677.npy episode_9678.npy\r\nepisode_1671.npy episode_2672.npy episode_3673.npy episode_4674.npy episode_5675.npy episode_6676.npy episode_7677.npy episode_8678.npy episode_9679.npy\r\nepisode_1672.npy episode_2673.npy episode_3674.npy episode_4675.npy episode_5676.npy episode_6677.npy episode_7678.npy episode_8679.npy episode_967.npy\r\nepisode_1673.npy episode_2674.npy episode_3675.npy episode_4676.npy episode_5677.npy episode_6678.npy episode_7679.npy episode_867.npy episode_9680.npy\r\nepisode_1674.npy episode_2675.npy episode_3676.npy episode_4677.npy episode_5678.npy episode_6679.npy episode_767.npy episode_8680.npy episode_9681.npy\r\nepisode_1675.npy episode_2676.npy episode_3677.npy episode_4678.npy episode_5679.npy episode_667.npy episode_7680.npy episode_8681.npy episode_9682.npy\r\nepisode_1676.npy episode_2677.npy episode_3678.npy episode_4679.npy episode_567.npy episode_6680.npy episode_7681.npy episode_8682.npy episode_9683.npy\r\nepisode_1677.npy episode_2678.npy episode_3679.npy episode_467.npy episode_5680.npy episode_6681.npy episode_7682.npy episode_8683.npy episode_9684.npy\r\nepisode_1678.npy episode_2679.npy episode_367.npy episode_4680.npy episode_5681.npy episode_6682.npy episode_7683.npy episode_8684.npy episode_9685.npy\r\nepisode_1679.npy episode_267.npy episode_3680.npy episode_4681.npy episode_5682.npy episode_6683.npy episode_7684.npy episode_8685.npy episode_9686.npy\r\nepisode_167.npy episode_2680.npy episode_3681.npy episode_4682.npy episode_5683.npy episode_6684.npy episode_7685.npy episode_8686.npy episode_9687.npy\r\nepisode_1680.npy episode_2681.npy episode_3682.npy episode_4683.npy episode_5684.npy episode_6685.npy episode_7686.npy episode_8687.npy episode_9688.npy\r\nepisode_1681.npy episode_2682.npy episode_3683.npy episode_4684.npy episode_5685.npy episode_6686.npy episode_7687.npy episode_8688.npy episode_9689.npy\r\nepisode_1682.npy episode_2683.npy episode_3684.npy episode_4685.npy episode_5686.npy episode_6687.npy episode_7688.npy episode_8689.npy episode_968.npy\r\nepisode_1683.npy episode_2684.npy episode_3685.npy episode_4686.npy episode_5687.npy episode_6688.npy episode_7689.npy episode_868.npy episode_9690.npy\r\nepisode_1684.npy episode_2685.npy episode_3686.npy episode_4687.npy episode_5688.npy episode_6689.npy episode_768.npy episode_8690.npy episode_9691.npy\r\nepisode_1685.npy episode_2686.npy episode_3687.npy episode_4688.npy episode_5689.npy episode_668.npy episode_7690.npy episode_8691.npy episode_9692.npy\r\nepisode_1686.npy episode_2687.npy episode_3688.npy episode_4689.npy episode_568.npy episode_6690.npy episode_7691.npy episode_8692.npy episode_9693.npy\r\nepisode_1687.npy episode_2688.npy episode_3689.npy episode_468.npy episode_5690.npy episode_6691.npy episode_7692.npy episode_8693.npy episode_9694.npy\r\nepisode_1688.npy episode_2689.npy episode_368.npy episode_4690.npy episode_5691.npy episode_6692.npy episode_7693.npy episode_8694.npy episode_9695.npy\r\nepisode_1689.npy episode_268.npy episode_3690.npy episode_4691.npy episode_5692.npy episode_6693.npy episode_7694.npy episode_8695.npy episode_9696.npy\r\nepisode_168.npy episode_2690.npy episode_3691.npy episode_4692.npy episode_5693.npy episode_6694.npy episode_7695.npy episode_8696.npy episode_9697.npy\r\nepisode_1690.npy episode_2691.npy episode_3692.npy episode_4693.npy episode_5694.npy episode_6695.npy episode_7696.npy episode_8697.npy episode_9698.npy\r\nepisode_1691.npy episode_2692.npy episode_3693.npy episode_4694.npy episode_5695.npy episode_6696.npy episode_7697.npy episode_8698.npy episode_9699.npy\r\nepisode_1692.npy episode_2693.npy episode_3694.npy episode_4695.npy episode_5696.npy episode_6697.npy episode_7698.npy episode_8699.npy episode_969.npy\r\nepisode_1693.npy episode_2694.npy episode_3695.npy episode_4696.npy episode_5697.npy episode_6698.npy episode_7699.npy episode_869.npy episode_96.npy\r\nepisode_1694.npy episode_2695.npy episode_3696.npy episode_4697.npy episode_5698.npy episode_6699.npy episode_769.npy episode_86.npy episode_9700.npy\r\nepisode_1695.npy episode_2696.npy episode_3697.npy episode_4698.npy episode_5699.npy episode_669.npy episode_76.npy episode_8700.npy episode_9701.npy\r\nepisode_1696.npy episode_2697.npy episode_3698.npy episode_4699.npy episode_569.npy episode_66.npy episode_7700.npy episode_8701.npy episode_9702.npy\r\nepisode_1697.npy episode_2698.npy episode_3699.npy episode_469.npy episode_56.npy episode_6700.npy episode_7701.npy episode_8702.npy episode_9703.npy\r\nepisode_1698.npy episode_2699.npy episode_369.npy episode_46.npy episode_5700.npy episode_6701.npy episode_7702.npy episode_8703.npy episode_9704.npy\r\nepisode_1699.npy episode_269.npy episode_36.npy episode_4700.npy episode_5701.npy episode_6702.npy episode_7703.npy episode_8704.npy episode_9705.npy\r\nepisode_169.npy episode_26.npy episode_3700.npy episode_4701.npy episode_5702.npy episode_6703.npy episode_7704.npy episode_8705.npy episode_9706.npy\r\nepisode_16.npy episode_2700.npy episode_3701.npy episode_4702.npy episode_5703.npy episode_6704.npy episode_7705.npy episode_8706.npy episode_9707.npy\r\nepisode_1700.npy episode_2701.npy episode_3702.npy episode_4703.npy episode_5704.npy episode_6705.npy episode_7706.npy episode_8707.npy episode_9708.npy\r\nepisode_1701.npy episode_2702.npy episode_3703.npy episode_4704.npy episode_5705.npy episode_6706.npy episode_7707.npy episode_8708.npy episode_9709.npy\r\nepisode_1702.npy episode_2703.npy episode_3704.npy episode_4705.npy episode_5706.npy episode_6707.npy episode_7708.npy episode_8709.npy episode_970.npy\r\nepisode_1703.npy episode_2704.npy episode_3705.npy episode_4706.npy episode_5707.npy episode_6708.npy episode_7709.npy episode_870.npy episode_9710.npy\r\nepisode_1704.npy episode_2705.npy episode_3706.npy episode_4707.npy episode_5708.npy episode_6709.npy episode_770.npy episode_8710.npy episode_9711.npy\r\nepisode_1705.npy episode_2706.npy episode_3707.npy episode_4708.npy episode_5709.npy episode_670.npy episode_7710.npy episode_8711.npy episode_9712.npy\r\nepisode_1706.npy episode_2707.npy episode_3708.npy episode_4709.npy episode_570.npy episode_6710.npy episode_7711.npy episode_8712.npy episode_9713.npy\r\nepisode_1707.npy episode_2708.npy episode_3709.npy episode_470.npy episode_5710.npy episode_6711.npy episode_7712.npy episode_8713.npy episode_9714.npy\r\nepisode_1708.npy episode_2709.npy episode_370.npy episode_4710.npy episode_5711.npy episode_6712.npy episode_7713.npy episode_8714.npy episode_9715.npy\r\nepisode_1709.npy episode_270.npy episode_3710.npy episode_4711.npy episode_5712.npy episode_6713.npy episode_7714.npy episode_8715.npy episode_9716.npy\r\nepisode_170.npy episode_2710.npy episode_3711.npy episode_4712.npy episode_5713.npy episode_6714.npy episode_7715.npy episode_8716.npy episode_9717.npy\r\nepisode_1710.npy episode_2711.npy episode_3712.npy episode_4713.npy episode_5714.npy episode_6715.npy episode_7716.npy episode_8717.npy episode_9718.npy\r\nepisode_1711.npy episode_2712.npy episode_3713.npy episode_4714.npy episode_5715.npy episode_6716.npy episode_7717.npy episode_8718.npy episode_9719.npy\r\nepisode_1712.npy episode_2713.npy episode_3714.npy episode_4715.npy episode_5716.npy episode_6717.npy episode_7718.npy episode_8719.npy episode_971.npy\r\nepisode_1713.npy episode_2714.npy episode_3715.npy episode_4716.npy episode_5717.npy episode_6718.npy episode_7719.npy episode_871.npy episode_9720.npy\r\nepisode_1714.npy episode_2715.npy episode_3716.npy episode_4717.npy episode_5718.npy episode_6719.npy episode_771.npy episode_8720.npy episode_9721.npy\r\nepisode_1715.npy episode_2716.npy episode_3717.npy episode_4718.npy episode_5719.npy episode_671.npy episode_7720.npy episode_8721.npy episode_9722.npy\r\nepisode_1716.npy episode_2717.npy episode_3718.npy episode_4719.npy episode_571.npy episode_6720.npy episode_7721.npy episode_8722.npy episode_9723.npy\r\nepisode_1717.npy episode_2718.npy episode_3719.npy episode_471.npy episode_5720.npy episode_6721.npy episode_7722.npy episode_8723.npy episode_9724.npy\r\nepisode_1718.npy episode_2719.npy episode_371.npy episode_4720.npy episode_5721.npy episode_6722.npy episode_7723.npy episode_8724.npy episode_9725.npy\r\nepisode_1719.npy episode_271.npy episode_3720.npy episode_4721.npy episode_5722.npy episode_6723.npy episode_7724.npy episode_8725.npy episode_9726.npy\r\nepisode_171.npy episode_2720.npy episode_3721.npy episode_4722.npy episode_5723.npy episode_6724.npy episode_7725.npy episode_8726.npy episode_9727.npy\r\nepisode_1720.npy episode_2721.npy episode_3722.npy episode_4723.npy episode_5724.npy episode_6725.npy episode_7726.npy episode_8727.npy episode_9728.npy\r\nepisode_1721.npy episode_2722.npy episode_3723.npy episode_4724.npy episode_5725.npy episode_6726.npy episode_7727.npy episode_8728.npy episode_9729.npy\r\nepisode_1722.npy episode_2723.npy episode_3724.npy episode_4725.npy episode_5726.npy episode_6727.npy episode_7728.npy episode_8729.npy episode_972.npy\r\nepisode_1723.npy episode_2724.npy episode_3725.npy episode_4726.npy episode_5727.npy episode_6728.npy episode_7729.npy episode_872.npy episode_9730.npy\r\nepisode_1724.npy episode_2725.npy episode_3726.npy episode_4727.npy episode_5728.npy episode_6729.npy episode_772.npy episode_8730.npy episode_9731.npy\r\nepisode_1725.npy episode_2726.npy episode_3727.npy episode_4728.npy episode_5729.npy episode_672.npy episode_7730.npy episode_8731.npy episode_9732.npy\r\nepisode_1726.npy episode_2727.npy episode_3728.npy episode_4729.npy episode_572.npy episode_6730.npy episode_7731.npy episode_8732.npy episode_9733.npy\r\nepisode_1727.npy episode_2728.npy episode_3729.npy episode_472.npy episode_5730.npy episode_6731.npy episode_7732.npy episode_8733.npy episode_9734.npy\r\nepisode_1728.npy episode_2729.npy episode_372.npy episode_4730.npy episode_5731.npy episode_6732.npy episode_7733.npy episode_8734.npy episode_9735.npy\r\nepisode_1729.npy episode_272.npy episode_3730.npy episode_4731.npy episode_5732.npy episode_6733.npy episode_7734.npy episode_8735.npy episode_9736.npy\r\nepisode_172.npy episode_2730.npy episode_3731.npy episode_4732.npy episode_5733.npy episode_6734.npy episode_7735.npy episode_8736.npy episode_9737.npy\r\nepisode_1730.npy episode_2731.npy episode_3732.npy episode_4733.npy episode_5734.npy episode_6735.npy episode_7736.npy episode_8737.npy episode_9738.npy\r\nepisode_1731.npy episode_2732.npy episode_3733.npy episode_4734.npy episode_5735.npy episode_6736.npy episode_7737.npy episode_8738.npy episode_9739.npy\r\nepisode_1732.npy episode_2733.npy episode_3734.npy episode_4735.npy episode_5736.npy episode_6737.npy episode_7738.npy episode_8739.npy episode_973.npy\r\nepisode_1733.npy episode_2734.npy episode_3735.npy episode_4736.npy episode_5737.npy episode_6738.npy episode_7739.npy episode_873.npy episode_9740.npy\r\nepisode_1734.npy episode_2735.npy episode_3736.npy episode_4737.npy episode_5738.npy episode_6739.npy episode_773.npy episode_8740.npy episode_9741.npy\r\nepisode_1735.npy episode_2736.npy episode_3737.npy episode_4738.npy episode_5739.npy episode_673.npy episode_7740.npy episode_8741.npy episode_9742.npy\r\nepisode_1736.npy episode_2737.npy episode_3738.npy episode_4739.npy episode_573.npy episode_6740.npy episode_7741.npy episode_8742.npy episode_9743.npy\r\nepisode_1737.npy episode_2738.npy episode_3739.npy episode_473.npy episode_5740.npy episode_6741.npy episode_7742.npy episode_8743.npy episode_9744.npy\r\nepisode_1738.npy episode_2739.npy episode_373.npy episode_4740.npy episode_5741.npy episode_6742.npy episode_7743.npy episode_8744.npy episode_9745.npy\r\nepisode_1739.npy episode_273.npy episode_3740.npy episode_4741.npy episode_5742.npy episode_6743.npy episode_7744.npy episode_8745.npy episode_9746.npy\r\nepisode_173.npy episode_2740.npy episode_3741.npy episode_4742.npy episode_5743.npy episode_6744.npy episode_7745.npy episode_8746.npy episode_9747.npy\r\nepisode_1740.npy episode_2741.npy episode_3742.npy episode_4743.npy episode_5744.npy episode_6745.npy episode_7746.npy episode_8747.npy episode_9748.npy\r\nepisode_1741.npy episode_2742.npy episode_3743.npy episode_4744.npy episode_5745.npy episode_6746.npy episode_7747.npy episode_8748.npy episode_9749.npy\r\nepisode_1742.npy episode_2743.npy episode_3744.npy episode_4745.npy episode_5746.npy episode_6747.npy episode_7748.npy episode_8749.npy episode_974.npy\r\nepisode_1743.npy episode_2744.npy episode_3745.npy episode_4746.npy episode_5747.npy episode_6748.npy episode_7749.npy episode_874.npy episode_9750.npy\r\nepisode_1744.npy episode_2745.npy episode_3746.npy episode_4747.npy episode_5748.npy episode_6749.npy episode_774.npy episode_8750.npy episode_9751.npy\r\nepisode_1745.npy episode_2746.npy episode_3747.npy episode_4748.npy episode_5749.npy episode_674.npy episode_7750.npy episode_8751.npy episode_9752.npy\r\nepisode_1746.npy episode_2747.npy episode_3748.npy episode_4749.npy episode_574.npy episode_6750.npy episode_7751.npy episode_8752.npy episode_9753.npy\r\nepisode_1747.npy episode_2748.npy episode_3749.npy episode_474.npy episode_5750.npy episode_6751.npy episode_7752.npy episode_8753.npy episode_9754.npy\r\nepisode_1748.npy episode_2749.npy episode_374.npy episode_4750.npy episode_5751.npy episode_6752.npy episode_7753.npy episode_8754.npy episode_9755.npy\r\nepisode_1749.npy episode_274.npy episode_3750.npy episode_4751.npy episode_5752.npy episode_6753.npy episode_7754.npy episode_8755.npy episode_9756.npy\r\nepisode_174.npy episode_2750.npy episode_3751.npy episode_4752.npy episode_5753.npy episode_6754.npy episode_7755.npy episode_8756.npy episode_9757.npy\r\nepisode_1750.npy episode_2751.npy episode_3752.npy episode_4753.npy episode_5754.npy episode_6755.npy episode_7756.npy episode_8757.npy episode_9758.npy\r\nepisode_1751.npy episode_2752.npy episode_3753.npy episode_4754.npy episode_5755.npy episode_6756.npy episode_7757.npy episode_8758.npy episode_9759.npy\r\nepisode_1752.npy episode_2753.npy episode_3754.npy episode_4755.npy episode_5756.npy episode_6757.npy episode_7758.npy episode_8759.npy episode_975.npy\r\nepisode_1753.npy episode_2754.npy episode_3755.npy episode_4756.npy episode_5757.npy episode_6758.npy episode_7759.npy episode_875.npy episode_9760.npy\r\nepisode_1754.npy episode_2755.npy episode_3756.npy episode_4757.npy episode_5758.npy episode_6759.npy episode_775.npy episode_8760.npy episode_9761.npy\r\nepisode_1755.npy episode_2756.npy episode_3757.npy episode_4758.npy episode_5759.npy episode_675.npy episode_7760.npy episode_8761.npy episode_9762.npy\r\nepisode_1756.npy episode_2757.npy episode_3758.npy episode_4759.npy episode_575.npy episode_6760.npy episode_7761.npy episode_8762.npy episode_9763.npy\r\nepisode_1757.npy episode_2758.npy episode_3759.npy episode_475.npy episode_5760.npy episode_6761.npy episode_7762.npy episode_8763.npy episode_9764.npy\r\nepisode_1758.npy episode_2759.npy episode_375.npy episode_4760.npy episode_5761.npy episode_6762.npy episode_7763.npy episode_8764.npy episode_9765.npy\r\nepisode_1759.npy episode_275.npy episode_3760.npy episode_4761.npy episode_5762.npy episode_6763.npy episode_7764.npy episode_8765.npy episode_9766.npy\r\nepisode_175.npy episode_2760.npy episode_3761.npy episode_4762.npy episode_5763.npy episode_6764.npy episode_7765.npy episode_8766.npy episode_9767.npy\r\nepisode_1760.npy episode_2761.npy episode_3762.npy episode_4763.npy episode_5764.npy episode_6765.npy episode_7766.npy episode_8767.npy episode_9768.npy\r\nepisode_1761.npy episode_2762.npy episode_3763.npy episode_4764.npy episode_5765.npy episode_6766.npy episode_7767.npy episode_8768.npy episode_9769.npy\r\nepisode_1762.npy episode_2763.npy episode_3764.npy episode_4765.npy episode_5766.npy episode_6767.npy episode_7768.npy episode_8769.npy episode_976.npy\r\nepisode_1763.npy episode_2764.npy episode_3765.npy episode_4766.npy episode_5767.npy episode_6768.npy episode_7769.npy episode_876.npy episode_9770.npy\r\nepisode_1764.npy episode_2765.npy episode_3766.npy episode_4767.npy episode_5768.npy episode_6769.npy episode_776.npy episode_8770.npy episode_9771.npy\r\nepisode_1765.npy episode_2766.npy episode_3767.npy episode_4768.npy episode_5769.npy episode_676.npy episode_7770.npy episode_8771.npy episode_9772.npy\r\nepisode_1766.npy episode_2767.npy episode_3768.npy episode_4769.npy episode_576.npy episode_6770.npy episode_7771.npy episode_8772.npy episode_9773.npy\r\nepisode_1767.npy episode_2768.npy episode_3769.npy episode_476.npy episode_5770.npy episode_6771.npy episode_7772.npy episode_8773.npy episode_9774.npy\r\nepisode_1768.npy episode_2769.npy episode_376.npy episode_4770.npy episode_5771.npy episode_6772.npy episode_7773.npy episode_8774.npy episode_9775.npy\r\nepisode_1769.npy episode_276.npy episode_3770.npy episode_4771.npy episode_5772.npy episode_6773.npy episode_7774.npy episode_8775.npy episode_9776.npy\r\nepisode_176.npy episode_2770.npy episode_3771.npy episode_4772.npy episode_5773.npy episode_6774.npy episode_7775.npy episode_8776.npy episode_9777.npy\r\nepisode_1770.npy episode_2771.npy episode_3772.npy episode_4773.npy episode_5774.npy episode_6775.npy episode_7776.npy episode_8777.npy episode_9778.npy\r\nepisode_1771.npy episode_2772.npy episode_3773.npy episode_4774.npy episode_5775.npy episode_6776.npy episode_7777.npy episode_8778.npy episode_9779.npy\r\nepisode_1772.npy episode_2773.npy episode_3774.npy episode_4775.npy episode_5776.npy episode_6777.npy episode_7778.npy episode_8779.npy episode_977.npy\r\nepisode_1773.npy episode_2774.npy episode_3775.npy episode_4776.npy episode_5777.npy episode_6778.npy episode_7779.npy episode_877.npy episode_9780.npy\r\nepisode_1774.npy episode_2775.npy episode_3776.npy episode_4777.npy episode_5778.npy episode_6779.npy episode_777.npy episode_8780.npy episode_9781.npy\r\nepisode_1775.npy episode_2776.npy episode_3777.npy episode_4778.npy episode_5779.npy episode_677.npy episode_7780.npy episode_8781.npy episode_9782.npy\r\nepisode_1776.npy episode_2777.npy episode_3778.npy episode_4779.npy episode_577.npy episode_6780.npy episode_7781.npy episode_8782.npy episode_9783.npy\r\nepisode_1777.npy episode_2778.npy episode_3779.npy episode_477.npy episode_5780.npy episode_6781.npy episode_7782.npy episode_8783.npy episode_9784.npy\r\nepisode_1778.npy episode_2779.npy episode_377.npy episode_4780.npy episode_5781.npy episode_6782.npy episode_7783.npy episode_8784.npy episode_9785.npy\r\nepisode_1779.npy episode_277.npy episode_3780.npy episode_4781.npy episode_5782.npy episode_6783.npy episode_7784.npy episode_8785.npy episode_9786.npy\r\nepisode_177.npy episode_2780.npy episode_3781.npy episode_4782.npy episode_5783.npy episode_6784.npy episode_7785.npy episode_8786.npy episode_9787.npy\r\nepisode_1780.npy episode_2781.npy episode_3782.npy episode_4783.npy episode_5784.npy episode_6785.npy episode_7786.npy episode_8787.npy episode_9788.npy\r\nepisode_1781.npy episode_2782.npy episode_3783.npy episode_4784.npy episode_5785.npy episode_6786.npy episode_7787.npy episode_8788.npy episode_9789.npy\r\nepisode_1782.npy episode_2783.npy episode_3784.npy episode_4785.npy episode_5786.npy episode_6787.npy episode_7788.npy episode_8789.npy episode_978.npy\r\nepisode_1783.npy episode_2784.npy episode_3785.npy episode_4786.npy episode_5787.npy episode_6788.npy episode_7789.npy episode_878.npy episode_9790.npy\r\nepisode_1784.npy episode_2785.npy episode_3786.npy episode_4787.npy episode_5788.npy episode_6789.npy episode_778.npy episode_8790.npy episode_9791.npy\r\nepisode_1785.npy episode_2786.npy episode_3787.npy episode_4788.npy episode_5789.npy episode_678.npy episode_7790.npy episode_8791.npy episode_9792.npy\r\nepisode_1786.npy episode_2787.npy episode_3788.npy episode_4789.npy episode_578.npy episode_6790.npy episode_7791.npy episode_8792.npy episode_9793.npy\r\nepisode_1787.npy episode_2788.npy episode_3789.npy episode_478.npy episode_5790.npy episode_6791.npy episode_7792.npy episode_8793.npy episode_9794.npy\r\nepisode_1788.npy episode_2789.npy episode_378.npy episode_4790.npy episode_5791.npy episode_6792.npy episode_7793.npy episode_8794.npy episode_9795.npy\r\nepisode_1789.npy episode_278.npy episode_3790.npy episode_4791.npy episode_5792.npy episode_6793.npy episode_7794.npy episode_8795.npy episode_9796.npy\r\nepisode_178.npy episode_2790.npy episode_3791.npy episode_4792.npy episode_5793.npy episode_6794.npy episode_7795.npy episode_8796.npy episode_9797.npy\r\nepisode_1790.npy episode_2791.npy episode_3792.npy episode_4793.npy episode_5794.npy episode_6795.npy episode_7796.npy episode_8797.npy episode_9798.npy\r\nepisode_1791.npy episode_2792.npy episode_3793.npy episode_4794.npy episode_5795.npy episode_6796.npy episode_7797.npy episode_8798.npy episode_9799.npy\r\nepisode_1792.npy episode_2793.npy episode_3794.npy episode_4795.npy episode_5796.npy episode_6797.npy episode_7798.npy episode_8799.npy episode_979.npy\r\nepisode_1793.npy episode_2794.npy episode_3795.npy episode_4796.npy episode_5797.npy episode_6798.npy episode_7799.npy episode_879.npy episode_97.npy\r\nepisode_1794.npy episode_2795.npy episode_3796.npy episode_4797.npy episode_5798.npy episode_6799.npy episode_779.npy episode_87.npy episode_9800.npy\r\nepisode_1795.npy episode_2796.npy episode_3797.npy episode_4798.npy episode_5799.npy episode_679.npy episode_77.npy episode_8800.npy episode_9801.npy\r\nepisode_1796.npy episode_2797.npy episode_3798.npy episode_4799.npy episode_579.npy episode_67.npy episode_7800.npy episode_8801.npy episode_9802.npy\r\nepisode_1797.npy episode_2798.npy episode_3799.npy episode_479.npy episode_57.npy episode_6800.npy episode_7801.npy episode_8802.npy episode_9803.npy\r\nepisode_1798.npy episode_2799.npy episode_379.npy episode_47.npy episode_5800.npy episode_6801.npy episode_7802.npy episode_8803.npy episode_9804.npy\r\nepisode_1799.npy episode_279.npy episode_37.npy episode_4800.npy episode_5801.npy episode_6802.npy episode_7803.npy episode_8804.npy episode_9805.npy\r\nepisode_179.npy episode_27.npy episode_3800.npy episode_4801.npy episode_5802.npy episode_6803.npy episode_7804.npy episode_8805.npy episode_9806.npy\r\nepisode_17.npy episode_2800.npy episode_3801.npy episode_4802.npy episode_5803.npy episode_6804.npy episode_7805.npy episode_8806.npy episode_9807.npy\r\nepisode_1800.npy episode_2801.npy episode_3802.npy episode_4803.npy episode_5804.npy episode_6805.npy episode_7806.npy episode_8807.npy episode_9808.npy\r\nepisode_1801.npy episode_2802.npy episode_3803.npy episode_4804.npy episode_5805.npy episode_6806.npy episode_7807.npy episode_8808.npy episode_9809.npy\r\nepisode_1802.npy episode_2803.npy episode_3804.npy episode_4805.npy episode_5806.npy episode_6807.npy episode_7808.npy episode_8809.npy episode_980.npy\r\nepisode_1803.npy episode_2804.npy episode_3805.npy episode_4806.npy episode_5807.npy episode_6808.npy episode_7809.npy episode_880.npy episode_9810.npy\r\nepisode_1804.npy episode_2805.npy episode_3806.npy episode_4807.npy episode_5808.npy episode_6809.npy episode_780.npy episode_8810.npy episode_9811.npy\r\nepisode_1805.npy episode_2806.npy episode_3807.npy episode_4808.npy episode_5809.npy episode_680.npy episode_7810.npy episode_8811.npy episode_9812.npy\r\nepisode_1806.npy episode_2807.npy episode_3808.npy episode_4809.npy episode_580.npy episode_6810.npy episode_7811.npy episode_8812.npy episode_9813.npy\r\nepisode_1807.npy episode_2808.npy episode_3809.npy episode_480.npy episode_5810.npy episode_6811.npy episode_7812.npy episode_8813.npy episode_9814.npy\r\nepisode_1808.npy episode_2809.npy episode_380.npy episode_4810.npy episode_5811.npy episode_6812.npy episode_7813.npy episode_8814.npy episode_9815.npy\r\nepisode_1809.npy episode_280.npy episode_3810.npy episode_4811.npy episode_5812.npy episode_6813.npy episode_7814.npy episode_8815.npy episode_9816.npy\r\nepisode_180.npy episode_2810.npy episode_3811.npy episode_4812.npy episode_5813.npy episode_6814.npy episode_7815.npy episode_8816.npy episode_9817.npy\r\nepisode_1810.npy episode_2811.npy episode_3812.npy episode_4813.npy episode_5814.npy episode_6815.npy episode_7816.npy episode_8817.npy episode_9818.npy\r\nepisode_1811.npy episode_2812.npy episode_3813.npy episode_4814.npy episode_5815.npy episode_6816.npy episode_7817.npy episode_8818.npy episode_9819.npy\r\nepisode_1812.npy episode_2813.npy episode_3814.npy episode_4815.npy episode_5816.npy episode_6817.npy episode_7818.npy episode_8819.npy episode_981.npy\r\nepisode_1813.npy episode_2814.npy episode_3815.npy episode_4816.npy episode_5817.npy episode_6818.npy episode_7819.npy episode_881.npy episode_9820.npy\r\nepisode_1814.npy episode_2815.npy episode_3816.npy episode_4817.npy episode_5818.npy episode_6819.npy episode_781.npy episode_8820.npy episode_9821.npy\r\nepisode_1815.npy episode_2816.npy episode_3817.npy episode_4818.npy episode_5819.npy episode_681.npy episode_7820.npy episode_8821.npy episode_9822.npy\r\nepisode_1816.npy episode_2817.npy episode_3818.npy episode_4819.npy episode_581.npy episode_6820.npy episode_7821.npy episode_8822.npy episode_9823.npy\r\nepisode_1817.npy episode_2818.npy episode_3819.npy episode_481.npy episode_5820.npy episode_6821.npy episode_7822.npy episode_8823.npy episode_9824.npy\r\nepisode_1818.npy episode_2819.npy episode_381.npy episode_4820.npy episode_5821.npy episode_6822.npy episode_7823.npy episode_8824.npy episode_9825.npy\r\nepisode_1819.npy episode_281.npy episode_3820.npy episode_4821.npy episode_5822.npy episode_6823.npy episode_7824.npy episode_8825.npy episode_9826.npy\r\nepisode_181.npy episode_2820.npy episode_3821.npy episode_4822.npy episode_5823.npy episode_6824.npy episode_7825.npy episode_8826.npy episode_9827.npy\r\nepisode_1820.npy episode_2821.npy episode_3822.npy episode_4823.npy episode_5824.npy episode_6825.npy episode_7826.npy episode_8827.npy episode_9828.npy\r\nepisode_1821.npy episode_2822.npy episode_3823.npy episode_4824.npy episode_5825.npy episode_6826.npy episode_7827.npy episode_8828.npy episode_9829.npy\r\nepisode_1822.npy episode_2823.npy episode_3824.npy episode_4825.npy episode_5826.npy episode_6827.npy episode_7828.npy episode_8829.npy episode_982.npy\r\nepisode_1823.npy episode_2824.npy episode_3825.npy episode_4826.npy episode_5827.npy episode_6828.npy episode_7829.npy episode_882.npy episode_9830.npy\r\nepisode_1824.npy episode_2825.npy episode_3826.npy episode_4827.npy episode_5828.npy episode_6829.npy episode_782.npy episode_8830.npy episode_9831.npy\r\nepisode_1825.npy episode_2826.npy episode_3827.npy episode_4828.npy episode_5829.npy episode_682.npy episode_7830.npy episode_8831.npy episode_9832.npy\r\nepisode_1826.npy episode_2827.npy episode_3828.npy episode_4829.npy episode_582.npy episode_6830.npy episode_7831.npy episode_8832.npy episode_9833.npy\r\nepisode_1827.npy episode_2828.npy episode_3829.npy episode_482.npy episode_5830.npy episode_6831.npy episode_7832.npy episode_8833.npy episode_9834.npy\r\nepisode_1828.npy episode_2829.npy episode_382.npy episode_4830.npy episode_5831.npy episode_6832.npy episode_7833.npy episode_8834.npy episode_9835.npy\r\nepisode_1829.npy episode_282.npy episode_3830.npy episode_4831.npy episode_5832.npy episode_6833.npy episode_7834.npy episode_8835.npy episode_9836.npy\r\nepisode_182.npy episode_2830.npy episode_3831.npy episode_4832.npy episode_5833.npy episode_6834.npy episode_7835.npy episode_8836.npy episode_9837.npy\r\nepisode_1830.npy episode_2831.npy episode_3832.npy episode_4833.npy episode_5834.npy episode_6835.npy episode_7836.npy episode_8837.npy episode_9838.npy\r\nepisode_1831.npy episode_2832.npy episode_3833.npy episode_4834.npy episode_5835.npy episode_6836.npy episode_7837.npy episode_8838.npy episode_9839.npy\r\nepisode_1832.npy episode_2833.npy episode_3834.npy episode_4835.npy episode_5836.npy episode_6837.npy episode_7838.npy episode_8839.npy episode_983.npy\r\nepisode_1833.npy episode_2834.npy episode_3835.npy episode_4836.npy episode_5837.npy episode_6838.npy episode_7839.npy episode_883.npy episode_9840.npy\r\nepisode_1834.npy episode_2835.npy episode_3836.npy episode_4837.npy episode_5838.npy episode_6839.npy episode_783.npy episode_8840.npy episode_9841.npy\r\nepisode_1835.npy episode_2836.npy episode_3837.npy episode_4838.npy episode_5839.npy episode_683.npy episode_7840.npy episode_8841.npy episode_9842.npy\r\nepisode_1836.npy episode_2837.npy episode_3838.npy episode_4839.npy episode_583.npy episode_6840.npy episode_7841.npy episode_8842.npy episode_9843.npy\r\nepisode_1837.npy episode_2838.npy episode_3839.npy episode_483.npy episode_5840.npy episode_6841.npy episode_7842.npy episode_8843.npy episode_9844.npy\r\nepisode_1838.npy episode_2839.npy episode_383.npy episode_4840.npy episode_5841.npy episode_6842.npy episode_7843.npy episode_8844.npy episode_9845.npy\r\nepisode_1839.npy episode_283.npy episode_3840.npy episode_4841.npy episode_5842.npy episode_6843.npy episode_7844.npy episode_8845.npy episode_9846.npy\r\nepisode_183.npy episode_2840.npy episode_3841.npy episode_4842.npy episode_5843.npy episode_6844.npy episode_7845.npy episode_8846.npy episode_9847.npy\r\nepisode_1840.npy episode_2841.npy episode_3842.npy episode_4843.npy episode_5844.npy episode_6845.npy episode_7846.npy episode_8847.npy episode_9848.npy\r\nepisode_1841.npy episode_2842.npy episode_3843.npy episode_4844.npy episode_5845.npy episode_6846.npy episode_7847.npy episode_8848.npy episode_9849.npy\r\nepisode_1842.npy episode_2843.npy episode_3844.npy episode_4845.npy episode_5846.npy episode_6847.npy episode_7848.npy episode_8849.npy episode_984.npy\r\nepisode_1843.npy episode_2844.npy episode_3845.npy episode_4846.npy episode_5847.npy episode_6848.npy episode_7849.npy episode_884.npy episode_9850.npy\r\nepisode_1844.npy episode_2845.npy episode_3846.npy episode_4847.npy episode_5848.npy episode_6849.npy episode_784.npy episode_8850.npy episode_9851.npy\r\nepisode_1845.npy episode_2846.npy episode_3847.npy episode_4848.npy episode_5849.npy episode_684.npy episode_7850.npy episode_8851.npy episode_9852.npy\r\nepisode_1846.npy episode_2847.npy episode_3848.npy episode_4849.npy episode_584.npy episode_6850.npy episode_7851.npy episode_8852.npy episode_9853.npy\r\nepisode_1847.npy episode_2848.npy episode_3849.npy episode_484.npy episode_5850.npy episode_6851.npy episode_7852.npy episode_8853.npy episode_9854.npy\r\nepisode_1848.npy episode_2849.npy episode_384.npy episode_4850.npy episode_5851.npy episode_6852.npy episode_7853.npy episode_8854.npy episode_9855.npy\r\nepisode_1849.npy episode_284.npy episode_3850.npy episode_4851.npy episode_5852.npy episode_6853.npy episode_7854.npy episode_8855.npy episode_9856.npy\r\nepisode_184.npy episode_2850.npy episode_3851.npy episode_4852.npy episode_5853.npy episode_6854.npy episode_7855.npy episode_8856.npy episode_9857.npy\r\nepisode_1850.npy episode_2851.npy episode_3852.npy episode_4853.npy episode_5854.npy episode_6855.npy episode_7856.npy episode_8857.npy episode_9858.npy\r\nepisode_1851.npy episode_2852.npy episode_3853.npy episode_4854.npy episode_5855.npy episode_6856.npy episode_7857.npy episode_8858.npy episode_9859.npy\r\nepisode_1852.npy episode_2853.npy episode_3854.npy episode_4855.npy episode_5856.npy episode_6857.npy episode_7858.npy episode_8859.npy episode_985.npy\r\nepisode_1853.npy episode_2854.npy episode_3855.npy episode_4856.npy episode_5857.npy episode_6858.npy episode_7859.npy episode_885.npy episode_9860.npy\r\nepisode_1854.npy episode_2855.npy episode_3856.npy episode_4857.npy episode_5858.npy episode_6859.npy episode_785.npy episode_8860.npy episode_9861.npy\r\nepisode_1855.npy episode_2856.npy episode_3857.npy episode_4858.npy episode_5859.npy episode_685.npy episode_7860.npy episode_8861.npy episode_9862.npy\r\nepisode_1856.npy episode_2857.npy episode_3858.npy episode_4859.npy episode_585.npy episode_6860.npy episode_7861.npy episode_8862.npy episode_9863.npy\r\nepisode_1857.npy episode_2858.npy episode_3859.npy episode_485.npy episode_5860.npy episode_6861.npy episode_7862.npy episode_8863.npy episode_9864.npy\r\nepisode_1858.npy episode_2859.npy episode_385.npy episode_4860.npy episode_5861.npy episode_6862.npy episode_7863.npy episode_8864.npy episode_9865.npy\r\nepisode_1859.npy episode_285.npy episode_3860.npy episode_4861.npy episode_5862.npy episode_6863.npy episode_7864.npy episode_8865.npy episode_9866.npy\r\nepisode_185.npy episode_2860.npy episode_3861.npy episode_4862.npy episode_5863.npy episode_6864.npy episode_7865.npy episode_8866.npy episode_9867.npy\r\nepisode_1860.npy episode_2861.npy episode_3862.npy episode_4863.npy episode_5864.npy episode_6865.npy episode_7866.npy episode_8867.npy episode_9868.npy\r\nepisode_1861.npy episode_2862.npy episode_3863.npy episode_4864.npy episode_5865.npy episode_6866.npy episode_7867.npy episode_8868.npy episode_9869.npy\r\nepisode_1862.npy episode_2863.npy episode_3864.npy episode_4865.npy episode_5866.npy episode_6867.npy episode_7868.npy episode_8869.npy episode_986.npy\r\nepisode_1863.npy episode_2864.npy episode_3865.npy episode_4866.npy episode_5867.npy episode_6868.npy episode_7869.npy episode_886.npy episode_9870.npy\r\nepisode_1864.npy episode_2865.npy episode_3866.npy episode_4867.npy episode_5868.npy episode_6869.npy episode_786.npy episode_8870.npy episode_9871.npy\r\nepisode_1865.npy episode_2866.npy episode_3867.npy episode_4868.npy episode_5869.npy episode_686.npy episode_7870.npy episode_8871.npy episode_9872.npy\r\nepisode_1866.npy episode_2867.npy episode_3868.npy episode_4869.npy episode_586.npy episode_6870.npy episode_7871.npy episode_8872.npy episode_9873.npy\r\nepisode_1867.npy episode_2868.npy episode_3869.npy episode_486.npy episode_5870.npy episode_6871.npy episode_7872.npy episode_8873.npy episode_9874.npy\r\nepisode_1868.npy episode_2869.npy episode_386.npy episode_4870.npy episode_5871.npy episode_6872.npy episode_7873.npy episode_8874.npy episode_9875.npy\r\nepisode_1869.npy episode_286.npy episode_3870.npy episode_4871.npy episode_5872.npy episode_6873.npy episode_7874.npy episode_8875.npy episode_9876.npy\r\nepisode_186.npy episode_2870.npy episode_3871.npy episode_4872.npy episode_5873.npy episode_6874.npy episode_7875.npy episode_8876.npy episode_9877.npy\r\nepisode_1870.npy episode_2871.npy episode_3872.npy episode_4873.npy episode_5874.npy episode_6875.npy episode_7876.npy episode_8877.npy episode_9878.npy\r\nepisode_1871.npy episode_2872.npy episode_3873.npy episode_4874.npy episode_5875.npy episode_6876.npy episode_7877.npy episode_8878.npy episode_9879.npy\r\nepisode_1872.npy episode_2873.npy episode_3874.npy episode_4875.npy episode_5876.npy episode_6877.npy episode_7878.npy episode_8879.npy episode_987.npy\r\nepisode_1873.npy episode_2874.npy episode_3875.npy episode_4876.npy episode_5877.npy episode_6878.npy episode_7879.npy episode_887.npy episode_9880.npy\r\nepisode_1874.npy episode_2875.npy episode_3876.npy episode_4877.npy episode_5878.npy episode_6879.npy episode_787.npy episode_8880.npy episode_9881.npy\r\nepisode_1875.npy episode_2876.npy episode_3877.npy episode_4878.npy episode_5879.npy episode_687.npy episode_7880.npy episode_8881.npy episode_9882.npy\r\nepisode_1876.npy episode_2877.npy episode_3878.npy episode_4879.npy episode_587.npy episode_6880.npy episode_7881.npy episode_8882.npy episode_9883.npy\r\nepisode_1877.npy episode_2878.npy episode_3879.npy episode_487.npy episode_5880.npy episode_6881.npy episode_7882.npy episode_8883.npy episode_9884.npy\r\nepisode_1878.npy episode_2879.npy episode_387.npy episode_4880.npy episode_5881.npy episode_6882.npy episode_7883.npy episode_8884.npy episode_9885.npy\r\nepisode_1879.npy episode_287.npy episode_3880.npy episode_4881.npy episode_5882.npy episode_6883.npy episode_7884.npy episode_8885.npy episode_9886.npy\r\nepisode_187.npy episode_2880.npy episode_3881.npy episode_4882.npy episode_5883.npy episode_6884.npy episode_7885.npy episode_8886.npy episode_9887.npy\r\nepisode_1880.npy episode_2881.npy episode_3882.npy episode_4883.npy episode_5884.npy episode_6885.npy episode_7886.npy episode_8887.npy episode_9888.npy\r\nepisode_1881.npy episode_2882.npy episode_3883.npy episode_4884.npy episode_5885.npy episode_6886.npy episode_7887.npy episode_8888.npy episode_9889.npy\r\nepisode_1882.npy episode_2883.npy episode_3884.npy episode_4885.npy episode_5886.npy episode_6887.npy episode_7888.npy episode_8889.npy episode_988.npy\r\nepisode_1883.npy episode_2884.npy episode_3885.npy episode_4886.npy episode_5887.npy episode_6888.npy episode_7889.npy episode_888.npy episode_9890.npy\r\nepisode_1884.npy episode_2885.npy episode_3886.npy episode_4887.npy episode_5888.npy episode_6889.npy episode_788.npy episode_8890.npy episode_9891.npy\r\nepisode_1885.npy episode_2886.npy episode_3887.npy episode_4888.npy episode_5889.npy episode_688.npy episode_7890.npy episode_8891.npy episode_9892.npy\r\nepisode_1886.npy episode_2887.npy episode_3888.npy episode_4889.npy episode_588.npy episode_6890.npy episode_7891.npy episode_8892.npy episode_9893.npy\r\nepisode_1887.npy episode_2888.npy episode_3889.npy episode_488.npy episode_5890.npy episode_6891.npy episode_7892.npy episode_8893.npy episode_9894.npy\r\nepisode_1888.npy episode_2889.npy episode_388.npy episode_4890.npy episode_5891.npy episode_6892.npy episode_7893.npy episode_8894.npy episode_9895.npy\r\nepisode_1889.npy episode_288.npy episode_3890.npy episode_4891.npy episode_5892.npy episode_6893.npy episode_7894.npy episode_8895.npy episode_9896.npy\r\nepisode_188.npy episode_2890.npy episode_3891.npy episode_4892.npy episode_5893.npy episode_6894.npy episode_7895.npy episode_8896.npy episode_9897.npy\r\nepisode_1890.npy episode_2891.npy episode_3892.npy episode_4893.npy episode_5894.npy episode_6895.npy episode_7896.npy episode_8897.npy episode_9898.npy\r\nepisode_1891.npy episode_2892.npy episode_3893.npy episode_4894.npy episode_5895.npy episode_6896.npy episode_7897.npy episode_8898.npy episode_9899.npy\r\nepisode_1892.npy episode_2893.npy episode_3894.npy episode_4895.npy episode_5896.npy episode_6897.npy episode_7898.npy episode_8899.npy episode_989.npy\r\nepisode_1893.npy episode_2894.npy episode_3895.npy episode_4896.npy episode_5897.npy episode_6898.npy episode_7899.npy episode_889.npy episode_98.npy\r\nepisode_1894.npy episode_2895.npy episode_3896.npy episode_4897.npy episode_5898.npy episode_6899.npy episode_789.npy episode_88.npy episode_9900.npy\r\nepisode_1895.npy episode_2896.npy episode_3897.npy episode_4898.npy episode_5899.npy episode_689.npy episode_78.npy episode_8900.npy episode_9901.npy\r\nepisode_1896.npy episode_2897.npy episode_3898.npy episode_4899.npy episode_589.npy episode_68.npy episode_7900.npy episode_8901.npy episode_9902.npy\r\nepisode_1897.npy episode_2898.npy episode_3899.npy episode_489.npy episode_58.npy episode_6900.npy episode_7901.npy episode_8902.npy episode_9903.npy\r\nepisode_1898.npy episode_2899.npy episode_389.npy episode_48.npy episode_5900.npy episode_6901.npy episode_7902.npy episode_8903.npy episode_9904.npy\r\nepisode_1899.npy episode_289.npy episode_38.npy episode_4900.npy episode_5901.npy episode_6902.npy episode_7903.npy episode_8904.npy episode_9905.npy\r\nepisode_189.npy episode_28.npy episode_3900.npy episode_4901.npy episode_5902.npy episode_6903.npy episode_7904.npy episode_8905.npy episode_9906.npy\r\nepisode_18.npy episode_2900.npy episode_3901.npy episode_4902.npy episode_5903.npy episode_6904.npy episode_7905.npy episode_8906.npy episode_9907.npy\r\nepisode_1900.npy episode_2901.npy episode_3902.npy episode_4903.npy episode_5904.npy episode_6905.npy episode_7906.npy episode_8907.npy episode_9908.npy\r\nepisode_1901.npy episode_2902.npy episode_3903.npy episode_4904.npy episode_5905.npy episode_6906.npy episode_7907.npy episode_8908.npy episode_9909.npy\r\nepisode_1902.npy episode_2903.npy episode_3904.npy episode_4905.npy episode_5906.npy episode_6907.npy episode_7908.npy episode_8909.npy episode_990.npy\r\nepisode_1903.npy episode_2904.npy episode_3905.npy episode_4906.npy episode_5907.npy episode_6908.npy episode_7909.npy episode_890.npy episode_9910.npy\r\nepisode_1904.npy episode_2905.npy episode_3906.npy episode_4907.npy episode_5908.npy episode_6909.npy episode_790.npy episode_8910.npy episode_9911.npy\r\nepisode_1905.npy episode_2906.npy episode_3907.npy episode_4908.npy episode_5909.npy episode_690.npy episode_7910.npy episode_8911.npy episode_9912.npy\r\nepisode_1906.npy episode_2907.npy episode_3908.npy episode_4909.npy episode_590.npy episode_6910.npy episode_7911.npy episode_8912.npy episode_9913.npy\r\nepisode_1907.npy episode_2908.npy episode_3909.npy episode_490.npy episode_5910.npy episode_6911.npy episode_7912.npy episode_8913.npy episode_9914.npy\r\nepisode_1908.npy episode_2909.npy episode_390.npy episode_4910.npy episode_5911.npy episode_6912.npy episode_7913.npy episode_8914.npy episode_9915.npy\r\nepisode_1909.npy episode_290.npy episode_3910.npy episode_4911.npy episode_5912.npy episode_6913.npy episode_7914.npy episode_8915.npy episode_9916.npy\r\nepisode_190.npy episode_2910.npy episode_3911.npy episode_4912.npy episode_5913.npy episode_6914.npy episode_7915.npy episode_8916.npy episode_9917.npy\r\nepisode_1910.npy episode_2911.npy episode_3912.npy episode_4913.npy episode_5914.npy episode_6915.npy episode_7916.npy episode_8917.npy episode_9918.npy\r\nepisode_1911.npy episode_2912.npy episode_3913.npy episode_4914.npy episode_5915.npy episode_6916.npy episode_7917.npy episode_8918.npy episode_9919.npy\r\nepisode_1912.npy episode_2913.npy episode_3914.npy episode_4915.npy episode_5916.npy episode_6917.npy episode_7918.npy episode_8919.npy episode_991.npy\r\nepisode_1913.npy episode_2914.npy episode_3915.npy episode_4916.npy episode_5917.npy episode_6918.npy episode_7919.npy episode_891.npy episode_9920.npy\r\nepisode_1914.npy episode_2915.npy episode_3916.npy episode_4917.npy episode_5918.npy episode_6919.npy episode_791.npy episode_8920.npy episode_9921.npy\r\nepisode_1915.npy episode_2916.npy episode_3917.npy episode_4918.npy episode_5919.npy episode_691.npy episode_7920.npy episode_8921.npy episode_9922.npy\r\nepisode_1916.npy episode_2917.npy episode_3918.npy episode_4919.npy episode_591.npy episode_6920.npy episode_7921.npy episode_8922.npy episode_9923.npy\r\nepisode_1917.npy episode_2918.npy episode_3919.npy episode_491.npy episode_5920.npy episode_6921.npy episode_7922.npy episode_8923.npy episode_9924.npy\r\nepisode_1918.npy episode_2919.npy episode_391.npy episode_4920.npy episode_5921.npy episode_6922.npy episode_7923.npy episode_8924.npy episode_9925.npy\r\nepisode_1919.npy episode_291.npy episode_3920.npy episode_4921.npy episode_5922.npy episode_6923.npy episode_7924.npy episode_8925.npy episode_9926.npy\r\nepisode_191.npy episode_2920.npy episode_3921.npy episode_4922.npy episode_5923.npy episode_6924.npy episode_7925.npy episode_8926.npy episode_9927.npy\r\nepisode_1920.npy episode_2921.npy episode_3922.npy episode_4923.npy episode_5924.npy episode_6925.npy episode_7926.npy episode_8927.npy episode_9928.npy\r\nepisode_1921.npy episode_2922.npy episode_3923.npy episode_4924.npy episode_5925.npy episode_6926.npy episode_7927.npy episode_8928.npy episode_9929.npy\r\nepisode_1922.npy episode_2923.npy episode_3924.npy episode_4925.npy episode_5926.npy episode_6927.npy episode_7928.npy episode_8929.npy episode_992.npy\r\nepisode_1923.npy episode_2924.npy episode_3925.npy episode_4926.npy episode_5927.npy episode_6928.npy episode_7929.npy episode_892.npy episode_9930.npy\r\nepisode_1924.npy episode_2925.npy episode_3926.npy episode_4927.npy episode_5928.npy episode_6929.npy episode_792.npy episode_8930.npy episode_9931.npy\r\nepisode_1925.npy episode_2926.npy episode_3927.npy episode_4928.npy episode_5929.npy episode_692.npy episode_7930.npy episode_8931.npy episode_9932.npy\r\nepisode_1926.npy episode_2927.npy episode_3928.npy episode_4929.npy episode_592.npy episode_6930.npy episode_7931.npy episode_8932.npy episode_9933.npy\r\nepisode_1927.npy episode_2928.npy episode_3929.npy episode_492.npy episode_5930.npy episode_6931.npy episode_7932.npy episode_8933.npy episode_9934.npy\r\nepisode_1928.npy episode_2929.npy episode_392.npy episode_4930.npy episode_5931.npy episode_6932.npy episode_7933.npy episode_8934.npy episode_9935.npy\r\nepisode_1929.npy episode_292.npy episode_3930.npy episode_4931.npy episode_5932.npy episode_6933.npy episode_7934.npy episode_8935.npy episode_9936.npy\r\nepisode_192.npy episode_2930.npy episode_3931.npy episode_4932.npy episode_5933.npy episode_6934.npy episode_7935.npy episode_8936.npy episode_9937.npy\r\nepisode_1930.npy episode_2931.npy episode_3932.npy episode_4933.npy episode_5934.npy episode_6935.npy episode_7936.npy episode_8937.npy episode_9938.npy\r\nepisode_1931.npy episode_2932.npy episode_3933.npy episode_4934.npy episode_5935.npy episode_6936.npy episode_7937.npy episode_8938.npy episode_9939.npy\r\nepisode_1932.npy episode_2933.npy episode_3934.npy episode_4935.npy episode_5936.npy episode_6937.npy episode_7938.npy episode_8939.npy episode_993.npy\r\nepisode_1933.npy episode_2934.npy episode_3935.npy episode_4936.npy episode_5937.npy episode_6938.npy episode_7939.npy episode_893.npy episode_9940.npy\r\nepisode_1934.npy episode_2935.npy episode_3936.npy episode_4937.npy episode_5938.npy episode_6939.npy episode_793.npy episode_8940.npy episode_9941.npy\r\nepisode_1935.npy episode_2936.npy episode_3937.npy episode_4938.npy episode_5939.npy episode_693.npy episode_7940.npy episode_8941.npy episode_9942.npy\r\nepisode_1936.npy episode_2937.npy episode_3938.npy episode_4939.npy episode_593.npy episode_6940.npy episode_7941.npy episode_8942.npy episode_9943.npy\r\nepisode_1937.npy episode_2938.npy episode_3939.npy episode_493.npy episode_5940.npy episode_6941.npy episode_7942.npy episode_8943.npy episode_9944.npy\r\nepisode_1938.npy episode_2939.npy episode_393.npy episode_4940.npy episode_5941.npy episode_6942.npy episode_7943.npy episode_8944.npy episode_9945.npy\r\nepisode_1939.npy episode_293.npy episode_3940.npy episode_4941.npy episode_5942.npy episode_6943.npy episode_7944.npy episode_8945.npy episode_9946.npy\r\nepisode_193.npy episode_2940.npy episode_3941.npy episode_4942.npy episode_5943.npy episode_6944.npy episode_7945.npy episode_8946.npy episode_9947.npy\r\nepisode_1940.npy episode_2941.npy episode_3942.npy episode_4943.npy episode_5944.npy episode_6945.npy episode_7946.npy episode_8947.npy episode_9948.npy\r\nepisode_1941.npy episode_2942.npy episode_3943.npy episode_4944.npy episode_5945.npy episode_6946.npy episode_7947.npy episode_8948.npy episode_9949.npy\r\nepisode_1942.npy episode_2943.npy episode_3944.npy episode_4945.npy episode_5946.npy episode_6947.npy episode_7948.npy episode_8949.npy episode_994.npy\r\nepisode_1943.npy episode_2944.npy episode_3945.npy episode_4946.npy episode_5947.npy episode_6948.npy episode_7949.npy episode_894.npy episode_9950.npy\r\nepisode_1944.npy episode_2945.npy episode_3946.npy episode_4947.npy episode_5948.npy episode_6949.npy episode_794.npy episode_8950.npy episode_9951.npy\r\nepisode_1945.npy episode_2946.npy episode_3947.npy episode_4948.npy episode_5949.npy episode_694.npy episode_7950.npy episode_8951.npy episode_9952.npy\r\nepisode_1946.npy episode_2947.npy episode_3948.npy episode_4949.npy episode_594.npy episode_6950.npy episode_7951.npy episode_8952.npy episode_9953.npy\r\nepisode_1947.npy episode_2948.npy episode_3949.npy episode_494.npy episode_5950.npy episode_6951.npy episode_7952.npy episode_8953.npy episode_9954.npy\r\nepisode_1948.npy episode_2949.npy episode_394.npy episode_4950.npy episode_5951.npy episode_6952.npy episode_7953.npy episode_8954.npy episode_9955.npy\r\nepisode_1949.npy episode_294.npy episode_3950.npy episode_4951.npy episode_5952.npy episode_6953.npy episode_7954.npy episode_8955.npy episode_9956.npy\r\nepisode_194.npy episode_2950.npy episode_3951.npy episode_4952.npy episode_5953.npy episode_6954.npy episode_7955.npy episode_8956.npy episode_9957.npy\r\nepisode_1950.npy episode_2951.npy episode_3952.npy episode_4953.npy episode_5954.npy episode_6955.npy episode_7956.npy episode_8957.npy episode_9958.npy\r\nepisode_1951.npy episode_2952.npy episode_3953.npy episode_4954.npy episode_5955.npy episode_6956.npy episode_7957.npy episode_8958.npy episode_9959.npy\r\nepisode_1952.npy episode_2953.npy episode_3954.npy episode_4955.npy episode_5956.npy episode_6957.npy episode_7958.npy episode_8959.npy episode_995.npy\r\nepisode_1953.npy episode_2954.npy episode_3955.npy episode_4956.npy episode_5957.npy episode_6958.npy episode_7959.npy episode_895.npy episode_9960.npy\r\nepisode_1954.npy episode_2955.npy episode_3956.npy episode_4957.npy episode_5958.npy episode_6959.npy episode_795.npy episode_8960.npy episode_9961.npy\r\nepisode_1955.npy episode_2956.npy episode_3957.npy episode_4958.npy episode_5959.npy episode_695.npy episode_7960.npy episode_8961.npy episode_9962.npy\r\nepisode_1956.npy episode_2957.npy episode_3958.npy episode_4959.npy episode_595.npy episode_6960.npy episode_7961.npy episode_8962.npy episode_9963.npy\r\nepisode_1957.npy episode_2958.npy episode_3959.npy episode_495.npy episode_5960.npy episode_6961.npy episode_7962.npy episode_8963.npy episode_9964.npy\r\nepisode_1958.npy episode_2959.npy episode_395.npy episode_4960.npy episode_5961.npy episode_6962.npy episode_7963.npy episode_8964.npy episode_9965.npy\r\nepisode_1959.npy episode_295.npy episode_3960.npy episode_4961.npy episode_5962.npy episode_6963.npy episode_7964.npy episode_8965.npy episode_9966.npy\r\nepisode_195.npy episode_2960.npy episode_3961.npy episode_4962.npy episode_5963.npy episode_6964.npy episode_7965.npy episode_8966.npy episode_9967.npy\r\nepisode_1960.npy episode_2961.npy episode_3962.npy episode_4963.npy episode_5964.npy episode_6965.npy episode_7966.npy episode_8967.npy episode_9968.npy\r\nepisode_1961.npy episode_2962.npy episode_3963.npy episode_4964.npy episode_5965.npy episode_6966.npy episode_7967.npy episode_8968.npy episode_9969.npy\r\nepisode_1962.npy episode_2963.npy episode_3964.npy episode_4965.npy episode_5966.npy episode_6967.npy episode_7968.npy episode_8969.npy episode_996.npy\r\nepisode_1963.npy episode_2964.npy episode_3965.npy episode_4966.npy episode_5967.npy episode_6968.npy episode_7969.npy episode_896.npy episode_9970.npy\r\nepisode_1964.npy episode_2965.npy episode_3966.npy episode_4967.npy episode_5968.npy episode_6969.npy episode_796.npy episode_8970.npy episode_9971.npy\r\nepisode_1965.npy episode_2966.npy episode_3967.npy episode_4968.npy episode_5969.npy episode_696.npy episode_7970.npy episode_8971.npy episode_9972.npy\r\nepisode_1966.npy episode_2967.npy episode_3968.npy episode_4969.npy episode_596.npy episode_6970.npy episode_7971.npy episode_8972.npy episode_9973.npy\r\nepisode_1967.npy episode_2968.npy episode_3969.npy episode_496.npy episode_5970.npy episode_6971.npy episode_7972.npy episode_8973.npy episode_9974.npy\r\nepisode_1968.npy episode_2969.npy episode_396.npy episode_4970.npy episode_5971.npy episode_6972.npy episode_7973.npy episode_8974.npy episode_9975.npy\r\nepisode_1969.npy episode_296.npy episode_3970.npy episode_4971.npy episode_5972.npy episode_6973.npy episode_7974.npy episode_8975.npy episode_9976.npy\r\nepisode_196.npy episode_2970.npy episode_3971.npy episode_4972.npy episode_5973.npy episode_6974.npy episode_7975.npy episode_8976.npy episode_9977.npy\r\nepisode_1970.npy episode_2971.npy episode_3972.npy episode_4973.npy episode_5974.npy episode_6975.npy episode_7976.npy episode_8977.npy episode_9978.npy\r\nepisode_1971.npy episode_2972.npy episode_3973.npy episode_4974.npy episode_5975.npy episode_6976.npy episode_7977.npy episode_8978.npy episode_9979.npy\r\nepisode_1972.npy episode_2973.npy episode_3974.npy episode_4975.npy episode_5976.npy episode_6977.npy episode_7978.npy episode_8979.npy episode_997.npy\r\nepisode_1973.npy episode_2974.npy episode_3975.npy episode_4976.npy episode_5977.npy episode_6978.npy episode_7979.npy episode_897.npy episode_9980.npy\r\nepisode_1974.npy episode_2975.npy episode_3976.npy episode_4977.npy episode_5978.npy episode_6979.npy episode_797.npy episode_8980.npy episode_9981.npy\r\nepisode_1975.npy episode_2976.npy episode_3977.npy episode_4978.npy episode_5979.npy episode_697.npy episode_7980.npy episode_8981.npy episode_9982.npy\r\nepisode_1976.npy episode_2977.npy episode_3978.npy episode_4979.npy episode_597.npy episode_6980.npy episode_7981.npy episode_8982.npy episode_9983.npy\r\nepisode_1977.npy episode_2978.npy episode_3979.npy episode_497.npy episode_5980.npy episode_6981.npy episode_7982.npy episode_8983.npy episode_9984.npy\r\nepisode_1978.npy episode_2979.npy episode_397.npy episode_4980.npy episode_5981.npy episode_6982.npy episode_7983.npy episode_8984.npy episode_9985.npy\r\nepisode_1979.npy episode_297.npy episode_3980.npy episode_4981.npy episode_5982.npy episode_6983.npy episode_7984.npy episode_8985.npy episode_9986.npy\r\nepisode_197.npy episode_2980.npy episode_3981.npy episode_4982.npy episode_5983.npy episode_6984.npy episode_7985.npy episode_8986.npy episode_9987.npy\r\nepisode_1980.npy episode_2981.npy episode_3982.npy episode_4983.npy episode_5984.npy episode_6985.npy episode_7986.npy episode_8987.npy episode_9988.npy\r\nepisode_1981.npy episode_2982.npy episode_3983.npy episode_4984.npy episode_5985.npy episode_6986.npy episode_7987.npy episode_8988.npy episode_9989.npy\r\nepisode_1982.npy episode_2983.npy episode_3984.npy episode_4985.npy episode_5986.npy episode_6987.npy episode_7988.npy episode_8989.npy episode_998.npy\r\nepisode_1983.npy episode_2984.npy episode_3985.npy episode_4986.npy episode_5987.npy episode_6988.npy episode_7989.npy episode_898.npy episode_9990.npy\r\nepisode_1984.npy episode_2985.npy episode_3986.npy episode_4987.npy episode_5988.npy episode_6989.npy episode_798.npy episode_8990.npy episode_9991.npy\r\nepisode_1985.npy episode_2986.npy episode_3987.npy episode_4988.npy episode_5989.npy episode_698.npy episode_7990.npy episode_8991.npy episode_9992.npy\r\nepisode_1986.npy episode_2987.npy episode_3988.npy episode_4989.npy episode_598.npy episode_6990.npy episode_7991.npy episode_8992.npy episode_9993.npy\r\nepisode_1987.npy episode_2988.npy episode_3989.npy episode_498.npy episode_5990.npy episode_6991.npy episode_7992.npy episode_8993.npy episode_9994.npy\r\nepisode_1988.npy episode_2989.npy episode_398.npy episode_4990.npy episode_5991.npy episode_6992.npy episode_7993.npy episode_8994.npy episode_9995.npy\r\nepisode_1989.npy episode_298.npy episode_3990.npy episode_4991.npy episode_5992.npy episode_6993.npy episode_7994.npy episode_8995.npy episode_9996.npy\r\nepisode_198.npy episode_2990.npy episode_3991.npy episode_4992.npy episode_5993.npy episode_6994.npy episode_7995.npy episode_8996.npy episode_9997.npy\r\nepisode_1990.npy episode_2991.npy episode_3992.npy episode_4993.npy episode_5994.npy episode_6995.npy episode_7996.npy episode_8997.npy episode_9998.npy\r\nepisode_1991.npy episode_2992.npy episode_3993.npy episode_4994.npy episode_5995.npy episode_6996.npy episode_7997.npy episode_8998.npy episode_9999.npy\r\nepisode_1992.npy episode_2993.npy episode_3994.npy episode_4995.npy episode_5996.npy episode_6997.npy episode_7998.npy episode_8999.npy episode_999.npy\r\nepisode_1993.npy episode_2994.npy episode_3995.npy episode_4996.npy episode_5997.npy episode_6998.npy episode_7999.npy episode_899.npy episode_99.npy\r\nepisode_1994.npy episode_2995.npy episode_3996.npy episode_4997.npy episode_5998.npy episode_6999.npy episode_799.npy episode_89.npy episode_9.npy\r\nepisode_1995.npy episode_2996.npy episode_3997.npy episode_4998.npy episode_5999.npy episode_699.npy episode_79.npy episode_8.npy metadata.npy\r\nepisode_1996.npy episode_2997.npy episode_3998.npy episode_4999.npy episode_599.npy episode_69.npy episode_7.npy episode_9000.npy\r\nepisode_1997.npy episode_2998.npy episode_3999.npy episode_499.npy episode_59.npy episode_6.npy episode_8000.npy episode_9001.npy\r\nepisode_1998.npy episode_2999.npy episode_399.npy episode_49.npy episode_5.npy episode_7000.npy episode_8001.npy episode_9002.npy\r\nepisode_1999.npy episode_299.npy episode_39.npy episode_4.npy episode_6000.npy episode_7001.npy episode_8002.npy episode_9003.npy\r\nepisode_199.npy episode_29.npy episode_3.npy episode_5000.npy episode_6001.npy episode_7002.npy episode_8003.npy episode_9004.npy\r\nepisode_19.npy episode_2.npy episode_4000.npy episode_5001.npy episode_6002.npy episode_7003.npy episode_8004.npy episode_9005.npy\r\nepisode_1.npy episode_3000.npy episode_4001.npy episode_5002.npy episode_6003.npy episode_7004.npy episode_8005.npy episode_9006.npy\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes",,terminal_output +464,977938,"TERMINAL",0,0,"cd ..",,terminal_command +465,982659,"TERMINAL",0,0,"pwd",,terminal_command +466,986248,"TERMINAL",0,0,"srun",,terminal_focus +467,987223,"TERMINAL",0,0,"cd",,terminal_output +468,987383,"TERMINAL",0,0," ",,terminal_output +469,987508,"TERMINAL",0,0,".",,terminal_output +470,987693,"TERMINAL",0,0,".",,terminal_output +471,987785,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0401:~/Projects[?2004h(jasmine) [tum_cte0515@hkn0401 Projects]$ ",,terminal_output +472,988797,"TERMINAL",0,0,"cd",,terminal_output +473,989014,"TERMINAL",0,0," ",,terminal_output +474,989248,"TERMINAL",0,0,"j",,terminal_output +475,989368,"TERMINAL",0,0,"a",,terminal_output +476,989528,"TERMINAL",0,0,"",,terminal_output +477,990012,"TERMINAL",0,0,"f",,terminal_output +478,990173,"TERMINAL",0,0,"ar/",,terminal_output +479,990644,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0401:~/Projects/jafar[?2004h(jasmine) [tum_cte0515@hkn0401 jafar]$ ",,terminal_output +480,991039,"TERMINAL",0,0,"l",,terminal_output +481,991144,"TERMINAL",0,0,"s\r\n[?2004l\r",,terminal_output +482,991249,"TERMINAL",0,0,"generate_dataset.py genie.py models README.md sample.py train_dynamics.py train_tokenizer.py wandb\r\ngeneration_1757423819.8842306.gif LICENSE __pycache__ requirements.txt slurm train_lam.py utils\r\n]0;tum_cte0515@hkn0401:~/Projects/jafar[?2004h(jasmine) [tum_cte0515@hkn0401 jafar]$ ",,terminal_output +483,992852,"TERMINAL",0,0,"p",,terminal_output +484,993057,"TERMINAL",0,0,"y",,terminal_output +485,993412,"TERMINAL",0,0,"t",,terminal_output +486,993469,"TERMINAL",0,0,"h",,terminal_output +487,993652,"TERMINAL",0,0,"o",,terminal_output +488,993838,"TERMINAL",0,0,"n",,terminal_output +489,993919,"TERMINAL",0,0," ",,terminal_output +490,994713,"TERMINAL",0,0,"g",,terminal_output +491,994814,"TERMINAL",0,0,"e",,terminal_output +492,994865,"TERMINAL",0,0,"n",,terminal_output +493,994964,"TERMINAL",0,0,"",,terminal_output +494,995740,"TERMINAL",0,0,"e",,terminal_output +495,995954,"TERMINAL",0,0,"rat",,terminal_output +496,997237,"TERMINAL",0,0,"a",,terminal_output +497,997493,"TERMINAL",0,0,"",,terminal_output +498,997966,"TERMINAL",0,0,"",,terminal_output +499,998960,"TERMINAL",0,0,"t",,terminal_output +500,1000030,"TERMINAL",0,0,"´",,terminal_output +501,1000302,"TERMINAL",0,0,"",,terminal_output +502,1000470,"TERMINAL",0,0,"",,terminal_output +503,1001812,"TERMINAL",0,0,"",,terminal_output +504,1002549,"TERMINAL",0,0,"e",,terminal_output +505,1002812,"TERMINAL",0,0,"_dataset.py ",,terminal_output +506,1006198,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",0,0,"",python,tab +507,1011150,"TERMINAL",0,0,"-",,terminal_output +508,1011243,"TERMINAL",0,0,"-",,terminal_output +509,1011594,"TERMINAL",0,0,"o",,terminal_output +510,1011824,"TERMINAL",0,0,"u",,terminal_output +511,1012146,"TERMINAL",0,0,"t",,terminal_output +512,1013869,"TERMINAL",0,0,"p",,terminal_output +513,1014105,"TERMINAL",0,0,"u",,terminal_output +514,1014163,"TERMINAL",0,0,"t",,terminal_output +515,1014435,"TERMINAL",0,0,"_",,terminal_output +516,1014639,"TERMINAL",0,0,"d",,terminal_output +517,1014745,"TERMINAL",0,0,"i",,terminal_output +518,1014901,"TERMINAL",0,0,"r",,terminal_output +519,1016364,"TERMINAL",0,0,"#",,terminal_output +520,1017318,"TERMINAL",0,0,"",,terminal_output +521,1017591,"TERMINAL",0,0,"=",,terminal_output +522,1018273,"TERMINAL",0,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun",,terminal_output +523,1018958,"TERMINAL",0,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun""",,terminal_output +524,1019653,"TERMINAL",0,0,"",,terminal_output +525,1020445,"TERMINAL",0,0,"/",,terminal_output +526,1022087,"TERMINAL",0,0,"c",,terminal_output +527,1022149,"TERMINAL",0,0,"o",,terminal_output +528,1022413,"TERMINAL",0,0,"i",,terminal_output +529,1022988,"TERMINAL",0,0,"",,terminal_output +530,1023072,"TERMINAL",0,0,"",,terminal_output +531,1023291,"TERMINAL",0,0,"",,terminal_output +532,1023531,"TERMINAL",0,0,"n",,terminal_output +533,1023720,"TERMINAL",0,0,"p",,terminal_output +534,1023804,"TERMINAL",0,0,"y",,terminal_output +535,1024052,"TERMINAL",0,0,"_",,terminal_output +536,1024620,"TERMINAL",0,0,"t",,terminal_output +537,1024748,"TERMINAL",0,0,"e",,terminal_output +538,1024826,"TERMINAL",0,0,"s",,terminal_output +539,1024920,"TERMINAL",0,0,"t",,terminal_output +540,1028194,"TERMINAL",0,0," ",,terminal_output +541,1028522,"TERMINAL",0,0,"-",,terminal_output +542,1028610,"TERMINAL",0,0,"-",,terminal_output +543,1029933,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",0,0,"",python,tab +544,1031721,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",328,0,"",python,selection_mouse +545,1032116,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",323,12,"num_episodes",python,selection_mouse +546,1035680,"TERMINAL",0,0,"num_episodes",,terminal_output +547,1036824,"TERMINAL",0,0,"num_episodes=1",,terminal_output +548,1037029,"TERMINAL",0,0,"00",,terminal_output +549,1037674,"TERMINAL",0,0,"\r\n[?2004l\rTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py"", line 9, in \r\n from gym3 import types_np\r\nModuleNotFoundError: No module named 'gym3'\r\n]0;tum_cte0515@hkn0401:~/Projects/jafar[?2004h(jasmine) [tum_cte0515@hkn0401 jafar]$ ",,terminal_output +550,1041404,"TERMINAL",0,0,"d",,terminal_output +551,1041552,"TERMINAL",0,0,"e",,terminal_output +552,1041911,"TERMINAL",0,0,"a",,terminal_output +553,1042007,"TERMINAL",0,0,"c",,terminal_output +554,1042108,"TERMINAL",0,0,"t",,terminal_output +555,1042325,"TERMINAL",0,0,"i",,terminal_output +556,1042544,"TERMINAL",0,0,"va",,terminal_output +557,1042746,"TERMINAL",0,0,"t",,terminal_output +558,1042883,"TERMINAL",0,0,"e",,terminal_output +559,1042976,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0401:~/Projects/jafar[?2004h[tum_cte0515@hkn0401 jafar]$ ",,terminal_output +560,1043415,"TERMINAL",0,0,"so",,terminal_output +561,1043529,"TERMINAL",0,0,"u",,terminal_output +562,1043597,"TERMINAL",0,0,"r",,terminal_output +563,1043816,"TERMINAL",0,0,"c",,terminal_output +564,1043930,"TERMINAL",0,0,"e",,terminal_output +565,1044028,"TERMINAL",0,0," ",,terminal_output +566,1044100,"TERMINAL",0,0,".",,terminal_output +567,1044220,"TERMINAL",0,0,"v",,terminal_output +568,1044304,"TERMINAL",0,0,"env/",,terminal_output +569,1044698,"TERMINAL",0,0,"b",,terminal_output +570,1044793,"TERMINAL",0,0,"in/",,terminal_output +571,1044989,"TERMINAL",0,0,"ac",,terminal_output +572,1045197,"TERMINAL",0,0,"tivate",,terminal_output +573,1045678,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0401:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0401 jafar]$ ",,terminal_output +574,1046147,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_output +575,1046276,"TERMINAL",0,0,"deactivate",,terminal_output +576,1047014,"TERMINAL",0,0,"python generate_dataset.py --output_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test --num_episodes=100",,terminal_output +577,1047671,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +578,1053347,"TERMINAL",0,0,"Gym has been unmaintained since 2022 and does not support NumPy 2.0 amongst other critical functionality.\r\nPlease upgrade to Gymnasium, the maintained drop-in replacement of Gym, or contact the authors of your software and request that they upgrade.\r\nUsers of this version of Gym should be able to simply replace 'import gym' with 'import gymnasium as gym' in the vast majority of cases.\r\nSee the migration guide at https://gymnasium.farama.org/introduction/migration_guide/ for additional information.\r\n",,terminal_output +579,1062382,"TERMINAL",0,0,"Episode 0 completed, length: 910\r\n",,terminal_output +580,1062517,"TERMINAL",0,0,"Episode 1 completed, length: 653\r\n",,terminal_output +581,1062705,"TERMINAL",0,0,"Episode 2 completed, length: 1000\r\n",,terminal_output +582,1062834,"TERMINAL",0,0,"Episode 3 completed, length: 715\r\n",,terminal_output +583,1062960,"TERMINAL",0,0,"Episode 4 completed, length: 634\r\n",,terminal_output +584,1063084,"TERMINAL",0,0,"Episode 5 completed, length: 527\r\n",,terminal_output +585,1063364,"TERMINAL",0,0,"Episode 6 completed, length: 1000\r\n",,terminal_output +586,1063589,"TERMINAL",0,0,"Episode 7 completed, length: 1000\r\n",,terminal_output +587,1063647,"TERMINAL",0,0,"Episode 8 completed, length: 1000\r\n",,terminal_output +588,1063726,"TERMINAL",0,0,"Episode 9 completed, length: 448\r\nEpisode 10 completed, length: 109\r\n",,terminal_output +589,1063946,"TERMINAL",0,0,"Episode 11 completed, length: 948\r\n",,terminal_output +590,1064209,"TERMINAL",0,0,"Episode 12 completed, length: 1000\r\n",,terminal_output +591,1064338,"TERMINAL",0,0,"Episode 13 completed, length: 356\r\n",,terminal_output +592,1064558,"TERMINAL",0,0,"Episode 14 completed, length: 1000\r\nEpisode 15 completed, length: 214\r\n",,terminal_output +593,1064653,"TERMINAL",0,0,"Episode 16 completed, length: 424\r\n",,terminal_output +594,1064818,"TERMINAL",0,0,"Episode 17 completed, length: 624\r\n",,terminal_output +595,1064877,"TERMINAL",0,0,"Episode 18 completed, length: 829\r\n",,terminal_output +596,1064975,"TERMINAL",0,0,"Episode 19 completed, length: 1000\r\n",,terminal_output +597,1065152,"TERMINAL",0,0,"Episode 20 completed, length: 775\r\n",,terminal_output +598,1065215,"TERMINAL",0,0,"Episode 21 completed, length: 1000\r\n",,terminal_output +599,1065485,"TERMINAL",0,0,"Episode 22 completed, length: 1000\r\n",,terminal_output +600,1065567,"TERMINAL",0,0,"Episode 23 completed, length: 1000\r\n",,terminal_output +601,1065628,"TERMINAL",0,0,"Episode 24 completed, length: 383\r\n",,terminal_output +602,1065778,"TERMINAL",0,0,"Episode 25 completed, length: 1000\r\nEpisode 26 completed, length: 301\r\n",,terminal_output +603,1065904,"TERMINAL",0,0,"Episode 27 completed, length: 1000\r\nEpisode 28 completed, length: 562\r\nEpisode 29 completed, length: 138\r\n",,terminal_output +604,1066001,"TERMINAL",0,0,"Episode 30 completed, length: 282\r\n",,terminal_output +605,1066437,"TERMINAL",0,0,"Episode 31 completed, length: 863\r\nEpisode 32 completed, length: 321\r\nEpisode 33 completed, length: 1000\r\nEpisode 34 completed, length: 288\r\n",,terminal_output +606,1066591,"TERMINAL",0,0,"Episode 35 completed, length: 850\r\n",,terminal_output +607,1066715,"TERMINAL",0,0,"Episode 36 completed, length: 242\r\nEpisode 37 completed, length: 719\r\n",,terminal_output +608,1066829,"TERMINAL",0,0,"Episode 38 completed, length: 681\r\n",,terminal_output +609,1067055,"TERMINAL",0,0,"Episode 39 completed, length: 1000\r\n",,terminal_output +610,1067327,"TERMINAL",0,0,"Episode 40 completed, length: 1000\r\n",,terminal_output +611,1067773,"TERMINAL",0,0,"Episode 41 completed, length: 1000\r\nEpisode 42 completed, length: 157\r\nEpisode 43 completed, length: 433\r\nEpisode 44 completed, length: 166\r\nEpisode 45 completed, length: 250\r\n",,terminal_output +612,1067991,"TERMINAL",0,0,"Episode 46 completed, length: 1000\r\nEpisode 47 completed, length: 1000\r\nEpisode 48 completed, length: 54\r\nEpisode 49 completed, length: 361\r\nEpisode 50 completed, length: 248\r\n",,terminal_output +613,1068057,"TERMINAL",0,0,"Episode 51 completed, length: 223\r\nEpisode 52 completed, length: 57\r\n",,terminal_output +614,1068256,"TERMINAL",0,0,"Episode 53 completed, length: 988\r\n",,terminal_output +615,1068627,"TERMINAL",0,0,"Episode 54 completed, length: 1000\r\nEpisode 55 completed, length: 197\r\n",,terminal_output +616,1068722,"TERMINAL",0,0,"Episode 56 completed, length: 342\r\n",,terminal_output +617,1068953,"TERMINAL",0,0,"Episode 57 completed, length: 1000\r\n",,terminal_output +618,1069105,"TERMINAL",0,0,"Episode 58 completed, length: 1000\r\n",,terminal_output +619,1069184,"TERMINAL",0,0,"Episode 59 completed, length: 1000\r\n",,terminal_output +620,1069502,"TERMINAL",0,0,"Episode 60 completed, length: 1000\r\n",,terminal_output +621,1069732,"TERMINAL",0,0,"Episode 61 completed, length: 1000\r\nEpisode 62 completed, length: 298\r\n",,terminal_output +622,1070053,"TERMINAL",0,0,"Episode 63 completed, length: 1000\r\nEpisode 64 completed, length: 221\r\nEpisode 65 completed, length: 62\r\n",,terminal_output +623,1070182,"TERMINAL",0,0,"Episode 66 completed, length: 849\r\nEpisode 67 completed, length: 186\r\nEpisode 68 completed, length: 730\r\n",,terminal_output +624,1070233,"TERMINAL",0,0,"Episode 69 completed, length: 328\r\n",,terminal_output +625,1070467,"TERMINAL",0,0,"Episode 70 completed, length: 1000\r\n",,terminal_output +626,1070966,"TERMINAL",0,0,"Episode 71 completed, length: 1000\r\nEpisode 72 completed, length: 365\r\nEpisode 73 completed, length: 251\r\nEpisode 74 completed, length: 79\r\n",,terminal_output +627,1071388,"TERMINAL",0,0,"Episode 75 completed, length: 575\r\n",,terminal_output +628,1071468,"TERMINAL",0,0,"Episode 76 completed, length: 1000\r\n",,terminal_output +629,1071728,"TERMINAL",0,0,"Episode 77 completed, length: 927\r\nEpisode 78 completed, length: 211\r\nEpisode 79 completed, length: 287\r\n",,terminal_output +630,1071868,"TERMINAL",0,0,"Episode 80 completed, length: 423\r\nEpisode 81 completed, length: 502\r\n",,terminal_output +631,1071966,"TERMINAL",0,0,"Episode 82 completed, length: 841\r\n",,terminal_output +632,1072069,"TERMINAL",0,0,"Episode 83 completed, length: 645\r\n",,terminal_output +633,1072607,"TERMINAL",0,0,"Episode 84 completed, length: 669\r\nEpisode 85 completed, length: 293\r\nEpisode 86 completed, length: 262\r\nEpisode 87 completed, length: 381\r\nEpisode 88 completed, length: 570\r\nEpisode 89 completed, length: 238\r\nEpisode 90 completed, length: 1000\r\nEpisode 91 completed, length: 306\r\n",,terminal_output +634,1072748,"TERMINAL",0,0,"Episode 92 completed, length: 1000\r\n",,terminal_output +635,1072984,"TERMINAL",0,0,"Episode 93 completed, length: 1000\r\n",,terminal_output +636,1073127,"TERMINAL",0,0,"Episode 94 completed, length: 460\r\nEpisode 95 completed, length: 189\r\n",,terminal_output +637,1073500,"TERMINAL",0,0,"Episode 96 completed, length: 1000\r\n",,terminal_output +638,1073567,"TERMINAL",0,0,"Episode 97 completed, length: 599\r\n",,terminal_output +639,1073817,"TERMINAL",0,0,"Episode 98 completed, length: 1000\r\nEpisode 99 completed, length: 69\r\nDataset generated with 100 valid episodes\r\n",,terminal_output +640,1073993,"TERMINAL",0,0,"]0;tum_cte0515@hkn0401:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0401 jafar]$ ",,terminal_output +641,1075879,"TERMINAL",0,0,"d",,terminal_output +642,1076042,"TERMINAL",0,0,"e",,terminal_output +643,1076228,"TERMINAL",0,0,"a",,terminal_output +644,1076297,"TERMINAL",0,0,"c",,terminal_output +645,1076411,"TERMINAL",0,0,"t",,terminal_output +646,1076568,"TERMINAL",0,0,"i",,terminal_output +647,1076745,"TERMINAL",0,0,"v",,terminal_output +648,1076828,"TERMINAL",0,0,"at",,terminal_output +649,1076889,"TERMINAL",0,0,"e",,terminal_output +650,1077510,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0401:~/Projects/jafar[?2004h[tum_cte0515@hkn0401 jafar]$ ",,terminal_output +651,1077909,"TERMINAL",0,0,"c",,terminal_output +652,1078007,"TERMINAL",0,0,"d",,terminal_output +653,1078161,"TERMINAL",0,0," ",,terminal_output +654,1078232,"TERMINAL",0,0,".",,terminal_output +655,1078426,"TERMINAL",0,0,".",,terminal_output +656,1078526,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0401:~/Projects[?2004h[tum_cte0515@hkn0401 Projects]$ ",,terminal_output +657,1078708,"TERMINAL",0,0,"l",,terminal_output +658,1078801,"TERMINAL",0,0,"s",,terminal_output +659,1078873,"TERMINAL",0,0,"\r\n[?2004l\rcheckpoints jafar jasmine jasmine_jobs jasmine_jobs_2 jasmine_jobs_3 tmp\r\n]0;tum_cte0515@hkn0401:~/Projects[?2004h[tum_cte0515@hkn0401 Projects]$ ",,terminal_output +660,1079411,"TERMINAL",0,0,"cd",,terminal_output +661,1079543,"TERMINAL",0,0," ",,terminal_output +662,1079892,"TERMINAL",0,0,"j",,terminal_output +663,1079987,"TERMINAL",0,0,"a",,terminal_output +664,1080227,"TERMINAL",0,0,"m",,terminal_output +665,1080620,"TERMINAL",0,0,"s",,terminal_output +666,1080692,"TERMINAL",0,0,"m",,terminal_output +667,1080784,"TERMINAL",0,0,"ine",,terminal_output +668,1081052,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h[tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +669,1081115,"TERMINAL",0,0,"l",,terminal_output +670,1081216,"TERMINAL",0,0,"s",,terminal_output +671,1081301,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +672,1081444,"TERMINAL",0,0,"ali-old-branch.diff diff2.diff frame.png input_pipeline killer.sh logs overfit_dir.zip README.md scripts_cremers tests wandb\r\ndata diff.diff frames jasmine LICENSE message.md __pycache__ requirements-franz.txt slurm utils\r\ndebug frame-knoms.png gifs killer_partition.sh log.log models pyproject.toml samples test.py uv.lock\r\n]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h[tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +673,1083662,"TERMINAL",0,0,"ls",,terminal_output +674,1083754,"TERMINAL",0,0,"cd jasmine",,terminal_output +675,1083987,"TERMINAL",0,0,"ls",,terminal_output +676,1084101,"TERMINAL",0,0,"cd ..",,terminal_output +677,1084206,"TERMINAL",0,0,"deactivate",,terminal_output +678,1084404,"TERMINAL",0,0,"python generate_dataset.py --output_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test --num_episodes=100",,terminal_output +679,1084560,"TERMINAL",0,0,"\rsource .venv/bin/activate",,terminal_output +680,1084646,"TERMINAL",0,0,"deactivate",,terminal_output +681,1085311,"TERMINAL",0,0,"python generate_dataset.py --output_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test --num_episodes=100",,terminal_output +682,1085435,"TERMINAL",0,0,"\rls",,terminal_output +683,1085964,"TERMINAL",0,0,"cd jafar/",,terminal_output +684,1086312,"TERMINAL",0,0,"ls",,terminal_output +685,1086595,"TERMINAL",0,0,"cd jafar/",,terminal_output +686,1086746,"TERMINAL",0,0,"..",,terminal_output +687,1086991,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +688,1087242,"TERMINAL",0,0,"\rource .venv/bin/activate",,terminal_output +689,1087830,"TERMINAL",0,0,"h slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +690,1089193,"TERMINAL",0,0,"bash",,terminal_focus +691,1090273,"TERMINAL",0,0,"ls",,terminal_command +692,1090304,"TERMINAL",0,0,"]633;C",,terminal_output +693,1090394,"TERMINAL",0,0,"array_records coinrun_episodes_10m_chunked coinrun_episodes_10m_gt_actions_split coinrun_episodes_500m_gt_actions_split\r\ncoinrun_episodes coinrun_episodes_10m_gt_actions coinrun_episodes_10m_gt_actions_split_test coinrun_episodes_test\r\ncoinrun_episodes_10m coinrun_episodes_10m_gt_actions_distinct_seed coinrun_episodes_10m_split npy_test\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun",,terminal_output +694,1096669,"TERMINAL",0,0,"cd npy_test/",,terminal_command +695,1097421,"TERMINAL",0,0,"pwd",,terminal_command +696,1100492,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +697,1103081,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,0,"",shellscript,selection_mouse +698,1104197,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,85,"",shellscript,content +699,1104210,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",657,0,"",shellscript,selection_command +700,1104547,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,0,"",shellscript,selection_command +701,1104828,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",658,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test",shellscript,content +702,1106455,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",759,0,"",shellscript,selection_mouse +703,1107239,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",758,0,"",shellscript,selection_command +704,1108066,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",757,0,"",shellscript,selection_command +705,1108203,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",758,0,"",shellscript,selection_command +706,1109276,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",758,85,"",shellscript,content +707,1109279,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",757,0,"",shellscript,selection_command +708,1109628,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",758,0,"",shellscript,selection_command +709,1110319,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",758,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test",shellscript,content +710,1110962,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",834,0,"",shellscript,selection_command +711,1113680,"TERMINAL",0,0,"srun",,terminal_focus +712,1114963,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --val_interval 2 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +713,1115091,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +714,1115208,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +715,1119169,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +716,1125595,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +717,1126028,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +718,1126955,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_121919-cdtzddag\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/cdtzddag\r\nParameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +719,1128418,"TERMINAL",0,0,"============\r\n\r\n /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test \r\n\r\n============\r\n",,terminal_output +720,1129106,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +721,1133843,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 745, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 587, in main\r\n first_batch = next(dataloader_train)\r\n ^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 557, in \r\n videos_sharding, local_data=elem[""videos""]\r\n ~~~~^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/array.py"", line 386, in __getitem__\r\n return indexing.rewriting_take(self, idx)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/numpy/indexing.py"", line 655, in rewriting_take\r\n treedef, static_idx, dynamic_idx = split_index_for_jit(idx, arr.shape)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/numpy/indexing.py"", line 751, in split_index_for_jit\r\n raise TypeError(f""JAX does not support string indexing; got {idx=}"")\r\nTypeError: JAX does not support string indexing; got idx=('videos',)\r\n",,terminal_output +722,1134472,"TERMINAL",0,0,"wandb: \r\nwandb: 🚀 View run coinrun-dyn-dev-3536670 at: https://wandb.ai/instant-uv/jafar/runs/cdtzddag\r\nwandb: Find logs at: ../../../../../hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_121919-cdtzddag/logs\r\n",,terminal_output +723,1134637,"TERMINAL",0,0,"W1002 12:19:27.778152 1470987 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugproto job_name: ""jax_worker"": CANCELLED: CANCELLED\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_status:1, grpc_message:""CANCELLED""} [type.googleapis.com/tensorflow.DerivedStatus='']\r\n",,terminal_output +724,1135310,"TERMINAL",0,0,"srun: error: hkn0401: task 0: Exited with exit code 1\r\n]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h[tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +725,1152022,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"import os\n\n\nos.environ.setdefault(""XLA_PYTHON_CLIENT_MEM_FRACTION"", ""0.98"")\n\nfrom dataclasses import dataclass, field\nimport itertools\nfrom typing import cast, Optional\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport flax.nnx as nnx\nfrom torch.utils.data import DataLoader\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader_torch import get_dataloader\nfrom utils.train_utils import (\n get_lr_schedule,\n count_parameters_by_component,\n print_mem_stats,\n print_compiled_memory_stats,\n print_compiled_cost_analysis,\n)\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n # Tokenizer\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_type: str = ""maskgit"" # supported options: maskgit, causal\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n z_loss_weight: float = 0.0\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n use_gt_actions: bool = False\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n eval_full_frame: bool = False\n val_maskgit_steps: int = 25\n val_temperature: float = 1\n val_sample_argmax: bool = False\n wandb_id: str = """"\n\n\ndef build_model(args: Args, rng: jax.Array) -> tuple[Genie, jax.Array]:\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_actions=args.num_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n use_gt_actions=args.use_gt_actions,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n decode=False,\n rngs=rngs,\n )\n if args.use_gt_actions:\n assert (\n not args.lam_checkpoint\n ), ""Cannot use LAM when using ground-truth actions.""\n else:\n assert genie.lam is not None\n del genie.lam.decoder\n return genie, rng\n\n\ndef build_optimizer(genie: Genie, args: Args) -> nnx.ModelAndOptimizer:\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.param_dtype, # moments in full precision\n )\n optimizer = nnx.ModelAndOptimizer(genie, tx)\n return optimizer\n\n\ndef build_mesh_and_sharding(\n num_devices: int,\n) -> tuple[Mesh, NamedSharding, NamedSharding, NamedSharding]:\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n actions_sharding = NamedSharding(mesh, PartitionSpec(""data"", None))\n return mesh, replicated_sharding, videos_sharding, actions_sharding\n\n\ndef shard_optimizer_states(\n optimizer: nnx.ModelAndOptimizer, replicated_sharding: NamedSharding\n) -> None:\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n\ndef build_dataloader(args: Args, data_dir: str) -> DataLoader:\n return get_dataloader(data_dir, args.seq_len, args.batch_size)\n\n\ndef build_checkpoint_manager(args: Args) -> Optional[ocp.CheckpointManager]:\n if args.restore_ckpt or args.save_ckpt:\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n return checkpoint_manager\n else:\n return None\n\n\ndef restore_or_initialize_components(\n args: Args,\n checkpoint_manager: Optional[ocp.CheckpointManager],\n optimizer: nnx.ModelAndOptimizer,\n rng: jax.Array,\n replicated_sharding: NamedSharding,\n restore_step: Optional[int] = None,\n) -> tuple[\n int,\n nnx.ModelAndOptimizer,\n jax.Array,\n]:\n step = 0\n if checkpoint_manager and restore_step is None:\n restore_step = checkpoint_manager.latest_step()\n if args.restore_ckpt:\n assert checkpoint_manager is not None\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n )\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(), args=restore_args\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n rng, _rng = jax.random.split(rng)\n optimizer = restore_genie_components(optimizer, replicated_sharding, _rng, args)\n return step, optimizer, rng\n\n\ndef _calculate_top_k_accuracy(\n token_logits_BTNV: jax.Array,\n video_tokens_BTN: jax.Array,\n mask_BTN: jax.Array,\n k: int,\n) -> jax.Array:\n _, topk_indices_BTNK = jax.lax.top_k(token_logits_BTNV, k)\n topk_correct = jnp.any(\n topk_indices_BTNK == video_tokens_BTN[..., jnp.newaxis], axis=-1\n )\n topk_acc = (mask_BTN * topk_correct).sum() / mask_BTN.sum()\n return topk_acc\n\n\ndef _calculate_step_metrics(\n outputs: dict[str, jax.Array],\n gt: jax.Array,\n num_actions: int,\n num_patch_latents: int,\n) -> tuple[jax.Array, dict]:\n mask_BTN = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask_BTN * ce_loss).sum() / mask_BTN.sum()\n z_val = jax.nn.logsumexp(outputs[""token_logits""], axis=-1)\n z_loss_metric = (mask_BTN * (z_val**2)).sum() / mask_BTN.sum()\n\n masked_token_top_1_acc = _calculate_top_k_accuracy(\n outputs[""token_logits""], outputs[""video_tokens""], mask_BTN, 1\n )\n masked_token_top_2_acc = _calculate_top_k_accuracy(\n outputs[""token_logits""], outputs[""video_tokens""], mask_BTN, 2\n )\n masked_token_top_5_acc = _calculate_top_k_accuracy(\n outputs[""token_logits""], outputs[""video_tokens""], mask_BTN, 5\n )\n masked_token_top_16_acc = _calculate_top_k_accuracy(\n outputs[""token_logits""], outputs[""video_tokens""], mask_BTN, 16\n )\n\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n gt_val = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt_val, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt_val, recon)).mean()\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]),\n size=num_patch_latents,\n fill_value=0,\n )\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_top1_accuracy=masked_token_top_1_acc,\n masked_token_top2_accuracy=masked_token_top_2_acc,\n masked_token_top5_accuracy=masked_token_top_5_acc,\n masked_token_top16_accuracy=masked_token_top_16_acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n z_loss=z_loss_metric,\n psnr=psnr,\n ssim=ssim,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n if ""lam_indices"" in outputs.keys():\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]),\n size=num_actions,\n fill_value=0,\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n metrics[""codebook_usage_lam""] = codebook_usage_lam\n return ce_loss, metrics\n\n\ndef main(args: Args) -> None:\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n genie, rng = build_model(args, rng)\n _, params, _ = nnx.split(genie, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n optimizer = build_optimizer(genie, args)\n del genie\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n _, replicated_sharding, videos_sharding, actions_sharding = build_mesh_and_sharding(\n num_devices\n )\n\n shard_optimizer_states(optimizer, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n checkpoint_manager = build_checkpoint_manager(args)\n\n # --- Create DataLoaderIterator from dataloader ---\n print(""============\n\n"", args.data_dir, ""\n\n============"")\n train_iterator = build_dataloader(args, args.data_dir)\n val_iterator = None\n if args.val_data_dir:\n val_iterator = build_dataloader(args, args.val_data_dir)\n\n # --- Restore checkpoint ---\n step, optimizer, rng = restore_or_initialize_components(\n args,\n checkpoint_manager,\n optimizer,\n rng,\n replicated_sharding,\n )\n\n # --- Define loss and train step (close over args) ---\n def dynamics_loss_fn(\n model: Genie,\n inputs: dict,\n ) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs)\n ce_loss, metrics = _calculate_step_metrics(\n outputs, gt, args.num_actions, args.num_patch_latents\n )\n z_loss = metrics[""z_loss""]\n total_loss = ce_loss + args.z_loss_weight * z_loss\n metrics[""total_loss""] = total_loss\n return total_loss, (outputs[""recon""], metrics)\n\n @nnx.jit(donate_argnums=0)\n def train_step(\n optimizer: nnx.ModelAndOptimizer, inputs: dict\n ) -> tuple[jax.Array, jax.Array, dict]:\n def loss_fn(model: Genie) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n model.train()\n return dynamics_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\n optimizer.model\n )\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return loss, recon, metrics\n\n @nnx.jit\n def val_step(genie: Genie, inputs: dict) -> dict:\n """"""Evaluate model and compute metrics""""""\n genie.eval()\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n (loss, (recon, metrics)) = dynamics_loss_fn(genie, inputs)\n val_output = {""loss"": loss, ""recon"": recon, ""metrics"": metrics}\n\n # --- Evaluate full frame prediction (sampling) ---\n if args.eval_full_frame:\n inputs[""videos""] = gt.astype(args.dtype)\n tokenizer_outputs = genie.tokenizer.vq_encode(\n inputs[""videos""], training=False\n )\n tokens_full_frame = tokenizer_outputs[""indices""]\n lam_indices_E = None\n if not args.use_gt_actions:\n lam_indices_E = genie.vq_encode(inputs, training=False)\n inputs[""latent_actions""] = lam_indices_E\n inputs[""videos""] = inputs[""videos""][\n :, :-1\n ] # remove last frame for generation\n recon_full_frame, logits_full_frame = genie.sample(\n inputs,\n args.seq_len,\n args.val_temperature,\n args.val_sample_argmax,\n args.val_maskgit_steps,\n )\n # Calculate metrics for the last frame only\n step_outputs = {\n ""recon"": recon_full_frame[:, -1],\n ""token_logits"": logits_full_frame[:, -1],\n ""video_tokens"": tokens_full_frame[:, -1],\n ""mask"": jnp.ones_like(tokens_full_frame[:, -1]),\n }\n if lam_indices_E is not None:\n lam_indices_B = lam_indices_E.reshape((-1, args.seq_len - 1))[:, -1]\n step_outputs[""lam_indices""] = lam_indices_B\n\n loss_full_frame, metrics_full_frame = _calculate_step_metrics(\n step_outputs, gt[:, -1], args.num_actions, args.num_patch_latents\n )\n val_output.update(\n {\n ""loss_full_frame"": loss_full_frame,\n ""recon_full_frame"": recon_full_frame,\n ""metrics_full_frame"": metrics_full_frame,\n }\n )\n return val_output\n\n def calculate_validation_metrics(val_dataloader, genie, rng):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n loss_full_frame_per_step = []\n metrics_full_frame_per_step = []\n batch = None\n recon = None\n recon_full_frame = None\n for batch in val_dataloader:\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n val_outputs = val_step(genie, batch)\n loss_per_step.append(val_outputs[""loss""])\n metrics_per_step.append(val_outputs[""metrics""])\n recon = val_outputs[""recon""]\n if args.eval_full_frame:\n loss_full_frame_per_step.append(val_outputs[""loss_full_frame""])\n metrics_full_frame_per_step.append(val_outputs[""metrics_full_frame""])\n recon_full_frame = val_outputs[""recon_full_frame""]\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(\n f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}""\n )\n\n val_metrics = {\n f""val_{key}"": np.mean([float(m[key]) for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n val_metrics[""val_loss""] = np.mean(loss_per_step)\n if args.eval_full_frame:\n val_metrics_full_frame = {\n f""val_full_frame_{key}"": np.mean(\n [float(m[key]) for m in metrics_full_frame_per_step]\n )\n for key in metrics_full_frame_per_step[0].keys()\n }\n val_metrics.update(val_metrics_full_frame)\n val_metrics[""val_full_frame_loss""] = np.mean(loss_full_frame_per_step)\n return val_metrics, batch, recon, recon_full_frame\n\n # --- TRAIN LOOP ---\n dataloader_train = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in train_iterator\n )\n dataloader_val = None\n if val_iterator:\n dataloader_val = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in val_iterator\n )\n if jax.process_index() == 0:\n first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader_train = itertools.chain([first_batch], dataloader_train)\n print(f""Starting training from step {step}..."")\n first_step = step\n while step < args.num_steps:\n for batch in dataloader_train:\n # --- Train step ---\n rng, _rng_mask = jax.random.split(rng, 2)\n batch[""rng""] = _rng_mask\n loss, recon, metrics = train_step(optimizer, batch)\n if step == first_step:\n print_mem_stats(""After params initialized"")\n step += 1\n\n # --- Validation loss ---\n val_results = {}\n if dataloader_val and step % args.val_interval == 0:\n rng, _rng_mask_val = jax.random.split(rng, 2)\n print(""Calculating validation metrics..."")\n val_metrics, val_gt_batch, val_recon, val_recon_full_frame = (\n calculate_validation_metrics(\n dataloader_val, optimizer.model, _rng_mask_val\n )\n )\n print(f""Step {step}, validation loss: {val_metrics['val_loss']}"")\n val_results = {\n ""metrics"": val_metrics,\n ""gt_batch"": val_gt_batch,\n ""recon"": val_recon,\n ""full_frame"": val_recon_full_frame,\n }\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n log_dict = {""loss"": loss, ""step"": step, **metrics}\n if val_results:\n log_dict.update(val_results[""metrics""])\n wandb.log(log_dict)\n if step % args.log_image_interval == 0:\n gt_seq = batch[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if val_results:\n val_results[""gt_seq_val""] = (\n val_results[""gt_batch""][""videos""][0].astype(jnp.float32)\n / 255.0\n )\n val_results[""recon_seq_val""] = val_results[""recon""][0].clip(\n 0, 1\n )\n val_comparison_seq = jnp.concatenate(\n (val_results[""gt_seq_val""], val_results[""recon_seq_val""]),\n axis=1,\n )\n val_results[""val_comparison_seq""] = einops.rearrange(\n val_comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if args.eval_full_frame:\n val_results[""full_frame_seq_val""] = val_results[\n ""full_frame""\n ][0].clip(0, 1)\n val_results[""val_full_frame_comparison_seq""] = (\n jnp.concatenate(\n (\n val_results[""gt_seq_val""],\n val_results[""full_frame_seq_val""],\n ),\n axis=1,\n )\n )\n val_results[""val_full_frame_comparison_seq""] = (\n einops.rearrange(\n val_results[""val_full_frame_comparison_seq""] * 255,\n ""t h w c -> h (t w) c"",\n )\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n if val_results:\n log_images.update(\n dict(\n val_image=wandb.Image(\n np.asarray(\n val_results[""gt_seq_val""][args.seq_len - 1]\n )\n ),\n val_recon=wandb.Image(\n np.asarray(\n val_results[""recon_seq_val""][\n args.seq_len - 1\n ]\n )\n ),\n val_true_vs_recon=wandb.Image(\n np.asarray(\n val_results[""val_comparison_seq""].astype(\n np.uint8\n )\n )\n ),\n )\n )\n if args.eval_full_frame:\n log_images.update(\n dict(\n val_full_frame=wandb.Image(\n np.asarray(\n val_results[""full_frame_seq_val""][\n args.seq_len - 1\n ]\n )\n ),\n val_true_vs_full_frame=wandb.Image(\n np.asarray(\n val_results[\n ""val_full_frame_comparison_seq""\n ].astype(np.uint8)\n )\n ),\n )\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n assert checkpoint_manager is not None\n optimizer_state = nnx.state(optimizer)\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n )\n checkpoint_manager.save(step, args=ckpt_manager_args)\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n if checkpoint_manager:\n checkpoint_manager.close()\n\n\nif __name__ == ""__main__"":\n args = tyro.cli(Args)\n main(args)\n",python,tab +726,1152025,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19429,0,"",python,selection_command +727,1153693,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19474,0,"",python,selection_mouse +728,1154652,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,0,"",python,selection_mouse +729,1154740,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,1,"[",python,selection_mouse +730,1154779,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,2,"[""",python,selection_mouse +731,1154809,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,3,"[""v",python,selection_mouse +732,1154878,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,4,"[""vi",python,selection_mouse +733,1154920,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,5,"[""vid",python,selection_mouse +734,1154920,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,6,"[""vide",python,selection_mouse +735,1155001,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,7,"[""video",python,selection_mouse +736,1155033,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,8,"[""videos",python,selection_mouse +737,1155086,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,9,"[""videos""",python,selection_mouse +738,1155098,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,10,"[""videos""]",python,selection_mouse +739,1157462,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,10,"",python,content +740,1157470,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19476,0,"",python,selection_command +741,1158333,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19491,0,"",python,selection_command +742,1159143,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19516,0,"",python,selection_command +743,1162987,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19493,24," ""actions"": (",python,selection_command +744,1163228,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19493,80," ""actions"": (\n jax.make_array_from_process_local_data(",python,selection_command +745,1163359,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19493,134," ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]",python,selection_command +746,1163501,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19493,152," ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )",python,selection_command +747,1163642,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19493,191," ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions",python,selection_command +748,1163848,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19493,217," ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None",python,selection_command +749,1164076,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19493,232," ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),",python,selection_command +750,1164412,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19493,233,"",python,content +751,1164444,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19501,0,"",python,selection_command +752,1164998,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19486,0,"",python,selection_command +753,1200944,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19768,0,"",python,selection_mouse +754,1200945,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19767,0,"",python,selection_command +755,1201044,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19767,1,",",python,selection_mouse +756,1201044,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19768,0,"",python,selection_command +757,1202172,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19736,0,"",python,selection_mouse +758,1203640,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,0,"",python,selection_mouse +759,1203745,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,1," ",python,selection_mouse +760,1203759,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,2," e",python,selection_mouse +761,1203776,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,3," el",python,selection_mouse +762,1203791,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,5," elem",python,selection_mouse +763,1203814,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,7," elem[""",python,selection_mouse +764,1203832,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,8," elem[""v",python,selection_mouse +765,1203867,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,9," elem[""vi",python,selection_mouse +766,1203946,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,10," elem[""vid",python,selection_mouse +767,1203976,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,11," elem[""vide",python,selection_mouse +768,1204023,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,12," elem[""video",python,selection_mouse +769,1204043,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,13," elem[""videos",python,selection_mouse +770,1204078,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,14," elem[""videos""",python,selection_mouse +771,1204112,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,15," elem[""videos""]",python,selection_mouse +772,1204662,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19734,15,"",python,content +773,1204698,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19733,0,"",python,selection_command +774,1204962,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19733,1,"",python,content +775,1204963,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19732,0,"",python,selection_command +776,1206007,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19751,0,"",python,selection_command +777,1206131,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19780,0,"",python,selection_command +778,1206712,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19753,28," ""actions"": (",python,selection_command +779,1206911,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19753,88," ""actions"": (\n jax.make_array_from_process_local_data(",python,selection_command +780,1207066,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19753,146," ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]",python,selection_command +781,1207203,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19753,168," ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )",python,selection_command +782,1207325,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19753,211," ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions",python,selection_command +783,1207512,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19753,241," ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None",python,selection_command +784,1207628,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19753,260," ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),",python,selection_command +785,1207947,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19753,261,"",python,content +786,1207984,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19765,0,"",python,selection_command +787,1209121,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19746,0,"",python,selection_command +788,1209598,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19752,0,"",python,selection_command +789,1209678,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19751,1,"",python,content +790,1209915,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19750,0,"",python,selection_command +791,1210097,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19714,0,"",python,selection_command +792,1210240,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19648,0,"",python,selection_command +793,1210398,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19630,0,"",python,selection_command +794,1210597,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19607,0,"",python,selection_command +795,1210902,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19586,0,"",python,selection_command +796,1211079,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19560,0,"",python,selection_command +797,1211193,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19542,0,"",python,selection_command +798,1211393,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19519,0,"",python,selection_command +799,1211506,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19501,0,"",python,selection_command +800,1211653,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19491,0,"",python,selection_command +801,1211936,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19492,0,"",python,selection_command +802,1212034,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19491,1,"",python,content +803,1212200,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19490,0,"",python,selection_command +804,1215037,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +805,1215134,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --val_interval 2 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +806,1215263,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +807,1215396,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +808,1219118,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +809,1225558,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +810,1226071,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +811,1226908,"TERMINAL",0,0,"wandb: creating run\r\nwandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_122059-y4usbd13\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/y4usbd13\r\n",,terminal_output +812,1227037,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +813,1228392,"TERMINAL",0,0,"============\r\n\r\n /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test \r\n\r\n============\r\n",,terminal_output +814,1229146,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +815,1249084,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +816,1249534,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19523,0,"",python,selection_mouse +817,1250961,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 731, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 575, in main\r\n compiled = train_step.lower(optimizer, first_batch).compile()\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/compilation.py"", line 476, in lower\r\n lowered = self.jitted_fn.lower(*pure_args, **pure_kwargs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/compilation.py"", line 126, in __call__\r\n out = self.f(*args, **kwargs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 446, in train_step\r\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/graph.py"", line 2045, in update_context_manager_wrapper\r\n return f(*args, **kwargs)\r\n ^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/autodiff.py"", line 163, in grad_wrapper\r\n fn_out = gradded_fn(*pure_args)\r\n ^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/autodiff.py"", line 88, in __call__\r\n out = self.f(*args)\r\n ^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 444, in loss_fn\r\n return dynamics_loss_fn(model, inputs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 429, in dynamics_loss_fn\r\n outputs = model(inputs)\r\n ^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py"", line 158, in __call__\r\n tokenizer_outputs = self.tokenizer.vq_encode(videos_BTHWC, training=False)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py"", line 115, in vq_encode\r\n x_BTNL = self.encoder(patch_BTNP)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py"", line 250, in __call__\r\n x_BTNI = self.input_norm1(x_BTNI)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/nn/normalization.py"", line 505, in __call__\r\n return _normalize(\r\n ^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/nn/normalization.py"", line 172, in _normalize\r\n scale = scale.reshape(feature_shape)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/numpy/array_methods.py"", line 1147, in meth\r\n return getattr(self.aval, name).fun(self, *args, **kwargs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/numpy/array_methods.py"", line 316, in _reshape\r\n newshape = _compute_newshape(self, args[0] if len(args) == 1 else args)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/numpy/array_methods.py"", line 477, in _compute_newshape\r\n raise TypeError(f""cannot reshape array of shape {arr.shape} (size {arr.size}) ""\r\nTypeError: cannot reshape array of shape (768,) (size 768) into shape [1, 1, 1, 48] (size 48)\r\n--------------------\r\nFor simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n",,terminal_output +818,1251635,"TERMINAL",0,0,"wandb: \r\nwandb: 🚀 View run coinrun-dyn-dev-3536670 at: https://wandb.ai/instant-uv/jafar/runs/y4usbd13\r\nwandb: Find logs at: ../../../../../hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_122059-y4usbd13/logs\r\n",,terminal_output +819,1251782,"TERMINAL",0,0,"W1002 12:21:24.963197 1471562 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugonly job_name: ""jax_worker"": CANCELLED: CANCELLED\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_status:1, grpc_message:""CANCELLED""} [type.googleapis.com/tensorflow.DerivedStatus='']\r\n",,terminal_output +820,1252239,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20017,0,"",python,selection_command +821,1252670,"TERMINAL",0,0,"srun: error: hkn0401: task 0: Exited with exit code 1\r\n]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h[tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +822,1252783,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",18437,0,"",python,selection_command +823,1252936,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",13437,0,"",python,selection_command +824,1254222,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",13433,65,"",python,content +825,1254235,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",13437,0,"",python,selection_command +826,1307820,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +827,1369119,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",0,0,"from typing import Dict\n\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.nnx as nnx\nimport orbax.checkpoint as ocp\n\nfrom models.dynamics import DynamicsMaskGIT, DynamicsCausal\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\n\nclass Genie(nnx.Module):\n """"""Genie model""""""\n\n def __init__(\n self,\n in_dim: int,\n tokenizer_dim: int,\n tokenizer_ffn_dim: int,\n latent_patch_dim: int,\n num_patch_latents: int,\n patch_size: int,\n tokenizer_num_blocks: int,\n tokenizer_num_heads: int,\n lam_dim: int,\n lam_ffn_dim: int,\n latent_action_dim: int,\n num_actions: int,\n lam_patch_size: int,\n lam_num_blocks: int,\n lam_num_heads: int,\n lam_co_train: bool,\n use_gt_actions: bool,\n dyna_type: str,\n dyna_dim: int,\n dyna_ffn_dim: int,\n dyna_num_blocks: int,\n dyna_num_heads: int,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n dropout: float = 0.0,\n mask_limit: float = 0.0,\n ):\n # --- Tokenizer ---\n self.in_dim = in_dim\n self.tokenizer_dim = tokenizer_dim\n self.tokenizer_ffn_dim = tokenizer_ffn_dim\n self.latent_patch_dim = latent_patch_dim\n self.num_patch_latents = num_patch_latents\n self.patch_size = patch_size\n self.tokenizer_num_blocks = tokenizer_num_blocks\n self.tokenizer_num_heads = tokenizer_num_heads\n # --- LAM ---\n self.lam_dim = lam_dim\n self.lam_ffn_dim = lam_ffn_dim\n self.latent_action_dim = latent_action_dim\n self.num_actions = num_actions\n self.lam_patch_size = lam_patch_size\n self.lam_num_blocks = lam_num_blocks\n self.lam_num_heads = lam_num_heads\n self.lam_co_train = lam_co_train\n self.use_gt_actions = use_gt_actions\n # --- Dynamics ---\n self.dyna_type = dyna_type\n self.dyna_dim = dyna_dim\n self.dyna_ffn_dim = dyna_ffn_dim\n self.dyna_num_blocks = dyna_num_blocks\n self.dyna_num_heads = dyna_num_heads\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.dropout = dropout\n self.mask_limit = mask_limit\n self.decode = decode\n\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n ffn_dim=self.tokenizer_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n if self.use_gt_actions:\n self.action_embed = nnx.Embed(\n self.num_actions, self.latent_action_dim, rngs=rngs\n )\n self.lam = None\n else:\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n ffn_dim=self.lam_ffn_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.action_embed = None\n if self.dyna_type == ""maskgit"":\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n elif self.dyna_type == ""causal"":\n self.dynamics = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n rngs=rngs,\n )\n else:\n raise ValueError(f""Invalid dynamics type: {self.dyna_type}"")\n\n def __call__(\n self,\n batch: Dict[str, jax.Array],\n ) -> Dict[str, jax.Array]:\n videos_BTHWC = batch[""videos""]\n tokenizer_outputs = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_indices_BTN = tokenizer_outputs[""indices""]\n latent_actions_BTm11L = None\n action_embeddings_BTm11L = None\n if self.use_gt_actions:\n assert self.action_embed is not None\n action_indices_E = None\n action_embeddings_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n action_embeddings_BTm11L = action_embeddings_BT1L[:, :-1]\n else:\n assert self.lam is not None\n lam_outputs = self.lam.vq_encode(videos_BTHWC, training=False)\n z_q_BTm11L = lam_outputs[""z_q""]\n action_indices_E = lam_outputs[""indices""]\n latent_actions_BTm11L = jax.lax.cond(\n self.lam_co_train,\n lambda: z_q_BTm11L,\n lambda: jax.lax.stop_gradient(z_q_BTm11L),\n )\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(token_indices_BTN),\n latent_actions=(\n action_embeddings_BTm11L\n if self.use_gt_actions\n else latent_actions_BTm11L\n ),\n )\n outputs[""mask_rng""] = batch[""rng""]\n dyna_logits_BTNV, dyna_mask = self.dynamics(outputs)\n outputs[""token_logits""] = dyna_logits_BTNV\n outputs[""mask""] = dyna_mask\n mle_indices_BTN = jnp.argmax(outputs[""token_logits""], axis=-1)\n H, W = batch[""videos""].shape[2:4]\n outputs[""recon""] = self.tokenizer.decode(mle_indices_BTN, (H, W))\n if action_indices_E is not None:\n outputs[""lam_indices""] = action_indices_E\n return outputs\n\n def sample(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n maskgit_steps: int = 25,\n ) -> tuple[jax.Array, jax.Array]:\n if self.dyna_type == ""maskgit"":\n return self.sample_maskgit(\n batch, seq_len, maskgit_steps, temperature, sample_argmax\n )\n elif self.dyna_type == ""causal"":\n return self.sample_causal(batch, seq_len, temperature, sample_argmax)\n else:\n raise ValueError(f""Dynamics model type unknown: {self.dyna_type}"")\n\n def sample_maskgit(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n steps: int = 25,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n P: S * N\n """"""\n assert isinstance(self.dynamics, DynamicsMaskGIT)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n init_logits_BSNV = jnp.zeros(\n shape=(*token_idxs_BSN.shape, self.num_patch_latents)\n )\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, :-1]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n # --- Extract submodule state ---\n dynamics_state = nnx.state(self.dynamics)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def maskgit_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, mask_BSN, action_tokens_EL = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_maskgit = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_maskgit, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_maskgit.patch_embed(token_idxs_BSN)\n mask_token_111M = dynamics_maskgit.mask_token.value\n mask_expanded_BSN1 = mask_BSN[..., None]\n vid_embed_BSNM = jnp.where(\n mask_expanded_BSN1, mask_token_111M, vid_embed_BSNM\n )\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_maskgit.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n vid_embed_BSNM += act_embed_BS1M\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (steps * 2))\n step_temp = temperature * (1.0 - unmasked_ratio)\n final_logits_BSNV = dynamics_maskgit.transformer(vid_embed_BSNM) / step_temp\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_BSN = jnp.argmax(final_logits_BSNV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_BSN = jax.random.categorical(_rng, final_logits_BSNV)\n gather_fn = jax.vmap(jax.vmap(jax.vmap(lambda x, y: x[y])))\n final_token_probs_BSN = gather_fn(\n jax.nn.softmax(final_logits_BSNV), sampled_token_idxs_BSN\n )\n final_token_probs_BSN += ~mask_BSN\n # Update masked tokens and logits only\n token_idxs_BSN = jnp.where(mask_BSN, sampled_token_idxs_BSN, token_idxs_BSN)\n logits_BSNV = jnp.where(\n jnp.expand_dims(mask_BSN, -1), final_logits_BSNV, logits_BSNV\n )\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n final_token_probs_flat_BP = einops.rearrange(\n final_token_probs_BSN, ""b s n -> b (s n)""\n )\n idx_mask_P = (\n jnp.arange(final_token_probs_flat_BP.shape[-1])\n <= N - num_unmasked_tokens\n )\n sorted_idxs_BP = jnp.argsort(final_token_probs_flat_BP, axis=-1)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask_P))\n mask_flat_BP = einops.rearrange(mask_BSN, ""b s n -> b (s n)"")\n new_mask_flat_BP = mask_update_fn(mask_flat_BP, sorted_idxs_BP)\n new_mask_BSN = einops.rearrange(new_mask_flat_BP, ""b (s n) -> b s n"", n=N)\n\n new_carry = (\n rng,\n token_idxs_BSN,\n logits_BSNV,\n new_mask_BSN,\n action_tokens_EL,\n )\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # Mask current frame (i.e., t == step_t)\n mask_S = jnp.arange(seq_len) == step_t\n mask_BSN = jnp.broadcast_to(mask_S[None, :, None], (B, seq_len, N)).astype(\n bool\n )\n masked_token_idxs_BSN = current_token_idxs_BSN * ~mask_BSN\n masked_logits_BSNV = current_logits_BSNV * jnp.expand_dims(~mask_BSN, -1)\n\n # --- Initialize and run MaskGIT loop ---\n init_carry_maskgit = (\n step_rng,\n masked_token_idxs_BSN,\n masked_logits_BSNV,\n mask_BSN,\n action_tokens_EL,\n )\n final_carry_maskgit = maskgit_step_fn(init_carry_maskgit, jnp.arange(steps))\n updated_token_idxs_BSN = final_carry_maskgit[1]\n updated_logits_BSNV = final_carry_maskgit[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, init_logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def sample_causal(\n self,\n batch: Dict[str, jax.Array],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> tuple[jax.Array, jax.Array]:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by\n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size\n T: number of input (conditioning) frames\n N: number of patches per frame\n M: model dimension\n S: sequence length\n H: height\n W: width\n E: B * (S - 1)\n """"""\n assert isinstance(self.dynamics, DynamicsCausal)\n # --- Encode videos and actions ---\n videos_BTHWC = batch[""videos""]\n tokenizer_out = self.tokenizer.vq_encode(videos_BTHWC, training=False)\n token_idxs_BTN = tokenizer_out[""indices""]\n B, T, N = token_idxs_BTN.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs_BTN.dtype)\n token_idxs_BSN = jnp.concatenate([token_idxs_BTN, pad], axis=1)\n logits_BSNV = jnp.zeros((*token_idxs_BSN.shape, self.num_patch_latents))\n dynamics_state = nnx.state(self.dynamics)\n\n if self.use_gt_actions:\n assert self.action_embed is not None\n latent_actions_BT1L = self.action_embed(batch[""actions""]).reshape(\n *batch[""actions""].shape[:2], 1, self.latent_action_dim\n )\n latent_actions_BTm11L = latent_actions_BT1L[:, :-1]\n action_tokens_EL = latent_actions_BTm11L.reshape(-1, self.latent_action_dim)\n else:\n assert self.lam is not None\n latent_actions_E = batch[""latent_actions""]\n action_tokens_EL = self.lam.vq.get_codes(latent_actions_E)\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def causal_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array],\n step_n: jax.Array,\n ) -> tuple[jax.Array, jax.Array, jax.Array, jax.Array, jax.Array]:\n rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t = carry\n S, N = token_idxs_BSN.shape[1:]\n L = action_tokens_EL.shape[-1]\n\n # We need to reconstruct the submodule inside scan body to prevent trace context mismatches\n dynamics_causal = DynamicsCausal(\n model_dim=self.dyna_dim,\n ffn_dim=self.dyna_ffn_dim,\n num_latents=self.num_patch_latents,\n latent_action_dim=self.latent_action_dim,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=self.decode,\n rngs=nnx.Rngs(0),\n )\n nnx.update(dynamics_causal, dynamics_state)\n\n # --- Construct + encode video ---\n vid_embed_BSNM = dynamics_causal.patch_embed(token_idxs_BSN)\n\n # --- Predict transition ---\n action_tokens_BSm1L = jnp.reshape(action_tokens_EL, (B, S - 1, L))\n act_embed_BSm1M = dynamics_causal.action_up(action_tokens_BSm1L)\n act_embed_BSM = jnp.pad(act_embed_BSm1M, ((0, 0), (1, 0), (0, 0)))\n act_embed_BS1M = jnp.reshape(\n act_embed_BSM, (B, S, 1, act_embed_BSM.shape[-1])\n )\n vid_embed_BSNp1M = jnp.concatenate([act_embed_BS1M, vid_embed_BSNM], axis=2)\n final_logits_BTNp1V = (\n dynamics_causal.transformer(vid_embed_BSNp1M, (step_t, step_n))\n / temperature\n )\n final_logits_BV = final_logits_BTNp1V[:, step_t, step_n, :]\n\n # --- Sample new tokens for final frame ---\n if sample_argmax:\n sampled_token_idxs_B = jnp.argmax(final_logits_BV, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs_B = jax.random.categorical(_rng, final_logits_BV)\n # Update next tokens only\n token_idxs_BSN = token_idxs_BSN.at[:, step_t, step_n].set(\n sampled_token_idxs_B\n )\n logits_BSNV = logits_BSNV.at[:, step_t, step_n].set(final_logits_BV)\n\n new_carry = (rng, token_idxs_BSN, logits_BSNV, action_tokens_EL, step_t)\n return new_carry\n\n @nnx.scan(in_axes=(nnx.Carry, 0), out_axes=nnx.Carry)\n def generation_step_fn(\n carry: tuple[jax.Array, jax.Array, jax.Array], step_t: jax.Array\n ) -> tuple[jax.Array, jax.Array, jax.Array]:\n rng, current_token_idxs_BSN, current_logits_BSNV = carry\n rng, step_rng = jax.random.split(rng)\n\n # --- Initialize and run causal loop ---\n init_carry_causal = (\n step_rng,\n current_token_idxs_BSN,\n current_logits_BSNV,\n action_tokens_EL,\n step_t,\n )\n final_carry_causal = causal_step_fn(init_carry_causal, jnp.arange(N))\n updated_token_idxs_BSN = final_carry_causal[1]\n updated_logits_BSNV = final_carry_causal[2]\n new_carry = (rng, updated_token_idxs_BSN, updated_logits_BSNV)\n return new_carry\n\n # --- Run the autoregressive generation using jax.lax.scan ---\n initial_carry = (batch[""rng""], token_idxs_BSN, logits_BSNV)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry = generation_step_fn(initial_carry, timesteps_to_scan)\n final_token_idxs_BSN = final_carry[1]\n final_logits_BSNV = final_carry[2]\n\n # --- Decode all tokens at once at the end ---\n H, W = batch[""videos""].shape[2:4]\n final_frames_BSHWC = self.tokenizer.decode(\n final_token_idxs_BSN,\n video_hw=(H, W),\n )\n return final_frames_BSHWC, final_logits_BSNV\n\n def vq_encode(self, batch: Dict[str, jax.Array], training: bool) -> jax.Array:\n # --- Preprocess videos ---\n assert self.lam is not None\n video_BTHWC = batch[""videos""]\n lam_output: Dict[str, jax.Array] = self.lam.vq_encode(\n video_BTHWC, training=training\n )\n lam_indices_E = lam_output[""indices""]\n return lam_indices_E\n\n\n# FIXME (f.srambical): add conversion script for old checkpoints\ndef restore_genie_components(\n optimizer: nnx.ModelAndOptimizer,\n sharding: jax.sharding.NamedSharding,\n rng: jax.Array,\n args,\n) -> nnx.ModelAndOptimizer:\n """"""Restore pre-trained Genie components""""""\n rng_tokenizer, rng_lam = jax.random.split(rng)\n rngs_tokenizer = nnx.Rngs(rng_tokenizer)\n rngs_lam = nnx.Rngs(rng_lam)\n\n tx = optimizer.tx\n model = optimizer.model\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n tokenizer_checkpoint_manager = ocp.CheckpointManager(\n directory=args.tokenizer_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n ffn_dim=args.tokenizer_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_tokenizer,\n )\n dummy_tokenizer_optimizer = nnx.ModelAndOptimizer(dummy_tokenizer, tx)\n dummy_tokenizer_optimizer_state = nnx.state(dummy_tokenizer_optimizer)\n abstract_sharded_tokenizer_optimizer_state = _create_abstract_sharded_pytree(\n dummy_tokenizer_optimizer_state, sharding\n )\n restored_tokenizer = tokenizer_checkpoint_manager.restore(\n step=tokenizer_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_tokenizer_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_tokenizer_optimizer.model, restored_tokenizer.model)\n model.tokenizer = dummy_tokenizer_optimizer.model\n tokenizer_checkpoint_manager.close()\n\n if args.lam_checkpoint:\n lam_checkpoint_manager = ocp.CheckpointManager(\n directory=args.lam_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n ffn_dim=args.lam_ffn_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs_lam,\n )\n dummy_lam_optimizer = nnx.ModelAndOptimizer(dummy_lam, tx)\n dummy_lam_optimizer_state = nnx.state(dummy_lam_optimizer)\n abstract_sharded_lam_optimizer_state = _create_abstract_sharded_pytree(\n dummy_lam_optimizer_state, sharding\n )\n restored_lam_optimizer = lam_checkpoint_manager.restore(\n step=lam_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore( # type: ignore\n abstract_sharded_lam_optimizer_state # type: ignore\n ),\n ),\n )[""model_state""]\n nnx.update(dummy_lam_optimizer.model, restored_lam_optimizer.model)\n model.lam = dummy_lam_optimizer.model\n # Remove the LAM decoder to save memory and avoid unnecessary computation.\n del model.lam.decoder\n lam_checkpoint_manager.close()\n\n # Reinitialize the optimizer states\n optimizer = nnx.ModelAndOptimizer(model, tx)\n return optimizer\n\n\ndef _create_abstract_sharded_pytree(\n pytree_template: nnx.GraphState, sharding_spec: jax.sharding.NamedSharding\n) -> jax.Array:\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)\n",python,tab +828,1369120,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",5425,0,"",python,selection_command +829,1372036,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",5470,0,"",python,selection_mouse +830,1373187,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",5472,0,"",python,selection_mouse +831,1374265,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",5464,0,"",python,selection_mouse +832,1374445,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",2447,0,"",python,selection_command +833,1375329,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",2464,0,"",python,selection_mouse +834,1376186,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",2459,14,"TokenizerVQVAE",python,selection_mouse +835,1376757,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",2464,0,"",python,selection_mouse +836,1379645,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",247,0,"",python,selection_mouse +837,1381166,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",404,0,"",python,selection_mouse +838,1381254,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",403,0,"",python,selection_command +839,1381254,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",403,1,",",python,selection_mouse +840,1381255,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",404,0,"",python,selection_command +841,1394601,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",2462,0,"",python,selection_mouse +842,1402348,"jasmine/models/tokenizer.py",0,0,"from typing import Dict, Tuple\n\nimport flax.nnx as nnx\nimport jax.numpy as jnp\nimport jax\n\nfrom utils.preprocess import patchify, unpatchify\nfrom utils.nn import STTransformer, VectorQuantizer\n\n\nclass TokenizerVQVAE(nnx.Module):\n """"""\n ST-ViVit VQ-VAE\n\n Dimension keys:\n B: batch size\n T: sequence length\n N: number of patches per frame\n L: latent dimension\n D: B * T * N\n H: height\n W: width\n C: number of channels\n P: patch token dimension (patch_size^2 * C)\n """"""\n\n def __init__(\n self,\n in_dim: int,\n model_dim: int,\n ffn_dim: int,\n latent_dim: int,\n num_latents: int,\n patch_size: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n codebook_dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n ):\n self.in_dim = in_dim\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.latent_dim = latent_dim\n self.num_latents = num_latents\n self.patch_size = patch_size\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.codebook_dropout = codebook_dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.encoder = STTransformer(\n self.in_dim * self.patch_size**2,\n self.model_dim,\n self.ffn_dim,\n self.latent_dim,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.vq = VectorQuantizer(\n self.latent_dim,\n self.num_latents,\n self.codebook_dropout,\n self.dtype,\n rngs=rngs,\n )\n self.out_dim = self.in_dim * self.patch_size**2\n self.decoder = STTransformer(\n self.latent_dim,\n self.model_dim,\n self.ffn_dim,\n self.out_dim,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n\n def __call__(\n self, batch: Dict[str, jax.Array], training: bool = True\n ) -> Dict[str, jax.Array]:\n H, W = batch[""videos""].shape[2:4]\n videos_BTHWC = batch[""videos""]\n outputs = self.vq_encode(videos_BTHWC, training)\n z_q_BTNL = outputs[""z_q""]\n recon_BTHWC = self.decoder(z_q_BTNL)\n recon_BTHWC = recon_BTHWC.astype(jnp.float32)\n recon_BTHWC = nnx.sigmoid(recon_BTHWC)\n recon_BTHWC = recon_BTHWC.astype(self.dtype)\n recon_BTHWC = unpatchify(recon_BTHWC, self.patch_size, H, W)\n outputs[""recon""] = recon_BTHWC\n return outputs\n\n def vq_encode(\n self, videos: jax.Array, training: bool = True\n ) -> Dict[str, jax.Array]:\n # --- Preprocess + encode ---\n B, T = videos.shape[:2]\n patch_BTNP = patchify(videos, self.patch_size)\n N = patch_BTNP.shape[2]\n x_BTNL = self.encoder(patch_BTNP)\n\n # --- Vector quantize ---\n x_DL = x_BTNL.reshape(B * T * N, self.latent_dim)\n z_q_DL, z_DL, emb_DL, indices_D = self.vq(x_DL, training)\n z_q_BTNL = z_q_DL.reshape(B, T, N, self.latent_dim)\n indices_BTN = indices_D.reshape(B, T, N)\n return dict(z_q=z_q_BTNL, z=z_DL, emb=emb_DL, indices=indices_BTN)\n\n def decode(self, indices_BTN: jax.Array, video_hw: Tuple[int, int]) -> jax.Array:\n z_BTNL = self.vq.codebook[indices_BTN]\n recon_BTNP = self.decoder(z_BTNL)\n recon_BTNP = recon_BTNP.astype(jnp.float32)\n recon_BTNP = nnx.sigmoid(recon_BTNP)\n recon_BTNP = recon_BTNP.astype(self.dtype)\n return unpatchify(recon_BTNP, self.patch_size, *video_hw)\n",python,tab +843,1408128,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",0,0,"from typing import Dict, Tuple\n\nimport flax.nnx as nnx\nimport jax.numpy as jnp\nimport jax\n\nfrom utils.preprocess import patchify, unpatchify\nfrom utils.nn import STTransformer, VectorQuantizer\n\n\nclass TokenizerVQVAE(nnx.Module):\n """"""\n ST-ViVit VQ-VAE\n\n Dimension keys:\n B: batch size\n T: sequence length\n N: number of patches per frame\n L: latent dimension\n D: B * T * N\n H: height\n W: width\n C: number of channels\n P: patch token dimension (patch_size^2 * C)\n """"""\n\n def __init__(\n self,\n in_dim: int,\n model_dim: int,\n ffn_dim: int,\n latent_dim: int,\n num_latents: int,\n patch_size: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n codebook_dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n ):\n self.in_dim = in_dim\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.latent_dim = latent_dim\n self.num_latents = num_latents\n self.patch_size = patch_size\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.codebook_dropout = codebook_dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n\n self.encoder = STTransformer(\n self.in_dim * self.patch_size**2,\n self.model_dim,\n self.ffn_dim,\n self.latent_dim,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n self.vq = VectorQuantizer(\n self.latent_dim,\n self.num_latents,\n self.codebook_dropout,\n self.dtype,\n rngs=rngs,\n )\n self.out_dim = self.in_dim * self.patch_size**2\n self.decoder = STTransformer(\n self.latent_dim,\n self.model_dim,\n self.ffn_dim,\n self.out_dim,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n )\n\n def __call__(\n self, batch: Dict[str, jax.Array], training: bool = True\n ) -> Dict[str, jax.Array]:\n H, W = batch[""videos""].shape[2:4]\n videos_BTHWC = batch[""videos""]\n outputs = self.vq_encode(videos_BTHWC, training)\n z_q_BTNL = outputs[""z_q""]\n recon_BTHWC = self.decoder(z_q_BTNL)\n recon_BTHWC = recon_BTHWC.astype(jnp.float32)\n recon_BTHWC = nnx.sigmoid(recon_BTHWC)\n recon_BTHWC = recon_BTHWC.astype(self.dtype)\n recon_BTHWC = unpatchify(recon_BTHWC, self.patch_size, H, W)\n outputs[""recon""] = recon_BTHWC\n return outputs\n\n def vq_encode(\n self, videos: jax.Array, training: bool = True\n ) -> Dict[str, jax.Array]:\n # --- Preprocess + encode ---\n B, T = videos.shape[:2]\n patch_BTNP = patchify(videos, self.patch_size)\n N = patch_BTNP.shape[2]\n x_BTNL = self.encoder(patch_BTNP)\n\n # --- Vector quantize ---\n x_DL = x_BTNL.reshape(B * T * N, self.latent_dim)\n z_q_DL, z_DL, emb_DL, indices_D = self.vq(x_DL, training)\n z_q_BTNL = z_q_DL.reshape(B, T, N, self.latent_dim)\n indices_BTN = indices_D.reshape(B, T, N)\n return dict(z_q=z_q_BTNL, z=z_DL, emb=emb_DL, indices=indices_BTN)\n\n def decode(self, indices_BTN: jax.Array, video_hw: Tuple[int, int]) -> jax.Array:\n z_BTNL = self.vq.codebook[indices_BTN]\n recon_BTNP = self.decoder(z_BTNL)\n recon_BTNP = recon_BTNP.astype(jnp.float32)\n recon_BTNP = nnx.sigmoid(recon_BTNP)\n recon_BTNP = recon_BTNP.astype(self.dtype)\n return unpatchify(recon_BTNP, self.patch_size, *video_hw)\n",python,tab +844,1408129,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",3320,0,"",python,selection_command +845,1411065,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",3345,0,"",python,selection_mouse +846,1411302,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1441,0,"",python,selection_command +847,1415882,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",0,0,"import math\nfrom typing import Tuple, Callable, List\n\nfrom flax import nnx\nimport jax\nimport jax.numpy as jnp\nimport einops\n\n\ndef _get_spatiotemporal_positional_encoding(d_model: int, max_len: int = 5000):\n """"""\n Creates a function that applies separate sinusoidal positional encodings to the temporal and spatial dimensions.\n """"""\n pe = jnp.zeros((max_len, d_model))\n position = jnp.arange(0, max_len, dtype=jnp.float32)[:, None]\n div_term = jnp.exp(jnp.arange(0, d_model, 2) * (-math.log(10000.0) / d_model))\n pe = pe.at[:, 0::2].set(jnp.sin(position * div_term))\n pe = pe.at[:, 1::2].set(jnp.cos(position * div_term))\n\n def _encode(x: jax.Array) -> jax.Array:\n """"""\n Args:\n x: The input tensor of shape (Batch, Time, Space, Dimension).\n\n Returns:\n The input tensor with positional encodings added.\n """"""\n assert x.ndim == 4, f""Input must be 4-dimensional, but got shape {x.shape}""\n\n num_timesteps = x.shape[1]\n num_spatial_patches = x.shape[2]\n\n # Temporal positional encoding: (1, T, 1, D)\n temporal_pe = pe[None, :num_timesteps, None, :]\n x = x + temporal_pe\n\n # Spatial positional encoding: (1, 1, S, D)\n spatial_pe = pe[None, None, :num_spatial_patches, :]\n x = x + spatial_pe\n\n return x\n\n return _encode\n\n\nclass STBlock(nnx.Module):\n def __init__(\n self,\n dim: int,\n ffn_dim: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n sow_weights: bool,\n sow_activations: bool,\n ):\n self.dim = dim\n self.ffn_dim = ffn_dim\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.sow_weights = sow_weights\n self.sow_activations = sow_activations\n\n self.spatial_norm = nnx.LayerNorm(\n num_features=self.dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.spatial_attention = nnx.MultiHeadAttention(\n num_heads=self.num_heads,\n in_features=self.dim,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=False\n ),\n rngs=rngs,\n decode=False,\n )\n\n self.temporal_norm = nnx.LayerNorm(\n num_features=self.dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.temporal_attention = nnx.MultiHeadAttention(\n num_heads=self.num_heads,\n in_features=self.dim,\n qkv_features=self.dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=True\n ),\n rngs=rngs,\n decode=False,\n )\n\n self.ffn_norm = nnx.LayerNorm(\n num_features=self.dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.ffn_dense1 = nnx.Linear(\n in_features=self.dim,\n out_features=self.ffn_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.ffn_dense2 = nnx.Linear(\n in_features=self.ffn_dim,\n out_features=self.dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n @nnx.remat\n def __call__(self, x_BTNM: jax.Array) -> jax.Array:\n # --- Spatial attention ---\n z_BTNM = self.spatial_norm(x_BTNM)\n z_BTNM = self.spatial_attention(z_BTNM, sow_weights=self.sow_weights)\n x_BTNM = x_BTNM + z_BTNM\n\n # --- Temporal attention ---\n x_BNTM = x_BTNM.swapaxes(1, 2)\n z_BNTM = self.temporal_norm(x_BNTM)\n z_BNTM = self.temporal_attention(z_BNTM, sow_weights=self.sow_weights)\n x_BNTM = x_BNTM + z_BNTM\n x_BTNM = x_BNTM.swapaxes(1, 2)\n\n # --- Feedforward ---\n z_BTNM = self.ffn_norm(x_BTNM)\n z_BTND = self.ffn_dense1(z_BTNM)\n z_BTND = jax.nn.gelu(z_BTND)\n z_BTNM = self.ffn_dense2(z_BTND)\n x_BTNM = x_BTNM + z_BTNM\n if self.sow_activations:\n self.sow(nnx.Intermediate, ""activations"", x_BTNM)\n return x_BTNM\n\n\nclass STTransformer(nnx.Module):\n """"""\n Dimension keys:\n B: batch size\n T: number of frames\n N: number of patches per frame\n I: number of input features\n M: model dimension\n D: FFN dimension\n V: vocabulary size\n """"""\n\n def __init__(\n self,\n input_dim: int,\n model_dim: int,\n ffn_dim: int,\n out_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n rngs: nnx.Rngs,\n sow_weights: bool = False,\n sow_activations: bool = False,\n sow_logits: bool = False,\n max_len: int = 5000,\n ):\n self.input_dim = input_dim\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.out_dim = out_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.sow_logits = sow_logits\n self.sow_weights = sow_weights\n self.sow_activations = sow_activations\n\n self.input_norm1 = nnx.LayerNorm(\n num_features=self.input_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.input_dense = nnx.Linear(\n in_features=self.input_dim,\n out_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.input_norm2 = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n\n self.pos_enc = _get_spatiotemporal_positional_encoding(\n self.model_dim, max_len=max_len\n )\n\n self.blocks = []\n for _ in range(self.num_blocks):\n self.blocks.append(\n STBlock(\n dim=self.model_dim,\n ffn_dim=self.ffn_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n rngs=rngs,\n sow_weights=self.sow_weights,\n sow_activations=self.sow_activations,\n )\n )\n\n self.output_dense = nnx.Linear(\n in_features=self.model_dim,\n out_features=self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(self, x_BTNI: jax.Array) -> jax.Array:\n x_BTNI = self.input_norm1(x_BTNI)\n x_BTNM = self.input_dense(x_BTNI)\n x_BTNM = self.input_norm2(x_BTNM)\n x_BTNM = self.pos_enc(x_BTNM)\n for block in self.blocks:\n x_BTNM = block(x_BTNM)\n\n x_BTNV = self.output_dense(x_BTNM)\n if self.sow_logits:\n self.sow(nnx.Intermediate, ""logits"", x_BTNV)\n return x_BTNV\n\n\nclass TransformerBlock(nnx.Module):\n def __init__(\n self,\n model_dim: int,\n ffn_dim: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n sow_weights: bool,\n sow_activations: bool,\n ):\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.decode = decode\n self.sow_weights = sow_weights\n self.sow_activations = sow_activations\n\n self.temporal_norm = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.spatial_norm = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.ffn_norm = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.temporal_attention = nnx.MultiHeadAttention(\n num_heads=self.num_heads,\n in_features=self.model_dim,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=True\n ),\n rngs=rngs,\n decode=self.decode,\n )\n self.spatial_attention = nnx.MultiHeadAttention(\n num_heads=self.num_heads,\n in_features=self.model_dim,\n qkv_features=self.model_dim,\n dropout_rate=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n attention_fn=_create_flash_attention_fn(\n self.use_flash_attention, is_causal=True\n ),\n rngs=rngs,\n decode=self.decode,\n )\n self.ffn_dense1 = nnx.Linear(\n in_features=self.model_dim,\n out_features=self.ffn_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.ffn_dense2 = nnx.Linear(\n in_features=self.ffn_dim,\n out_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n @nnx.remat\n def __call__(\n self, x_BTNM: jax.Array, pos_index: Tuple[jax.Array, jax.Array] | None = None\n ) -> jax.Array:\n # --- Spatial attention ---\n B, T, N, M = x_BTNM.shape\n z_FNM = einops.rearrange(x_BTNM, ""b t n m -> (b t) n m"")\n z_FNM = self.spatial_norm(z_FNM)\n z_FNM = self.spatial_attention(z_FNM, sow_weights=self.sow_weights)\n z_BTNM = einops.rearrange(z_FNM, ""(b t) n m -> b t n m"", t=T)\n x_BTNM = x_BTNM + z_BTNM\n # --- Temporal attention ---\n z_PTM = einops.rearrange(x_BTNM, ""b t n m -> (b n) t m"")\n z_PTM = self.temporal_norm(z_PTM)\n z_PTM = self.temporal_attention(z_PTM, sow_weights=self.sow_weights)\n z_BTNM = einops.rearrange(z_PTM, ""(b n) t m -> b t n m"", n=N)\n x_BTNM = x_BTNM + z_BTNM\n # --- Feedforward ---\n z_BTNM = self.ffn_norm(x_BTNM)\n z_BTND = self.ffn_dense1(z_BTNM)\n z_BTND = jax.nn.gelu(z_BTND)\n z_BTNM = self.ffn_dense2(z_BTND)\n x_BTNM = x_BTNM + z_BTNM\n if self.sow_activations:\n self.sow(nnx.Intermediate, ""activations"", x_BTNM)\n\n return x_BTNM\n\n\nclass Transformer(nnx.Module):\n """"""\n Dimension keys:\n B: batch size\n T: number of frames\n N: number of patches per frame\n I: number of input features\n M: model dimension\n D: FFN dimension\n V: vocabulary size\n F: number of frames in batch\n P: number of patch positions in batch\n """"""\n\n def __init__(\n self,\n input_dim: int,\n model_dim: int,\n ffn_dim: int,\n out_dim: int,\n num_blocks: int,\n num_heads: int,\n dropout: float,\n param_dtype: jnp.dtype,\n dtype: jnp.dtype,\n use_flash_attention: bool,\n decode: bool,\n rngs: nnx.Rngs,\n sow_logits: bool = False,\n sow_weights: bool = False,\n sow_activations: bool = False,\n max_len: int = 5000,\n ):\n self.input_dim = input_dim\n self.model_dim = model_dim\n self.ffn_dim = ffn_dim\n self.out_dim = out_dim\n self.num_blocks = num_blocks\n self.num_heads = num_heads\n self.dropout = dropout\n self.param_dtype = param_dtype\n self.dtype = dtype\n self.use_flash_attention = use_flash_attention\n self.sow_logits = sow_logits\n self.sow_weights = sow_weights\n self.sow_activations = sow_activations\n\n self.input_norm1 = nnx.LayerNorm(\n num_features=self.input_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n self.input_dense = nnx.Linear(\n in_features=self.input_dim,\n out_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n self.input_norm2 = nnx.LayerNorm(\n num_features=self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.param_dtype, # layer norm in full precision\n rngs=rngs,\n )\n\n self.pos_enc = _get_spatiotemporal_positional_encoding(\n self.model_dim, max_len=max_len\n )\n\n self.blocks: List[TransformerBlock] = []\n for _ in range(self.num_blocks):\n self.blocks.append(\n TransformerBlock(\n model_dim=self.model_dim,\n ffn_dim=self.ffn_dim,\n num_heads=self.num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n decode=decode,\n sow_weights=self.sow_weights,\n sow_activations=self.sow_activations,\n rngs=rngs,\n )\n )\n self.output_dense = nnx.Linear(\n in_features=self.model_dim,\n out_features=self.out_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n rngs=rngs,\n )\n\n def __call__(\n self, x_BTNI: jax.Array, pos_index: Tuple[jax.Array, jax.Array] | None = None\n ) -> jax.Array:\n x_BTNI = self.input_norm1(x_BTNI)\n x_BTNM = self.input_dense(x_BTNI)\n x_BTNM = self.input_norm2(x_BTNM)\n x_BTNM = self.pos_enc(x_BTNM)\n for block in self.blocks:\n x_BTNM = block(x_BTNM, pos_index)\n\n x_BTNV = self.output_dense(x_BTNM)\n if self.sow_logits:\n self.sow(nnx.Intermediate, ""logits"", x_BTNV)\n return x_BTNV\n\n\ndef normalize(x: jax.Array) -> jax.Array:\n return x / (jnp.linalg.norm(x, ord=2, axis=-1, keepdims=True) + 1e-8)\n\n\nclass VectorQuantizer(nnx.Module):\n """"""\n Dimension keys:\n D: B * T * N\n K: number of latents\n L: latent dimension\n """"""\n\n def __init__(\n self,\n latent_dim: int,\n num_latents: int,\n dropout: float,\n dtype: jnp.dtype,\n rngs: nnx.Rngs,\n ):\n self.latent_dim = latent_dim\n self.num_latents = num_latents\n self.dropout = dropout\n self.dtype = dtype\n\n self.codebook = nnx.Param(\n normalize(\n nnx.initializers.normal(stddev=1)(\n rngs.params(), (self.num_latents, self.latent_dim)\n )\n )\n )\n self.drop = nnx.Dropout(self.dropout, rngs=rngs)\n\n def __call__(\n self, x_DL: jax.Array, training: bool\n ) -> Tuple[jax.Array, jax.Array, jax.Array, jax.Array]:\n # --- Compute distances ---\n x_DL = x_DL.astype(self.dtype)\n codebook = self.codebook.value.astype(self.dtype)\n\n x_DL = normalize(x_DL)\n normalized_codebook_KL = normalize(codebook)\n distance_DK = -jnp.matmul(x_DL, normalized_codebook_KL.T)\n if training:\n distance_DK = self.drop(distance_DK)\n\n # --- Get indices and embeddings ---\n indices_D = jnp.argmin(distance_DK, axis=-1)\n z_DL = codebook[indices_D]\n\n # --- Straight through estimator ---\n z_q_DL = x_DL + jax.lax.stop_gradient(z_DL - x_DL)\n return z_q_DL, z_DL, x_DL, indices_D\n\n def get_codes(self, indices_E: jax.Array) -> jax.Array:\n return self.codebook[indices_E]\n\n\ndef _create_flash_attention_fn(use_flash_attention: bool, is_causal: bool) -> Callable:\n """"""\n Create an attention function that uses flash attention if enabled.\n\n flax.nnx.MultiHeadAttention provides tensors with shape (batch..., length, num_heads, head_dim),\n but jax.nn.dot_product_attention expects (batch, length, num_heads, head_dim). We reshape to\n ensure compatibility. cuDNN's flash attention additionally requires a sequence length that\n is a multiple of 4. We pad the sequence length to the nearest multiple of 4 and mask\n accordingly. Note that cuDNN requires the mask to be broadcast before calling the attention\n function due to strict shape checking.\n """"""\n\n def attention_fn(\n query_BTHD, key_BSHD, value_BSHD, bias=None, mask_B111=None, **kwargs\n ):\n implementation = ""cudnn"" if use_flash_attention else None\n\n def _merge_batch_dims(x):\n return einops.rearrange(x, ""... l h k -> (...) l h k"")\n\n def _pad(x, pad_size):\n return jnp.pad(x, ((0, 0), (0, pad_size), (0, 0), (0, 0)))\n\n original_shape = query_BTHD.shape\n T = query_BTHD.shape[-3]\n S = key_BSHD.shape[-3]\n\n # Pad to nearest multiple of 4\n Q = ((T + 3) // 4) * 4\n pad_size_Q = Q - T\n K = ((S + 3) // 4) * 4\n pad_size_K = K - S\n\n query_BQHD = _pad(_merge_batch_dims(query_BTHD), pad_size_Q)\n key_BKHD = _pad(_merge_batch_dims(key_BSHD), pad_size_K)\n value_BKHD = _pad(_merge_batch_dims(value_BSHD), pad_size_K)\n\n attention_mask = jnp.ones((Q, K), dtype=jnp.bool_)\n attention_mask = attention_mask.at[T:, :].set(False)\n attention_mask = attention_mask.at[:, S:].set(False)\n\n mask_11TS = attention_mask[jnp.newaxis, jnp.newaxis, :, :]\n\n bias_4d = (\n jnp.pad(\n _merge_batch_dims(bias),\n ((0, 0), (0, 0), (0, pad_size_Q), (0, pad_size_K)),\n )\n if bias is not None\n else None\n )\n\n # NOTE: jax.nn.dot_product_attention does not support dropout\n output_4d = jax.nn.dot_product_attention(\n query=query_BQHD,\n key=key_BKHD,\n value=value_BKHD,\n bias=bias_4d,\n mask=mask_11TS,\n implementation=implementation,\n is_causal=is_causal,\n )\n return output_4d[..., :T, :, :].reshape(original_shape)\n\n return attention_fn\n",python,tab +848,1415883,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7815,0,"",python,selection_command +849,1416992,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7885,0,"",python,selection_mouse +850,1417103,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7879,11,"input_dense",python,selection_mouse +851,1417454,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7846,0,"",python,selection_mouse +852,1417545,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7790,56,"jax.Array) -> jax.Array:\n x_BTNI = self.input_nor",python,selection_mouse +853,1418005,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7846,0,"",python,selection_mouse +854,1418066,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7837,11,"input_norm1",python,selection_mouse +855,1418840,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7851,0,"",python,selection_mouse +856,1418991,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7851,1,"B",python,selection_mouse +857,1419027,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7851,2,"BT",python,selection_mouse +858,1419060,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7851,3,"BTN",python,selection_mouse +859,1419093,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7851,4,"BTNI",python,selection_mouse +860,1420148,"TERMINAL",0,0,"bash",,terminal_focus +861,1424043,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5067,0,"",python,selection_mouse +862,1424221,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5067,6,"number",python,selection_mouse +863,1424493,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5067,7,"number ",python,selection_mouse +864,1424523,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5067,9,"number of",python,selection_mouse +865,1424597,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5038,35,"patches per frame\n I: number",python,selection_mouse +866,1424738,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5046,27,"per frame\n I: number",python,selection_mouse +867,1424766,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5049,24," frame\n I: number",python,selection_mouse +868,1424802,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5050,23,"frame\n I: number",python,selection_mouse +869,1424803,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5067,24,"number of input features",python,selection_mouse +870,1425253,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5091,0,"",python,selection_mouse +871,1425257,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5090,0,"",python,selection_command +872,1425434,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5083,8,"features",python,selection_mouse +873,1425465,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5084,7,"eatures",python,selection_command +874,1425639,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5084,0,"",python,selection_mouse +875,1425670,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5083,1,"f",python,selection_mouse +876,1425700,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5082,2," f",python,selection_mouse +877,1425711,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5077,7,"input f",python,selection_mouse +878,1425843,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5076,8," input f",python,selection_mouse +879,1425868,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5074,10,"of input f",python,selection_mouse +880,1425944,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5084,34,"eatures\n M: model dimension",python,selection_mouse +881,1425945,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5084,25,"eatures\n M: model ",python,selection_mouse +882,1425968,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5084,24,"eatures\n M: model",python,selection_mouse +883,1426385,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5067,17,"number of input f",python,selection_mouse +884,1426469,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5066,18," number of input f",python,selection_mouse +885,1426814,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5066,0,"",python,selection_mouse +886,1426905,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5066,1," ",python,selection_mouse +887,1427026,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5066,7," number",python,selection_mouse +888,1427113,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5066,8," number ",python,selection_mouse +889,1427126,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5066,10," number of",python,selection_mouse +890,1427141,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5066,11," number of ",python,selection_mouse +891,1427161,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5066,16," number of input",python,selection_mouse +892,1427541,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5082,0,"",python,selection_mouse +893,1432029,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +894,1434101,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1468,1,"4",python,selection_command +895,1434162,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1468,2,"48",python,selection_command +896,1436105,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1720,2,"48",python,selection_command +897,1436693,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",2027,2,"48",python,selection_command +898,1437362,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1468,2,"48",python,selection_command +899,1438044,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1720,2,"48",python,selection_command +900,1438258,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",2027,2,"48",python,selection_command +901,1439160,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1468,2,"48",python,selection_command +902,1439332,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1720,2,"48",python,selection_command +903,1439876,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",2027,2,"48",python,selection_command +904,1448359,"TERMINAL",0,0,"srun",,terminal_focus +905,1468784,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",0,0,"",python,tab +906,1472346,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1497,0,"",python,selection_mouse +907,1472531,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1497,10,"patch_size",python,selection_mouse +908,1473337,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1509,0,"",python,selection_mouse +909,1474468,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,0,"",python,selection_mouse +910,1474608,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,1,"s",python,selection_mouse +911,1474608,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,3,"sel",python,selection_mouse +912,1474643,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,4,"self",python,selection_mouse +913,1474677,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,5,"self.",python,selection_mouse +914,1474678,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,6,"self.p",python,selection_mouse +915,1474710,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,7,"self.pa",python,selection_mouse +916,1474749,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,8,"self.pat",python,selection_mouse +917,1474750,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,9,"self.patc",python,selection_mouse +918,1474786,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,10,"self.patch",python,selection_mouse +919,1474787,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,11,"self.patch_",python,selection_mouse +920,1474787,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1465,27,"\n self.in_dim * ",python,selection_mouse +921,1475079,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,18,"self.patch_size**2",python,selection_mouse +922,1475188,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1492,19,"self.patch_size**2,",python,selection_mouse +923,1476359,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1510,0,"",python,selection_mouse +924,1484348,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +925,1485273,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1400,0,"",shellscript,selection_mouse +926,1485305,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1399,0,"",shellscript,selection_command +927,1485681,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1382,0,"",shellscript,selection_mouse +928,1485682,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1381,0,"",shellscript,selection_command +929,1488049,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1382,0,"\n ",shellscript,content +930,1488349,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1387,0,"-",shellscript,content +931,1488350,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1388,0,"",shellscript,selection_keyboard +932,1488453,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1388,0,"-",shellscript,content +933,1488454,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1389,0,"",shellscript,selection_keyboard +934,1488698,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1389,0,"p",shellscript,content +935,1488699,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1390,0,"",shellscript,selection_keyboard +936,1488819,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1390,0,"a",shellscript,content +937,1488820,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1391,0,"",shellscript,selection_keyboard +938,1489054,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1391,0,"r",shellscript,content +939,1489055,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1392,0,"",shellscript,selection_keyboard +940,1489498,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1391,1,"",shellscript,content +941,1489591,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1391,0,"t",shellscript,content +942,1489592,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1392,0,"",shellscript,selection_keyboard +943,1489716,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1392,0,"c",shellscript,content +944,1489717,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1393,0,"",shellscript,selection_keyboard +945,1489793,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1393,0,"h",shellscript,content +946,1489794,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1394,0,"",shellscript,selection_keyboard +947,1490156,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1394,0,"_",shellscript,content +948,1490157,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1395,0,"",shellscript,selection_keyboard +949,1490374,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1395,0,"s",shellscript,content +950,1490375,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1396,0,"",shellscript,selection_keyboard +951,1490827,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1396,0,"i",shellscript,content +952,1490828,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1397,0,"",shellscript,selection_keyboard +953,1490963,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1397,0,"z",shellscript,content +954,1490964,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1398,0,"",shellscript,selection_keyboard +955,1491062,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1398,0,"e",shellscript,content +956,1491063,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1399,0,"",shellscript,selection_keyboard +957,1491856,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1399,0,"=",shellscript,content +958,1491857,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1400,0,"",shellscript,selection_keyboard +959,1493187,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1400,0,"4",shellscript,content +960,1493188,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1401,0,"",shellscript,selection_keyboard +961,1493695,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1401,0," ",shellscript,content +962,1493696,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1402,0,"",shellscript,selection_keyboard +963,1494110,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1402,0,"\",shellscript,content +964,1494111,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1403,0,"",shellscript,selection_keyboard +965,1494418,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1402,0,"",shellscript,selection_command +966,1497811,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +967,1498083,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --val_interval 2 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +968,1498235,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +969,1498317,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +970,1501164,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",0,0,"",python,tab +971,1501964,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +972,1503610,"jasmine/train_dynamics.py",0,0,"",python,tab +973,1505343,"jasmine/train_dynamics.py",5302,0,"",python,selection_mouse +974,1506748,"jasmine/train_dynamics.py",0,0,"",python,selection_command +975,1507291,"jasmine/train_dynamics.py",10,0,"",python,selection_command +976,1507826,"jasmine/train_dynamics.py",11,0,"",python,selection_command +977,1507826,"jasmine/train_dynamics.py",12,0,"",python,selection_command +978,1507841,"jasmine/train_dynamics.py",76,0,"",python,selection_command +979,1507873,"jasmine/train_dynamics.py",77,0,"",python,selection_command +980,1507910,"jasmine/train_dynamics.py",118,0,"",python,selection_command +981,1507943,"jasmine/train_dynamics.py",135,0,"",python,selection_command +982,1507975,"jasmine/train_dynamics.py",169,0,"",python,selection_command +983,1508007,"jasmine/train_dynamics.py",170,0,"",python,selection_command +984,1508040,"jasmine/train_dynamics.py",184,0,"",python,selection_command +985,1508072,"jasmine/train_dynamics.py",244,0,"",python,selection_command +986,1508102,"jasmine/train_dynamics.py",303,0,"",python,selection_command +987,1508132,"jasmine/train_dynamics.py",316,0,"",python,selection_command +988,1508160,"jasmine/train_dynamics.py",347,0,"",python,selection_command +989,1508189,"jasmine/train_dynamics.py",366,0,"",python,selection_command +990,1508216,"jasmine/train_dynamics.py",387,0,"",python,selection_command +991,1508245,"jasmine/train_dynamics.py",398,0,"",python,selection_command +992,1508273,"jasmine/train_dynamics.py",422,0,"",python,selection_command +993,1508309,"jasmine/train_dynamics.py",434,0,"",python,selection_command +994,1508336,"jasmine/train_dynamics.py",447,0,"",python,selection_command +995,1508365,"jasmine/train_dynamics.py",470,0,"",python,selection_command +996,1508396,"jasmine/train_dynamics.py",510,0,"",python,selection_command +997,1508422,"jasmine/train_dynamics.py",511,0,"",python,selection_command +998,1508478,"jasmine/train_dynamics.py",561,0,"",python,selection_command +999,1508479,"jasmine/train_dynamics.py",611,0,"",python,selection_command +1000,1508482,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +1001,1508508,"jasmine/train_dynamics.py",643,0,"",python,selection_command +1002,1508568,"jasmine/train_dynamics.py",664,0,"",python,selection_command +1003,1508597,"jasmine/train_dynamics.py",699,0,"",python,selection_command +1004,1508629,"jasmine/train_dynamics.py",720,0,"",python,selection_command +1005,1508630,"jasmine/train_dynamics.py",753,0,"",python,selection_command +1006,1508662,"jasmine/train_dynamics.py",787,0,"",python,selection_command +1007,1508691,"jasmine/train_dynamics.py",789,0,"",python,selection_command +1008,1508720,"jasmine/train_dynamics.py",790,0,"",python,selection_command +1009,1508775,"jasmine/train_dynamics.py",791,0,"",python,selection_command +1010,1508776,"jasmine/train_dynamics.py",802,0,"",python,selection_command +1011,1508807,"jasmine/train_dynamics.py",814,0,"",python,selection_command +1012,1508808,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +1013,1508869,"jasmine/train_dynamics.py",831,0,"",python,selection_command +1014,1508898,"jasmine/train_dynamics.py",860,0,"",python,selection_command +1015,1508899,"jasmine/train_dynamics.py",878,0,"",python,selection_command +1016,1508930,"jasmine/train_dynamics.py",900,0,"",python,selection_command +1017,1508958,"jasmine/train_dynamics.py",928,0,"",python,selection_command +1018,1508986,"jasmine/train_dynamics.py",955,0,"",python,selection_command +1019,1509047,"jasmine/train_dynamics.py",982,0,"",python,selection_command +1020,1509079,"jasmine/train_dynamics.py",1005,0,"",python,selection_command +1021,1509110,"jasmine/train_dynamics.py",1033,0,"",python,selection_command +1022,1509137,"jasmine/train_dynamics.py",1064,0,"",python,selection_command +1023,1509168,"jasmine/train_dynamics.py",1083,0,"",python,selection_command +1024,1509197,"jasmine/train_dynamics.py",1108,0,"",python,selection_command +1025,1509227,"jasmine/train_dynamics.py",1133,0,"",python,selection_command +1026,1509228,"jasmine/train_dynamics.py",1158,0,"",python,selection_command +1027,1509294,"jasmine/train_dynamics.py",1185,0,"",python,selection_command +1028,1509295,"jasmine/train_dynamics.py",1214,0,"",python,selection_command +1029,1509326,"jasmine/train_dynamics.py",1297,0,"",python,selection_command +1030,1509355,"jasmine/train_dynamics.py",1303,0,"",python,selection_command +1031,1509385,"jasmine/train_dynamics.py",1332,0,"",python,selection_command +1032,1509414,"jasmine/train_dynamics.py",1392,0,"",python,selection_command +1033,1509445,"jasmine/train_dynamics.py",1408,0,"",python,selection_command +1034,1509475,"jasmine/train_dynamics.py",1437,0,"",python,selection_command +1035,1509628,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_122541-td4jceyg\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/td4jceyg\r\n",,terminal_output +1036,1509757,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +1037,1509951,"jasmine/train_dynamics.py",1471,0,"",python,selection_command +1038,1510459,"jasmine/train_dynamics.py",1502,0,"",python,selection_command +1039,1510494,"jasmine/train_dynamics.py",1536,0,"",python,selection_command +1040,1510554,"jasmine/train_dynamics.py",1560,0,"",python,selection_command +1041,1510591,"jasmine/train_dynamics.py",1594,0,"",python,selection_command +1042,1510624,"jasmine/train_dynamics.py",1627,0,"",python,selection_command +1043,1511361,"jasmine/train_dynamics.py",1662,0,"",python,selection_command +1044,1511860,"jasmine/train_dynamics.py",1672,0,"",python,selection_command +1045,1511893,"jasmine/train_dynamics.py",1695,0,"",python,selection_command +1046,1511917,"jasmine/train_dynamics.py",1723,0,"",python,selection_command +1047,1511926,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +1048,1511949,"jasmine/train_dynamics.py",1755,0,"",python,selection_command +1049,1511981,"jasmine/train_dynamics.py",1780,0,"",python,selection_command +1050,1512014,"jasmine/train_dynamics.py",1809,0,"",python,selection_command +1051,1512050,"jasmine/train_dynamics.py",1837,0,"",python,selection_command +1052,1512080,"jasmine/train_dynamics.py",1864,0,"",python,selection_command +1053,1512110,"jasmine/train_dynamics.py",1893,0,"",python,selection_command +1054,1512141,"jasmine/train_dynamics.py",1908,0,"",python,selection_command +1055,1512173,"jasmine/train_dynamics.py",1977,0,"",python,selection_command +1056,1512204,"jasmine/train_dynamics.py",2001,0,"",python,selection_command +1057,1512238,"jasmine/train_dynamics.py",2030,0,"",python,selection_command +1058,1512271,"jasmine/train_dynamics.py",2059,0,"",python,selection_command +1059,1512305,"jasmine/train_dynamics.py",2087,0,"",python,selection_command +1060,1512333,"jasmine/train_dynamics.py",2112,0,"",python,selection_command +1061,1512334,"jasmine/train_dynamics.py",2140,0,"",python,selection_command +1062,1512366,"jasmine/train_dynamics.py",2171,0,"",python,selection_command +1063,1512394,"jasmine/train_dynamics.py",2201,0,"",python,selection_command +1064,1512425,"jasmine/train_dynamics.py",2226,0,"",python,selection_command +1065,1512455,"jasmine/train_dynamics.py",2263,0,"",python,selection_command +1066,1512488,"jasmine/train_dynamics.py",2296,0,"",python,selection_command +1067,1512520,"jasmine/train_dynamics.py",2310,0,"",python,selection_command +1068,1512553,"jasmine/train_dynamics.py",2332,0,"",python,selection_command +1069,1512582,"jasmine/train_dynamics.py",2353,0,"",python,selection_command +1070,1512615,"jasmine/train_dynamics.py",2375,0,"",python,selection_command +1071,1512648,"jasmine/train_dynamics.py",2408,0,"",python,selection_command +1072,1514573,"jasmine/train_dynamics.py",2474,0,"",python,selection_command +1073,1517644,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 730, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 574, in main\r\n compiled = train_step.lower(optimizer, first_batch).compile()\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/compilation.py"", line 476, in lower\r\n lowered = self.jitted_fn.lower(*pure_args, **pure_kwargs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/compilation.py"", line 126, in __call__\r\n out = self.f(*args, **kwargs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 445, in train_step\r\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/graph.py"", line 2045, in update_context_manager_wrapper\r\n return f(*args, **kwargs)\r\n ^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/autodiff.py"", line 163, in grad_wrapper\r\n fn_out = gradded_fn(*pure_args)\r\n ^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/autodiff.py"", line 88, in __call__\r\n out = self.f(*args)\r\n ^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 443, in loss_fn\r\n return dynamics_loss_fn(model, inputs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 428, in dynamics_loss_fn\r\n outputs = model(inputs)\r\n ^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py"", line 158, in __call__\r\n tokenizer_outputs = self.tokenizer.vq_encode(videos_BTHWC, training=False)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py"", line 115, in vq_encode\r\n x_BTNL = self.encoder(patch_BTNP)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py"", line 250, in __call__\r\n x_BTNI = self.input_norm1(x_BTNI)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/nn/normalization.py"", line 505, in __call__\r\n return _normalize(\r\n ^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/nn/normalization.py"", line 172, in _normalize\r\n scale = scale.reshape(feature_shape)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/numpy/array_methods.py"", line 1147, in meth\r\n return getattr(self.aval, name).fun(self, *args, **kwargs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/numpy/array_methods.py"", line 316, in _reshape\r\n newshape = _compute_newshape(self, args[0] if len(args) == 1 else args)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/numpy/array_methods.py"", line 477, in _compute_newshape\r\n raise TypeError(f""cannot reshape array of shape {arr.shape} (size {arr.size}) ""\r\nTypeError: cannot reshape array of shape (768,) (size 768) into shape [1, 1, 1, 48] (size 48)\r\n--------------------\r\nFor simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n",,terminal_output +1074,1518258,"TERMINAL",0,0,"wandb: \r\nwandb: 🚀 View run coinrun-dyn-dev-3536670 at: https://wandb.ai/instant-uv/jafar/runs/td4jceyg\r\nwandb: Find logs at: ../../../../../hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_122541-td4jceyg/logs\r\n",,terminal_output +1075,1523424,"TERMINAL",0,0,"W1002 12:25:56.565449 1472303 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugproto job_name: ""jax_worker"": UNAVAILABLE: Cancelling all calls\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_message:""Cancelling all calls"", grpc_status:14}\r\n",,terminal_output +1076,1524251,"TERMINAL",0,0,"srun: error: hkn0401: task 0: Exited with exit code 1\r\n]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h[tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +1077,1531308,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1078,1531309,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",14719,0,"",python,selection_command +1079,1532728,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",14762,0,"",python,selection_mouse +1080,1533017,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",14762,6,"inputs",python,selection_mouse +1081,1534971,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",14747,0,"",python,selection_mouse +1082,1535116,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",14738,16,"dynamics_loss_fn",python,selection_mouse +1083,1547450,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19663,0,"",python,selection_mouse +1084,1549480,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19667,0,"",python,selection_mouse +1085,1549481,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19666,0,"",python,selection_command +1086,1552054,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19667,0,"",python,selection_command +1087,1552178,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19667,0,",",python,content +1088,1552180,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19668,0,"",python,selection_keyboard +1089,1552276,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19668,0," ",python,content +1090,1552277,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19669,0,"",python,selection_keyboard +1091,1552589,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19669,0,"e",python,content +1092,1552590,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19670,0,"",python,selection_keyboard +1093,1552697,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19670,0,"l",python,content +1094,1552699,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19671,0,"",python,selection_keyboard +1095,1552834,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19671,0,"e",python,content +1096,1552836,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19672,0,"",python,selection_keyboard +1097,1552934,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19672,0,"m",python,content +1098,1552936,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19673,0,"",python,selection_keyboard +1099,1553961,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19673,0,"=",python,content +1100,1554562,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19673,1,"",python,content +1101,1554692,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19672,1,"",python,content +1102,1554836,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19671,1,"",python,content +1103,1554977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19670,1,"",python,content +1104,1555074,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19669,1,"",python,content +1105,1555993,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19669,0,"l",python,content +1106,1555994,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19670,0,"",python,selection_keyboard +1107,1556485,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19670,0,"ocal_data=elem",python,content +1108,1557130,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19683,0,"",python,selection_command +1109,1559594,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19763,0,"",python,selection_mouse +1110,1559601,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19762,0,"",python,selection_command +1111,1560237,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19735,0,"",python,selection_mouse +1112,1564610,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --val_interval 2 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +1113,1564727,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +1114,1564850,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +1115,1568565,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +1116,1575259,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +1117,1575387,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +1118,1576219,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_122648-c720yljw\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/c720yljw\r\n",,terminal_output +1119,1576380,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +1120,1578510,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +1121,1584071,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 730, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 574, in main\r\n compiled = train_step.lower(optimizer, first_batch).compile()\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/compilation.py"", line 476, in lower\r\n lowered = self.jitted_fn.lower(*pure_args, **pure_kwargs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/compilation.py"", line 126, in __call__\r\n out = self.f(*args, **kwargs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 445, in train_step\r\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/graph.py"", line 2045, in update_context_manager_wrapper\r\n return f(*args, **kwargs)\r\n ^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/autodiff.py"", line 163, in grad_wrapper\r\n fn_out = gradded_fn(*pure_args)\r\n ^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/transforms/autodiff.py"", line 88, in __call__\r\n out = self.f(*args)\r\n ^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 443, in loss_fn\r\n return dynamics_loss_fn(model, inputs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 428, in dynamics_loss_fn\r\n outputs = model(inputs)\r\n ^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py"", line 158, in __call__\r\n tokenizer_outputs = self.tokenizer.vq_encode(videos_BTHWC, training=False)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py"", line 115, in vq_encode\r\n x_BTNL = self.encoder(patch_BTNP)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py"", line 250, in __call__\r\n x_BTNI = self.input_norm1(x_BTNI)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/nn/normalization.py"", line 505, in __call__\r\n return _normalize(\r\n ^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/flax/nnx/nn/normalization.py"", line 172, in _normalize\r\n scale = scale.reshape(feature_shape)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/numpy/array_methods.py"", line 1147, in meth\r\n return getattr(self.aval, name).fun(self, *args, **kwargs)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/numpy/array_methods.py"", line 316, in _reshape\r\n newshape = _compute_newshape(self, args[0] if len(args) == 1 else args)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/numpy/array_methods.py"", line 477, in _compute_newshape\r\n raise TypeError(f""cannot reshape array of shape {arr.shape} (size {arr.size}) ""\r\nTypeError: cannot reshape array of shape (768,) (size 768) into shape [1, 1, 1, 48] (size 48)\r\n--------------------\r\nFor simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n",,terminal_output +1122,1584649,"TERMINAL",0,0,"wandb: \r\nwandb: 🚀 View run coinrun-dyn-dev-3536670 at: https://wandb.ai/instant-uv/jafar/runs/c720yljw\r\nwandb: Find logs at: ../../../../../hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_122648-c720yljw/logs\r\n",,terminal_output +1123,1584788,"TERMINAL",0,0,"W1002 12:26:57.975221 1472833 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugproto job_name: ""jax_worker"": UNAVAILABLE: failed to connect to all addresses; last error: UNKNOWN: ipv4:10.0.1.33:63262: Failed to connect to remote host: Connection refused\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_status:14, grpc_message:""failed to connect to all addresses; last error: UNKNOWN: ipv4:10.0.1.33:63262: Failed to connect to remote host: Connection refused""}\r\n",,terminal_output +1124,1585674,"TERMINAL",0,0,"srun: error: hkn0401: task 0: Exited with exit code 1\r\n]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h[tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +1125,1588228,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1126,1598122,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +1127,1602888,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1378,0,"",shellscript,selection_mouse +1128,1603023,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1377,3,"110",shellscript,selection_mouse +1129,1604138,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1378,0,"",shellscript,selection_mouse +1130,1604184,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1377,3,"110",shellscript,selection_mouse +1131,1614627,"TERMINAL",0,0,"p",,terminal_output +1132,1614760,"TERMINAL",0,0,"y",,terminal_output +1133,1614934,"TERMINAL",0,0,"t",,terminal_output +1134,1615002,"TERMINAL",0,0,"h",,terminal_output +1135,1615153,"TERMINAL",0,0,"o",,terminal_output +1136,1615210,"TERMINAL",0,0,"n",,terminal_output +1137,1615489,"TERMINAL",0,0,"\r\n[?2004l\rPython 3.9.18 (main, Sep 4 2025, 00:00:00) \r\n[GCC 11.4.1 20231218 (Red Hat 11.4.1-3)] on linux\r\nType ""help"", ""copyright"", ""credits"" or ""license"" for more information.\r\n>>> ",,terminal_output +1138,1616415,"TERMINAL",0,0,"4",,terminal_output +1139,1616714,"TERMINAL",0,0,"*",,terminal_output +1140,1616836,"TERMINAL",0,0,"*",,terminal_output +1141,1617479,"TERMINAL",0,0,"2",,terminal_output +1142,1617980,"TERMINAL",0,0,"\r\n16\r\n>>> ",,terminal_output +1143,1619657,"TERMINAL",0,0,"1",,terminal_output +1144,1620227,"TERMINAL",0,0,"6",,terminal_output +1145,1620482,"TERMINAL",0,0,"*",,terminal_output +1146,1620733,"TERMINAL",0,0,"*",,terminal_output +1147,1620833,"TERMINAL",0,0,"2",,terminal_output +1148,1620914,"TERMINAL",0,0,"\r\n256\r\n>>> ",,terminal_output +1149,1639343,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",0,0,"",python,tab +1150,1639344,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",7815,0,"",python,selection_command +1151,1643338,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",13508,0,"",python,selection_mouse +1152,1643499,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",13501,11,"input_norm1",python,selection_mouse +1153,1644071,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",13525,0,"",python,selection_mouse +1154,1644238,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",13519,9,"LayerNorm",python,selection_mouse +1155,1644918,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",13509,0,"",python,selection_mouse +1156,1645034,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",13501,11,"input_norm1",python,selection_mouse +1157,1645491,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",13519,0,"",python,selection_mouse +1158,1645635,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",13519,9,"LayerNorm",python,selection_mouse +1159,1646263,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",13503,0,"",python,selection_mouse +1160,1647271,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5082,0,"",python,selection_command +1161,1649953,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",6417,0,"",python,selection_mouse +1162,1650103,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",6412,9,"input_dim",python,selection_mouse +1163,1652324,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5699,0,"",python,selection_mouse +1164,1652478,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5698,9,"model_dim",python,selection_mouse +1165,1652688,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5665,0,"",python,selection_mouse +1166,1652927,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",5663,9,"input_dim",python,selection_mouse +1167,1659485,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",0,0,"",python,tab +1168,1661654,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1479,0,"",python,selection_mouse +1169,1661771,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1478,4,"self",python,selection_mouse +1170,1662317,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1484,0,"",python,selection_mouse +1171,1662408,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",1483,6,"in_dim",python,selection_mouse +1172,1667591,"TERMINAL",0,0,"bash",,terminal_focus +1173,1669704,"TERMINAL",0,0,"dev",,terminal_command +1174,1672047,"TERMINAL",0,0,"python",,terminal_command +1175,1672124,"TERMINAL",0,0,"]633;CPython 3.9.18 (main, Sep 4 2025, 00:00:00) \r\n[GCC 11.4.1 20231218 (Red Hat 11.4.1-3)] on linux\r\nType ""help"", ""copyright"", ""credits"" or ""license"" for more information.\r\n>>> ",,terminal_output +1176,1673379,"TERMINAL",0,0,"4",,terminal_output +1177,1673630,"TERMINAL",0,0,"*",,terminal_output +1178,1673866,"TERMINAL",0,0,"*",,terminal_output +1179,1674040,"TERMINAL",0,0,"2",,terminal_output +1180,1674169,"TERMINAL",0,0,"\r\n16\r\n>>> ",,terminal_output +1181,1679109,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1182,1690966,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",0,0,"",python,tab +1183,1696467,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1184,1701514,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",3119,0,"",python,selection_mouse +1185,1701660,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",3109,14,"image_channels",python,selection_mouse +1186,1702344,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",3141,0,"",python,selection_mouse +1187,1702756,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",3106,0,"",python,selection_mouse +1188,1705865,"TERMINAL",0,0,"8",,terminal_output +1189,1706703,"TERMINAL",0,0,"",,terminal_output +1190,1707104,"TERMINAL",0,0,"1",,terminal_output +1191,1707429,"TERMINAL",0,0,"6",,terminal_output +1192,1707733,"TERMINAL",0,0," ",,terminal_output +1193,1707983,"TERMINAL",0,0,"*",,terminal_output +1194,1708108,"TERMINAL",0,0,"*",,terminal_output +1195,1709029,"TERMINAL",0,0,"2",,terminal_output +1196,1709097,"TERMINAL",0,0," ",,terminal_output +1197,1709830,"TERMINAL",0,0,"*",,terminal_output +1198,1710468,"TERMINAL",0,0,"",,terminal_output +1199,1711003,"TERMINAL",0,0,"\r\n256\r\n>>> ",,terminal_output +1200,1712764,"TERMINAL",0,0,"16 **2 ",,terminal_output +1201,1713201,"TERMINAL",0,0,"",,terminal_output +1202,1713284,"TERMINAL",0,0,"",,terminal_output +1203,1714254,"TERMINAL",0,0,"[1@3",,terminal_output +1204,1714518,"TERMINAL",0,0,"[1@ ",,terminal_output +1205,1714680,"TERMINAL",0,0,"[1@*",,terminal_output +1206,1714740,"TERMINAL",0,0,"[1@ ",,terminal_output +1207,1715067,"TERMINAL",0,0,"\r\n768\r\n>>> ",,terminal_output +1208,1721144,"TERMINAL",0,0,"1",,terminal_output +1209,1721547,"TERMINAL",0,0,"6",,terminal_output +1210,1721827,"TERMINAL",0,0," ",,terminal_output +1211,1722105,"TERMINAL",0,0,"*",,terminal_output +1212,1722563,"TERMINAL",0,0,"3",,terminal_output +1213,1722644,"TERMINAL",0,0,"\r\n48\r\n>>> ",,terminal_output +1214,1726558,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",3442,0,"",python,selection_mouse +1215,1728479,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",3386,0,"",python,selection_mouse +1216,1732721,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",3643,0,"",python,selection_command +1217,1733751,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",3339,0,"",python,selection_command +1218,1734102,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",3323,0,"",python,selection_command +1219,1734461,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1788,0,"",python,selection_command +1220,1736668,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1540,0,"",python,selection_command +1221,1738108,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",3663,0,"",python,selection_command +1222,1739246,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1540,0,"",python,selection_command +1223,1741077,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1510,0,"",python,selection_mouse +1224,1741220,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1506,17,"num_patch_latents",python,selection_mouse +1225,1741505,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1546,0,"",python,selection_mouse +1226,1741628,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1540,10,"patch_size",python,selection_mouse +1227,1742424,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1559,0,"",python,selection_mouse +1228,1742427,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1558,0,"",python,selection_command +1229,1742460,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1558,1,"4",python,selection_mouse +1230,1742461,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1559,0,"",python,selection_command +1231,1742579,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1558,1,"4",python,selection_mouse +1232,1742580,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1554,5,"t = 4",python,selection_mouse +1233,1742580,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1551,8," int = 4",python,selection_mouse +1234,1742614,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1548,11,"ze: int = 4",python,selection_mouse +1235,1742615,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1546,13,"size: int = 4",python,selection_mouse +1236,1742648,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1544,15,"h_size: int = 4",python,selection_mouse +1237,1742649,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1542,17,"tch_size: int = 4",python,selection_mouse +1238,1742681,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1539,20," patch_size: int = 4",python,selection_mouse +1239,1742681,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1538,21," patch_size: int = 4",python,selection_mouse +1240,1742711,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1537,22," patch_size: int = 4",python,selection_mouse +1241,1742747,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1559,1,"\n",python,selection_mouse +1242,1743126,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1540,0,"",python,selection_mouse +1243,1743486,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1542,0,"",python,selection_mouse +1244,1743905,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1545,0,"",python,selection_mouse +1245,1744060,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1540,10,"patch_size",python,selection_mouse +1246,1744739,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1559,0,"",python,selection_mouse +1247,1744743,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",1558,0,"",python,selection_command +1248,1746194,"jasmine/models/tokenizer.py",0,0,"",python,tab +1249,1752834,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +1250,1755944,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1524,0,"",shellscript,selection_mouse +1251,1756316,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1611,0,"",shellscript,selection_mouse +1252,1756693,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1749,0,"",shellscript,selection_mouse +1253,1757022,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1889,0,"",shellscript,selection_mouse +1254,1757431,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1936,0,"",shellscript,selection_mouse +1255,1757733,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1796,0,"",shellscript,selection_mouse +1256,1760735,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1389,0,"",shellscript,selection_command +1257,1763044,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1383,21,"",shellscript,content +1258,1763078,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1387,0,"",shellscript,selection_command +1259,1763955,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1389,0,"patch_size=4 \\n --",shellscript,content +1260,1763956,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1389,0,"",shellscript,selection_command +1261,1767777,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=00:20:00\n#SBATCH --partition=dev_accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_test\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_test/train\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_test/val\n\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nexport PYTHONUNBUFFERED=1\n\nsrun python jasmine/train_tokenizer.py \\n --save_ckpt \\n --image_height=64 \\n --image_width=64 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=120 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=50 \\n --log_checkpoint_interval=2 \\n --log \\n --name=coinrun-tokenizer-dataset-test-$slurm_job_id \\n --tags tokenizer coinrun dev \\n --entity instant-uv \\n --project jafar \\n --warmup_steps 0 \\n --wsd_decay_steps 0 \\n --num_steps 10 \\n --data_dir $array_records_dir_train \\n --val_data_dir $array_records_dir_val \\n --val_interval 3 \\n --val_steps 5\n",shellscript,tab +1262,1769712,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1564,0,"",shellscript,selection_mouse +1263,1770216,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1546,0,"",shellscript,selection_mouse +1264,1770221,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1545,0,"",shellscript,selection_command +1265,1771311,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1546,0,"\n ",shellscript,content +1266,1771850,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1551,0,"-",shellscript,content +1267,1771851,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1552,0,"",shellscript,selection_keyboard +1268,1771994,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1552,0,"-",shellscript,content +1269,1771995,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1553,0,"",shellscript,selection_keyboard +1270,1772165,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1553,0,"p",shellscript,content +1271,1772166,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1554,0,"",shellscript,selection_keyboard +1272,1772735,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1554,0,"atch_size=4 \",shellscript,content +1273,1774067,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1449,0,"",shellscript,selection_mouse +1274,1782926,"TERMINAL",0,0,"srun",,terminal_focus +1275,1785447,"TERMINAL",0,0,"sh",,terminal_output +1276,1785568,"TERMINAL",0,0," ",,terminal_output +1277,1785840,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",,terminal_output +1278,1787019,"TERMINAL",0,0,"\r\nKeyboardInterrupt\r\n>>> ",,terminal_output +1279,1787415,"TERMINAL",0,0,"\r\n",,terminal_output +1280,1787577,"TERMINAL",0,0,"]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h[tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +1281,1788679,"TERMINAL",0,0,"s",,terminal_output +1282,1788767,"TERMINAL",0,0,"o",,terminal_output +1283,1788853,"TERMINAL",0,0,"u",,terminal_output +1284,1788907,"TERMINAL",0,0,"r",,terminal_output +1285,1789095,"TERMINAL",0,0,"c",,terminal_output +1286,1789146,"TERMINAL",0,0,"e",,terminal_output +1287,1789242,"TERMINAL",0,0," ",,terminal_output +1288,1789422,"TERMINAL",0,0,".",,terminal_output +1289,1789506,"TERMINAL",0,0,"",,terminal_output +1290,1789658,"TERMINAL",0,0,"b",,terminal_output +1291,1789790,"TERMINAL",0,0,"",,terminal_output +1292,1790279,"TERMINAL",0,0,"",,terminal_output +1293,1790868,"TERMINAL",0,0,"v",,terminal_output +1294,1790951,"TERMINAL",0,0,"e",,terminal_output +1295,1791131,"TERMINAL",0,0,"nv/",,terminal_output +1296,1791499,"TERMINAL",0,0,"n",,terminal_output +1297,1791579,"TERMINAL",0,0,"",,terminal_output +1298,1792070,"TERMINAL",0,0,"",,terminal_output +1299,1792373,"TERMINAL",0,0,"b",,terminal_output +1300,1792455,"TERMINAL",0,0,"in/",,terminal_output +1301,1792842,"TERMINAL",0,0,"a",,terminal_output +1302,1792968,"TERMINAL",0,0,"c",,terminal_output +1303,1793124,"TERMINAL",0,0,"tivate",,terminal_output +1304,1793567,"TERMINAL",0,0,"\r\n[?2004l\r]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +1305,1794547,"TERMINAL",0,0,"s",,terminal_output +1306,1794606,"TERMINAL",0,0,"h",,terminal_output +1307,1794698,"TERMINAL",0,0," ",,terminal_output +1308,1795006,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",,terminal_output +1309,1795236,"TERMINAL",0,0,"\rslurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/tokenizer/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/tokenizer/%x_%j.log\r\n#SBATCH --job-name=train_tokenizer_1e-4\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_test\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_test/train\r\n# array_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_test/val\r\n\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_tokenizer.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=120 \\r\n --init_lr=0 \\r\n --patch_size=4 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --log \\r\n --name=coinrun-tokenizer-dataset-test-$slurm_job_id \\r\n --tags tokenizer coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --val_data_dir $array_records_dir_val \\r\n --val_interval 3 \\r\n --val_steps 5\r\n",,terminal_output +1310,1795451,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +1311,1795577,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +1312,1798488,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +1313,1802514,"TERMINAL",0,0,"Counting all components: ['decoder', 'encoder', 'vq']\r\n",,terminal_output +1314,1802865,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +1315,1803720,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_123036-5adyg0sl\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-tokenizer-dataset-test-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/5adyg0sl\r\n",,terminal_output +1316,1803873,"TERMINAL",0,0,"Parameter counts:\r\n{'decoder': 16858736, 'encoder': 16858752, 'vq': 32768, 'total': 33750256}\r\n",,terminal_output +1317,1847132,"TERMINAL",0,0,"Total memory size: 29.3 GB, Output size: 0.5 GB, Temp size: 28.8 GB, Argument size: 0.4 GB, Host temp size: 0.0 GB.\r\nFLOPs: 1.335e+13, Bytes: 7.598e+11 (707.6 GB), Intensity: 17.6 FLOPs/byte\r\nStarting training from step 0...\r\n",,terminal_output +1318,1847341,"TERMINAL",0,0,"\r\nMemstats: After params initialized:\r\n\tUsing (GB) 0.52 / 38.7 (1.343669%) on cuda:0\r\n",,terminal_output +1319,1864260,"TERMINAL",0,0,"WARNING:absl:[process=0][thread=MainThread][operation_id=1] _SignalingThread.join() waiting for signals ([, ]) blocking the main thread will slow down blocking save times. This is likely due to main thread calling result() on a CommitFuture.\r\n",,terminal_output +1320,1864668,"TERMINAL",0,0,"WARNING:absl:[process=0][thread=MainThread][operation_id=1] _SignalingThread.join() waiting for signals ([, ]) blocking the main thread will slow down blocking save times. This is likely due to main thread calling result() on a CommitFuture.\r\n",,terminal_output +1321,1866372,"TERMINAL",0,0,"Saved checkpoint at step 2\r\n",,terminal_output +1322,1866556,"TERMINAL",0,0,"Calculating validation metrics...\r\n",,terminal_output +1323,1880142,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",0,0,"",shellscript,tab +1324,1882644,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1185,0,"",shellscript,selection_mouse +1325,1882743,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1185,1,"/",shellscript,selection_mouse +1326,1882744,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1185,2,"/h",shellscript,selection_mouse +1327,1882760,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1185,4,"/hkf",shellscript,selection_mouse +1328,1882780,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1185,8,"/hkfs/wo",shellscript,selection_mouse +1329,1882828,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1169,16,"\nCHECKPOINT_DIR=",shellscript,selection_mouse +1330,1882872,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1168,17,"\n\nCHECKPOINT_DIR=",shellscript,selection_mouse +1331,1882910,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1169,16,"\nCHECKPOINT_DIR=",shellscript,selection_mouse +1332,1883288,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1185,97,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/",shellscript,selection_mouse +1333,1883297,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1185,96,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name",shellscript,selection_mouse +1334,1883331,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1169,16,"\nCHECKPOINT_DIR=",shellscript,selection_mouse +1335,1883407,"TERMINAL",0,0,"Step 3, validation loss: 0.19625379145145416\r\n",,terminal_output +1336,1883764,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1185,88,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$",shellscript,selection_mouse +1337,1883839,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1185,87,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/",shellscript,selection_mouse +1338,1883919,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1185,86,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer",shellscript,selection_mouse +1339,1883958,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1185,85,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenize",shellscript,selection_mouse +1340,1884348,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",1185,86,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer",shellscript,selection_mouse +1341,1887488,"TERMINAL",0,0,"python",,terminal_focus +1342,1888443,"TERMINAL",0,0,"\r\n",,terminal_output +1343,1888466,"TERMINAL",0,0,"]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +1344,1900096,"TERMINAL",0,0,"WARNING:absl:[process=0][thread=MainThread][operation_id=2] _SignalingThread.join() waiting for signals ([, ]) blocking the main thread will slow down blocking save times. This is likely due to main thread calling result() on a CommitFuture.\r\nWARNING:absl:[process=0][thread=MainThread][operation_id=2] _SignalingThread.join() waiting for signals ([, ]) blocking the main thread will slow down blocking save times. This is likely due to main thread calling result() on a CommitFuture.\r\n",,terminal_output +1345,1902314,"TERMINAL",0,0,"Saved checkpoint at step 4\r\n",,terminal_output +1346,1904271,"TERMINAL",0,0,"Calculating validation metrics...\r\n",,terminal_output +1347,1908958,"TERMINAL",0,0,"Step 6, validation loss: 0.17459146678447723\r\nWARNING:absl:[process=0][thread=MainThread][operation_id=3] _SignalingThread.join() waiting for signals ([, ]) blocking the main thread will slow down blocking save times. This is likely due to main thread calling result() on a CommitFuture.\r\nWARNING:absl:[process=0][thread=MainThread][operation_id=3] _SignalingThread.join() waiting for signals ([, ]) blocking the main thread will slow down blocking save times. This is likely due to main thread calling result() on a CommitFuture.\r\n",,terminal_output +1348,1909365,"TERMINAL",0,0,"Saved checkpoint at step 6\r\n",,terminal_output +1349,1910115,"TERMINAL",0,0,"WARNING:absl:[process=0][thread=MainThread][operation_id=4] _SignalingThread.join() waiting for signals ([, ]) blocking the main thread will slow down blocking save times. This is likely due to main thread calling result() on a CommitFuture.\r\nWARNING:absl:[process=0][thread=MainThread][operation_id=4] _SignalingThread.join() waiting for signals ([, ]) blocking the main thread will slow down blocking save times. This is likely due to main thread calling result() on a CommitFuture.\r\n",,terminal_output +1350,1910141,"TERMINAL",0,0,"cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670",,terminal_command +1351,1910145,"TERMINAL",0,0,"]633;C]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670",,terminal_output +1352,1910860,"TERMINAL",0,0,"ls",,terminal_command +1353,1910893,"TERMINAL",0,0,"]633;C000002 000004 000006 000008.orbax-checkpoint-tmp\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670",,terminal_output +1354,1913477,"TERMINAL",0,0,"Saved checkpoint at step 8\r\n",,terminal_output +1355,1913621,"TERMINAL",0,0,"Calculating validation metrics...\r\n",,terminal_output +1356,1915027,"TERMINAL",0,0,"WARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670/000008 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670/000008) to end with "".orbax-checkpoint-tmp"".\r\nWARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670/000002 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670/000002) to end with "".orbax-checkpoint-tmp"".\r\nWARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670/000006 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670/000006) to end with "".orbax-checkpoint-tmp"".\r\nWARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670/000004 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670/000004) to end with "".orbax-checkpoint-tmp"".\r\n",,terminal_output +1357,1915921,"TERMINAL",0,0,"srun",,terminal_focus +1358,1917133,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3536670.5 task 0: running\r\n",,terminal_output +1359,1917600,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.5\r\nsrun: forcing job termination\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-12:\r\nProcess SpawnProcess-11:\r\nProcess SpawnProcess-9:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-16:\r\nProcess SpawnProcess-10:\r\nProcess SpawnProcess-13:\r\nProcess SpawnProcess-15:\r\nProcess SpawnProcess-14:\r\nProcess SpawnProcess-4:\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\nTraceback (most recent call last):\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\nKeyboardInterrupt\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\nTraceback (most recent call last):\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\nTraceback (most recent call last):\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\nTraceback (most recent call last):\r\nTraceback (most recent call last):\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\nTraceback (most recent call last):\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\nTraceback (most recent call last):\r\nslurmstepd: error: *** STEP 3536670.5 ON hkn0401 CANCELLED AT 2025-10-02T12:32:30 ***\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 314, in _bootstrap\r\n self.run()\r\n File ""/usr/lib64/python3.12/multiprocessing/process.py"", line 108, in run\r\n self._target(*self._args, **self._kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/grain_pool.py"", line 261, in _worker_loop\r\n if not multiprocessing_common.add_element_to_queue( # pytype: disable=wrong-arg-types\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/grain/_src/python/multiprocessing_common.py"", line 54, in add_element_to_queue\r\n elements_queue.put(element, timeout=_QUEUE_WAIT_TIMEOUT_SECONDS)\r\n File ""/usr/lib64/python3.12/multiprocessing/queues.py"", line 89, in put\r\n if not self._sem.acquire(block, timeout):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\n",,terminal_output +1360,1917769,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.5\r\nsrun: job abort in progress\r\n",,terminal_output +1361,1918010,"TERMINAL",0,0,"bash",,terminal_focus +1362,1918048,"TERMINAL",0,0,"]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +1363,1919031,"TERMINAL",0,0,"ls",,terminal_command +1364,1924925,"TERMINAL",0,0,"pwd",,terminal_command +1365,1929389,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",0,0,"",shellscript,tab +1366,1931701,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +1367,1934091,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1175,0,"",shellscript,selection_mouse +1368,1934128,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1174,0,"",shellscript,selection_command +1369,1934208,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1174,1,"5",shellscript,selection_mouse +1370,1934208,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1175,0,"",shellscript,selection_command +1371,1934285,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1172,3,"955",shellscript,selection_mouse +1372,1934319,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1167,8,"/3528955",shellscript,selection_mouse +1373,1934320,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1162,13,"fault/3528955",shellscript,selection_mouse +1374,1934354,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1157,18,"er_default/3528955",shellscript,selection_mouse +1375,1934355,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1151,24,"okenizer_default/3528955",shellscript,selection_mouse +1376,1934392,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1147,28,"in_tokenizer_default/3528955",shellscript,selection_mouse +1377,1934392,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1143,32,"/train_tokenizer_default/3528955",shellscript,selection_mouse +1378,1934430,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1138,37,"nizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1379,1934431,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1136,39,"kenizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1380,1934431,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1134,41,"tokenizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1381,1934468,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1132,43,"s/tokenizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1382,1934469,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1175,1,"\n",shellscript,selection_mouse +1383,1934576,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1175,18,"\n\nenv | grep SLURM",shellscript,selection_mouse +1384,1934690,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1175,1,"\n",shellscript,selection_mouse +1385,1935080,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1049,126,"ckpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1386,1935107,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1050,125,"kpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1387,1935141,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1051,124,"point=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1388,1935180,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1052,123,"oint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1389,1935237,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1053,122,"int=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1390,1935251,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1054,121,"nt=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1391,1935370,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1055,120,"t=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1392,1935395,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1056,119,"=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1393,1935601,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1057,118,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/train_tokenizer_default/3528955",shellscript,selection_mouse +1394,1936551,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1057,118,"",shellscript,content +1395,1936570,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1056,0,"",shellscript,selection_command +1396,1936923,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1057,0,"",shellscript,selection_command +1397,1937404,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1057,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670",shellscript,content +1398,1937965,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1162,0,"",shellscript,selection_command +1399,1942334,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",0,0,"",shellscript,tab +1400,1947092,"TERMINAL",0,0,"srun",,terminal_focus +1401,1948252,"TERMINAL",0,0,"s",,terminal_output +1402,1948692,"TERMINAL",0,0,"",,terminal_output +1403,1948841,"TERMINAL",0,0,"",,terminal_output +1404,1949090,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",,terminal_output +1405,1949227,"TERMINAL",0,0,"\rource .venv/bin/activate",,terminal_output +1406,1949912,"TERMINAL",0,0,"python",,terminal_output +1407,1950239,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +1408,1951683,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --val_interval 2 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +1409,1951832,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +1410,1951962,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +1411,1955788,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +1412,1962150,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +1413,1962503,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +1414,1963331,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_123315-4ycgzjbl\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/4ycgzjbl\r\n",,terminal_output +1415,1963492,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +1416,1965604,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +1417,2043573,"TERMINAL",0,0,"Total memory size: 23.3 GB, Output size: 0.9 GB, Temp size: 22.3 GB, Argument size: 0.9 GB, Host temp size: 0.0 GB.\r\nFLOPs: 1.367e+13, Bytes: 6.890e+11 (641.7 GB), Intensity: 19.8 FLOPs/byte\r\nStarting training from step 0...\r\n",,terminal_output +1418,2044260,"TERMINAL",0,0,"\r\nMemstats: After params initialized:\r\n\tUsing (GB) 1.15 / 38.7 (2.971576%) on cuda:0\r\n",,terminal_output +1419,2223011,"TERMINAL",0,0,"bash",,terminal_focus +1420,2242025,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1421,2258437,"TERMINAL",0,0,"srun",,terminal_focus +1422,2260641,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1423,2270163,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1424,2274533,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20668,0,"",python,selection_mouse +1425,2274568,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20667,0,"",python,selection_command +1426,2274684,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20667,1,"1",python,selection_mouse +1427,2274715,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20668,0,"",python,selection_command +1428,2275338,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20581,0,"",python,selection_mouse +1429,2275729,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20615,0,"",python,selection_mouse +1430,2275888,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20603,15,"print_mem_stats",python,selection_mouse +1431,2289174,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1432,2295164,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +1433,2328417,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3536670.6 task 0: running\r\n",,terminal_output +1434,2328733,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.6\r\nsrun: forcing job termination\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 730, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 581, in main\r\n while step < args.num_steps:\r\n ^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\nWARNING:asyncio:socket.send() raised exception.\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3536670.6 ON hkn0401 CANCELLED AT 2025-10-02T12:39:21 ***\r\n",,terminal_output +1435,2329194,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.6\r\nsrun: job abort in progress\r\n^Csrun: sending Ctrl-C to StepId=3536670.6\r\nsrun: job abort in progress\r\n]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +1436,2334481,"jasmine/train_dynamics.py",0,0,"",python,tab +1437,2338789,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +1438,2340065,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1881,0,"",shellscript,selection_mouse +1439,2340103,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1880,0,"",shellscript,selection_command +1440,2341651,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1721,0,"",shellscript,selection_mouse +1441,2341652,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1720,0,"",shellscript,selection_command +1442,2342093,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1674,0,"",shellscript,selection_mouse +1443,2342094,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1673,0,"",shellscript,selection_command +1444,2342461,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1700,0,"",shellscript,selection_mouse +1445,2342462,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1699,0,"",shellscript,selection_command +1446,2343202,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1905,0,"",shellscript,selection_mouse +1447,2343205,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1904,0,"",shellscript,selection_command +1448,2343549,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1923,0,"",shellscript,selection_mouse +1449,2343553,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1922,0,"",shellscript,selection_command +1450,2343638,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1922,1,"5",shellscript,selection_mouse +1451,2343644,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1923,0,"",shellscript,selection_command +1452,2344077,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1881,0,"",shellscript,selection_mouse +1453,2344141,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1880,0,"",shellscript,selection_command +1454,2344885,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1794,0,"",shellscript,selection_mouse +1455,2345265,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1846,0,"",shellscript,selection_mouse +1456,2346071,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1858,0,"\n ",shellscript,content +1457,2347065,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1863,0,"-",shellscript,content +1458,2347066,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1864,0,"",shellscript,selection_keyboard +1459,2347182,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1864,0,"-",shellscript,content +1460,2347183,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1865,0,"",shellscript,selection_keyboard +1461,2347382,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1865,0,"l",shellscript,content +1462,2347383,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1866,0,"",shellscript,selection_keyboard +1463,2347564,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1866,0,"o",shellscript,content +1464,2347564,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1867,0,"",shellscript,selection_keyboard +1465,2349462,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1867,0,"g",shellscript,content +1466,2349464,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1868,0,"",shellscript,selection_keyboard +1467,2351287,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1868,0,"_",shellscript,content +1468,2351289,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1869,0,"",shellscript,selection_keyboard +1469,2356323,"jasmine/train_dynamics.py",0,0,"",python,tab +1470,2360698,"jasmine/train_dynamics.py",2487,0,"",python,selection_mouse +1471,2360834,"jasmine/train_dynamics.py",2478,12,"log_interval",python,selection_mouse +1472,2365083,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +1473,2366379,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1865,4,"",shellscript,content +1474,2366714,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1865,0,"log_interval",shellscript,content +1475,2367692,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1877,0,"=",shellscript,content +1476,2367693,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1878,0,"",shellscript,selection_keyboard +1477,2367764,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1878,0,"!",shellscript,content +1478,2367765,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1879,0,"",shellscript,selection_keyboard +1479,2368125,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1879,0," ",shellscript,content +1480,2368125,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1880,0,"",shellscript,selection_keyboard +1481,2368398,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1879,1,"",shellscript,content +1482,2368534,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1878,1,"",shellscript,content +1483,2368621,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1878,0,"1",shellscript,content +1484,2368621,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1879,0,"",shellscript,selection_keyboard +1485,2369063,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1879,0," ",shellscript,content +1486,2369064,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1880,0,"",shellscript,selection_keyboard +1487,2369296,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1880,0,"\",shellscript,content +1488,2369297,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1881,0,"",shellscript,selection_keyboard +1489,2369835,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1880,0,"",shellscript,selection_command +1490,2369963,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1879,0,"",shellscript,selection_command +1491,2370113,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1878,0,"",shellscript,selection_command +1492,2370463,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1877,1,"",shellscript,content +1493,2370488,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1877,0," ",shellscript,content +1494,2370489,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1878,0,"",shellscript,selection_keyboard +1495,2371554,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1877,0,"",shellscript,selection_command +1496,2375951,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +1497,2376127,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --log_interval 1 \\r\n --val_interval 2 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +1498,2376271,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +1499,2376362,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +1500,2380230,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +1501,2386611,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +1502,2387241,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +1503,2387929,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_124020-66kv8l9p\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/66kv8l9p\r\n",,terminal_output +1504,2388084,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +1505,2390202,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +1506,2402114,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1507,2403830,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20668,0,"",python,selection_mouse +1508,2403831,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20667,0,"",python,selection_command +1509,2404724,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20668,0,"\n ",python,content +1510,2404922,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20681,0,"p",python,content +1511,2404923,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20682,0,"",python,selection_keyboard +1512,2405036,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20682,0,"r",python,content +1513,2405037,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20683,0,"",python,selection_keyboard +1514,2405173,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20683,0,"i",python,content +1515,2405174,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20684,0,"",python,selection_keyboard +1516,2405264,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20684,0,"n",python,content +1517,2405266,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20685,0,"",python,selection_keyboard +1518,2405332,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20685,0,"t",python,content +1519,2405333,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20686,0,"",python,selection_keyboard +1520,2405887,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20686,0,"()",python,content +1521,2405889,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20687,0,"",python,selection_keyboard +1522,2406127,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20687,0,"""""",python,content +1523,2406128,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20688,0,"",python,selection_keyboard +1524,2407132,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20688,0,"S",python,content +1525,2407134,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20689,0,"",python,selection_keyboard +1526,2407580,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20690,0,", step",python,content +1527,2407580,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20689,0,"tep",python,content +1528,2407584,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20699,0,"",python,selection_command +1529,2408305,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20698,0,"",python,selection_command +1530,2410470,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3536670.7 task 0: running\r\n",,terminal_output +1531,2410787,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.7\r\nsrun: forcing job termination\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3536670.7 ON hkn0401 CANCELLED AT 2025-10-02T12:40:43 ***\r\n",,terminal_output +1532,2411128,"TERMINAL",0,0,"]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +1533,2412108,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +1534,2412371,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --log_interval 1 \\r\n --val_interval 2 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +1535,2412514,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +1536,2412596,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +1537,2416525,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +1538,2422826,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +1539,2423168,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +1540,2423878,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_124056-an7mepk5\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/an7mepk5\r\n",,terminal_output +1541,2424076,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +1542,2426157,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +1543,2495085,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1544,2499517,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",15021,0,"",python,selection_mouse +1545,2503442,"TERMINAL",0,0,"Total memory size: 23.3 GB, Output size: 0.9 GB, Temp size: 22.3 GB, Argument size: 0.9 GB, Host temp size: 0.0 GB.\r\nFLOPs: 7.842e+11, Bytes: 6.638e+11 (618.2 GB), Intensity: 1.2 FLOPs/byte\r\nStarting training from step 0...\r\n",,terminal_output +1546,2503931,"TERMINAL",0,0,"\r\nMemstats: After params initialized:\r\n\tUsing (GB) 1.21 / 38.7 (3.126615%) on cuda:0\r\nStep 1\r\n",,terminal_output +1547,2506762,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",17036,0,"",python,selection_command +1548,2510174,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11831,0,"",python,selection_command +1549,2510986,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11781,0,"",python,selection_command +1550,2511149,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11756,0,"",python,selection_command +1551,2511313,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11719,0,"",python,selection_command +1552,2512260,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11743,0,"\n ",python,content +1553,2512577,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11748,0,"a",python,content +1554,2512578,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11749,0,"",python,selection_keyboard +1555,2512678,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11749,0,"s",python,content +1556,2512680,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11750,0,"",python,selection_keyboard +1557,2512887,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11750,0,"s",python,content +1558,2512889,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11751,0,"",python,selection_keyboard +1559,2513188,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11751,0,"e",python,content +1560,2513190,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11752,0,"",python,selection_keyboard +1561,2513308,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11752,0,"r",python,content +1562,2513309,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11753,0,"",python,selection_keyboard +1563,2513683,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11753,0,"t",python,content +1564,2513685,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11754,0,"",python,selection_keyboard +1565,2515358,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11754,0," num_devices == 1, ""This script is not meant to be run on multiple devices.""",python,content +1566,2515928,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11829,0,"",python,selection_command +1567,2519556,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",11918,0,"",python,selection_command +1568,2529805,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",13481,0,"",python,selection_command +1569,2530274,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",13505,0,"",python,selection_command +1570,2530553,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",13547,0,"",python,selection_command +1571,2530820,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",13658,0,"",python,selection_command +1572,2531063,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",17476,0,"",python,selection_command +1573,2531244,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",17753,0,"",python,selection_command +1574,2534466,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19383,0,"",python,selection_command +1575,2534680,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19393,0,"",python,selection_command +1576,2538564,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19383,0,"",python,selection_command +1577,2543934,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,0,"",python,selection_mouse +1578,2544112,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,13,"= (\n {",python,selection_mouse +1579,2544142,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,33,"= (\n {\n ""videos",python,selection_mouse +1580,2544191,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,93,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n v",python,selection_mouse +1581,2544192,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,138,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )",python,selection_mouse +1582,2544192,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,148,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }",python,selection_mouse +1583,2544206,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,159,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n fo",python,selection_mouse +1584,2544219,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,157,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n ",python,selection_mouse +1585,2544233,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,155,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n ",python,selection_mouse +1586,2544251,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,187,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in train_iterator\n ",python,selection_mouse +1587,2544278,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,185,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in train_iterator\n ",python,selection_mouse +1588,2544288,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,190,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in train_iterator\n )\n",python,selection_mouse +1589,2544471,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,191,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in train_iterator\n )\n ",python,selection_mouse +1590,2544522,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,192,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in train_iterator\n )\n ",python,selection_mouse +1591,2544544,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,187,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in train_iterator\n ",python,selection_mouse +1592,2544609,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,188,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in train_iterator\n ",python,selection_mouse +1593,2544663,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,189,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in train_iterator\n )",python,selection_mouse +1594,2544759,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,156,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n ",python,selection_mouse +1595,2544773,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,157,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n ",python,selection_mouse +1596,2544790,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,158,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n f",python,selection_mouse +1597,2544814,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,159,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n fo",python,selection_mouse +1598,2544832,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,161,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for ",python,selection_mouse +1599,2544874,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,162,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for e",python,selection_mouse +1600,2545009,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,163,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for el",python,selection_mouse +1601,2545043,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,164,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for ele",python,selection_mouse +1602,2545077,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,165,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem",python,selection_mouse +1603,2545113,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,166,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem ",python,selection_mouse +1604,2545254,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,189,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in train_iterator\n )",python,selection_mouse +1605,2545353,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,199,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in train_iterator\n )\n datal",python,selection_mouse +1606,2545407,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,189,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in train_iterator\n )",python,selection_mouse +1607,2545720,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,163,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for el",python,selection_mouse +1608,2545737,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,164,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for ele",python,selection_mouse +1609,2545772,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,165,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem",python,selection_mouse +1610,2545812,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,166,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem ",python,selection_mouse +1611,2545853,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,167,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem i",python,selection_mouse +1612,2545971,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,168,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in",python,selection_mouse +1613,2546164,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,169,"= (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem\n )\n }\n for elem in ",python,selection_mouse +1614,2547156,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,169,"",python,content +1615,2548088,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19389,0,"",python,selection_command +1616,2549837,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19375,0,"= ",python,content +1617,2549841,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19377,0,"",python,selection_command +1618,2551759,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19376,0,"",python,selection_command +1619,2551926,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19396,0,"",python,selection_command +1620,2552319,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19392,6,"",python,content +1621,2552327,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19396,0,"",python,selection_command +1622,2552466,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19358,0,"",python,selection_command +1623,2553312,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19396,0,"",python,selection_command +1624,2553541,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19422,0,"",python,selection_command +1625,2553878,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19443,0,"",python,selection_command +1626,2554548,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19465,0,"val_iterator",python,content +1627,2554548,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19464,1,"",python,content +1628,2556106,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19481,0,"",python,selection_command +1629,2556709,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,80,"",python,content +1630,2556713,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19497,0,"",python,selection_command +1631,2557118,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,71,"",python,content +1632,2557123,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19489,0,"",python,selection_command +1633,2557551,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,51,"",python,content +1634,2557558,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19485,0,"",python,selection_command +1635,2558166,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,10,"",python,content +1636,2558190,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19481,0,"",python,selection_command +1637,2558571,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19443,0,"",python,selection_command +1638,2558831,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19422,0,"",python,selection_command +1639,2559808,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19443,0,"",python,selection_command +1640,2560003,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19481,0,"",python,selection_command +1641,2562748,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1642,2563012,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",18997,0,"",python,selection_mouse +1643,2580924,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1644,2583617,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19370,0,"",python,selection_mouse +1645,2583778,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19358,16,"dataloader_train",python,selection_mouse +1646,2714580,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3536670.8 task 0: running\r\n",,terminal_output +1647,2714831,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.8\r\nsrun: forcing job termination\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3536670.8 ON hkn0401 CANCELLED AT 2025-10-02T12:45:47 ***\r\n",,terminal_output +1648,2714957,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.8\r\nsrun: job abort in progress\r\n",,terminal_output +1649,2715107,"TERMINAL",0,0,"]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +1650,2715488,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +1651,2715740,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --log_interval 1 \\r\n --val_interval 2 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +1652,2715869,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +1653,2715950,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +1654,2719530,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +1655,2726102,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +1656,2726440,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +1657,2727187,"TERMINAL",0,0,"wandb: creating run\r\n",,terminal_output +1658,2727286,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_124559-6ee9v61v\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/6ee9v61v\r\n",,terminal_output +1659,2727424,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +1660,2729500,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +1661,2733383,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 718, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 559, in main\r\n first_batch = next(dataloader_train)\r\n ^^^^^^^^^^^^^^^^^^^^^^\r\nTypeError: 'DataLoader' object is not an iterator\r\n",,terminal_output +1662,2734117,"TERMINAL",0,0,"wandb: \r\nwandb: 🚀 View run coinrun-dyn-dev-3536670 at: https://wandb.ai/instant-uv/jafar/runs/6ee9v61v\r\nwandb: Find logs at: ../../../../../hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_124559-6ee9v61v/logs\r\n",,terminal_output +1663,2739242,"TERMINAL",0,0,"W1002 12:46:12.423430 1479842 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugonly job_name: ""jax_worker"": UNAVAILABLE: Cancelling all calls\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_message:""Cancelling all calls"", grpc_status:14}\r\n",,terminal_output +1664,2740189,"TERMINAL",0,0,"srun: error: hkn0401: task 0: Exited with exit code 1\r\n]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +1665,2742788,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1666,2742789,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,0,"",python,selection_command +1667,2744192,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19547,0,"",python,selection_mouse +1668,2745553,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1669,2745554,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19385,0,"",python,selection_mouse +1670,2746249,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1671,2747125,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",13553,0,"",python,selection_mouse +1672,2747340,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6173,0,"",python,selection_command +1673,2748269,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6253,0,"",python,selection_mouse +1674,2752571,"jasmine/utils/dataloader_torch.py",0,0,"",python,tab +1675,2756919,"jasmine/utils/dataloader_torch.py",899,0,"",python,selection_mouse +1676,2757076,"jasmine/utils/dataloader_torch.py",895,14,"get_dataloader",python,selection_mouse +1677,2758244,"jasmine/utils/dataloader_torch.py",1003,0,"",python,selection_mouse +1678,2758414,"jasmine/utils/dataloader_torch.py",999,10,"DataLoader",python,selection_mouse +1679,2760156,"jasmine/utils/dataloader_torch.py",1103,0,"",python,selection_mouse +1680,2760298,"jasmine/utils/dataloader_torch.py",1100,10,"collate_fn",python,selection_mouse +1681,2767743,"jasmine/utils/dataloader_torch.py",1070,0,"",python,selection_mouse +1682,2767873,"jasmine/utils/dataloader_torch.py",1067,7,"shuffle",python,selection_mouse +1683,2768413,"jasmine/utils/dataloader_torch.py",1075,0,"",python,selection_mouse +1684,2768553,"jasmine/utils/dataloader_torch.py",1075,4,"True",python,selection_mouse +1685,2768791,"jasmine/utils/dataloader_torch.py",1076,0,"",python,selection_mouse +1686,2769133,"jasmine/utils/dataloader_torch.py",1042,0,"",python,selection_mouse +1687,2769267,"jasmine/utils/dataloader_torch.py",1036,10,"batch_size",python,selection_mouse +1688,2769850,"jasmine/utils/dataloader_torch.py",1052,0,"",python,selection_mouse +1689,2769996,"jasmine/utils/dataloader_torch.py",1047,10,"batch_size",python,selection_mouse +1690,2770587,"jasmine/utils/dataloader_torch.py",1022,0,"",python,selection_mouse +1691,2770741,"jasmine/utils/dataloader_torch.py",1019,7,"dataset",python,selection_mouse +1692,2771064,"jasmine/utils/dataloader_torch.py",1040,0,"",python,selection_mouse +1693,2771212,"jasmine/utils/dataloader_torch.py",1036,10,"batch_size",python,selection_mouse +1694,2771709,"jasmine/utils/dataloader_torch.py",1049,0,"",python,selection_mouse +1695,2771867,"jasmine/utils/dataloader_torch.py",1047,10,"batch_size",python,selection_mouse +1696,2772298,"jasmine/utils/dataloader_torch.py",1074,0,"",python,selection_mouse +1697,2772462,"jasmine/utils/dataloader_torch.py",1067,7,"shuffle",python,selection_mouse +1698,2772936,"jasmine/utils/dataloader_torch.py",1077,0,"",python,selection_mouse +1699,2773092,"jasmine/utils/dataloader_torch.py",1075,4,"True",python,selection_mouse +1700,2773477,"jasmine/utils/dataloader_torch.py",1094,0,"",python,selection_mouse +1701,2773657,"jasmine/utils/dataloader_torch.py",1089,10,"collate_fn",python,selection_mouse +1702,2774062,"jasmine/utils/dataloader_torch.py",1105,0,"",python,selection_mouse +1703,2774177,"jasmine/utils/dataloader_torch.py",1100,10,"collate_fn",python,selection_mouse +1704,2774678,"jasmine/utils/dataloader_torch.py",1096,0,"",python,selection_mouse +1705,2774920,"jasmine/utils/dataloader_torch.py",1089,10,"collate_fn",python,selection_mouse +1706,2775333,"jasmine/utils/dataloader_torch.py",1101,0,"",python,selection_mouse +1707,2775454,"jasmine/utils/dataloader_torch.py",1100,10,"collate_fn",python,selection_mouse +1708,2775969,"jasmine/utils/dataloader_torch.py",1075,0,"",python,selection_mouse +1709,2776105,"jasmine/utils/dataloader_torch.py",1075,4,"True",python,selection_mouse +1710,2776500,"jasmine/utils/dataloader_torch.py",1073,0,"",python,selection_mouse +1711,2776601,"jasmine/utils/dataloader_torch.py",1072,1,"l",python,selection_mouse +1712,2776766,"jasmine/utils/dataloader_torch.py",1072,0,"",python,selection_mouse +1713,2791224,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1714,2791597,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1715,2792875,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1716,2793109,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1717,2793482,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1718,2793740,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1719,2794026,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1720,2794317,"jasmine/utils/dataloader_torch.py",797,0,"",python,selection_mouse +1721,2794327,"jasmine/utils/dataloader_torch.py",796,0,"",python,selection_command +1722,2794345,"jasmine/utils/dataloader_torch.py",796,1,":",python,selection_mouse +1723,2794346,"jasmine/utils/dataloader_torch.py",797,0,"",python,selection_command +1724,2794648,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1725,2794911,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1726,2795193,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1727,2795498,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1728,2795814,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1729,2796106,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1730,2796461,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1731,2796768,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1732,2797160,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1733,2797589,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1734,2798001,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1735,2798426,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1736,2798761,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1737,2799331,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1738,2799779,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1739,2800142,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1740,2801034,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1741,2801401,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1742,2801764,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1743,2802089,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1744,2802407,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_mouse +1745,2802710,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_mouse +1746,2802828,"jasmine/utils/dataloader_torch.py",773,1,"\n",python,selection_mouse +1747,2807592,"jasmine/utils/dataloader_torch.py",1005,0,"",python,selection_mouse +1748,2809735,".venv/lib64/python3.12/site-packages/torch/utils/data/dataloader.py",0,0,"# mypy: allow-untyped-defs\nr""""""Definition of the DataLoader and associated iterators that subclass _BaseDataLoaderIter.\n\nTo support these two classes, in `./_utils` we define many utility methods and\nfunctions to be run in multiprocessing. E.g., the data loading worker loop is\nin `./_utils/worker.py`.\n""""""\nfrom __future__ import annotations\n\nimport functools\nimport itertools\nimport logging\nimport multiprocessing as python_multiprocessing\nimport os\nimport queue\nimport threading\nimport warnings\nfrom typing import Any, Callable, Generic, Optional, TYPE_CHECKING, TypeVar, Union\nfrom typing_extensions import Self\n\nimport torch\nimport torch.distributed as dist\nimport torch.utils.data.graph_settings\nfrom torch._utils import ExceptionWrapper\nfrom torch.utils.data import _utils\nfrom torch.utils.data.datapipes.datapipe import (\n _IterDataPipeSerializationWrapper,\n _MapDataPipeSerializationWrapper,\n IterDataPipe,\n MapDataPipe,\n)\nfrom torch.utils.data.dataset import Dataset, IterableDataset\nfrom torch.utils.data.sampler import (\n BatchSampler,\n RandomSampler,\n Sampler,\n SequentialSampler,\n)\n\n\nif TYPE_CHECKING:\n from collections.abc import Iterable\n\n__all__ = [\n ""DataLoader"",\n ""get_worker_info"",\n ""default_collate"",\n ""default_convert"",\n]\n\n\n_T = TypeVar(""_T"")\n_T_co = TypeVar(""_T_co"", covariant=True)\n_worker_init_fn_t = Callable[[int], None]\n\n# Ideally we would parameterize `DataLoader` by the return type of `collate_fn`, but there is currently no way to have that\n# type parameter set to a default value if the user doesn't pass in a custom 'collate_fn'.\n# See https://github.com/python/mypy/issues/3737.\n_collate_fn_t = Callable[[list[_T]], Any]\n\n\n# These functions used to be defined in this file. However, it was moved to\n# _utils/collate.py. Although it is rather hard to access this from user land\n# (one has to explicitly directly `import torch.utils.data.dataloader`), there\n# probably is user code out there using it. This aliasing maintains BC in this\n# aspect.\ndefault_collate: _collate_fn_t = _utils.collate.default_collate\ndefault_convert = _utils.collate.default_convert\n\nget_worker_info = _utils.worker.get_worker_info\n\nlogger = logging.getLogger(__name__)\n\n\nclass _DatasetKind:\n Map = 0\n Iterable = 1\n\n @staticmethod\n def create_fetcher(kind, dataset, auto_collation, collate_fn, drop_last):\n if kind == _DatasetKind.Map:\n return _utils.fetch._MapDatasetFetcher(\n dataset, auto_collation, collate_fn, drop_last\n )\n else:\n return _utils.fetch._IterableDatasetFetcher(\n dataset, auto_collation, collate_fn, drop_last\n )\n\n\nclass _InfiniteConstantSampler(Sampler):\n r""""""Analogous to ``itertools.repeat(None, None)``.\n\n Used as sampler for :class:`~torch.utils.data.IterableDataset`.\n """"""\n\n def __iter__(self):\n while True:\n yield None\n\n\ndef _get_distributed_settings():\n if dist.is_available() and dist.is_initialized():\n return dist.get_world_size(), dist.get_rank()\n else:\n return 1, 0\n\n\ndef _sharding_worker_init_fn(worker_init_fn, world_size, rank_id, worker_id):\n global_worker_id = worker_id\n info = torch.utils.data.get_worker_info()\n assert info is not None\n total_workers = info.num_workers\n datapipe = info.dataset\n assert isinstance(datapipe, (IterDataPipe, MapDataPipe))\n # To distribute elements across distributed process evenly, we should shard data on distributed\n # processes first then shard on worker processes\n total_workers *= world_size\n global_worker_id = global_worker_id * world_size + rank_id\n # For BC, use default SHARDING_PRIORITIES\n torch.utils.data.graph_settings.apply_sharding(\n datapipe, total_workers, global_worker_id\n )\n if worker_init_fn is not None:\n worker_init_fn(worker_id)\n\n\ndef _share_dist_seed(generator, pg):\n _shared_seed = torch.empty((), dtype=torch.int64).random_(generator=generator)\n if isinstance(pg, dist.ProcessGroup):\n dist.broadcast(_shared_seed, src=0, group=pg)\n return _shared_seed.item()\n\n\nclass DataLoader(Generic[_T_co]):\n r""""""\n Data loader combines a dataset and a sampler, and provides an iterable over the given dataset.\n\n The :class:`~torch.utils.data.DataLoader` supports both map-style and\n iterable-style datasets with single- or multi-process loading, customizing\n loading order and optional automatic batching (collation) and memory pinning.\n\n See :py:mod:`torch.utils.data` documentation page for more details.\n\n Args:\n dataset (Dataset): dataset from which to load the data.\n batch_size (int, optional): how many samples per batch to load\n (default: ``1``).\n shuffle (bool, optional): set to ``True`` to have the data reshuffled\n at every epoch (default: ``False``).\n sampler (Sampler or Iterable, optional): defines the strategy to draw\n samples from the dataset. Can be any ``Iterable`` with ``__len__``\n implemented. If specified, :attr:`shuffle` must not be specified.\n batch_sampler (Sampler or Iterable, optional): like :attr:`sampler`, but\n returns a batch of indices at a time. Mutually exclusive with\n :attr:`batch_size`, :attr:`shuffle`, :attr:`sampler`,\n and :attr:`drop_last`.\n num_workers (int, optional): how many subprocesses to use for data\n loading. ``0`` means that the data will be loaded in the main process.\n (default: ``0``)\n collate_fn (Callable, optional): merges a list of samples to form a\n mini-batch of Tensor(s). Used when using batched loading from a\n map-style dataset.\n pin_memory (bool, optional): If ``True``, the data loader will copy Tensors\n into device/CUDA pinned memory before returning them. If your data elements\n are a custom type, or your :attr:`collate_fn` returns a batch that is a custom type,\n see the example below.\n drop_last (bool, optional): set to ``True`` to drop the last incomplete batch,\n if the dataset size is not divisible by the batch size. If ``False`` and\n the size of dataset is not divisible by the batch size, then the last batch\n will be smaller. (default: ``False``)\n timeout (numeric, optional): if positive, the timeout value for collecting a batch\n from workers. Should always be non-negative. (default: ``0``)\n worker_init_fn (Callable, optional): If not ``None``, this will be called on each\n worker subprocess with the worker id (an int in ``[0, num_workers - 1]``) as\n input, after seeding and before data loading. (default: ``None``)\n multiprocessing_context (str or multiprocessing.context.BaseContext, optional): If\n ``None``, the default\n `multiprocessing context `_ # noqa: D401\n of your operating system will\n be used. (default: ``None``)\n generator (torch.Generator, optional): If not ``None``, this RNG will be used\n by RandomSampler to generate random indexes and multiprocessing to generate\n ``base_seed`` for workers. (default: ``None``)\n prefetch_factor (int, optional, keyword-only arg): Number of batches loaded\n in advance by each worker. ``2`` means there will be a total of\n 2 * num_workers batches prefetched across all workers. (default value depends\n on the set value for num_workers. If value of num_workers=0 default is ``None``.\n Otherwise, if value of ``num_workers > 0`` default is ``2``).\n persistent_workers (bool, optional): If ``True``, the data loader will not shut down\n the worker processes after a dataset has been consumed once. This allows to\n maintain the workers `Dataset` instances alive. (default: ``False``)\n pin_memory_device (str, optional): the device to :attr:`pin_memory` on if ``pin_memory`` is\n ``True``. If not given, the current :ref:`accelerator` will be the\n default. This argument is discouraged and subject to deprecated.\n in_order (bool, optional): If ``False``, the data loader will not enforce that batches\n are returned in a first-in, first-out order. Only applies when ``num_workers > 0``. (default: ``True``)\n\n\n .. warning:: If the ``spawn`` start method is used, :attr:`worker_init_fn`\n cannot be an unpicklable object, e.g., a lambda function. See\n :ref:`multiprocessing-best-practices` on more details related\n to multiprocessing in PyTorch.\n\n .. warning:: ``len(dataloader)`` heuristic is based on the length of the sampler used.\n When :attr:`dataset` is an :class:`~torch.utils.data.IterableDataset`,\n it instead returns an estimate based on ``len(dataset) / batch_size``, with proper\n rounding depending on :attr:`drop_last`, regardless of multi-process loading\n configurations. This represents the best guess PyTorch can make because PyTorch\n trusts user :attr:`dataset` code in correctly handling multi-process\n loading to avoid duplicate data.\n\n However, if sharding results in multiple workers having incomplete last batches,\n this estimate can still be inaccurate, because (1) an otherwise complete batch can\n be broken into multiple ones and (2) more than one batch worth of samples can be\n dropped when :attr:`drop_last` is set. Unfortunately, PyTorch can not detect such\n cases in general.\n\n See `Dataset Types`_ for more details on these two types of datasets and how\n :class:`~torch.utils.data.IterableDataset` interacts with\n `Multi-process data loading`_.\n\n .. warning:: See :ref:`reproducibility`, and :ref:`dataloader-workers-random-seed`, and\n :ref:`data-loading-randomness` notes for random seed related questions.\n\n .. warning:: Setting `in_order` to `False` can harm reproducibility and may lead to a skewed data\n distribution being fed to the trainer in cases with imbalanced data.\n """"""\n\n dataset: Dataset[_T_co]\n batch_size: Optional[int]\n num_workers: int\n pin_memory: bool\n drop_last: bool\n timeout: float\n sampler: Union[Sampler, Iterable]\n pin_memory_device: str\n prefetch_factor: Optional[int]\n _iterator: Optional[_BaseDataLoaderIter]\n __initialized = False\n\n def __init__(\n self,\n dataset: Dataset[_T_co],\n batch_size: Optional[int] = 1,\n shuffle: Optional[bool] = None,\n sampler: Union[Sampler, Iterable, None] = None,\n batch_sampler: Union[Sampler[list], Iterable[list], None] = None,\n num_workers: int = 0,\n collate_fn: Optional[_collate_fn_t] = None,\n pin_memory: bool = False,\n drop_last: bool = False,\n timeout: float = 0,\n worker_init_fn: Optional[_worker_init_fn_t] = None,\n multiprocessing_context=None,\n generator=None,\n *,\n prefetch_factor: Optional[int] = None,\n persistent_workers: bool = False,\n pin_memory_device: str = """",\n in_order: bool = True,\n ) -> None:\n torch._C._log_api_usage_once(""python.data_loader"")\n\n if num_workers < 0:\n raise ValueError(\n ""num_workers option should be non-negative; ""\n ""use num_workers=0 to disable multiprocessing.""\n )\n\n if timeout < 0:\n raise ValueError(""timeout option should be non-negative"")\n\n if num_workers == 0 and prefetch_factor is not None:\n raise ValueError(\n ""prefetch_factor option could only be specified in multiprocessing.""\n ""let num_workers > 0 to enable multiprocessing, otherwise set prefetch_factor to None.""\n )\n elif num_workers > 0 and prefetch_factor is None:\n prefetch_factor = 2\n elif prefetch_factor is not None and prefetch_factor < 0:\n raise ValueError(""prefetch_factor option should be non-negative"")\n\n if persistent_workers and num_workers == 0:\n raise ValueError(""persistent_workers option needs num_workers > 0"")\n\n self.dataset = dataset\n self.num_workers = num_workers\n self.prefetch_factor = prefetch_factor\n self.pin_memory = pin_memory\n self.pin_memory_device = pin_memory_device\n self.timeout = timeout\n self.worker_init_fn = worker_init_fn\n self.multiprocessing_context = multiprocessing_context\n self.in_order = in_order\n\n # Adds forward compatibilities so classic DataLoader can work with DataPipes:\n # _DataPipeSerializationWrapper container makes it easier to serialize without redefining pickler\n if isinstance(self.dataset, IterDataPipe):\n self.dataset = _IterDataPipeSerializationWrapper(self.dataset)\n elif isinstance(self.dataset, MapDataPipe):\n self.dataset = _MapDataPipeSerializationWrapper(self.dataset)\n\n # Arg-check dataset related before checking samplers because we want to\n # tell users that iterable-style datasets are incompatible with custom\n # samplers first, so that they don't learn that this combo doesn't work\n # after spending time fixing the custom sampler errors.\n if isinstance(dataset, IterableDataset):\n self._dataset_kind = _DatasetKind.Iterable\n # NOTE [ Custom Samplers and IterableDataset ]\n #\n # `IterableDataset` does not support custom `batch_sampler` or\n # `sampler` since the key is irrelevant (unless we support\n # generator-style dataset one day...).\n #\n # For `sampler`, we always create a dummy sampler. This is an\n # infinite sampler even when the dataset may have an implemented\n # finite `__len__` because in multi-process data loading, naive\n # settings will return duplicated data (which may be desired), and\n # thus using a sampler with length matching that of dataset will\n # cause data lost (you may have duplicates of the first couple\n # batches, but never see anything afterwards). Therefore,\n # `Iterabledataset` always uses an infinite sampler, an instance of\n # `_InfiniteConstantSampler` defined above.\n #\n # A custom `batch_sampler` essentially only controls the batch size.\n # However, it is unclear how useful it would be since an iterable-style\n # dataset can handle that within itself. Moreover, it is pointless\n # in multi-process data loading as the assignment order of batches\n # to workers is an implementation detail so users can not control\n # how to batchify each worker's iterable. Thus, we disable this\n # option. If this turns out to be useful in future, we can re-enable\n # this, and support custom samplers that specify the assignments to\n # specific workers.\n if isinstance(dataset, IterDataPipe):\n if shuffle is not None:\n dataset = torch.utils.data.graph_settings.apply_shuffle_settings(\n dataset, shuffle=shuffle\n )\n # We cannot check `shuffle is not None` here, since previously `shuffle=False` was the default.\n elif shuffle not in {False, None}:\n raise ValueError(\n f""DataLoader with IterableDataset: expected unspecified shuffle option, but got shuffle={shuffle}""\n )\n\n if sampler is not None:\n # See NOTE [ Custom Samplers and IterableDataset ]\n raise ValueError(\n f""DataLoader with IterableDataset: expected unspecified sampler option, but got sampler={sampler}""\n )\n elif batch_sampler is not None:\n # See NOTE [ Custom Samplers and IterableDataset ]\n raise ValueError(\n ""DataLoader with IterableDataset: expected unspecified ""\n f""batch_sampler option, but got batch_sampler={batch_sampler}""\n )\n else:\n shuffle = bool(shuffle)\n self._dataset_kind = _DatasetKind.Map\n\n if sampler is not None and shuffle:\n raise ValueError(""sampler option is mutually exclusive with shuffle"")\n\n if batch_sampler is not None:\n # auto_collation with custom batch_sampler\n if batch_size != 1 or shuffle or sampler is not None or drop_last:\n raise ValueError(\n ""batch_sampler option is mutually exclusive ""\n ""with batch_size, shuffle, sampler, and ""\n ""drop_last""\n )\n batch_size = None\n drop_last = False\n elif batch_size is None:\n # no auto_collation\n if drop_last:\n raise ValueError(\n ""batch_size=None option disables auto-batching ""\n ""and is mutually exclusive with drop_last""\n )\n\n if sampler is None: # give default samplers\n if self._dataset_kind == _DatasetKind.Iterable:\n # See NOTE [ Custom Samplers and IterableDataset ]\n sampler = _InfiniteConstantSampler()\n else: # map-style\n if shuffle:\n sampler = RandomSampler(dataset, generator=generator) # type: ignore[arg-type]\n else:\n sampler = SequentialSampler(dataset) # type: ignore[arg-type]\n\n if batch_size is not None and batch_sampler is None:\n # auto_collation without custom batch_sampler\n batch_sampler = BatchSampler(sampler, batch_size, drop_last)\n\n self.batch_size = batch_size\n self.drop_last = drop_last\n self.sampler = sampler\n self.batch_sampler = batch_sampler\n self.generator = generator\n\n if collate_fn is None:\n if self._auto_collation:\n collate_fn = _utils.collate.default_collate\n else:\n collate_fn = _utils.collate.default_convert\n\n self.collate_fn = collate_fn\n self.persistent_workers = persistent_workers\n\n self.__initialized = True\n self._IterableDataset_len_called = (\n None # See NOTE [ IterableDataset and __len__ ]\n )\n\n self._iterator = None\n\n self.check_worker_number_rationality()\n\n torch.set_vital(""Dataloader"", ""enabled"", ""True"") # type: ignore[attr-defined]\n\n def _get_iterator(self) -> _BaseDataLoaderIter:\n if self.num_workers == 0:\n return _SingleProcessDataLoaderIter(self)\n else:\n self.check_worker_number_rationality()\n return _MultiProcessingDataLoaderIter(self)\n\n @property\n def multiprocessing_context(self):\n return self.__multiprocessing_context\n\n @multiprocessing_context.setter\n def multiprocessing_context(self, multiprocessing_context):\n if multiprocessing_context is not None:\n if self.num_workers > 0:\n if isinstance(multiprocessing_context, str):\n valid_start_methods = torch.multiprocessing.get_all_start_methods()\n if multiprocessing_context not in valid_start_methods:\n raise ValueError(\n ""multiprocessing_context option ""\n f""should specify a valid start method in {valid_start_methods!r}, but got ""\n f""multiprocessing_context={multiprocessing_context!r}""\n )\n multiprocessing_context = torch.multiprocessing.get_context(\n multiprocessing_context\n )\n\n if not isinstance(\n multiprocessing_context, python_multiprocessing.context.BaseContext\n ):\n raise TypeError(\n ""multiprocessing_context option should be a valid context ""\n ""object or a string specifying the start method, but got ""\n f""multiprocessing_context={multiprocessing_context}""\n )\n else:\n raise ValueError(\n ""multiprocessing_context can only be used with ""\n ""multi-process loading (num_workers > 0), but got ""\n f""num_workers={self.num_workers}""\n )\n\n self.__multiprocessing_context = multiprocessing_context\n\n def __setattr__(self, attr, val):\n if self.__initialized and attr in (\n ""batch_size"",\n ""batch_sampler"",\n ""sampler"",\n ""drop_last"",\n ""dataset"",\n ""persistent_workers"",\n ):\n raise ValueError(\n f""{attr} attribute should not be set after {self.__class__.__name__} is initialized""\n )\n\n super().__setattr__(attr, val)\n\n def __iter__(self) -> _BaseDataLoaderIter:\n # When using a single worker the returned iterator should be\n # created everytime to avoid resetting its state\n # However, in the case of a multiple workers iterator\n # the iterator is only created once in the lifetime of the\n # DataLoader object so that workers can be reused\n if self.persistent_workers and self.num_workers > 0:\n if self._iterator is None:\n self._iterator = self._get_iterator()\n else:\n self._iterator._reset(self)\n return self._iterator\n else:\n return self._get_iterator()\n\n @property\n def _auto_collation(self):\n return self.batch_sampler is not None\n\n @property\n def _index_sampler(self):\n # The actual sampler used for generating indices for `_DatasetFetcher`\n # (see _utils/fetch.py) to read data at each time. This would be\n # `.batch_sampler` if in auto-collation mode, and `.sampler` otherwise.\n # We can't change `.sampler` and `.batch_sampler` attributes for BC\n # reasons.\n if self._auto_collation:\n return self.batch_sampler\n else:\n return self.sampler\n\n def __len__(self) -> int:\n if self._dataset_kind == _DatasetKind.Iterable:\n # NOTE [ IterableDataset and __len__ ]\n #\n # For `IterableDataset`, `__len__` could be inaccurate when one naively\n # does multi-processing data loading, since the samples will be duplicated.\n # However, no real use case should be actually using that behavior, so\n # it should count as a user error. We should generally trust user\n # code to do the proper thing (e.g., configure each replica differently\n # in `__iter__`), and give us the correct `__len__` if they choose to\n # implement it (this will still throw if the dataset does not implement\n # a `__len__`).\n #\n # To provide a further warning, we track if `__len__` was called on the\n # `DataLoader`, save the returned value in `self._len_called`, and warn\n # if the iterator ends up yielding more than this number of samples.\n\n # Cannot statically verify that dataset is Sized\n length = self._IterableDataset_len_called = len(self.dataset) # type: ignore[assignment, arg-type]\n if (\n self.batch_size is not None\n ): # IterableDataset doesn't allow custom sampler or batch_sampler\n from math import ceil\n\n if self.drop_last:\n length = length // self.batch_size\n else:\n length = ceil(length / self.batch_size)\n return length\n else:\n return len(self._index_sampler)\n\n def check_worker_number_rationality(self):\n # This function check whether the dataloader's worker number is rational based on\n # current system's resource. Current rule is that if the number of workers this\n # Dataloader will create is bigger than the number of logical cpus that is allowed to\n # use, than we will pop up a warning to let user pay attention.\n #\n # eg. If current system has 2 physical CPUs with 16 cores each. And each core support 2\n # threads, then the total logical cpus here is 2 * 16 * 2 = 64. Let's say current\n # DataLoader process can use half of them which is 32, then the rational max number of\n # worker that initiated from this process is 32.\n # Now, let's say the created DataLoader has num_works = 40, which is bigger than 32.\n # So the warning message is triggered to notify the user to lower the worker number if\n # necessary.\n #\n #\n # [Note] Please note that this function repects `cpuset` only when os.sched_getaffinity is\n # available (available in most of Linux system, but not OSX and Windows).\n # When os.sched_getaffinity is not available, os.cpu_count() is called instead, but\n # it doesn't repect cpuset.\n # We don't take threading into account since each worker process is single threaded\n # at this time.\n #\n # We don't set any threading flags (eg. OMP_NUM_THREADS, MKL_NUM_THREADS, etc)\n # other than `torch.set_num_threads` to 1 in the worker process, if the passing\n # in functions use 3rd party modules that rely on those threading flags to determine\n # how many thread to create (eg. numpy, etc), then it is caller's responsibility to\n # set those flags correctly.\n def _create_warning_msg(num_worker_suggest, num_worker_created, cpuset_checked):\n suggested_max_worker_msg = (\n (\n (\n ""Our suggested max number of worker in current system is {}{}, which is smaller ""\n ""than what this DataLoader is going to create.""\n ).format(\n num_worker_suggest,\n (\n """"\n if cpuset_checked\n else "" (`cpuset` is not taken into account)""\n ),\n )\n )\n if num_worker_suggest is not None\n else (\n ""DataLoader is not able to compute a suggested max number of worker in current system.""\n )\n )\n\n warn_msg = (\n f""This DataLoader will create {num_worker_created} worker processes in total. {suggested_max_worker_msg} ""\n ""Please be aware that excessive worker creation might get DataLoader running slow or even freeze, ""\n ""lower the worker number to avoid potential slowness/freeze if necessary.""\n )\n return warn_msg\n\n if not self.num_workers or self.num_workers == 0:\n return\n\n # try to compute a suggested max number of worker based on system's resource\n max_num_worker_suggest = None\n cpuset_checked = False\n if hasattr(os, ""sched_getaffinity""):\n try:\n max_num_worker_suggest = len(os.sched_getaffinity(0))\n cpuset_checked = True\n except Exception:\n pass\n if max_num_worker_suggest is None:\n # os.cpu_count() could return Optional[int]\n # get cpu count first and check None in order to satisfy mypy check\n cpu_count = os.cpu_count()\n if cpu_count is not None:\n max_num_worker_suggest = cpu_count\n\n if max_num_worker_suggest is None:\n warnings.warn(\n _create_warning_msg(\n max_num_worker_suggest, self.num_workers, cpuset_checked\n )\n )\n return\n\n if self.num_workers > max_num_worker_suggest:\n warnings.warn(\n _create_warning_msg(\n max_num_worker_suggest, self.num_workers, cpuset_checked\n )\n )\n\n\nclass _BaseDataLoaderIter:\n def __init__(self, loader: DataLoader) -> None:\n self._dataset = loader.dataset\n self._shared_seed = None\n self._pg = None\n if isinstance(self._dataset, IterDataPipe):\n if dist.is_available() and dist.is_initialized():\n self._pg = dist.new_group(backend=""gloo"")\n self._shared_seed = _share_dist_seed(loader.generator, self._pg)\n shared_rng = torch.Generator()\n shared_rng.manual_seed(self._shared_seed)\n self._dataset = torch.utils.data.graph_settings.apply_random_seed(\n self._dataset, shared_rng\n )\n self._dataset_kind = loader._dataset_kind\n self._IterableDataset_len_called = loader._IterableDataset_len_called\n self._auto_collation = loader._auto_collation\n self._drop_last = loader.drop_last\n self._index_sampler = loader._index_sampler\n self._num_workers = loader.num_workers\n ws, rank = _get_distributed_settings()\n self._world_size = ws\n self._rank = rank\n # If pin_memory_device not set, default behaviour is current accelerator.\n # If pin_memory_device is set but pin_memory is not set, the default\n # behaviour false.\n if len(loader.pin_memory_device) == 0:\n if loader.pin_memory and not torch.accelerator.is_available():\n warn_msg = (\n ""'pin_memory' argument is set as true but no accelerator is found, ""\n ""then device pinned memory won't be used.""\n )\n warnings.warn(warn_msg)\n\n self._pin_memory = loader.pin_memory and torch.accelerator.is_available()\n self._pin_memory_device = None\n # Currently, pin_memory would raise error on the MPS backend (see\n # https://github.com/pytorch/pytorch/issues/86060), so forcibly\n # disable pin_memory on MPS. Remove this restriction once pinned\n # memory allocation for MPS is fixed.\n if (\n self._pin_memory\n and (acc := torch.accelerator.current_accelerator()) is not None\n and acc.type == ""mps""\n ):\n self._pin_memory = False\n warn_msg = (\n ""'pin_memory' argument is set as true but not supported on MPS now, ""\n ""then device pinned memory won't be used.""\n )\n warnings.warn(warn_msg)\n else:\n if not loader.pin_memory:\n warn_msg = (\n ""'pin_memory_device' is set but 'pin_memory' argument is not set, ""\n ""then device pinned memory won't be used.""\n ""please set 'pin_memory' to true, if you need to use the device pin memory""\n )\n warnings.warn(warn_msg)\n\n self._pin_memory = loader.pin_memory\n self._pin_memory_device = loader.pin_memory_device\n self._timeout = loader.timeout\n self._collate_fn = loader.collate_fn\n self._sampler_iter = iter(self._index_sampler)\n self._base_seed = (\n torch.empty((), dtype=torch.int64)\n .random_(generator=loader.generator)\n .item()\n )\n self._persistent_workers = loader.persistent_workers\n self._num_yielded = 0\n self._profile_name = f""enumerate(DataLoader)#{self.__class__.__name__}.__next__""\n\n def __iter__(self) -> Self:\n return self\n\n def _reset(self, loader, first_iter=False):\n self._sampler_iter = iter(self._index_sampler)\n self._num_yielded = 0\n self._IterableDataset_len_called = loader._IterableDataset_len_called\n if isinstance(self._dataset, IterDataPipe):\n self._shared_seed = _share_dist_seed(loader.generator, self._pg)\n shared_rng = torch.Generator()\n shared_rng.manual_seed(self._shared_seed)\n self._dataset = torch.utils.data.graph_settings.apply_random_seed(\n self._dataset, shared_rng\n )\n\n def _next_index(self):\n return next(self._sampler_iter) # may raise StopIteration\n\n def _next_data(self):\n raise NotImplementedError\n\n def __next__(self) -> Any:\n with torch.autograd.profiler.record_function(self._profile_name):\n if self._sampler_iter is None:\n # TODO(https://github.com/pytorch/pytorch/issues/76750)\n self._reset() # type: ignore[call-arg]\n data = self._next_data()\n self._num_yielded += 1\n if (\n self._dataset_kind == _DatasetKind.Iterable\n and self._IterableDataset_len_called is not None\n and self._num_yielded > self._IterableDataset_len_called\n ):\n warn_msg = (\n f""Length of IterableDataset {self._dataset} was reported to be {self._IterableDataset_len_called}""\n f""(when accessing len(dataloader)), but {self._num_yielded} samples have been fetched. ""\n )\n if self._num_workers > 0:\n warn_msg += (\n ""For multiprocessing data-loading, this could be caused by not properly configuring the ""\n ""IterableDataset replica at each worker. Please see ""\n ""https://pytorch.org/docs/stable/data.html#torch.utils.data.IterableDataset for examples.""\n )\n warnings.warn(warn_msg)\n return data\n\n def __len__(self) -> int:\n return len(self._index_sampler)\n\n def __getstate__(self):\n # TODO: add limited pickling support for sharing an iterator\n # across multiple threads for HOGWILD.\n # Probably the best way to do this is by moving the sample pushing\n # to a separate thread and then just sharing the data queue\n # but signalling the end is tricky without a non-blocking API\n raise NotImplementedError(""{} cannot be pickled"", self.__class__.__name__)\n\n\nclass _SingleProcessDataLoaderIter(_BaseDataLoaderIter):\n def __init__(self, loader):\n super().__init__(loader)\n assert self._timeout == 0\n assert self._num_workers == 0\n\n # Adds forward compatibilities so classic DataLoader can work with DataPipes:\n # Taking care of distributed sharding\n if isinstance(self._dataset, (IterDataPipe, MapDataPipe)):\n # For BC, use default SHARDING_PRIORITIES\n torch.utils.data.graph_settings.apply_sharding(\n self._dataset, self._world_size, self._rank\n )\n\n self._dataset_fetcher = _DatasetKind.create_fetcher(\n self._dataset_kind,\n self._dataset,\n self._auto_collation,\n self._collate_fn,\n self._drop_last,\n )\n\n def _next_data(self):\n index = self._next_index() # may raise StopIteration\n data = self._dataset_fetcher.fetch(index) # may raise StopIteration\n if self._pin_memory:\n data = _utils.pin_memory.pin_memory(data, self._pin_memory_device)\n return data\n\n\nclass _MultiProcessingDataLoaderIter(_BaseDataLoaderIter):\n r""""""Iterates once over the DataLoader's dataset, as specified by the sampler.""""""\n\n # NOTE [ Data Loader Multiprocessing Shutdown Logic ]\n #\n # Preliminary:\n #\n # Our data model looks like this (queues are indicated with curly brackets):\n #\n # main process ||\n # | ||\n # {index_queue} ||\n # | ||\n # worker processes || DATA\n # | ||\n # {worker_result_queue} || FLOW\n # | ||\n # pin_memory_thread of main process || DIRECTION\n # | ||\n # {data_queue} ||\n # | ||\n # data output \/\n #\n # P.S. `worker_result_queue` and `pin_memory_thread` part may be omitted if\n # `pin_memory=False`.\n #\n #\n # Terminating multiprocessing logic requires very careful design. In\n # particular, we need to make sure that\n #\n # 1. The iterator gracefully exits the workers when its last reference is\n # gone or it is depleted.\n #\n # In this case, the workers should be gracefully exited because the\n # main process may still need to continue to run, and we want cleaning\n # up code in the workers to be executed (e.g., releasing GPU memory).\n # Naturally, we implement the shutdown logic in `__del__` of\n # DataLoaderIterator.\n #\n # We delay the discussion on the logic in this case until later.\n #\n # 2. The iterator exits the workers when the loader process and/or worker\n # processes exits normally or with error.\n #\n # We set all workers and `pin_memory_thread` to have `daemon=True`.\n #\n # You may ask, why can't we make the workers non-daemonic, and\n # gracefully exit using the same logic as we have in `__del__` when the\n # iterator gets deleted (see 1 above)?\n #\n # First of all, `__del__` is **not** guaranteed to be called when\n # interpreter exits. Even if it is called, by the time it executes,\n # many Python core library resources may already be freed, and even\n # simple things like acquiring an internal lock of a queue may hang.\n # Therefore, in this case, we actually need to prevent `__del__` from\n # being executed, and rely on the automatic termination of daemonic\n # children.\n #\n # Thus, we register an `atexit` hook that sets a global flag\n # `_utils.python_exit_status`. Since `atexit` hooks are executed in the\n # reverse order of registration, we are guaranteed that this flag is\n # set before library resources we use are freed (which, at least in\n # CPython, is done via an `atexit` handler defined in\n # `multiprocessing/util.py`\n # https://github.com/python/cpython/blob/c606624af8d4cb3b4a052fb263bb983b3f87585b/Lib/multiprocessing/util.py#L320-L362\n # registered when an object requiring this mechanism is first\n # created, e.g., `mp.Queue`\n # https://github.com/python/cpython/blob/c606624af8d4cb3b4a052fb263bb983b3f87585b/Lib/multiprocessing/context.py#L100-L103\n # https://github.com/python/cpython/blob/c606624af8d4cb3b4a052fb263bb983b3f87585b/Lib/multiprocessing/queues.py#L29\n # )\n #\n # So in `__del__`, we check if `_utils.python_exit_status` is set or\n # `None` (freed), and perform no-op if so.\n #\n # However, simply letting library clean-up codes run can also be bad,\n # because such codes (i.e., `multiprocessing.util._exit_function()`)\n # include join putting threads for `mp.Queue`, which can be blocking.\n # Hence, the main process putting threads are called with\n # `cancel_join_thread` at creation. See later section\n # [ 3b. A process won't hang when putting into a queue; ]\n # for more details.\n #\n # Here are two example cases where library clean-up codes can run\n # before `__del__` is called:\n #\n # 1. If we hold onto a reference to the iterator, it more often\n # than not tries to do `multiprocessing` library cleaning before\n # clearing the alive referenced objects (https://github.com/pytorch/pytorch/issues/48666)\n # and thus prevents our cleaning-up code to run first.\n #\n # 2. A similar issue araises when a `DataLoader` is used in a subprocess.\n # When a process ends, it shuts the all its daemonic children\n # down with a SIGTERM (instead of joining them without a timeout).\n # Simiarly for threads, but by a different mechanism. This fact,\n # together with a few implementation details of multiprocessing, forces\n # us to make workers daemonic. All of our problems arise when a\n # DataLoader is used in a subprocess, and are caused by multiprocessing\n # code which looks more or less like this:\n #\n # try:\n # your_function_using_a_dataloader()\n # finally:\n # multiprocessing.util._exit_function()\n #\n # The joining/termination mentioned above happens inside\n # `_exit_function()`. Now, if `your_function_using_a_dataloader()`\n # throws, the stack trace stored in the exception will prevent the\n # frame which uses `DataLoaderIter` to be freed. If the frame has any\n # reference to the `DataLoaderIter` (e.g., in a method of the iter),\n # its `__del__`, which starts the shutdown procedure, will not be\n # called. That, in turn, means that workers aren't notified. Attempting\n # to join in `_exit_function` will then result in a hang.\n #\n # For context, `_exit_function` is also registered as an `atexit` call.\n # So it is unclear to me (@ssnl) why this is needed in a finally block.\n # The code dates back to 2008 and there is no comment on the original\n # PEP 371 or patch https://bugs.python.org/issue3050 (containing both\n # the finally block and the `atexit` registration) that explains this.\n #\n #\n # Finally, another choice is to just shutdown workers with logic in 1\n # above whenever we see an error in `next`. This isn't ideal because\n # a. It prevents users from using try-catch to resume data loading.\n # b. It doesn't prevent hanging if users have references to the\n # iterator.\n #\n # 3. All processes exit if any of them die unexpectedly by fatal signals.\n #\n # As shown above, the workers are set as daemonic children of the main\n # process. However, automatic cleaning-up of such child processes only\n # happens if the parent process exits gracefully (e.g., not via fatal\n # signals like SIGKILL). So we must ensure that each process will exit\n # even the process that should send/receive data to/from it were\n # killed, i.e.,\n #\n # a. A process won't hang when getting from a queue.\n #\n # Even with carefully designed data dependencies (i.e., a `put()`\n # always corresponding to a `get()`), hanging on `get()` can still\n # happen when data in queue is corrupted (e.g., due to\n # `cancel_join_thread` or unexpected exit).\n #\n # For child exit, we set a timeout whenever we try to get data\n # from `data_queue`, and check the workers' status on each timeout\n # and error.\n # See `_DataLoaderiter._get_batch()` and\n # `_DataLoaderiter._try_get_data()` for details.\n #\n # Additionally, for child exit on non-Windows platforms, we also\n # register a SIGCHLD handler (which is supported on Windows) on\n # the main process, which checks if any of the workers fail in the\n # (Python) handler. This is more efficient and faster in detecting\n # worker failures, compared to only using the above mechanism.\n # See `DataLoader.cpp` and `_utils/signal_handling.py` for details.\n #\n # For `.get()` calls where the sender(s) is not the workers, we\n # guard them with timeouts, and check the status of the sender\n # when timeout happens:\n # + in the workers, the `_utils.worker.ManagerWatchdog` class\n # checks the status of the main process.\n # + if `pin_memory=True`, when getting from `pin_memory_thread`,\n # check `pin_memory_thread` status periodically until `.get()`\n # returns or see that `pin_memory_thread` died.\n #\n # b. A process won't hang when putting into a queue;\n #\n # We use `mp.Queue` which has a separate background thread to put\n # objects from an unbounded buffer array. The background thread is\n # daemonic and usually automatically joined when the process\n # *exits*.\n #\n # In case that the receiver has ended abruptly while\n # reading from the pipe, the join will hang forever. The usual\n # solution for this in Python is calling `q.cancel_join_thread`,\n # which prevents automatically joining it when finalizing\n # (exiting).\n #\n # Nonetheless, `cancel_join_thread` must only be called when the\n # queue is **not** going to be read from or write into by another\n # process, because it may hold onto a lock or leave corrupted data\n # in the queue, leading other readers/writers to hang.\n #\n # Hence,\n # + For worker processes, we only do so (for their output\n # queues, i.e., `worker_result_queue`) before exiting.\n # + For `pin_memory_thread`, its output queue `data_queue` is a\n # `queue.Queue` that does blocking `put` if the queue is full.\n # So there is no above problem, but as a result, in\n # `_pin_memory_loop`, we do need to wrap the `put` in a loop\n # that breaks not only upon success, but also when the main\n # process stops reading, i.e., is shutting down.\n # + For loader process, we `cancel_join_thread()` for all\n # `_index_queues` because the whole purpose of workers and\n # `pin_memory_thread` is to serve the loader process. If\n # loader process is already exiting, we don't really care if\n # the queues are corrupted.\n #\n #\n # Now let's get back to 1:\n # how we gracefully exit the workers when the last reference to the\n # iterator is gone.\n #\n # To achieve this, we implement the following logic along with the design\n # choices mentioned above:\n #\n # `workers_done_event`:\n # A `multiprocessing.Event` shared among the main process and all worker\n # processes. This is used to signal the workers that the iterator is\n # shutting down. After it is set, they will not send processed data to\n # queues anymore, and only wait for the final `None` before exiting.\n # `done_event` isn't strictly needed. I.e., we can just check for `None`\n # from the input queue, but it allows us to skip wasting resources\n # processing data if we are already shutting down.\n #\n # `pin_memory_thread_done_event`:\n # A `threading.Event` for a similar purpose to that of\n # `workers_done_event`, but is for the `pin_memory_thread`. The reason\n # that separate events are needed is that `pin_memory_thread` reads from\n # the output queue of the workers. But the workers, upon seeing that\n # `workers_done_event` is set, only wants to see the final `None`, and is\n # not required to flush all data in the output queue (e.g., it may call\n # `cancel_join_thread` on that queue if its `IterableDataset` iterator\n # happens to exhaust coincidentally, which is out of the control of the\n # main process). Thus, since we will exit `pin_memory_thread` before the\n # workers (see below), two separete events are used.\n #\n # NOTE: In short, the protocol is that the main process will set these\n # `done_event`s and then the corresponding processes/threads a `None`,\n # and that they may exit at any time after receiving the `None`.\n #\n # NOTE: Using `None` as the final signal is valid, since normal data will\n # always be a 2-tuple with the 1st element being the index of the data\n # transferred (different from dataset index/key), and the 2nd being\n # either the dataset key or the data sample (depending on which part\n # of the data model the queue is at).\n #\n # [ worker processes ]\n # While loader process is alive:\n # Get from `index_queue`.\n # If get anything else,\n # Check `workers_done_event`.\n # If set, continue to next iteration\n # i.e., keep getting until see the `None`, then exit.\n # Otherwise, process data:\n # If is fetching from an `IterableDataset` and the iterator\n # is exhausted, send an `_IterableDatasetStopIteration`\n # object to signal iteration end. The main process, upon\n # receiving such an object, will send `None` to this\n # worker and not use the corresponding `index_queue`\n # anymore.\n # If timed out,\n # No matter `workers_done_event` is set (still need to see `None`)\n # or not, must continue to next iteration.\n # (outside loop)\n # If `workers_done_event` is set, (this can be False with `IterableDataset`)\n # `data_queue.cancel_join_thread()`. (Everything is ending here:\n # main process won't read from it;\n # other workers will also call\n # `cancel_join_thread`.)\n #\n # [ pin_memory_thread ]\n # # No need to check main thread. If this thread is alive, the main loader\n # # thread must be alive, because this thread is set as daemonic.\n # While `pin_memory_thread_done_event` is not set:\n # Get from `worker_result_queue`.\n # If timed out, continue to get in the next iteration.\n # Otherwise, process data.\n # While `pin_memory_thread_done_event` is not set:\n # Put processed data to `data_queue` (a `queue.Queue` with blocking put)\n # If timed out, continue to put in the next iteration.\n # Otherwise, break, i.e., continuing to the out loop.\n #\n # NOTE: we don't check the status of the main thread because\n # 1. if the process is killed by fatal signal, `pin_memory_thread`\n # ends.\n # 2. in other cases, either the cleaning-up in __del__ or the\n # automatic exit of daemonic thread will take care of it.\n # This won't busy-wait either because `.get(timeout)` does not\n # busy-wait.\n #\n # [ main process ]\n # In the DataLoader Iter's `__del__`\n # b. Exit `pin_memory_thread`\n # i. Set `pin_memory_thread_done_event`.\n # ii Put `None` in `worker_result_queue`.\n # iii. Join the `pin_memory_thread`.\n # iv. `worker_result_queue.cancel_join_thread()`.\n #\n # c. Exit the workers.\n # i. Set `workers_done_event`.\n # ii. Put `None` in each worker's `index_queue`.\n # iii. Join the workers.\n # iv. Call `.cancel_join_thread()` on each worker's `index_queue`.\n #\n # NOTE: (c) is better placed after (b) because it may leave corrupted\n # data in `worker_result_queue`, which `pin_memory_thread`\n # reads from, in which case the `pin_memory_thread` can only\n # happen at timing out, which is slow. Nonetheless, same thing\n # happens if a worker is killed by signal at unfortunate times,\n # but in other cases, we are better off having a non-corrupted\n # `worker_result_queue` for `pin_memory_thread`.\n #\n # NOTE: If `pin_memory=False`, there is no `pin_memory_thread` and (b)\n # can be omitted\n #\n # NB: `done_event`s isn't strictly needed. E.g., we can just check for\n # `None` from `index_queue`, but it allows us to skip wasting resources\n # processing indices already in `index_queue` if we are already shutting\n # down.\n\n def __init__(self, loader):\n super().__init__(loader)\n\n self._prefetch_factor = loader.prefetch_factor\n self._in_order = loader.in_order\n\n assert self._num_workers > 0\n assert self._prefetch_factor > 0\n\n if loader.multiprocessing_context is None:\n multiprocessing_context = torch.multiprocessing\n else:\n multiprocessing_context = loader.multiprocessing_context\n\n self._worker_init_fn = loader.worker_init_fn\n\n # Adds forward compatibilities so classic DataLoader can work with DataPipes:\n # Additional worker init function will take care of sharding in MP and Distributed\n if isinstance(self._dataset, (IterDataPipe, MapDataPipe)):\n self._worker_init_fn = functools.partial(\n _sharding_worker_init_fn,\n self._worker_init_fn,\n self._world_size,\n self._rank,\n )\n\n # No certainty which module multiprocessing_context is\n self._worker_result_queue = multiprocessing_context.Queue() # type: ignore[var-annotated]\n self._worker_pids_set = False\n self._shutdown = False\n self._workers_done_event = multiprocessing_context.Event()\n\n self._index_queues = []\n self._workers = []\n for i in range(self._num_workers):\n # No certainty which module multiprocessing_context is\n index_queue = multiprocessing_context.Queue() # type: ignore[var-annotated]\n # Need to `cancel_join_thread` here!\n # See sections (2) and (3b) above.\n index_queue.cancel_join_thread()\n w = multiprocessing_context.Process(\n target=_utils.worker._worker_loop,\n args=(\n self._dataset_kind,\n self._dataset,\n index_queue,\n self._worker_result_queue,\n self._workers_done_event,\n self._auto_collation,\n self._collate_fn,\n self._drop_last,\n self._base_seed,\n self._worker_init_fn,\n i,\n self._num_workers,\n self._persistent_workers,\n self._shared_seed,\n ),\n )\n w.daemon = True\n # NB: Process.start() actually take some time as it needs to\n # start a process and pass the arguments over via a pipe.\n # Therefore, we only add a worker to self._workers list after\n # it started, so that we do not call .join() if program dies\n # before it starts, and __del__ tries to join but will get:\n # AssertionError: can only join a started process.\n w.start()\n self._index_queues.append(index_queue)\n self._workers.append(w)\n\n if self._pin_memory:\n self._pin_memory_thread_done_event = threading.Event()\n\n # Queue is not type-annotated\n self._data_queue = queue.Queue() # type: ignore[var-annotated]\n current_device = -1\n if self._pin_memory_device == ""cuda"":\n current_device = torch.cuda.current_device()\n elif self._pin_memory_device == ""xpu"":\n current_device = torch.xpu.current_device()\n elif self._pin_memory_device == torch._C._get_privateuse1_backend_name():\n custom_device_mod = getattr(\n torch, torch._C._get_privateuse1_backend_name()\n )\n current_device = custom_device_mod.current_device()\n elif self._pin_memory_device is None:\n current_device = torch.accelerator.current_device_index()\n pin_memory_thread = threading.Thread(\n target=_utils.pin_memory._pin_memory_loop,\n args=(\n self._worker_result_queue,\n self._data_queue,\n current_device,\n self._pin_memory_thread_done_event,\n self._pin_memory_device,\n ),\n )\n pin_memory_thread.daemon = True\n pin_memory_thread.start()\n # Similar to workers (see comment above), we only register\n # pin_memory_thread once it is started.\n self._pin_memory_thread = pin_memory_thread\n else:\n self._data_queue = self._worker_result_queue # type: ignore[assignment]\n\n # In some rare cases, persistent workers (daemonic processes)\n # would be terminated before `__del__` of iterator is invoked\n # when main process exits\n # It would cause failure when pin_memory_thread tries to read\n # corrupted data from worker_result_queue\n # atexit is used to shutdown thread and child processes in the\n # right sequence before main process exits\n if self._persistent_workers and self._pin_memory:\n import atexit\n\n for w in self._workers:\n atexit.register(_MultiProcessingDataLoaderIter._clean_up_worker, w)\n\n # .pid can be None only before process is spawned (not the case, so ignore)\n _utils.signal_handling._set_worker_pids(id(self), tuple(w.pid for w in self._workers)) # type: ignore[misc]\n _utils.signal_handling._set_SIGCHLD_handler()\n self._worker_pids_set = True\n self._reset(loader, first_iter=True)\n\n def _reset(self, loader, first_iter=False):\n super()._reset(loader, first_iter)\n self._send_idx = 0 # idx of the next task to be sent to workers\n self._rcvd_idx = 0 # idx of the next task to be returned in __next__\n # information about data not yet yielded, i.e., tasks w/ indices in range [rcvd_idx, send_idx).\n # map: task idx => - (worker_id,) if data isn't fetched (outstanding)\n # \ (worker_id, data) if data is already fetched (out-of-order)\n self._task_info = {}\n self._tasks_outstanding = (\n 0 # always equal to count(v for v in task_info.values() if len(v) == 1)\n )\n # A list of booleans representing whether each worker still has work to\n # do, i.e., not having exhausted its iterable dataset object. It always\n # contains all `True`s if not using an iterable-style dataset\n # (i.e., if kind != Iterable).\n # Not that this indicates that a worker still has work to do *for this epoch*.\n # It does not mean that a worker is dead. In case of `_persistent_workers`,\n # the worker will be reset to available in the next epoch.\n self._workers_status = [True for i in range(self._num_workers)]\n # A list of integers representing how many tasks are outstanding for each worker\n # Incremented when a task is dispatched to the worker\n # Decremented when that data has been given to the main thread\n # Each worker should have at most self._prefetch_factor tasks outstanding\n self._workers_num_tasks = [0 for i in range(self._num_workers)]\n # Reset the worker queue cycle so it resumes next epoch at worker 0\n self._worker_queue_idx_cycle = itertools.cycle(range(self._num_workers))\n # We resume the prefetching in case it was enabled\n if not first_iter:\n for idx in range(self._num_workers):\n self._index_queues[idx].put(\n _utils.worker._ResumeIteration(self._shared_seed)\n )\n resume_iteration_cnt = self._num_workers\n while resume_iteration_cnt > 0:\n return_idx, return_data = self._get_data()\n if isinstance(return_idx, _utils.worker._ResumeIteration):\n assert return_data is None\n resume_iteration_cnt -= 1\n # prime the prefetch loop\n for _ in range(self._prefetch_factor * self._num_workers):\n self._try_put_index()\n\n def _try_get_data(self, timeout=_utils.MP_STATUS_CHECK_INTERVAL):\n # Tries to fetch data from `self._data_queue` once for a given timeout.\n # This can also be used as inner loop of fetching without timeout, with\n # the sender status as the loop condition.\n #\n # This raises a `RuntimeError` if any worker died expectedly. This error\n # can come from either the SIGCHLD handler in `_utils/signal_handling.py`\n # (only for non-Windows platforms), or the manual check below on errors\n # and timeouts.\n #\n # Returns a 2-tuple:\n # (bool: whether successfully get data, any: data if successful else None)\n try:\n data = self._data_queue.get(timeout=timeout)\n return (True, data)\n except Exception as e:\n # At timeout and error, we manually check whether any worker has\n # failed. Note that this is the only mechanism for Windows to detect\n # worker failures.\n failed_workers = []\n for worker_id, w in enumerate(self._workers):\n if self._workers_status[worker_id] and not w.is_alive():\n failed_workers.append(w)\n self._mark_worker_as_unavailable(worker_id)\n if len(failed_workers) > 0:\n pids_str = "", "".join(str(w.pid) for w in failed_workers)\n raise RuntimeError(\n f""DataLoader worker (pid(s) {pids_str}) exited unexpectedly""\n ) from e\n if isinstance(e, queue.Empty):\n return (False, None)\n\n import errno\n import tempfile\n\n try:\n # Raise an exception if we are this close to the FDs limit.\n # Apparently, trying to open only one file is not a sufficient\n # test.\n # See NOTE [ DataLoader on Linux and open files limit ]\n fds_limit_margin = 10\n [tempfile.NamedTemporaryFile() for i in range(fds_limit_margin)]\n except OSError as e:\n if e.errno == errno.EMFILE:\n raise RuntimeError(\n ""Too many open files. Communication with the""\n "" workers is no longer possible. Please increase the""\n "" limit using `ulimit -n` in the shell or change the""\n "" sharing strategy by calling""\n "" `torch.multiprocessing.set_sharing_strategy('file_system')`""\n "" at the beginning of your code""\n ) from None\n raise\n\n # NOTE [ DataLoader on Linux and open files limit ]\n #\n # On Linux when DataLoader is used with multiprocessing we pass the data between\n # the root process and the workers through SHM files. We remove those files from\n # the filesystem as soon as they are created and keep them alive by\n # passing around their file descriptors through AF_UNIX sockets. (See\n # docs/source/multiprocessing.rst and 'Multiprocessing Technical Notes` in\n # the wiki (https://github.com/pytorch/pytorch/wiki).)\n #\n # This sometimes leads us to exceeding the open files limit. When that happens,\n # and the offending file descriptor is coming over a socket, the `socket` Python\n # package silently strips the file descriptor from the message, setting only the\n # `MSG_CTRUNC` flag (which might be a bit misleading since the manpage says that\n # it _indicates that some control data were discarded due to lack of space in\n # the buffer for ancillary data_). This might reflect the C implementation of\n # AF_UNIX sockets.\n #\n # This behaviour can be reproduced with the script and instructions at the\n # bottom of this note.\n #\n # When that happens, the standard Python `multiprocessing` (and not\n # `torch.multiprocessing`) raises a `RuntimeError: received 0 items of ancdata`\n #\n # Sometimes, instead of the FD being stripped, you may get an `OSError:\n # Too many open files`, both in the script below and in DataLoader. However,\n # this is rare and seems to be nondeterministic.\n #\n #\n # #!/usr/bin/env python3\n # import sys\n # import socket\n # import os\n # import array\n # import shutil\n # import socket\n #\n #\n # if len(sys.argv) != 4:\n # print(""Usage: "", sys.argv[0], "" tmp_dirname iteration (send|recv)"")\n # sys.exit(1)\n #\n # if __name__ == '__main__':\n # dirname = sys.argv[1]\n # sock_path = dirname + ""/sock""\n # iterations = int(sys.argv[2])\n # def dummy_path(i):\n # return dirname + ""/"" + str(i) + "".dummy""\n #\n #\n # if sys.argv[3] == 'send':\n # while not os.path.exists(sock_path):\n # pass\n # client = socket.socket(socket.AF_UNIX, socket.SOCK_DGRAM)\n # client.connect(sock_path)\n # for i in range(iterations):\n # fd = os.open(dummy_path(i), os.O_WRONLY | os.O_CREAT)\n # ancdata = array.array('i', [fd])\n # msg = bytes([i % 256])\n # print(""Sending fd "", fd, "" (iteration #"", i, "")"")\n # client.sendmsg([msg], [(socket.SOL_SOCKET, socket.SCM_RIGHTS, ancdata)])\n #\n #\n # else:\n # assert sys.argv[3] == 'recv'\n #\n # if os.path.exists(dirname):\n # raise Exception(""Directory exists"")\n #\n # os.mkdir(dirname)\n #\n # print(""Opening socket..."")\n # server = socket.socket(socket.AF_UNIX, socket.SOCK_DGRAM)\n # server.bind(sock_path)\n #\n # print(""Listening..."")\n # for i in range(iterations):\n # a = array.array('i')\n # msg, ancdata, flags, addr = server.recvmsg(1, socket.CMSG_SPACE(a.itemsize))\n # assert(len(ancdata) == 1)\n # cmsg_level, cmsg_type, cmsg_data = ancdata[0]\n # a.frombytes(cmsg_data)\n # print(""Received fd "", a[0], "" (iteration #"", i, "")"")\n #\n # shutil.rmtree(dirname)\n #\n # Steps to reproduce:\n #\n # 1. Run two shells and set lower file descriptor limit in the receiving one:\n # (shell1) ulimit -n 1020\n # (shell2) ulimit -n 1022\n #\n # 2. Run the script above with the `recv` option in the first shell\n # (shell1) ./test_socket.py sock_tmp 1017 recv\n #\n # 3. Run the script with the `send` option in the second shell:\n # (shell2) ./test_socket.py sock_tmp 1017 send\n\n def _get_data(self):\n # Fetches data from `self._data_queue`.\n #\n # We check workers' status every `MP_STATUS_CHECK_INTERVAL` seconds,\n # which we achieve by running `self._try_get_data(timeout=MP_STATUS_CHECK_INTERVAL)`\n # in a loop. This is the only mechanism to detect worker failures for\n # Windows. For other platforms, a SIGCHLD handler is also used for\n # worker failure detection.\n #\n # If `pin_memory=True`, we also need check if `pin_memory_thread` had\n # died at timeouts.\n if self._timeout > 0:\n success, data = self._try_get_data(self._timeout)\n if success:\n return data\n else:\n raise RuntimeError(\n f""DataLoader timed out after {self._timeout} seconds""\n )\n elif self._pin_memory:\n while self._pin_memory_thread.is_alive():\n success, data = self._try_get_data()\n if success:\n return data\n else:\n # while condition is false, i.e., pin_memory_thread died.\n raise RuntimeError(""Pin memory thread exited unexpectedly"")\n # In this case, `self._data_queue` is a `queue.Queue`,. But we don't\n # need to call `.task_done()` because we don't use `.join()`.\n else:\n while True:\n success, data = self._try_get_data()\n if success:\n return data\n\n def _next_data(self):\n while True:\n # If the worker responsible for `self._rcvd_idx` has already ended\n # and was unable to fulfill this task (due to exhausting an `IterableDataset`),\n # we try to advance `self._rcvd_idx` to find the next valid index.\n #\n # This part needs to run in the loop because both the `self._get_data()`\n # call and `_IterableDatasetStopIteration` check below can mark\n # extra worker(s) as dead.\n while self._rcvd_idx < self._send_idx:\n info = self._task_info.get(self._rcvd_idx, None)\n if info:\n worker_id = info[0]\n if (\n len(info) == 2 or self._workers_status[worker_id]\n ): # has data or is still active\n break\n del self._task_info[self._rcvd_idx]\n self._rcvd_idx += 1\n else:\n # no valid `self._rcvd_idx` is found (i.e., didn't break)\n if not self._persistent_workers:\n self._shutdown_workers()\n raise StopIteration\n\n # Now `self._rcvd_idx` is the batch index we want to fetch\n\n # Check if the next sample has already been generated\n if len(self._task_info[self._rcvd_idx]) == 2:\n worker_id, data = self._task_info.pop(self._rcvd_idx)\n self._rcvd_idx += 1\n return self._process_data(data, worker_id)\n\n assert not self._shutdown and self._tasks_outstanding > 0\n idx, data = self._get_data()\n self._tasks_outstanding -= 1\n if self._dataset_kind == _DatasetKind.Iterable:\n # Check for _IterableDatasetStopIteration\n if isinstance(data, _utils.worker._IterableDatasetStopIteration):\n if self._persistent_workers:\n self._workers_status[data.worker_id] = False\n else:\n self._mark_worker_as_unavailable(data.worker_id)\n self._try_put_index()\n continue\n\n if idx != self._rcvd_idx:\n if not self._in_order:\n # don't store it for later, process now\n # delete from self._task_info immediately\n # this keeps the object size manageable\n worker_id = self._task_info.pop(idx)[0]\n return self._process_data(data, worker_id)\n # store out-of-order samples\n self._task_info[idx] += (data,)\n else:\n worker_id = self._task_info.pop(idx)[0]\n self._rcvd_idx += 1\n return self._process_data(data, worker_id)\n\n def _try_put_index(self):\n max_tasks = self._prefetch_factor * self._num_workers\n assert self._tasks_outstanding < max_tasks\n\n try:\n index = self._next_index()\n except StopIteration:\n return\n for _ in range(self._num_workers): # find the next active worker, if any\n worker_queue_idx = next(self._worker_queue_idx_cycle)\n if self._workers_status[worker_queue_idx]:\n if self._in_order:\n break\n elif self._workers_num_tasks[worker_queue_idx] < max_tasks // sum(\n self._workers_status\n ):\n # when self._in_order is False, distribute work to a worker if it has capacity\n # _workers_status is updated only in this thread, so the sum is guaranteed > 0\n break\n else:\n # not found (i.e., didn't break)\n return\n\n self._index_queues[worker_queue_idx].put((self._send_idx, index)) # type: ignore[possibly-undefined]\n self._task_info[self._send_idx] = (worker_queue_idx,)\n self._workers_num_tasks[worker_queue_idx] += 1\n self._tasks_outstanding += 1\n self._send_idx += 1\n\n def _process_data(self, data, worker_idx):\n self._workers_num_tasks[worker_idx] -= 1\n self._try_put_index()\n if isinstance(data, ExceptionWrapper):\n data.reraise()\n return data\n\n def _mark_worker_as_unavailable(self, worker_id, shutdown=False):\n # Mark a worker as having finished its work e.g., due to\n # exhausting an `IterableDataset`. This should be used only when this\n # `_MultiProcessingDataLoaderIter` is going to continue running.\n\n assert self._workers_status[worker_id] or (\n self._persistent_workers and shutdown\n )\n\n # Signal termination to that specific worker.\n q = self._index_queues[worker_id]\n # Indicate that no more data will be put on this queue by the current\n # process.\n q.put(None)\n\n # Note that we don't actually join the worker here, nor do we remove the\n # worker's pid from C side struct because (1) joining may be slow, and\n # (2) since we don't join, the worker may still raise error, and we\n # prefer capturing those, rather than ignoring them, even though they\n # are raised after the worker has finished its job.\n # Joinning is deferred to `_shutdown_workers`, which it is called when\n # all workers finish their jobs (e.g., `IterableDataset` replicas) or\n # when this iterator is garbage collected.\n\n self._workers_status[worker_id] = False\n\n assert self._workers_done_event.is_set() == shutdown\n\n def _shutdown_workers(self):\n # Called when shutting down this `_MultiProcessingDataLoaderIter`.\n # See NOTE [ Data Loader Multiprocessing Shutdown Logic ] for details on\n # the logic of this function.\n if (\n _utils is None\n or _utils.python_exit_status is True\n or _utils.python_exit_status is None\n ):\n # See (2) of the note. If Python is shutting down, do no-op.\n return\n # Normal exit when last reference is gone / iterator is depleted.\n # See (1) and the second half of the note.\n if not self._shutdown:\n self._shutdown = True\n try:\n # Normal exit when last reference is gone / iterator is depleted.\n # See (1) and the second half of the note.\n\n # Exit `pin_memory_thread` first because exiting workers may leave\n # corrupted data in `worker_result_queue` which `pin_memory_thread`\n # reads from.\n if hasattr(self, ""_pin_memory_thread""):\n # Use hasattr in case error happens before we set the attribute.\n self._pin_memory_thread_done_event.set()\n # Send something to pin_memory_thread in case it is waiting\n # so that it can wake up and check `pin_memory_thread_done_event`\n self._worker_result_queue.put((None, None))\n self._pin_memory_thread.join()\n self._worker_result_queue.cancel_join_thread()\n self._worker_result_queue.close()\n\n # Exit workers now.\n self._workers_done_event.set()\n for worker_id in range(len(self._workers)):\n # Get number of workers from `len(self._workers)` instead of\n # `self._num_workers` in case we error before starting all\n # workers.\n # If we are using workers_status with persistent_workers\n # we have to shut it down because the worker is paused\n if self._persistent_workers or self._workers_status[worker_id]:\n self._mark_worker_as_unavailable(worker_id, shutdown=True)\n for w in self._workers:\n # We should be able to join here, but in case anything went\n # wrong, we set a timeout and if the workers fail to join,\n # they are killed in the `finally` block.\n w.join(timeout=_utils.MP_STATUS_CHECK_INTERVAL)\n for q in self._index_queues:\n q.cancel_join_thread()\n q.close()\n finally:\n # Even though all this function does is putting into queues that\n # we have called `cancel_join_thread` on, weird things can\n # happen when a worker is killed by a signal, e.g., hanging in\n # `Event.set()`. So we need to guard this with SIGCHLD handler,\n # and remove pids from the C side data structure only at the\n # end.\n #\n # FIXME: Unfortunately, for Windows, we are missing a worker\n # error detection mechanism here in this function, as it\n # doesn't provide a SIGCHLD handler.\n if self._worker_pids_set:\n _utils.signal_handling._remove_worker_pids(id(self))\n self._worker_pids_set = False\n for w in self._workers:\n if w.is_alive():\n # Existing mechanisms try to make the workers exit\n # peacefully, but in case that we unfortunately reach\n # here, which we shouldn't, (e.g., pytorch/pytorch#39570),\n # we kill the worker.\n w.terminate()\n\n # staticmethod is used to remove reference to `_MultiProcessingDataLoaderIter`\n @staticmethod\n def _clean_up_worker(w):\n try:\n w.join(timeout=_utils.MP_STATUS_CHECK_INTERVAL)\n finally:\n if w.is_alive():\n w.terminate()\n\n def __del__(self):\n self._shutdown_workers()\n",python,tab +1749,2809738,".venv/lib64/python3.12/site-packages/torch/utils/data/dataloader.py",4137,0,"",python,selection_command +1750,2882518,".venv/lib64/python3.12/site-packages/torch/utils/data/dataloader.py",20701,0,"",python,selection_mouse +1751,2882548,".venv/lib64/python3.12/site-packages/torch/utils/data/dataloader.py",20700,0,"",python,selection_command +1752,2885172,"jasmine/utils/dataloader_torch.py",0,0,"",python,tab +1753,2885928,"jasmine/utils/dataloader_torch.py",850,0,"",python,selection_mouse +1754,2885930,"jasmine/utils/dataloader_torch.py",849,0,"",python,selection_command +1755,2887269,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",0,0,"",python,tab +1756,2890699,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1757,2891509,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6696,0,"",python,selection_mouse +1758,2892340,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6299,0,"",python,selection_mouse +1759,2893537,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6262,0,"",python,selection_mouse +1760,2894090,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6257,0,"",python,selection_mouse +1761,2894561,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6253,0,"",python,selection_mouse +1762,2894706,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6243,14,"get_dataloader",python,selection_mouse +1763,2895374,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6269,0,"",python,selection_mouse +1764,2895979,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6260,0,"",python,selection_mouse +1765,2896128,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6258,8,"data_dir",python,selection_mouse +1766,2896643,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6275,0,"",python,selection_mouse +1767,2897223,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",6300,0,"",python,selection_mouse +1768,2899352,"jasmine/utils/dataloader_torch.py",0,0,"",python,tab +1769,2905445,"jasmine/utils/dataloader_torch.py",682,0,"",python,selection_mouse +1770,2905584,"jasmine/utils/dataloader_torch.py",681,7,"episode",python,selection_mouse +1771,2906842,"jasmine/utils/dataloader_torch.py",529,0,"",python,selection_mouse +1772,2906967,"jasmine/utils/dataloader_torch.py",527,4,"self",python,selection_mouse +1773,2907632,"jasmine/utils/dataloader_torch.py",534,0,"",python,selection_mouse +1774,2909793,"jasmine/utils/dataloader_torch.py",533,3,"idx",python,selection_mouse +1775,2948892,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1776,2955429,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1777,2955429,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,0,"",python,selection_command +1778,2957525,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,44," first_batch = next(dataloader_train)",python,selection_command +1779,2957758,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,93," first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore",python,selection_command +1780,2957923,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,163," first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()",python,selection_command +1781,2958049,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,227," first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())",python,selection_command +1782,2958202,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,290," first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())",python,selection_command +1783,2958344,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,344," first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training",python,selection_command +1784,2958721,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,420," first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader_train = itertools.chain([first_batch], dataloader_train)",python,selection_command +1785,2959267,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19518,0,"",python,selection_command +1786,2961305,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19863,0,"ä",python,content +1787,2961305,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19809,0,"ä",python,content +1788,2961305,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19746,0,"ä",python,content +1789,2961305,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19682,0,"ä",python,content +1790,2961305,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19612,0,"ä",python,content +1791,2961306,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19563,0,"ä",python,content +1792,2961306,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19518,0,"ä",python,content +1793,2961307,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19519,0,"",python,selection_keyboard +1794,2961493,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19870,0," ",python,content +1795,2961494,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19815,0," ",python,content +1796,2961494,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19751,0," ",python,content +1797,2961494,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19686,0," ",python,content +1798,2961494,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19615,0," ",python,content +1799,2961494,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19565,0," ",python,content +1800,2961494,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19519,0," ",python,content +1801,2961495,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19520,0,"",python,selection_keyboard +1802,2961863,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19519,0,"",python,selection_command +1803,2962780,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19874,2,"",python,content +1804,2962782,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19819,2,"",python,content +1805,2962799,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19754,2,"",python,content +1806,2962800,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19688,2,"",python,content +1807,2962802,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19616,2,"",python,content +1808,2962834,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19564,2,"",python,content +1809,2962836,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19518,2,"",python,content +1810,2964770,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,44," first_batch = next(dataloader_train)",python,selection_command +1811,2964967,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,93," first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore",python,selection_command +1812,2965118,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,163," first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()",python,selection_command +1813,2965283,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,227," first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())",python,selection_command +1814,2965422,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,290," first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())",python,selection_command +1815,2965552,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,344," first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training",python,selection_command +1816,2965902,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,420," first_batch = next(dataloader_train)\n first_batch[""rng""] = rng # type: ignore\n compiled = train_step.lower(optimizer, first_batch).compile()\n print_compiled_memory_stats(compiled.memory_analysis())\n print_compiled_cost_analysis(compiled.cost_analysis())\n # Do not skip the first batch during training\n dataloader_train = itertools.chain([first_batch], dataloader_train)",python,selection_command +1817,2966099,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19518,0,"",python,selection_command +1818,2966864,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19863,0,"#",python,content +1819,2966864,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19809,0,"#",python,content +1820,2966864,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19746,0,"#",python,content +1821,2966864,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19682,0,"#",python,content +1822,2966864,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19612,0,"#",python,content +1823,2966864,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19563,0,"#",python,content +1824,2966864,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19518,0,"#",python,content +1825,2966865,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19519,0,"",python,selection_keyboard +1826,2966928,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19870,0," ",python,content +1827,2966928,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19815,0," ",python,content +1828,2966928,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19751,0," ",python,content +1829,2966928,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19686,0," ",python,content +1830,2966928,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19615,0," ",python,content +1831,2966928,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19565,0," ",python,content +1832,2966928,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19519,0," ",python,content +1833,2966929,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19520,0,"",python,selection_keyboard +1834,2967235,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19519,0,"",python,selection_command +1835,2967381,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19486,0,"",python,selection_command +1836,2967664,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19509,0,"\n ",python,content +1837,2967885,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19518,0,"p",python,content +1838,2967886,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19519,0,"",python,selection_keyboard +1839,2968034,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19519,0,"a",python,content +1840,2968035,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19520,0,"",python,selection_keyboard +1841,2968195,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19520,0,"s",python,content +1842,2968196,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19521,0,"",python,selection_keyboard +1843,2968305,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19521,0,"s",python,content +1844,2968306,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19522,0,"",python,selection_keyboard +1845,2968607,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19521,0,"",python,selection_command +1846,2971363,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +1847,2972593,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --log_interval 1 \\r\n --val_interval 2 \\r\n --eval_full_frame \\r\n --val_steps 5\r\nSLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +1848,2972654,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +1849,2976190,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +1850,2976442,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1851,2977772,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19869,0,"",python,selection_mouse +1852,2978327,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19920,0,"",python,selection_mouse +1853,2978449,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19919,5,"chain",python,selection_mouse +1854,2980238,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20859,0,"",python,selection_mouse +1855,2981758,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20204,0,"",python,selection_mouse +1856,2982684,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +1857,2983022,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20093,0,"",python,selection_mouse +1858,2983156,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +1859,2983217,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20086,16,"dataloader_train",python,selection_mouse +1860,2984010,"TERMINAL",0,0,"wandb: creating run\r\nwandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_125016-6w71romt\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/6w71romt\r\n",,terminal_output +1861,2984170,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +1862,2986259,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +1863,2990640,"TERMINAL",0,0,"Starting training from step 0...\r\n",,terminal_output +1864,2990942,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 719, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 573, in main\r\n batch[""rng""] = _rng_mask\r\n ~~~~~^^^^^^^\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/numpy/array_methods.py"", line 617, in _unimplemented_setitem\r\n raise TypeError(msg.format(type(self)))\r\nTypeError: JAX arrays are immutable and do not support in-place item assignment. Instead of x[idx] = y, use x = x.at[idx].set(y) or another .at[] method: https://docs.jax.dev/en/latest/_autosummary/jax.numpy.ndarray.at.html\r\n",,terminal_output +1865,2991601,"TERMINAL",0,0,"wandb: \r\nwandb: 🚀 View run coinrun-dyn-dev-3536670 at: https://wandb.ai/instant-uv/jafar/runs/6w71romt\r\nwandb: Find logs at: ../../../../../hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_125016-6w71romt/logs\r\n",,terminal_output +1866,2991746,"TERMINAL",0,0,"W1002 12:50:24.899564 1480517 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugonly job_name: ""jax_worker"": CANCELLED: CANCELLED\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_message:""CANCELLED"", grpc_status:1} [type.googleapis.com/tensorflow.DerivedStatus='']\r\n",,terminal_output +1867,2992444,"TERMINAL",0,0,"srun: error: hkn0401: task 0: Exited with exit code 1\r\n]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +1868,2993477,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +1869,2994514,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19775,0,"",python,selection_mouse +1870,2994585,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19710,65,"ompiled_memory_stats(compiled.memory_analysis())\n # print_",python,selection_mouse +1871,2994831,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19977,0,"",python,selection_mouse +1872,2994945,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19899,78,"r_train = itertools.chain([first_batch], dataloader_train)\n print(f""Startin",python,selection_mouse +1873,2995147,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19844,0,"",python,selection_mouse +1874,2995272,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19843,4,"skip",python,selection_mouse +1875,2995564,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19900,0,"",python,selection_mouse +1876,2995744,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19890,16,"dataloader_train",python,selection_mouse +1877,2996215,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19897,0,"",python,selection_mouse +1878,2996284,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19890,16,"dataloader_train",python,selection_mouse +1879,2997437,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20355,0,"",python,selection_mouse +1880,2997570,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20343,15,"print_mem_stats",python,selection_mouse +1881,3005749,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19522,0,"",python,selection_mouse +1882,3005753,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19521,0,"",python,selection_command +1883,3006528,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",0,0,"",python,tab +1884,3006862,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",0,0,"",python,tab +1885,3013204,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",0,0,"",python,tab +1886,3014166,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",407,0,"",python,selection_mouse +1887,3014522,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",483,0,"",python,selection_mouse +1888,3019352,"jasmine/utils/dataloader_torch.py",0,0,"",python,tab +1889,3020292,"jasmine/utils/dataloader_torch.py",797,0,"",python,selection_mouse +1890,3020293,"jasmine/utils/dataloader_torch.py",796,0,"",python,selection_command +1891,3020294,"jasmine/utils/dataloader_torch.py",796,1,":",python,selection_mouse +1892,3020299,"jasmine/utils/dataloader_torch.py",797,0,"",python,selection_command +1893,3020551,"jasmine/utils/dataloader_torch.py",709,0,"",python,selection_mouse +1894,3020937,"jasmine/utils/dataloader_torch.py",772,0,"",python,selection_mouse +1895,3020938,"jasmine/utils/dataloader_torch.py",771,0,"",python,selection_command +1896,3020967,"jasmine/utils/dataloader_torch.py",771,1,"0",python,selection_mouse +1897,3020968,"jasmine/utils/dataloader_torch.py",772,0,"",python,selection_command +1898,3021381,"jasmine/utils/dataloader_torch.py",625,0,"",python,selection_mouse +1899,3021696,"jasmine/utils/dataloader_torch.py",698,0,"",python,selection_mouse +1900,3022368,"jasmine/utils/dataloader_torch.py",685,0,"",python,selection_mouse +1901,3022587,"jasmine/utils/dataloader_torch.py",681,7,"episode",python,selection_mouse +1902,3024160,"jasmine/utils/dataloader_torch.py",685,0,"",python,selection_mouse +1903,3024754,"jasmine/utils/dataloader_torch.py",684,0,"",python,selection_mouse +1904,3024996,"jasmine/utils/dataloader_torch.py",681,7,"episode",python,selection_mouse +1905,3026023,"jasmine/utils/dataloader_torch.py",534,0,"",python,selection_mouse +1906,3028139,"jasmine/utils/dataloader_torch.py",742,0,"",python,selection_mouse +1907,3033043,"jasmine/utils/dataloader_torch.py",742,0,"{",python,content +1908,3033045,"jasmine/utils/dataloader_torch.py",743,0,"",python,selection_keyboard +1909,3034591,"jasmine/utils/dataloader_torch.py",773,0,"}",python,content +1910,3034592,"jasmine/utils/dataloader_torch.py",743,0,"""videos"": ",python,content +1911,3034595,"jasmine/utils/dataloader_torch.py",784,0,"",python,selection_command +1912,3040272,"jasmine/utils/dataloader_torch.py",783,1,"",python,content +1913,3040273,"jasmine/utils/dataloader_torch.py",743,10,"",python,content +1914,3040893,"jasmine/utils/dataloader_torch.py",742,1,"",python,content +1915,3041846,"jasmine/utils/dataloader_torch.py",741,0,"",python,selection_command +1916,3041905,"jasmine/utils/dataloader_torch.py",773,0,"",python,selection_command +1917,3042123,"jasmine/utils/dataloader_torch.py",774,0,"",python,selection_command +1918,3042288,"jasmine/utils/dataloader_torch.py",789,0,"",python,selection_command +1919,3042598,"jasmine/utils/dataloader_torch.py",812,0,"",python,selection_command +1920,3042971,"jasmine/utils/dataloader_torch.py",865,0,"",python,selection_command +1921,3043562,"jasmine/utils/dataloader_torch.py",864,0,"",python,selection_command +1922,3043698,"jasmine/utils/dataloader_torch.py",863,0,"",python,selection_command +1923,3043838,"jasmine/utils/dataloader_torch.py",862,0,"",python,selection_command +1924,3047576,"jasmine/utils/dataloader_torch.py",862,0,"{",python,content +1925,3047578,"jasmine/utils/dataloader_torch.py",863,0,"",python,selection_keyboard +1926,3048787,"jasmine/utils/dataloader_torch.py",889,0,"}",python,content +1927,3048788,"jasmine/utils/dataloader_torch.py",863,0,"""videos"": ",python,content +1928,3048791,"jasmine/utils/dataloader_torch.py",900,0,"",python,selection_command +1929,3052392,"jasmine/train_dynamics.py",0,0,"",python,tab +1930,3053850,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/utils/nn.py",0,0,"",python,tab +1931,3054179,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",0,0,"",python,tab +1932,3056354,"jasmine/train_dynamics.py",0,0,"",python,tab +1933,3057764,"jasmine/train_dynamics.py",2499,0,"",python,selection_mouse +1934,3057784,"jasmine/train_dynamics.py",2498,0,"",python,selection_command +1935,3061917,"jasmine/train_dynamics.py",6179,0,"",python,selection_command +1936,3062221,"jasmine/train_dynamics.py",6220,0,"",python,selection_command +1937,3062564,"jasmine/train_dynamics.py",6247,0,"",python,selection_command +1938,3063208,"jasmine/train_dynamics.py",13481,0,"",python,selection_command +1939,3064523,"jasmine/train_dynamics.py",13505,0,"",python,selection_command +1940,3064894,"jasmine/train_dynamics.py",13547,0,"",python,selection_command +1941,3065246,"jasmine/train_dynamics.py",13658,0,"",python,selection_command +1942,3065804,"jasmine/train_dynamics.py",17476,0,"",python,selection_command +1943,3066641,"jasmine/train_dynamics.py",17753,0,"",python,selection_command +1944,3066968,"jasmine/train_dynamics.py",19358,0,"",python,selection_command +1945,3068409,"jasmine/train_dynamics.py",19396,0,"",python,selection_command +1946,3068576,"jasmine/train_dynamics.py",19422,0,"",python,selection_command +1947,3068743,"jasmine/train_dynamics.py",19443,0,"",python,selection_command +1948,3068898,"jasmine/train_dynamics.py",19481,0,"",python,selection_command +1949,3069029,"jasmine/train_dynamics.py",19514,0,"",python,selection_command +1950,3069377,"jasmine/train_dynamics.py",19527,0,"",python,selection_command +1951,3070005,"jasmine/train_dynamics.py",19523,46," # first_batch = next(dataloader_train)",python,selection_command +1952,3070206,"jasmine/train_dynamics.py",19523,97," # first_batch = next(dataloader_train)\n # first_batch[""rng""] = rng # type: ignore",python,selection_command +1953,3070344,"jasmine/train_dynamics.py",19523,169," # first_batch = next(dataloader_train)\n # first_batch[""rng""] = rng # type: ignore\n # compiled = train_step.lower(optimizer, first_batch).compile()",python,selection_command +1954,3070500,"jasmine/train_dynamics.py",19523,235," # first_batch = next(dataloader_train)\n # first_batch[""rng""] = rng # type: ignore\n # compiled = train_step.lower(optimizer, first_batch).compile()\n # print_compiled_memory_stats(compiled.memory_analysis())",python,selection_command +1955,3070630,"jasmine/train_dynamics.py",19523,300," # first_batch = next(dataloader_train)\n # first_batch[""rng""] = rng # type: ignore\n # compiled = train_step.lower(optimizer, first_batch).compile()\n # print_compiled_memory_stats(compiled.memory_analysis())\n # print_compiled_cost_analysis(compiled.cost_analysis())",python,selection_command +1956,3071144,"jasmine/train_dynamics.py",19763,0,"",python,selection_command +1957,3071629,"jasmine/train_dynamics.py",19828,0,"",python,selection_command +1958,3071772,"jasmine/train_dynamics.py",19884,0,"",python,selection_command +1959,3071932,"jasmine/train_dynamics.py",19962,0,"",python,selection_command +1960,3072093,"jasmine/train_dynamics.py",20014,0,"",python,selection_command +1961,3072766,"jasmine/train_dynamics.py",20086,0,"",python,selection_command +1962,3075143,"jasmine/train_dynamics.py",20096,0,"",python,selection_mouse +1963,3075304,"jasmine/train_dynamics.py",20086,16,"dataloader_train",python,selection_mouse +1964,3082775,"jasmine/train_dynamics.py",20086,15,"dataloader_trai",python,selection_command +1965,3085422,"jasmine/train_dynamics.py",20086,16,"dataloader_train",python,selection_command +1966,3090809,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +1967,3090937,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --log_interval 1 \\r\n --val_interval 2 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +1968,3091081,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +1969,3091211,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +1970,3095104,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +1971,3101550,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +1972,3101949,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +1973,3102719,"TERMINAL",0,0,"wandb: creating run\r\nwandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_125215-uwl3tafv\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/uwl3tafv\r\n",,terminal_output +1974,3102893,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +1975,3105019,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +1976,3108751,"TERMINAL",0,0,"Starting training from step 0...\r\n",,terminal_output +1977,3161039,"jasmine/train_dynamics.py",0,0,"",python,tab +1978,3161157,"jasmine/train_dynamics.py",11744,0,"",python,selection_command +1979,3182255,"TERMINAL",0,0,"\r\nMemstats: After params initialized:\r\n\tUsing (GB) 1.18 / 38.7 (3.049096%) on cuda:0\r\nStep 1\r\n",,terminal_output +1980,3214316,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3536670.11 task 0: running\r\n",,terminal_output +1981,3214414,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.11\r\nsrun: forcing job termination\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3536670.11 ON hkn0401 CANCELLED AT 2025-10-02T12:54:07 ***\r\n",,terminal_output +1982,3214635,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.11\r\nsrun: job abort in progress\r\n",,terminal_output +1983,3214787,"TERMINAL",0,0,"]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +1984,3217098,"jasmine/train_dynamics.py",0,0,"",python,tab +1985,3220866,"jasmine/utils/dataloader_torch.py",0,0,"",python,tab +1986,3222439,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +1987,3223601,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1897,0,"",shellscript,selection_mouse +1988,3224526,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1898,0,"",shellscript,selection_command +1989,3224656,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1899,0,"",shellscript,selection_command +1990,3224804,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1900,0,"",shellscript,selection_command +1991,3224962,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1901,0,"",shellscript,selection_command +1992,3225132,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1902,0,"",shellscript,selection_command +1993,3225512,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1901,1,"",shellscript,content +1994,3225998,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1901,0,"1",shellscript,content +1995,3225999,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1902,0,"",shellscript,selection_keyboard +1996,3226196,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1902,0,"0",shellscript,content +1997,3226197,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1903,0,"",shellscript,selection_keyboard +1998,3226406,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1902,0,"",shellscript,selection_command +1999,3230285,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1926,0,"",shellscript,selection_command +2000,3234732,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +2001,3235075,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --log_interval 1 \\r\n --val_interval 10 \\r\n --eval_full_frame \\r\n --val_steps 5\r\nSLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\nGpuFreq=control_disabled\r\n",,terminal_output +2002,3238767,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +2003,3245227,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +2004,3246004,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +2005,3246442,"TERMINAL",0,0,"wandb: creating run\r\nwandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_125438-kp34ikdq\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/kp34ikdq\r\n",,terminal_output +2006,3246522,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +2007,3248567,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +2008,3252689,"TERMINAL",0,0,"Starting training from step 0...\r\n",,terminal_output +2009,3326412,"TERMINAL",0,0,"\r\nMemstats: After params initialized:\r\n\tUsing (GB) 1.21 / 38.7 (3.126615%) on cuda:0\r\nStep 1\r\n",,terminal_output +2010,3367348,"TERMINAL",0,0,"Step 2\r\n",,terminal_output +2011,3369963,"TERMINAL",0,0,"Saved checkpoint at step 2\r\n",,terminal_output +2012,3370595,"TERMINAL",0,0,"Step 3\r\n",,terminal_output +2013,3372590,"TERMINAL",0,0,"Step 4\r\n",,terminal_output +2014,3384157,"TERMINAL",0,0,"Saved checkpoint at step 4\r\n",,terminal_output +2015,3384518,"TERMINAL",0,0,"Step 5\r\n",,terminal_output +2016,3386572,"TERMINAL",0,0,"Step 6\r\n",,terminal_output +2017,3396776,"TERMINAL",0,0,"Saved checkpoint at step 6\r\n",,terminal_output +2018,3398116,"TERMINAL",0,0,"Step 7\r\n",,terminal_output +2019,3400142,"TERMINAL",0,0,"Step 8\r\n",,terminal_output +2020,3402482,"TERMINAL",0,0,"Saved checkpoint at step 8\r\n",,terminal_output +2021,3403521,"TERMINAL",0,0,"Step 9\r\n",,terminal_output +2022,3404048,"TERMINAL",0,0,"WARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000008 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000008) to end with "".orbax-checkpoint-tmp"".\r\nWARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000002 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000002) to end with "".orbax-checkpoint-tmp"".\r\nWARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000006 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000006) to end with "".orbax-checkpoint-tmp"".\r\nWARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000004 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000004) to end with "".orbax-checkpoint-tmp"".\r\n",,terminal_output +2023,3405788,"TERMINAL",0,0,"Step 10\r\nCalculating validation metrics...\r\n",,terminal_output +2024,3408857,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3536670.12 task 0: running\r\n",,terminal_output +2025,3408967,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.12\r\nsrun: forcing job termination\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3536670.12 ON hkn0401 CANCELLED AT 2025-10-02T12:57:22 ***\r\n",,terminal_output +2026,3409108,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.12\r\nsrun: job abort in progress\r\n",,terminal_output +2027,3409351,"TERMINAL",0,0,"]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +2028,3412116,"jasmine/train_dynamics.py",0,0,"",python,tab +2029,3413272,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +2030,3415591,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,13,"",python,content +2031,3415617,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19518,0,"",python,selection_command +2032,3416371,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,46," # first_batch = next(dataloader_train)",python,selection_command +2033,3416559,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,97," # first_batch = next(dataloader_train)\n # first_batch[""rng""] = rng # type: ignore",python,selection_command +2034,3416693,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,169," # first_batch = next(dataloader_train)\n # first_batch[""rng""] = rng # type: ignore\n # compiled = train_step.lower(optimizer, first_batch).compile()",python,selection_command +2035,3416843,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,235," # first_batch = next(dataloader_train)\n # first_batch[""rng""] = rng # type: ignore\n # compiled = train_step.lower(optimizer, first_batch).compile()\n # print_compiled_memory_stats(compiled.memory_analysis())",python,selection_command +2036,3416970,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,300," # first_batch = next(dataloader_train)\n # first_batch[""rng""] = rng # type: ignore\n # compiled = train_step.lower(optimizer, first_batch).compile()\n # print_compiled_memory_stats(compiled.memory_analysis())\n # print_compiled_cost_analysis(compiled.cost_analysis())",python,selection_command +2037,3417132,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,356," # first_batch = next(dataloader_train)\n # first_batch[""rng""] = rng # type: ignore\n # compiled = train_step.lower(optimizer, first_batch).compile()\n # print_compiled_memory_stats(compiled.memory_analysis())\n # print_compiled_cost_analysis(compiled.cost_analysis())\n # # Do not skip the first batch during training",python,selection_command +2038,3417256,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,434," # first_batch = next(dataloader_train)\n # first_batch[""rng""] = rng # type: ignore\n # compiled = train_step.lower(optimizer, first_batch).compile()\n # print_compiled_memory_stats(compiled.memory_analysis())\n # print_compiled_cost_analysis(compiled.cost_analysis())\n # # Do not skip the first batch during training\n # dataloader_train = itertools.chain([first_batch], dataloader_train)",python,selection_command +2039,3417449,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19518,0,"",python,selection_command +2040,3417920,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19875,1,"",python,content +2041,3417920,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19819,1,"",python,content +2042,3417920,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19754,1,"",python,content +2043,3417920,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19688,1,"",python,content +2044,3417920,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19616,1,"",python,content +2045,3417920,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19565,1,"",python,content +2046,3417920,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19518,1,"",python,content +2047,3418056,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19869,1,"",python,content +2048,3418056,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19814,1,"",python,content +2049,3418056,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19750,1,"",python,content +2050,3418057,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19685,1,"",python,content +2051,3418057,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19614,1,"",python,content +2052,3418057,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19564,1,"",python,content +2053,3418057,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19518,1,"",python,content +2054,3418140,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19517,0,"",python,selection_command +2055,3421026,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +2056,3421122,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --log_interval 1 \\r\n --val_interval 10 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +2057,3421264,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +2058,3421398,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +2059,3422299,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +2060,3425473,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +2061,3431858,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +2062,3432212,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +2063,3433123,"TERMINAL",0,0,"wandb: creating run\r\nwandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_125745-7hvo8i22\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/7hvo8i22\r\n",,terminal_output +2064,3433181,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +2065,3434565,"TERMINAL",0,0,"WARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000008 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000008) to end with "".orbax-checkpoint-tmp"".\r\nWARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000006 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000006) to end with "".orbax-checkpoint-tmp"".\r\nWARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000004 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000004) to end with "".orbax-checkpoint-tmp"".\r\n",,terminal_output +2066,3435282,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +2067,3439505,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 718, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 559, in main\r\n first_batch = next(dataloader_train)\r\n ^^^^^^^^^^^^^^^^^^^^^^\r\nTypeError: 'DataLoader' object is not an iterator\r\n",,terminal_output +2068,3440683,"TERMINAL",0,0,"wandb: \r\nwandb: 🚀 View run coinrun-dyn-dev-3536670 at: https://wandb.ai/instant-uv/jafar/runs/7hvo8i22\r\nwandb: Find logs at: ../../../../../hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_125745-7hvo8i22/logs\r\n",,terminal_output +2069,3440818,"TERMINAL",0,0,"W1002 12:57:54.004564 1488753 pjrt_client.cc:1469] WatchJobStateAsync failed for task goo.gle/debugproto job_name: ""jax_worker"": CANCELLED: CANCELLED\r\nAdditional GRPC error information from remote target coordination_service while calling /tensorflow.CoordinationService/WatchJobState:\r\n:UNKNOWN:Error received from peer {grpc_status:1, grpc_message:""CANCELLED""} [type.googleapis.com/tensorflow.DerivedStatus='']\r\n",,terminal_output +2070,3441559,"TERMINAL",0,0,"srun: error: hkn0401: task 0: Exited with exit code 1\r\n]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +2071,3453144,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +2072,3453145,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19510,0,"",python,selection_command +2073,3454229,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19583,0,"",python,selection_mouse +2074,3455188,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19533,0,"",python,selection_mouse +2075,3455244,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19532,4,"next",python,selection_mouse +2076,3456355,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19533,0,"",python,selection_mouse +2077,3457185,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19532,4,"next",python,selection_mouse +2078,3464776,"TERMINAL",0,0,"\r(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +2079,3488053,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +2080,3489140,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19504,0,"",python,selection_mouse +2081,3489972,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19537,0,"",python,selection_mouse +2082,3490944,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19537,0,"i",python,content +2083,3490945,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19538,0,"",python,selection_keyboard +2084,3491186,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19538,0,"t",python,content +2085,3491188,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19539,0,"",python,selection_keyboard +2086,3491386,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19539,0,"e",python,content +2087,3491388,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19540,0,"",python,selection_keyboard +2088,3491469,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19540,0,"r",python,content +2089,3491470,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19541,0,"",python,selection_keyboard +2090,3492052,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19541,0,"(",python,content +2091,3492053,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19542,0,"",python,selection_keyboard +2092,3493472,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19559,0,")",python,content +2093,3493473,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19560,0,"",python,selection_keyboard +2094,3493702,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19559,0,"",python,selection_command +2095,3497041,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +2096,3497238,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --log_interval 1 \\r\n --val_interval 10 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +2097,3497330,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +2098,3497461,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +2099,3500987,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +2100,3507567,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +2101,3508178,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +2102,3508808,"TERMINAL",0,0,"wandb: creating run\r\nwandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_125901-b39vz7mw\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/b39vz7mw\r\n",,terminal_output +2103,3508895,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +2104,3510300,"TERMINAL",0,0,"WARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000008 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000008) to end with "".orbax-checkpoint-tmp"".\r\nWARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000006 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000006) to end with "".orbax-checkpoint-tmp"".\r\nWARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000004 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000004) to end with "".orbax-checkpoint-tmp"".\r\n",,terminal_output +2105,3511063,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +2106,3588605,"TERMINAL",0,0,"Total memory size: 23.3 GB, Output size: 0.9 GB, Temp size: 22.3 GB, Argument size: 0.9 GB, Host temp size: 0.0 GB.\r\nFLOPs: 7.851e+11, Bytes: 6.639e+11 (618.3 GB), Intensity: 1.2 FLOPs/byte\r\nStarting training from step 0...\r\n",,terminal_output +2107,3589382,"TERMINAL",0,0,"\r\nMemstats: After params initialized:\r\n\tUsing (GB) 1.21 / 38.7 (3.126615%) on cuda:0\r\nStep 1\r\n",,terminal_output +2108,3630242,"TERMINAL",0,0,"Step 2\r\n",,terminal_output +2109,3632551,"TERMINAL",0,0,"Saved checkpoint at step 2\r\n",,terminal_output +2110,3711255,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3536670.14 task 0: running\r\n",,terminal_output +2111,3711405,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.14\r\nsrun: forcing job termination\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 718, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 568, in main\r\n while step < args.num_steps:\r\n ^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3536670.14 ON hkn0401 CANCELLED AT 2025-10-02T13:02:24 ***\r\n",,terminal_output +2112,3711600,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.14\r\nsrun: job abort in progress\r\n",,terminal_output +2113,3711761,"TERMINAL",0,0,"]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +2114,3713272,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +2115,3722854,"jasmine/utils/dataloader_torch.py",0,0,"",python,tab +2116,3722855,"jasmine/utils/dataloader_torch.py",851,0,"",python,selection_command +2117,3723292,"jasmine/train_dynamics.py",0,0,"",python,tab +2118,3723293,"jasmine/train_dynamics.py",11744,0,"",python,selection_command +2119,3734488,"jasmine/train_dynamics.py",19354,37," dataloader_train = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in train_iterator\n )",python,content +2120,3736168,"jasmine/train_dynamics.py",19856,37," dataloader_val = (\n {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in val_iterator\n )",python,content +2121,3743430,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +2122,3744153,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=00:20:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670\r\n\r\nenv | grep SLURM\r\n\r\nexport PYTHONUNBUFFERED=1\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=2 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --log_interval 1 \\r\n --val_interval 10 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +2123,3744283,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1469415\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759399921\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759403521\r\nSLURM_PMI2_SRUN_PORT=44089\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536670\r\nSLURM_PTY_PORT=41791\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=44895\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536670\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=44895\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +2124,3744412,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +2125,3748197,"TERMINAL",0,0,"Running on 1 devices.\r\n",,terminal_output +2126,3754634,"TERMINAL",0,0,"Counting all components: ['dynamics', 'lam', 'tokenizer']\r\n",,terminal_output +2127,3755062,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +2128,3755818,"TERMINAL",0,0,"wandb: creating run\r\nwandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_130308-826jhtsl\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536670\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/826jhtsl\r\n",,terminal_output +2129,3755996,"TERMINAL",0,0,"Parameter counts:\r\n{'dynamics': 26555904, 'lam': 17640416, 'tokenizer': 33750256, 'total': 77946576}\r\n",,terminal_output +2130,3757417,"TERMINAL",0,0,"WARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000008 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000008) to end with "".orbax-checkpoint-tmp"".\r\nWARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000006 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000006) to end with "".orbax-checkpoint-tmp"".\r\nWARNING:absl:Path /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000004 could not be identified as a temporary checkpoint path using . Got error: Expected AtomicRenameTemporaryPath (/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/interactive/3536670/000004) to end with "".orbax-checkpoint-tmp"".\r\n",,terminal_output +2131,3758135,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +2132,3835568,"TERMINAL",0,0,"Total memory size: 23.3 GB, Output size: 0.9 GB, Temp size: 22.3 GB, Argument size: 0.9 GB, Host temp size: 0.0 GB.\r\nFLOPs: 2.074e+12, Bytes: 6.664e+11 (620.6 GB), Intensity: 3.1 FLOPs/byte\r\nStarting training from step 0...\r\n",,terminal_output +2133,3836060,"TERMINAL",0,0,"\r\nMemstats: After params initialized:\r\n\tUsing (GB) 1.16 / 38.7 (2.997416%) on cuda:0\r\nStep 1\r\n",,terminal_output +2134,3896307,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +2135,3902983,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",0,0,"",python,tab +2136,3921848,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",0,0,"""""""\nGenerates a dataset of random-action CoinRun episodes.\nEpisodes are saved individually as memory-mapped files for efficient loading.\n""""""\n\nfrom dataclasses import dataclass\n\nfrom gym3 import types_np\nimport numpy as np\nfrom procgen import ProcgenGym3Env\nimport tyro\nimport json\nimport os\nfrom jasmine_data.utils import save_chunks\n\n\n@dataclass\nclass Args:\n num_episodes_train: int = 10000\n num_episodes_val: int = 500\n num_episodes_test: int = 500\n output_dir: str = ""data/coinrun_episodes""\n min_episode_length: int = 1000\n max_episode_length: int = 1000\n chunk_size: int = 160\n chunks_per_file: int = 100\n seed: int = 0\n\n\nargs = tyro.cli(Args)\nassert (\n args.max_episode_length >= args.min_episode_length\n), ""Maximum episode length must be greater than or equal to minimum episode length.""\n\nif args.min_episode_length < args.chunk_size:\n print(\n ""Warning: Minimum episode length is smaller than chunk size. Note that episodes shorter than the chunk size will be discarded.""\n )\n\n\n# --- Generate episodes ---\ndef generate_episodes(num_episodes, split):\n episode_idx = 0\n episode_metadata = []\n obs_chunks = []\n act_chunks = []\n file_idx = 0\n output_dir_split = os.path.join(args.output_dir, split)\n while episode_idx < num_episodes:\n seed = np.random.randint(0, 10000)\n env = ProcgenGym3Env(num=1, env_name=""coinrun"", start_level=seed)\n\n observations_seq = []\n actions_seq = []\n episode_obs_chunks = []\n episode_act_chunks = []\n\n # --- Run episode ---\n step_t = 0\n first_obs = True\n for step_t in range(args.max_episode_length):\n _, obs, first = env.observe()\n action = types_np.sample(env.ac_space, bshape=(env.num,))\n env.act(action)\n observations_seq.append(obs[""rgb""])\n actions_seq.append(action)\n if len(observations_seq) == args.chunk_size:\n episode_obs_chunks.append(observations_seq)\n episode_act_chunks.append(actions_seq)\n observations_seq = []\n actions_seq = []\n if first and not first_obs:\n break\n first_obs = False\n\n # --- Save episode ---\n if step_t + 1 >= args.min_episode_length:\n if observations_seq:\n if len(observations_seq) < args.chunk_size:\n print(\n f""Warning: Inconsistent chunk_sizes. Episode has {len(observations_seq)} frames, ""\n f""which is smaller than the requested chunk_size: {args.chunk_size}. ""\n ""This might lead to performance degradation during training.""\n )\n episode_obs_chunks.append(observations_seq)\n episode_act_chunks.append(actions_seq)\n\n obs_chunks_data = [\n np.concatenate(seq, axis=0).astype(np.uint8)\n for seq in episode_obs_chunks\n ]\n act_chunks_data = [\n np.concatenate(act, axis=0) for act in episode_act_chunks\n ]\n obs_chunks.extend(obs_chunks_data)\n act_chunks.extend(act_chunks_data)\n\n ep_metadata, file_idx, obs_chunks, act_chunks = save_chunks(\n file_idx, args.chunks_per_file, output_dir_split, obs_chunks, act_chunks\n )\n episode_metadata.extend(ep_metadata)\n\n print(f""Episode {episode_idx} completed, length: {step_t + 1}."")\n episode_idx += 1\n else:\n print(f""Episode too short ({step_t + 1}), resampling..."")\n\n if len(obs_chunks) > 0:\n print(\n f""Warning: Dropping {len(obs_chunks)} chunks for consistent number of chunks per file."",\n ""Consider changing the chunk_size and chunks_per_file parameters to prevent data-loss."",\n )\n\n print(f""Done generating {split} split"")\n return episode_metadata\n\n\ndef get_action_space():\n env = ProcgenGym3Env(num=1, env_name=""coinrun"", start_level=0)\n return env.ac_space.eltype.n\n\n\ndef main():\n # Set random seed and create dataset directories\n np.random.seed(args.seed)\n # --- Generate episodes ---\n train_episode_metadata = generate_episodes(args.num_episodes_train, ""train"")\n val_episode_metadata = generate_episodes(args.num_episodes_val, ""val"")\n test_episode_metadata = generate_episodes(args.num_episodes_test, ""test"")\n\n # --- Save metadata ---\n metadata = {\n ""env"": ""coinrun"",\n ""num_actions"": get_action_space(),\n ""num_episodes_train"": args.num_episodes_train,\n ""num_episodes_val"": args.num_episodes_val,\n ""num_episodes_test"": args.num_episodes_test,\n ""avg_episode_len_train"": np.mean(\n [ep[""avg_seq_len""] for ep in train_episode_metadata]\n ),\n ""avg_episode_len_val"": np.mean(\n [ep[""avg_seq_len""] for ep in val_episode_metadata]\n ),\n ""avg_episode_len_test"": np.mean(\n [ep[""avg_seq_len""] for ep in test_episode_metadata]\n ),\n ""episode_metadata_train"": train_episode_metadata,\n ""episode_metadata_val"": val_episode_metadata,\n ""episode_metadata_test"": test_episode_metadata,\n }\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(f""Done generating dataset."")\n\n\nif __name__ == ""__main__"":\n main()\n",python,tab +2137,3925612,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",1365,0,"",python,selection_mouse +2138,3925884,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",1356,14,"ProcgenGym3Env",python,selection_mouse +2139,3927255,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",1364,0,"",python,selection_mouse +2140,4298899,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3536670.15 task 0: running\r\n",,terminal_output +2141,4299115,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.15\r\nsrun: forcing job termination\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 746, in \r\n main(args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 596, in main\r\n while step < args.num_steps:\r\n ^^^^^^^^^^^^^^^^^^^^^\r\nKeyboardInterrupt\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3536670.15 ON hkn0401 CANCELLED AT 2025-10-02T13:12:12 ***\r\n",,terminal_output +2142,4299202,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3536670.15\r\nsrun: job abort in progress\r\n",,terminal_output +2143,4299447,"TERMINAL",0,0,"]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +2144,4299829,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +2145,4307277,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19502,0,"",python,selection_mouse +2146,4307414,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19501,6,"videos",python,selection_mouse +2147,4308043,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19654,0,"",python,selection_mouse +2148,4308194,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19650,7,"actions",python,selection_mouse +2149,4308688,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19712,0,"",python,selection_mouse +2150,4308751,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19711,1,"t",python,selection_mouse +2151,4309322,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19711,1,"",python,content +2152,4309322,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19564,0,"t",python,content +2153,4310677,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19712,0,"t",python,content +2154,4310678,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19564,1,"",python,content +2155,4316863,"TERMINAL",0,0,"salloc: Job 3536670 has exceeded its time limit and its allocation has been revoked.\nslurmstepd: error: *** STEP 3536670.interactive ON hkn0401 CANCELLED AT 2025-10-02T13:12:30 DUE TO TIME LIMIT ***\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\n",,terminal_output +2156,4327458,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +2157,4333038,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +2158,4337585,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",86,0,"",shellscript,selection_mouse +2159,4340753,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",85,1,"",shellscript,content +2160,4340801,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",85,0,"4",shellscript,content +2161,4340802,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",86,0,"",shellscript,selection_keyboard +2162,4346510,"TERMINAL",0,0,"srun: error: hkn0401: task 0: Killed\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +2163,4354207,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_command +2164,4354220,"TERMINAL",0,0,"]633;C]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +2165,4355631,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_command +2166,4355661,"TERMINAL",0,0,"]633;CSubmitted batch job 3536940\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +2167,4357578,"TERMINAL",0,0,"queue",,terminal_command +2168,4357696,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1991.localdomain: Thu Oct 2 13:13:10 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3532383 accelerat train_to tum_cte0 R 1:33:53\t 1 hkn04253536940 dev_accel train_dy tum_cte0 PD\t0:00\t 1 (None)3533705 large generate tum_cte0 R 1-21:14:21\t 1 hkn1901",,terminal_output +2169,4358744,"TERMINAL",0,0,"142",,terminal_output +2170,4359742,"TERMINAL",0,0,"253",,terminal_output +2171,4360893,"TERMINAL",0,0,"364",,terminal_output +2172,4361812,"TERMINAL",0,0,"475",,terminal_output +2173,4363153,"TERMINAL",0,0,"597",,terminal_output +2174,4363883,"TERMINAL",0,0,"74:008",,terminal_output +2175,4364918,"TERMINAL",0,0,"81 Rhkn04019",,terminal_output +2176,4365989,"TERMINAL",0,0,"92130",,terminal_output +2177,4367109,"TERMINAL",0,0,"20321",,terminal_output +2178,4368067,"TERMINAL",0,0,"1432",,terminal_output +2179,4368878,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +2180,4524295,"TERMINAL",0,0,"queue",,terminal_command +2181,4524350,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1991.localdomain: Thu Oct 2 13:15:57 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3532383 accelerat train_to tum_cte0 R 1:36:40\t 1 hkn04253533705 large generate tum_cte0 R 1-21:17:08\t 1 hkn1901",,terminal_output +2182,4525337,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +2183,4531030,"TERMINAL",0,0,"logs",,terminal_command +2184,4536363,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +2185,4538673,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +2186,4553745,"TERMINAL",0,0,"cd coinrun/dynamics/",,terminal_command +2187,4554585,"TERMINAL",0,0,"ls",,terminal_command +2188,4554627,"TERMINAL",0,0,"]633;Ccausal maskgit train_dyn_single_gpu_3536940.log\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics",,terminal_output +2189,4560924,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/train_dyn_single_gpu_3536940.log",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=00:40:00\n#SBATCH --partition=dev_accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\n#SBATCH --job-name=train_dyn_single_gpu\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670\n\nenv | grep SLURM\n\nexport PYTHONUNBUFFERED=1\n\nsrun python jasmine/train_dynamics.py \\n --save_ckpt \\n --image_height=64 \\n --image_width=64 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=110 \\n --patch_size=4 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=50 \\n --log_checkpoint_interval=2 \\n --dyna_type=maskgit \\n --log \\n --name=coinrun-dyn-dev-$slurm_job_id \\n --tags dyn coinrun dev \\n --entity instant-uv \\n --project jafar \\n --warmup_steps 0 \\n --wsd_decay_steps 0 \\n --num_steps 10 \\n --data_dir $array_records_dir_train \\n --tokenizer_checkpoint $tokenizer_checkpoint \\n --val_data_dir $array_records_dir_val \\n --log_interval 1 \\n --val_interval 10 \\n --eval_full_frame \\n --val_steps 5\nSLURM_JOB_USER=tum_cte0515\nSLURM_TASKS_PER_NODE=4\nSLURM_JOB_UID=999226\nSLURM_TASK_PID=1497666\nSLURM_JOB_GPUS=0\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\nSLURMD_NODENAME=hkn0401\nSLURM_JOB_START_TIME=1759403598\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1759405998\nSLURM_CPUS_ON_NODE=24\nSLURM_JOB_CPUS_PER_NODE=24\nSLURM_GPUS_ON_NODE=1\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=dev_accelerated\nSLURM_TRES_PER_TASK=cpu=5\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=1\nSLURM_JOBID=3536940\nSLURM_JOB_QOS=normal\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=5\nSLURM_NTASKS=4\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_NODELIST=hkn0401\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=4\nSLURM_NNODES=1\nSLURM_SUBMIT_HOST=hkn1991.localdomain\nSLURM_JOB_ID=3536940\nSLURM_NODEID=0\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_dyn_single_gpu\nSLURM_NTASKS_PER_NODE=4\nSLURM_JOB_GID=502226\nSLURM_JOB_NODELIST=hkn0401\nGpuFreq=control_disabled\nW1002 13:13:50.189230 1497902 platform_util.cc:218] unable to create StreamExecutor for CUDA:1: CUDA error: Failed call to cuDeviceGet: CUDA_ERROR_INVALID_DEVICE: invalid device ordinal\nW1002 13:13:50.189230 1497901 platform_util.cc:218] unable to create StreamExecutor for CUDA:3: CUDA error: Failed call to cuDeviceGet: CUDA_ERROR_INVALID_DEVICE: invalid device ordinal\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 812, in backends\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 812, in backends\nW1002 13:13:50.189986 1497903 platform_util.cc:218] unable to create StreamExecutor for CUDA:2: CUDA error: Failed call to cuDeviceGet: CUDA_ERROR_INVALID_DEVICE: invalid device ordinal\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 812, in backends\n backend = _init_backend(platform)\n ^^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 896, in _init_backend\n backend = _init_backend(platform)\n backend = _init_backend(platform)\n ^^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 896, in _init_backend\n ^^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 896, in _init_backend\n backend = registration.factory()\n ^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 559, in make_pjrt_c_api_client\n backend = registration.factory()\n ^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 559, in make_pjrt_c_api_client\n return xla_client.make_c_api_client(\n backend = registration.factory()\n ^^^^^^^^^^^^^^^^^^^^^^\n return xla_client.make_c_api_client(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 559, in make_pjrt_c_api_client\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jaxlib/xla_client.py"", line 156, in make_c_api_client\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jaxlib/xla_client.py"", line 156, in make_c_api_client\n return xla_client.make_c_api_client(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jaxlib/xla_client.py"", line 156, in make_c_api_client\n return _xla.get_c_api_client(\n return _xla.get_c_api_client(\n ^^^^^^^^^^^^^^^^^^^^^^\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 746, in \n main(args)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 350, in main\n return _xla.get_c_api_client(\n num_devices = jax.device_count()\n ^^^^^^^^^^^^^^^^^^^^^^\n ^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 975, in device_count\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 746, in \n ^^^^^^^^^^^^^^^^^^^^^^\njaxlib._jax.XlaRuntimeError: INTERNAL: no supported devices found for platform CUDA\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 746, in \n main(args)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 350, in main\n num_devices = jax.device_count()\n return int(get_backend(backend).device_count())\n main(args)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 350, in main\n ^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 944, in get_backend\n num_devices = jax.device_count()\n ^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 975, in device_count\n ^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 975, in device_count\n return _get_backend_uncached(platform)\n return int(get_backend(backend).device_count())\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 923, in _get_backend_uncached\n ^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 944, in get_backend\n return int(get_backend(backend).device_count())\n ^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 944, in get_backend\n return _get_backend_uncached(platform)\n bs = backends()\n ^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 828, in backends\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 923, in _get_backend_uncached\n return _get_backend_uncached(platform)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n raise RuntimeError(err_msg)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 923, in _get_backend_uncached\n bs = backends()\nRuntimeError: Unable to initialize backend 'cuda': INTERNAL: no supported devices found for platform CUDA (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n ^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 828, in backends\n bs = backends()\n ^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 828, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': INTERNAL: no supported devices found for platform CUDA (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': INTERNAL: no supported devices found for platform CUDA (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 812, in backends\n backend = _init_backend(platform)\n ^^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 896, in _init_backend\n backend = registration.factory()\n ^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 559, in make_pjrt_c_api_client\n return xla_client.make_c_api_client(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jaxlib/xla_client.py"", line 156, in make_c_api_client\n return _xla.get_c_api_client(\n ^^^^^^^^^^^^^^^^^^^^^^\njaxlib._jax.XlaRuntimeError: INTERNAL: Getting local topologies failed: Error 1: GetKeyValue() timed out with key: cuda:local_topology/cuda/1 and duration: 2m\n\nError 2: GetKeyValue() timed out with key: cuda:local_topology/cuda/2 and duration: 2m\n\nError 3: GetKeyValue() timed out with key: cuda:local_topology/cuda/3 and duration: 2m\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 746, in \n main(args)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py"", line 350, in main\n num_devices = jax.device_count()\n ^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 975, in device_count\n return int(get_backend(backend).device_count())\n ^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 944, in get_backend\n return _get_backend_uncached(platform)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 923, in _get_backend_uncached\n bs = backends()\n ^^^^^^^^^^\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/jax/_src/xla_bridge.py"", line 828, in backends\n raise RuntimeError(err_msg)\nRuntimeError: Unable to initialize backend 'cuda': INTERNAL: Getting local topologies failed: Error 1: GetKeyValue() timed out with key: cuda:local_topology/cuda/1 and duration: 2m\n\nError 2: GetKeyValue() timed out with key: cuda:local_topology/cuda/2 and duration: 2m\n\nError 3: GetKeyValue() timed out with key: cuda:local_topology/cuda/3 and duration: 2m (you may need to uninstall the failing plugin package, or set JAX_PLATFORMS=cpu to skip this backend.)\nsrun: error: hkn0401: tasks 1-3: Exited with exit code 1\nsrun: error: hkn0401: task 0: Exited with exit code 1\n\n============================= JOB FEEDBACK =============================\n\nJob ID: 3536940\nCluster: hk\nUser/Group: tum_cte0515/hk-project-p0023960\nAccount: hk-project-p0023960\nState: FAILED (exit code 1)\nPartition: dev_accelerated\nNodes: 1\nCores per node: 24\nNodelist: hkn0401\nCPU Utilized: 00:00:17\nCPU Efficiency: 0.46% of 01:01:12 core-walltime\nJob Wall-clock time: 00:02:33\nStarttime: Thu Oct 2 13:13:18 2025\nEndtime: Thu Oct 2 13:15:51 2025\nMemory Utilized: 1.54 GB\nMemory Efficiency: 1.27% of 121.46 GB (121.46 GB/node)\nThe task which had the largest memory consumption differs by 105.63% from the average task max memory consumption\nEnergy Consumed: 57506 Joule / 15.9738888888889 Watthours\nAverage node power draw: 375.856209150327 Watt\n",log,tab +2190,4561193,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/train_dyn_single_gpu_3536940.log",836,0,"",log,selection_mouse +2191,4562699,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/train_dyn_single_gpu_3536940.log",15529,0,"",log,selection_command +2192,4568747,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +2193,4568748,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",66,0,"",shellscript,selection_mouse +2194,4568940,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",66,0,"A",shellscript,content +2195,4568942,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",67,0,"",shellscript,selection_keyboard +2196,4569070,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",66,1,"",shellscript,content +2197,4569565,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",65,1,"",shellscript,content +2198,4569671,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",65,0,"1",shellscript,content +2199,4569672,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",66,0,"",shellscript,selection_keyboard +2200,4570240,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",65,0,"",shellscript,selection_command +2201,4570911,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",152,0,"",shellscript,selection_mouse +2202,4570911,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",151,0,"",shellscript,selection_command +2203,4571447,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",173,0,"",shellscript,selection_mouse +2204,4571451,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",172,0,"",shellscript,selection_command +2205,4572159,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/train_dyn_single_gpu_3536940.log",0,0,"",log,tab +2206,4572160,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/train_dyn_single_gpu_3536940.log",12757,0,"",log,selection_mouse +2207,4573495,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/train_dyn_single_gpu_3536940.log",12863,0,"",log,selection_mouse +2208,4578208,"TERMINAL",0,0,"dev",,terminal_command +2209,4581648,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_command +2210,4581688,"TERMINAL",0,0,"]633;CSubmitted batch job 3536945\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +2211,4582924,"TERMINAL",0,0,"queue",,terminal_command +2212,4583011,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1991.localdomain: Thu Oct 2 13:16:56 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3532383 accelerat train_to tum_cte0 R 1:37:39\t 1 hkn04253536945 dev_accel train_dy tum_cte0 PD\t0:00\t 1 (None)3533705 large generate tum_cte0 R 1-21:18:07\t 1 hkn1901",,terminal_output +2213,4584071,"TERMINAL",0,0,"7408",,terminal_output +2214,4585212,"TERMINAL",0,0,"819",,terminal_output +2215,4586184,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +2216,4586212,"TERMINAL",0,0,"9210",,terminal_output +2217,4587250,"TERMINAL",0,0,"7:0031",,terminal_output +2218,4587683,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20245,0,"",python,selection_mouse +2219,4587684,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20244,0,"",python,selection_command +2220,4588256,"TERMINAL",0,0,"142",,terminal_output +2221,4588385,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19767,0,"",python,selection_mouse +2222,4588417,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19766,0,"",python,selection_command +2223,4589280,"TERMINAL",0,0,"253",,terminal_output +2224,4590428,"TERMINAL",0,0,"364",,terminal_output +2225,4590621,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19524,0,"",python,selection_mouse +2226,4590654,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19523,0,"",python,selection_command +2227,4591255,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19716,0,"",python,selection_mouse +2228,4591256,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19715,0,"",python,selection_command +2229,4591439,"TERMINAL",0,0,"475",,terminal_output +2230,4592040,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19802,0,"",python,selection_mouse +2231,4592045,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19801,0,"",python,selection_command +2232,4592418,"TERMINAL",0,0,"586",,terminal_output +2233,4593509,"TERMINAL",0,0,"697",,terminal_output +2234,4594504,"TERMINAL",0,0,"7508",,terminal_output +2235,4595553,"TERMINAL",0,0,"819",,terminal_output +2236,4596638,"TERMINAL",0,0,"9220",,terminal_output +2237,4597644,"TERMINAL",0,0,"1031",,terminal_output +2238,4598701,"TERMINAL",0,0,"142",,terminal_output +2239,4599749,"TERMINAL",0,0,"253",,terminal_output +2240,4599884,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19388,0,"",python,selection_mouse +2241,4599884,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19387,0,"",python,selection_command +2242,4600466,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19344,0,"",python,selection_mouse +2243,4600910,"TERMINAL",0,0,"364",,terminal_output +2244,4601038,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19368,0,"",python,selection_mouse +2245,4601186,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19358,16,"dataloader_train",python,selection_mouse +2246,4601866,"TERMINAL",0,0,"486",,terminal_output +2247,4602970,"TERMINAL",0,0,"697",,terminal_output +2248,4603947,"TERMINAL",0,0,"78:008",,terminal_output +2249,4605015,"TERMINAL",0,0,"81 Rhkn04019",,terminal_output +2250,4606012,"TERMINAL",0,0,"92130",,terminal_output +2251,4607152,"TERMINAL",0,0,"20321",,terminal_output +2252,4608103,"TERMINAL",0,0,"1432",,terminal_output +2253,4609171,"TERMINAL",0,0,"2543",,terminal_output +2254,4610209,"TERMINAL",0,0,"3654",,terminal_output +2255,4611401,"TERMINAL",0,0,"4765",,terminal_output +2256,4612358,"TERMINAL",0,0,"5876",,terminal_output +2257,4613367,"TERMINAL",0,0,"6987",,terminal_output +2258,4614418,"TERMINAL",0,0,"71098",,terminal_output +2259,4615428,"TERMINAL",0,0,"81109",,terminal_output +2260,4616473,"TERMINAL",0,0,"92140",,terminal_output +2261,4617537,"TERMINAL",0,0,"30321",,terminal_output +2262,4618568,"TERMINAL",0,0,"1432",,terminal_output +2263,4619721,"TERMINAL",0,0,"2543",,terminal_output +2264,4620665,"TERMINAL",0,0,"3654",,terminal_output +2265,4621902,"TERMINAL",0,0,"4765",,terminal_output +2266,4622750,"TERMINAL",0,0,"5876",,terminal_output +2267,4623813,"TERMINAL",0,0,"6987",,terminal_output +2268,4624846,"TERMINAL",0,0,"721209",,terminal_output +2269,4625892,"TERMINAL",0,0,"92150",,terminal_output +2270,4626940,"TERMINAL",0,0,"40321",,terminal_output +2271,4627987,"TERMINAL",0,0,"1432",,terminal_output +2272,4629036,"TERMINAL",0,0,"2543",,terminal_output +2273,4630077,"TERMINAL",0,0,"3654",,terminal_output +2274,4631127,"TERMINAL",0,0,"4765",,terminal_output +2275,4632185,"TERMINAL",0,0,"5876",,terminal_output +2276,4633228,"TERMINAL",0,0,"6987",,terminal_output +2277,4634272,"TERMINAL",0,0,"73098",,terminal_output +2278,4635385,"TERMINAL",0,0,"81309",,terminal_output +2279,4636654,"TERMINAL",0,0,"9219:00",,terminal_output +2280,4637675,"TERMINAL",0,0,"50321",,terminal_output +2281,4638772,"TERMINAL",0,0,"1432",,terminal_output +2282,4639784,"TERMINAL",0,0,"2543",,terminal_output +2283,4640825,"TERMINAL",0,0,"3765",,terminal_output +2284,4641941,"TERMINAL",0,0,"5876",,terminal_output +2285,4642972,"TERMINAL",0,0,"6987",,terminal_output +2286,4643984,"TERMINAL",0,0,"74098",,terminal_output +2287,4645068,"TERMINAL",0,0,"81409",,terminal_output +2288,4646224,"TERMINAL",0,0,"92110",,terminal_output +2289,4647167,"TERMINAL",0,0,"8:00321",,terminal_output +2290,4648144,"TERMINAL",0,0,"1432",,terminal_output +2291,4649205,"TERMINAL",0,0,"2543",,terminal_output +2292,4650339,"TERMINAL",0,0,"3654",,terminal_output +2293,4651274,"TERMINAL",0,0,"4765",,terminal_output +2294,4652318,"TERMINAL",0,0,"5876",,terminal_output +2295,4653406,"TERMINAL",0,0,"6987",,terminal_output +2296,4654433,"TERMINAL",0,0,"75098",,terminal_output +2297,4655475,"TERMINAL",0,0,"81509",,terminal_output +2298,4656504,"TERMINAL",0,0,"92120",,terminal_output +2299,4657577,"TERMINAL",0,0,"10321",,terminal_output +2300,4658627,"TERMINAL",0,0,"1432",,terminal_output +2301,4659649,"TERMINAL",0,0,"2543",,terminal_output +2302,4660785,"TERMINAL",0,0,"3654",,terminal_output +2303,4661799,"TERMINAL",0,0,"4765",,terminal_output +2304,4662826,"TERMINAL",0,0,"5876",,terminal_output +2305,4663826,"TERMINAL",0,0,"69:0098",,terminal_output +2306,4664898,"TERMINAL",0,0,"811:009",,terminal_output +2307,4666104,"TERMINAL",0,0,"92130",,terminal_output +2308,4666979,"TERMINAL",0,0,"20321",,terminal_output +2309,4668051,"TERMINAL",0,0,"1432",,terminal_output +2310,4669072,"TERMINAL",0,0,"2543",,terminal_output +2311,4670135,"TERMINAL",0,0,"3654",,terminal_output +2312,4671141,"TERMINAL",0,0,"4765",,terminal_output +2313,4672198,"TERMINAL",0,0,"5876",,terminal_output +2314,4673247,"TERMINAL",0,0,"6987",,terminal_output +2315,4674282,"TERMINAL",0,0,"71098",,terminal_output +2316,4675425,"TERMINAL",0,0,"81109",,terminal_output +2317,4676502,"TERMINAL",0,0,"92140",,terminal_output +2318,4677479,"TERMINAL",0,0,"30321",,terminal_output +2319,4679550,"TERMINAL",0,0,"1432",,terminal_output +2320,4679630,"TERMINAL",0,0,"2543",,terminal_output +2321,4680550,"TERMINAL",0,0,"3654",,terminal_output +2322,4681609,"TERMINAL",0,0,"4765",,terminal_output +2323,4682637,"TERMINAL",0,0,"5876",,terminal_output +2324,4683678,"TERMINAL",0,0,"6987",,terminal_output +2325,4684742,"TERMINAL",0,0,"72098",,terminal_output +2326,4685760,"TERMINAL",0,0,"81209",,terminal_output +2327,4686809,"TERMINAL",0,0,"92150",,terminal_output +2328,4687852,"TERMINAL",0,0,"40432",,terminal_output +2329,4688969,"TERMINAL",0,0,"2543",,terminal_output +2330,4689976,"TERMINAL",0,0,"3654",,terminal_output +2331,4691004,"TERMINAL",0,0,"4765",,terminal_output +2332,4691991,"TERMINAL",0,0,"5876",,terminal_output +2333,4693050,"TERMINAL",0,0,"6987",,terminal_output +2334,4694068,"TERMINAL",0,0,"73098",,terminal_output +2335,4695109,"TERMINAL",0,0,"81309",,terminal_output +2336,4696153,"TERMINAL",0,0,"92120:00",,terminal_output +2337,4697229,"TERMINAL",0,0,"50321",,terminal_output +2338,4698275,"TERMINAL",0,0,"1432",,terminal_output +2339,4699282,"TERMINAL",0,0,"2543",,terminal_output +2340,4700321,"TERMINAL",0,0,"3654",,terminal_output +2341,4701363,"TERMINAL",0,0,"4765",,terminal_output +2342,4702408,"TERMINAL",0,0,"5876",,terminal_output +2343,4703442,"TERMINAL",0,0,"6987",,terminal_output +2344,4704518,"TERMINAL",0,0,"74098",,terminal_output +2345,4705646,"TERMINAL",0,0,"81409",,terminal_output +2346,4706609,"TERMINAL",0,0,"92110",,terminal_output +2347,4707654,"TERMINAL",0,0,"9:00321",,terminal_output +2348,4708756,"TERMINAL",0,0,"1432",,terminal_output +2349,4709752,"TERMINAL",0,0,"2543",,terminal_output +2350,4710778,"TERMINAL",0,0,"3654",,terminal_output +2351,4711889,"TERMINAL",0,0,"4765",,terminal_output +2352,4712836,"TERMINAL",0,0,"5987",,terminal_output +2353,4713941,"TERMINAL",0,0,"75098",,terminal_output +2354,4714963,"TERMINAL",0,0,"81509",,terminal_output +2355,4715989,"TERMINAL",0,0,"92120",,terminal_output +2356,4716995,"TERMINAL",0,0,"10321",,terminal_output +2357,4718031,"TERMINAL",0,0,"1432",,terminal_output +2358,4719072,"TERMINAL",0,0,"2543",,terminal_output +2359,4720110,"TERMINAL",0,0,"3654",,terminal_output +2360,4721151,"TERMINAL",0,0,"4765",,terminal_output +2361,4722190,"TERMINAL",0,0,"5876",,terminal_output +2362,4723234,"TERMINAL",0,0,"6987",,terminal_output +2363,4724280,"TERMINAL",0,0,"740:0098",,terminal_output +2364,4725371,"TERMINAL",0,0,"812:009",,terminal_output +2365,4726347,"TERMINAL",0,0,"92130",,terminal_output +2366,4727390,"TERMINAL",0,0,"20321",,terminal_output +2367,4728481,"TERMINAL",0,0,"1432",,terminal_output +2368,4729505,"TERMINAL",0,0,"2543",,terminal_output +2369,4730527,"TERMINAL",0,0,"3654",,terminal_output +2370,4731607,"TERMINAL",0,0,"4765",,terminal_output +2371,4732608,"TERMINAL",0,0,"5876",,terminal_output +2372,4733655,"TERMINAL",0,0,"6987",,terminal_output +2373,4734698,"TERMINAL",0,0,"71098",,terminal_output +2374,4735762,"TERMINAL",0,0,"81109",,terminal_output +2375,4736878,"TERMINAL",0,0,"92140",,terminal_output +2376,4737825,"TERMINAL",0,0,"30432",,terminal_output +2377,4738865,"TERMINAL",0,0,"2543",,terminal_output +2378,4739974,"TERMINAL",0,0,"3654",,terminal_output +2379,4741314,"TERMINAL",0,0,"4765",,terminal_output +2380,4742115,"TERMINAL",0,0,"5876",,terminal_output +2381,4743169,"TERMINAL",0,0,"6987",,terminal_output +2382,4744148,"TERMINAL",0,0,"72098",,terminal_output +2383,4745116,"TERMINAL",0,0,"81209",,terminal_output +2384,4746196,"TERMINAL",0,0,"92150",,terminal_output +2385,4747219,"TERMINAL",0,0,"40321",,terminal_output +2386,4748345,"TERMINAL",0,0,"1432",,terminal_output +2387,4749287,"TERMINAL",0,0,"2543",,terminal_output +2388,4750395,"TERMINAL",0,0,"3654",,terminal_output +2389,4751420,"TERMINAL",0,0,"4765",,terminal_output +2390,4752410,"TERMINAL",0,0,"5876",,terminal_output +2391,4753573,"TERMINAL",0,0,"6987",,terminal_output +2392,4754484,"TERMINAL",0,0,"73098",,terminal_output +2393,4755623,"TERMINAL",0,0,"81309",,terminal_output +2394,4756564,"TERMINAL",0,0,"9211:00",,terminal_output +2395,4757613,"TERMINAL",0,0,"50321",,terminal_output +2396,4758662,"TERMINAL",0,0,"1432",,terminal_output +2397,4759814,"TERMINAL",0,0,"2543",,terminal_output +2398,4760842,"TERMINAL",0,0,"3654",,terminal_output +2399,4761813,"TERMINAL",0,0,"4765",,terminal_output +2400,4762896,"TERMINAL",0,0,"5987",,terminal_output +2401,4763914,"TERMINAL",0,0,"74098",,terminal_output +2402,4764965,"TERMINAL",0,0,"81409",,terminal_output +2403,4765964,"TERMINAL",0,0,"92110",,terminal_output +2404,4767016,"TERMINAL",0,0,"20:00321",,terminal_output +2405,4768072,"TERMINAL",0,0,"1432",,terminal_output +2406,4769110,"TERMINAL",0,0,"2543",,terminal_output +2407,4770155,"TERMINAL",0,0,"3654",,terminal_output +2408,4770329,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +2409,4771209,"TERMINAL",0,0,"4765",,terminal_output +2410,4772305,"TERMINAL",0,0,"5876",,terminal_output +2411,4773283,"TERMINAL",0,0,"6987",,terminal_output +2412,4774329,"TERMINAL",0,0,"75098",,terminal_output +2413,4774932,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19785,0,"",python,selection_mouse +2414,4774933,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19785,1,"i",python,selection_mouse +2415,4775178,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19786,0,"",python,selection_mouse +2416,4775467,"TERMINAL",0,0,"81509",,terminal_output +2417,4775600,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19767,0,"",python,selection_mouse +2418,4775624,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19766,0,"",python,selection_command +2419,4775941,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19798,0,"",python,selection_mouse +2420,4776084,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19788,14,"train_iterator",python,selection_mouse +2421,4776478,"TERMINAL",0,0,"92120",,terminal_output +2422,4776572,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19781,0,"",python,selection_mouse +2423,4776727,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19780,4,"elem",python,selection_mouse +2424,4777088,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19808,0,"",python,selection_mouse +2425,4777118,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19807,0,"",python,selection_command +2426,4777119,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19807,1,")",python,selection_mouse +2427,4777124,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19808,0,"",python,selection_command +2428,4777371,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19802,0,"",python,selection_mouse +2429,4777372,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19801,0,"",python,selection_command +2430,4777533,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19788,14,"train_iterator",python,selection_mouse +2431,4777564,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19789,13,"rain_iterator",python,selection_command +2432,4777565,"TERMINAL",0,0,"10321",,terminal_output +2433,4778267,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19788,14,"train_iterator",python,selection_mouse +2434,4778650,"TERMINAL",0,0,"1432",,terminal_output +2435,4779403,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19791,0,"",python,selection_mouse +2436,4779454,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19788,14,"train_iterator",python,selection_mouse +2437,4779604,"TERMINAL",0,0,"2543",,terminal_output +2438,4780594,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19794,0,"",python,selection_mouse +2439,4780663,"TERMINAL",0,0,"3654",,terminal_output +2440,4781698,"TERMINAL",0,0,"4765",,terminal_output +2441,4782737,"TERMINAL",0,0,"5876",,terminal_output +2442,4783783,"TERMINAL",0,0,"6987",,terminal_output +2443,4784534,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19808,0,"",python,selection_mouse +2444,4784565,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19807,0,"",python,selection_command +2445,4784866,"TERMINAL",0,0,"71:013:009",,terminal_output +2446,4785906,"TERMINAL",0,0,"92130",,terminal_output +2447,4786688,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19376,0,"",python,selection_mouse +2448,4786926,"TERMINAL",0,0,"20321",,terminal_output +2449,4787962,"TERMINAL",0,0,"1432",,terminal_output +2450,4789010,"TERMINAL",0,0,"2543",,terminal_output +2451,4789378,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19387,0,"",python,selection_command +2452,4790051,"TERMINAL",0,0,"3654",,terminal_output +2453,4790420,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,9," {",python,selection_command +2454,4790639,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,71," {\n ""videos"": jax.make_array_from_process_local_data(",python,selection_command +2455,4790782,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,130," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]",python,selection_command +2456,4790980,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,145," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),",python,selection_command +2457,4791097,"TERMINAL",0,0,"4765",,terminal_output +2458,4791382,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,170," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (",python,selection_command +2459,4791406,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,226," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(",python,selection_command +2460,4791443,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,280," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]",python,selection_command +2461,4791473,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,298," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )",python,selection_command +2462,4791509,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,337," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions",python,selection_command +2463,4791634,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,363," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None",python,selection_command +2464,4791792,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,378," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),",python,selection_command +2465,4791939,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,388," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }",python,selection_command +2466,4792086,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,423," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in train_iterator",python,selection_command +2467,4792164,"TERMINAL",0,0,"5876",,terminal_output +2468,4792210,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,429," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, local_data=elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in train_iterator\n )",python,selection_command +2469,4792513,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,430,"",python,content +2470,4792552,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19383,0,"",python,selection_command +2471,4792722,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19358,0,"",python,selection_command +2472,4792905,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19333,0,"",python,selection_command +2473,4793118,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19358,0,"",python,selection_command +2474,4793175,"TERMINAL",0,0,"6987",,terminal_output +2475,4793429,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19378,0,"",python,selection_command +2476,4793623,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19377,1,"",python,content +2477,4794246,"TERMINAL",0,0,"71098",,terminal_output +2478,4794457,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19377,0,"t",python,content +2479,4794458,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19378,0,"",python,selection_keyboard +2480,4794606,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19378,0,"r",python,content +2481,4794608,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,0,"",python,selection_keyboard +2482,4794821,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19379,0,"a",python,content +2483,4794823,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19380,0,"",python,selection_keyboard +2484,4794913,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19380,0,"i",python,content +2485,4794914,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19381,0,"",python,selection_keyboard +2486,4794986,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19381,0,"n",python,content +2487,4794987,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19382,0,"",python,selection_keyboard +2488,4795272,"TERMINAL",0,0,"81109",,terminal_output +2489,4795680,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19382,0,"_iterator",python,content +2490,4796298,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19390,0,"",python,selection_command +2491,4796332,"TERMINAL",0,0,"92140",,terminal_output +2492,4796384,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19416,0,"",python,selection_command +2493,4796661,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19437,0,"",python,selection_command +2494,4797083,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19464,0,"",python,selection_command +2495,4797344,"TERMINAL",0,0,"30321",,terminal_output +2496,4798403,"TERMINAL",0,0,"1432",,terminal_output +2497,4798759,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19465,0,"",python,selection_command +2498,4798986,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19464,1,"",python,content +2499,4799463,"TERMINAL",0,0,"2543",,terminal_output +2500,4800207,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19464,0,"val_iterator",python,content +2501,4800488,"TERMINAL",0,0,"3654",,terminal_output +2502,4800646,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19475,0,"",python,selection_command +2503,4800852,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19489,0,"",python,selection_command +2504,4801541,"TERMINAL",0,0,"4765",,terminal_output +2505,4801987,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,13," {",python,selection_command +2506,4802242,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,79," {\n ""videos"": jax.make_array_from_process_local_data(",python,selection_command +2507,4802546,"TERMINAL",0,0,"5876",,terminal_output +2508,4802900,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,131," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]",python,selection_command +2509,4803381,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,150," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),",python,selection_command +2510,4803418,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,179," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (",python,selection_command +2511,4803419,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,239," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(",python,selection_command +2512,4803453,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,297," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]",python,selection_command +2513,4803484,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,319," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )",python,selection_command +2514,4803608,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,362," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions",python,selection_command +2515,4803609,"TERMINAL",0,0,"6987",,terminal_output +2516,4803776,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,392," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None",python,selection_command +2517,4803913,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,411," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),",python,selection_command +2518,4804067,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,425," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }",python,selection_command +2519,4804213,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,462," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in val_iterator",python,selection_command +2520,4804558,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,472," {\n ""videos"": jax.make_array_from_process_local_data(\n videos_sharding, elem[""videos""]\n ),\n ""actions"": (\n jax.make_array_from_process_local_data(\n actions_sharding, elem[""actions""]\n )\n if args.use_gt_actions\n else None\n ),\n }\n for elem in val_iterator\n )",python,selection_command +2521,4804656,"TERMINAL",0,0,"72098",,terminal_output +2522,4804737,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19477,473,"",python,content +2523,4804742,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19481,0,"",python,selection_command +2524,4805768,"TERMINAL",0,0,"81209",,terminal_output +2525,4806459,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19443,0,"",python,selection_command +2526,4806534,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19422,0,"",python,selection_command +2527,4806679,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19396,0,"",python,selection_command +2528,4806770,"TERMINAL",0,0,"92150",,terminal_output +2529,4807757,"TERMINAL",0,0,"40321",,terminal_output +2530,4808098,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19363,0,"",python,selection_mouse +2531,4808296,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19358,16,"dataloader_train",python,selection_mouse +2532,4808906,"TERMINAL",0,0,"1432",,terminal_output +2533,4809876,"TERMINAL",0,0,"2654",,terminal_output +2534,4810890,"TERMINAL",0,0,"4765",,terminal_output +2535,4811959,"TERMINAL",0,0,"5876",,terminal_output +2536,4813046,"TERMINAL",0,0,"6987",,terminal_output +2537,4814027,"TERMINAL",0,0,"73098",,terminal_output +2538,4815053,"TERMINAL",0,0,"81309",,terminal_output +2539,4816169,"TERMINAL",0,0,"9212:00",,terminal_output +2540,4817128,"TERMINAL",0,0,"50321",,terminal_output +2541,4818165,"TERMINAL",0,0,"1432",,terminal_output +2542,4819233,"TERMINAL",0,0,"2543",,terminal_output +2543,4820260,"TERMINAL",0,0,"3654",,terminal_output +2544,4821298,"TERMINAL",0,0,"4765",,terminal_output +2545,4821651,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19374,0,"",python,selection_command +2546,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20078,3,"",python,content +2547,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20077,0,"o",python,content +2548,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20075,1,"a",python,content +2549,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20069,4,"in_it",python,content +2550,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20068,0,"r",python,content +2551,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20065,2,"",python,content +2552,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19932,3,"",python,content +2553,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19931,0,"o",python,content +2554,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19929,1,"a",python,content +2555,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19923,4,"in_it",python,content +2556,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19922,0,"r",python,content +2557,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19919,2,"",python,content +2558,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19882,3,"",python,content +2559,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19881,0,"o",python,content +2560,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19879,1,"a",python,content +2561,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19873,4,"in_it",python,content +2562,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19872,0,"r",python,content +2563,4821673,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19869,2,"",python,content +2564,4821674,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19555,3,"",python,content +2565,4821674,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19554,0,"o",python,content +2566,4821674,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19552,1,"a",python,content +2567,4821674,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19546,4,"in_it",python,content +2568,4821674,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19545,0,"r",python,content +2569,4821674,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19542,2,"",python,content +2570,4821674,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19371,3,"",python,content +2571,4821674,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19370,0,"o",python,content +2572,4821674,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19368,1,"a",python,content +2573,4821674,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19362,4,"in_it",python,content +2574,4821674,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19361,0,"r",python,content +2575,4821674,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19358,2,"",python,content +2576,4822367,"TERMINAL",0,0,"5876",,terminal_output +2577,4823409,"TERMINAL",0,0,"6987",,terminal_output +2578,4824091,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19451,0,"",python,selection_mouse +2579,4824209,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19445,14,"dataloader_val",python,selection_mouse +2580,4824431,"TERMINAL",0,0,"74098",,terminal_output +2581,4825798,"TERMINAL",0,0,"81409",,terminal_output +2582,4826881,"TERMINAL",0,0,"93211",,terminal_output +2583,4827900,"TERMINAL",0,0,"1:01432",,terminal_output +2584,4828958,"TERMINAL",0,0,"2543",,terminal_output +2585,4829967,"TERMINAL",0,0,"3654",,terminal_output +2586,4831077,"TERMINAL",0,0,"4765",,terminal_output +2587,4831938,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19459,0,"",python,selection_command +2588,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20830,1,"tir",python,content +2589,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20827,2,"",python,content +2590,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20822,3,"_it",python,content +2591,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20819,2,"",python,content +2592,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20817,1,"v",python,content +2593,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20506,1,"tir",python,content +2594,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20503,2,"",python,content +2595,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20498,3,"_it",python,content +2596,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20495,2,"",python,content +2597,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20493,1,"v",python,content +2598,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19458,1,"tir",python,content +2599,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19455,2,"",python,content +2600,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19450,3,"_it",python,content +2601,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19447,2,"",python,content +2602,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19445,1,"v",python,content +2603,4831977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19407,1,"tir",python,content +2604,4831978,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19404,2,"",python,content +2605,4831978,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19399,3,"_it",python,content +2606,4831978,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19396,2,"",python,content +2607,4831978,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19394,1,"v",python,content +2608,4832087,"TERMINAL",0,0,"5876",,terminal_output +2609,4833131,"TERMINAL",0,0,"6987",,terminal_output +2610,4833395,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19485,0,"",python,selection_mouse +2611,4833797,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19450,0,"",python,selection_mouse +2612,4833977,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19443,12,"val_iteratir",python,selection_mouse +2613,4834128,"TERMINAL",0,0,"75098",,terminal_output +2614,4835166,"TERMINAL",0,0,"81509",,terminal_output +2615,4836291,"TERMINAL",0,0,"92120",,terminal_output +2616,4837316,"TERMINAL",0,0,"10321",,terminal_output +2617,4838344,"TERMINAL",0,0,"1432",,terminal_output +2618,4839359,"TERMINAL",0,0,"2543",,terminal_output +2619,4839689,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19455,0,"",python,selection_command +2620,4839772,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20821,1,"o",python,content +2621,4839772,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20499,1,"o",python,content +2622,4839772,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19453,1,"o",python,content +2623,4839773,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19404,1,"o",python,content +2624,4840420,"TERMINAL",0,0,"3654",,terminal_output +2625,4840668,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19433,0,"",python,selection_command +2626,4840877,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19410,0,"",python,selection_command +2627,4841471,"TERMINAL",0,0,"4765",,terminal_output +2628,4841738,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19374,0,"",python,selection_command +2629,4842469,"TERMINAL",0,0,"5876",,terminal_output +2630,4843563,"TERMINAL",0,0,"6987",,terminal_output +2631,4844541,"TERMINAL",0,0,"72:0098",,terminal_output +2632,4844703,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19354,35," train_iterator = train_iterator",python,selection_command +2633,4844982,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19354,59," train_iterator = train_iterator\n val_iterator = None",python,selection_command +2634,4845103,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19354,80," train_iterator = train_iterator\n val_iterator = None\n if val_iterator:",python,selection_command +2635,4845280,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19354,116," train_iterator = train_iterator\n val_iterator = None\n if val_iterator:\n val_iterator = val_iterator",python,selection_command +2636,4845627,"TERMINAL",0,0,"814:009",,terminal_output +2637,4846298,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19354,117,"",python,content +2638,4846303,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19358,0,"",python,selection_command +2639,4846665,"TERMINAL",0,0,"92130",,terminal_output +2640,4847672,"TERMINAL",0,0,"20321",,terminal_output +2641,4848208,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19391,0,"",python,selection_command +2642,4848396,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19440,0,"",python,selection_command +2643,4848689,"TERMINAL",0,0,"1432",,terminal_output +2644,4848895,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19489,0,"",python,selection_command +2645,4848926,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19559,0,"",python,selection_command +2646,4848959,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19623,0,"",python,selection_command +2647,4848990,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19686,0,"",python,selection_command +2648,4849027,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19740,0,"",python,selection_command +2649,4849060,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19812,0,"",python,selection_command +2650,4849101,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19864,0,"",python,selection_command +2651,4849102,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19886,0,"",python,selection_command +2652,4849130,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19919,0,"",python,selection_command +2653,4849166,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19956,0,"",python,selection_command +2654,4849196,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19989,0,"",python,selection_command +2655,4849347,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20043,0,"",python,selection_command +2656,4849517,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",20080,0,"",python,selection_command +2657,4849737,"TERMINAL",0,0,"2543",,terminal_output +2658,4850801,"TERMINAL",0,0,"3654",,terminal_output +2659,4851860,"TERMINAL",0,0,"4876",,terminal_output +2660,4852271,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19945,0,"",python,selection_mouse +2661,4852397,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",19936,14,"train_iterator",python,selection_mouse +2662,4852987,"TERMINAL",0,0,"6987",,terminal_output +2663,4854113,"TERMINAL",0,0,"71098",,terminal_output +2664,4855019,"TERMINAL",0,0,"81109",,terminal_output +2665,4856017,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +2666,4858584,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +2667,4858600,"TERMINAL",0,0,"]633;Csalloc: Pending job allocation 3536954\r\nsalloc: job 3536954 queued and waiting for resources\r\n",,terminal_output +2668,4860581,"TERMINAL",0,0,"bash",,terminal_focus +2669,4862136,"TERMINAL",0,0,"queue",,terminal_command +2670,4862206,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1991.localdomain: Thu Oct 2 13:21:35 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3532383 accelerat train_to tum_cte0 R 1:42:18\t 1 hkn04253536954 dev_accel interact tum_cte0 PD\t0:00\t 1 (QOSMaxJobsPerUserLimit)3536945 dev_accel train_dy tum_cte0 R\t4:17\t 1 hkn04013533705 large generate tum_cte0 R 1-21:22:46\t 1 hkn1901",,terminal_output +2671,4863285,"TERMINAL",0,0,"6987",,terminal_output +2672,4864293,"TERMINAL",0,0,"72098",,terminal_output +2673,4865357,"TERMINAL",0,0,"81209",,terminal_output +2674,4866411,"TERMINAL",0,0,"92150",,terminal_output +2675,4867497,"TERMINAL",0,0,"40321",,terminal_output +2676,4868505,"TERMINAL",0,0,"1432",,terminal_output +2677,4869571,"TERMINAL",0,0,"2543",,terminal_output +2678,4870594,"TERMINAL",0,0,"3654",,terminal_output +2679,4871661,"TERMINAL",0,0,"4765",,terminal_output +2680,4872743,"TERMINAL",0,0,"5876",,terminal_output +2681,4873832,"TERMINAL",0,0,"6987",,terminal_output +2682,4874902,"TERMINAL",0,0,"73098",,terminal_output +2683,4875924,"TERMINAL",0,0,"82313:00",,terminal_output +2684,4876945,"TERMINAL",0,0,"50321",,terminal_output +2685,4877927,"TERMINAL",0,0,"1432",,terminal_output +2686,4883049,"TERMINAL",0,0,"2543365447655876",,terminal_output +2687,4883202,"TERMINAL",0,0,"6987",,terminal_output +2688,4884216,"TERMINAL",0,0,"74098",,terminal_output +2689,4885263,"TERMINAL",0,0,"81409",,terminal_output +2690,4886347,"TERMINAL",0,0,"92110",,terminal_output +2691,4887353,"TERMINAL",0,0,"2:00321",,terminal_output +2692,4888406,"TERMINAL",0,0,"1432",,terminal_output +2693,4889418,"TERMINAL",0,0,"2543",,terminal_output +2694,4890458,"TERMINAL",0,0,"3654",,terminal_output +2695,4891535,"TERMINAL",0,0,"4765",,terminal_output +2696,4892627,"TERMINAL",0,0,"5876",,terminal_output +2697,4893585,"TERMINAL",0,0,"6987",,terminal_output +2698,4894623,"TERMINAL",0,0,"75098",,terminal_output +2699,4895679,"TERMINAL",0,0,"81509",,terminal_output +2700,4896716,"TERMINAL",0,0,"92120",,terminal_output +2701,4897765,"TERMINAL",0,0,"10321",,terminal_output +2702,4898810,"TERMINAL",0,0,"1432",,terminal_output +2703,4899853,"TERMINAL",0,0,"2654",,terminal_output +2704,4900908,"TERMINAL",0,0,"4765",,terminal_output +2705,4901938,"TERMINAL",0,0,"5876",,terminal_output +2706,4903130,"TERMINAL",0,0,"6987",,terminal_output +2707,4904038,"TERMINAL",0,0,"73:0098",,terminal_output +2708,4905088,"TERMINAL",0,0,"815:009",,terminal_output +2709,4906213,"TERMINAL",0,0,"92130",,terminal_output +2710,4907172,"TERMINAL",0,0,"20321",,terminal_output +2711,4908227,"TERMINAL",0,0,"1432",,terminal_output +2712,4909275,"TERMINAL",0,0,"2543",,terminal_output +2713,4910325,"TERMINAL",0,0,"3654",,terminal_output +2714,4911446,"TERMINAL",0,0,"4765",,terminal_output +2715,4912477,"TERMINAL",0,0,"5876",,terminal_output +2716,4913503,"TERMINAL",0,0,"6987",,terminal_output +2717,4914498,"TERMINAL",0,0,"71098",,terminal_output +2718,4915553,"TERMINAL",0,0,"81109",,terminal_output +2719,4916634,"TERMINAL",0,0,"92140",,terminal_output +2720,4917657,"TERMINAL",0,0,"30321",,terminal_output +2721,4918742,"TERMINAL",0,0,"1432",,terminal_output +2722,4919735,"TERMINAL",0,0,"2543",,terminal_output +2723,4920877,"TERMINAL",0,0,"3654",,terminal_output +2724,4921825,"TERMINAL",0,0,"4876",,terminal_output +2725,4922976,"TERMINAL",0,0,"6987",,terminal_output +2726,4923947,"TERMINAL",0,0,"72098",,terminal_output +2727,4925177,"TERMINAL",0,0,"81209",,terminal_output +2728,4926049,"TERMINAL",0,0,"92150",,terminal_output +2729,4927057,"TERMINAL",0,0,"40321",,terminal_output +2730,4928117,"TERMINAL",0,0,"1432",,terminal_output +2731,4929189,"TERMINAL",0,0,"2543",,terminal_output +2732,4930296,"TERMINAL",0,0,"3654",,terminal_output +2733,4931279,"TERMINAL",0,0,"4765",,terminal_output +2734,4932346,"TERMINAL",0,0,"5876",,terminal_output +2735,4933340,"TERMINAL",0,0,"6987",,terminal_output +2736,4934385,"TERMINAL",0,0,"73098",,terminal_output +2737,4935451,"TERMINAL",0,0,"81309",,terminal_output +2738,4936497,"TERMINAL",0,0,"9214:00",,terminal_output +2739,4937558,"TERMINAL",0,0,"50321",,terminal_output +2740,4938597,"TERMINAL",0,0,"1432",,terminal_output +2741,4939641,"TERMINAL",0,0,"2543",,terminal_output +2742,4940714,"TERMINAL",0,0,"3654",,terminal_output +2743,4941739,"TERMINAL",0,0,"4765",,terminal_output +2744,4942774,"TERMINAL",0,0,"5876",,terminal_output +2745,4943862,"TERMINAL",0,0,"64098",,terminal_output +2746,4944866,"TERMINAL",0,0,"81409",,terminal_output +2747,4945963,"TERMINAL",0,0,"92110",,terminal_output +2748,4946953,"TERMINAL",0,0,"3:00321",,terminal_output +2749,4948039,"TERMINAL",0,0,"1432",,terminal_output +2750,4949036,"TERMINAL",0,0,"2543",,terminal_output +2751,4950083,"TERMINAL",0,0,"3654",,terminal_output +2752,4951128,"TERMINAL",0,0,"4765",,terminal_output +2753,4952164,"TERMINAL",0,0,"5876",,terminal_output +2754,4953211,"TERMINAL",0,0,"6987",,terminal_output +2755,4954256,"TERMINAL",0,0,"75098",,terminal_output +2756,4955383,"TERMINAL",0,0,"81509",,terminal_output +2757,4956322,"TERMINAL",0,0,"92120",,terminal_output +2758,4957437,"TERMINAL",0,0,"10321",,terminal_output +2759,4958456,"TERMINAL",0,0,"1432",,terminal_output +2760,4959480,"TERMINAL",0,0,"2543",,terminal_output +2761,4960515,"TERMINAL",0,0,"3654",,terminal_output +2762,4961536,"TERMINAL",0,0,"4765",,terminal_output +2763,4962599,"TERMINAL",0,0,"5876",,terminal_output +2764,4963677,"TERMINAL",0,0,"6987",,terminal_output +2765,4964665,"TERMINAL",0,0,"74:0098",,terminal_output +2766,4965725,"TERMINAL",0,0,"816:009",,terminal_output +2767,4966764,"TERMINAL",0,0,"92130",,terminal_output +2768,4967791,"TERMINAL",0,0,"20321",,terminal_output +2769,4968899,"TERMINAL",0,0,"1543",,terminal_output +2770,4969863,"TERMINAL",0,0,"3654",,terminal_output +2771,4970903,"TERMINAL",0,0,"4765",,terminal_output +2772,4971972,"TERMINAL",0,0,"5876",,terminal_output +2773,4972998,"TERMINAL",0,0,"6987",,terminal_output +2774,4974127,"TERMINAL",0,0,"71098",,terminal_output +2775,4975059,"TERMINAL",0,0,"81109",,terminal_output +2776,4976108,"TERMINAL",0,0,"92140",,terminal_output +2777,4977137,"TERMINAL",0,0,"30321",,terminal_output +2778,4978197,"TERMINAL",0,0,"1432",,terminal_output +2779,4979225,"TERMINAL",0,0,"2543",,terminal_output +2780,4980366,"TERMINAL",0,0,"3654",,terminal_output +2781,4981321,"TERMINAL",0,0,"4765",,terminal_output +2782,4982363,"TERMINAL",0,0,"5876",,terminal_output +2783,4983450,"TERMINAL",0,0,"6987",,terminal_output +2784,4984566,"TERMINAL",0,0,"72098",,terminal_output +2785,4985593,"TERMINAL",0,0,"81209",,terminal_output +2786,4986554,"TERMINAL",0,0,"92150",,terminal_output +2787,4987598,"TERMINAL",0,0,"40321",,terminal_output +2788,4988650,"TERMINAL",0,0,"1432",,terminal_output +2789,4989775,"TERMINAL",0,0,"2543",,terminal_output +2790,4990812,"TERMINAL",0,0,"3654",,terminal_output +2791,4991785,"TERMINAL",0,0,"4765",,terminal_output +2792,4992864,"TERMINAL",0,0,"5987",,terminal_output +2793,4993887,"TERMINAL",0,0,"73098",,terminal_output +2794,4994908,"TERMINAL",0,0,"81309",,terminal_output +2795,4996038,"TERMINAL",0,0,"9215:00",,terminal_output +2796,4997061,"TERMINAL",0,0,"50321",,terminal_output +2797,4998086,"TERMINAL",0,0,"1432",,terminal_output +2798,4999123,"TERMINAL",0,0,"2543",,terminal_output +2799,5000176,"TERMINAL",0,0,"3654",,terminal_output +2800,5001220,"TERMINAL",0,0,"4765",,terminal_output +2801,5002266,"TERMINAL",0,0,"5876",,terminal_output +2802,5003308,"TERMINAL",0,0,"6987",,terminal_output +2803,5004432,"TERMINAL",0,0,"74098",,terminal_output +2804,5005460,"TERMINAL",0,0,"81409",,terminal_output +2805,5006453,"TERMINAL",0,0,"92110",,terminal_output +2806,5007480,"TERMINAL",0,0,"4:00321",,terminal_output +2807,5008531,"TERMINAL",0,0,"1432",,terminal_output +2808,5009570,"TERMINAL",0,0,"2543",,terminal_output +2809,5010618,"TERMINAL",0,0,"3654",,terminal_output +2810,5011706,"TERMINAL",0,0,"4765",,terminal_output +2811,5012777,"TERMINAL",0,0,"5876",,terminal_output +2812,5013774,"TERMINAL",0,0,"6987",,terminal_output +2813,5014803,"TERMINAL",0,0,"75098",,terminal_output +2814,5015855,"TERMINAL",0,0,"825120",,terminal_output +2815,5016902,"TERMINAL",0,0,"10321",,terminal_output +2816,5018091,"TERMINAL",0,0,"1432",,terminal_output +2817,5019080,"TERMINAL",0,0,"2543",,terminal_output +2818,5020100,"TERMINAL",0,0,"3654",,terminal_output +2819,5021127,"TERMINAL",0,0,"4765",,terminal_output +2820,5022147,"TERMINAL",0,0,"5876",,terminal_output +2821,5023178,"TERMINAL",0,0,"6987",,terminal_output +2822,5024195,"TERMINAL",0,0,"75:0098",,terminal_output +2823,5025238,"TERMINAL",0,0,"817:009",,terminal_output +2824,5026344,"TERMINAL",0,0,"92130",,terminal_output +2825,5027323,"TERMINAL",0,0,"20321",,terminal_output +2826,5028392,"TERMINAL",0,0,"1432",,terminal_output +2827,5029421,"TERMINAL",0,0,"2543",,terminal_output +2828,5030453,"TERMINAL",0,0,"3654",,terminal_output +2829,5031489,"TERMINAL",0,0,"4765",,terminal_output +2830,5032544,"TERMINAL",0,0,"5876",,terminal_output +2831,5033588,"TERMINAL",0,0,"6987",,terminal_output +2832,5034619,"TERMINAL",0,0,"71098",,terminal_output +2833,5035786,"TERMINAL",0,0,"81109",,terminal_output +2834,5036761,"TERMINAL",0,0,"92140",,terminal_output +2835,5037747,"TERMINAL",0,0,"30321",,terminal_output +2836,5038790,"TERMINAL",0,0,"1432",,terminal_output +2837,5039737,"TERMINAL",0,0,"salloc",,terminal_focus +2838,5039882,"TERMINAL",0,0,"2654",,terminal_output +2839,5040888,"TERMINAL",0,0,"4765",,terminal_output +2840,5041974,"TERMINAL",0,0,"5876",,terminal_output +2841,5043020,"TERMINAL",0,0,"6987",,terminal_output +2842,5043988,"TERMINAL",0,0,"72098",,terminal_output +2843,5045121,"TERMINAL",0,0,"81209",,terminal_output +2844,5046110,"TERMINAL",0,0,"92150",,terminal_output +2845,5047135,"TERMINAL",0,0,"40321",,terminal_output +2846,5048264,"TERMINAL",0,0,"1432",,terminal_output +2847,5049206,"TERMINAL",0,0,"2543",,terminal_output +2848,5050245,"TERMINAL",0,0,"3654",,terminal_output +2849,5051285,"TERMINAL",0,0,"4765",,terminal_output +2850,5052360,"TERMINAL",0,0,"5876",,terminal_output +2851,5053364,"TERMINAL",0,0,"6987",,terminal_output +2852,5054510,"TERMINAL",0,0,"73098",,terminal_output +2853,5055643,"TERMINAL",0,0,"81309",,terminal_output +2854,5056486,"TERMINAL",0,0,"9216:00",,terminal_output +2855,5057534,"TERMINAL",0,0,"50321",,terminal_output +2856,5058564,"TERMINAL",0,0,"1432",,terminal_output +2857,5059603,"TERMINAL",0,0,"2543",,terminal_output +2858,5060642,"TERMINAL",0,0,"3654",,terminal_output +2859,5061715,"TERMINAL",0,0,"4765",,terminal_output +2860,5062724,"TERMINAL",0,0,"5876",,terminal_output +2861,5063782,"TERMINAL",0,0,"6987",,terminal_output +2862,5064805,"TERMINAL",0,0,"74098",,terminal_output +2863,5065875,"TERMINAL",0,0,"824110",,terminal_output +2864,5066892,"TERMINAL",0,0,"5:00321",,terminal_output +2865,5068022,"TERMINAL",0,0,"1432",,terminal_output +2866,5069052,"TERMINAL",0,0,"2543",,terminal_output +2867,5070051,"TERMINAL",0,0,"3654",,terminal_output +2868,5071205,"TERMINAL",0,0,"4765",,terminal_output +2869,5072123,"TERMINAL",0,0,"5876",,terminal_output +2870,5073152,"TERMINAL",0,0,"6987",,terminal_output +2871,5074280,"TERMINAL",0,0,"75098",,terminal_output +2872,5075238,"TERMINAL",0,0,"81509",,terminal_output +2873,5076271,"TERMINAL",0,0,"92120",,terminal_output +2874,5077346,"TERMINAL",0,0,"10321",,terminal_output +2875,5078473,"TERMINAL",0,0,"1432",,terminal_output +2876,5079502,"TERMINAL",0,0,"2543",,terminal_output +2877,5080526,"TERMINAL",0,0,"3654",,terminal_output +2878,5081483,"TERMINAL",0,0,"4765",,terminal_output +2879,5082525,"TERMINAL",0,0,"5876",,terminal_output +2880,5083591,"TERMINAL",0,0,"6987",,terminal_output +2881,5084596,"TERMINAL",0,0,"76:0098",,terminal_output +2882,5085738,"TERMINAL",0,0,"818:009",,terminal_output +2883,5086942,"TERMINAL",0,0,"92130",,terminal_output +2884,5087723,"TERMINAL",0,0,"20321",,terminal_output +2885,5088815,"TERMINAL",0,0,"1432",,terminal_output +2886,5089811,"TERMINAL",0,0,"2543",,terminal_output +2887,5090841,"TERMINAL",0,0,"3765",,terminal_output +2888,5091895,"TERMINAL",0,0,"5876",,terminal_output +2889,5093010,"TERMINAL",0,0,"6987",,terminal_output +2890,5094036,"TERMINAL",0,0,"71098",,terminal_output +2891,5094997,"TERMINAL",0,0,"81109",,terminal_output +2892,5096110,"TERMINAL",0,0,"92140",,terminal_output +2893,5097105,"TERMINAL",0,0,"30321",,terminal_output +2894,5098118,"TERMINAL",0,0,"1432",,terminal_output +2895,5099161,"TERMINAL",0,0,"2543",,terminal_output +2896,5100200,"TERMINAL",0,0,"3654",,terminal_output +2897,5101238,"TERMINAL",0,0,"4765",,terminal_output +2898,5102295,"TERMINAL",0,0,"5876",,terminal_output +2899,5103316,"TERMINAL",0,0,"6987",,terminal_output +2900,5104353,"TERMINAL",0,0,"72098",,terminal_output +2901,5105389,"TERMINAL",0,0,"81209",,terminal_output +2902,5106428,"TERMINAL",0,0,"92150",,terminal_output +2903,5107464,"TERMINAL",0,0,"40321",,terminal_output +2904,5108501,"TERMINAL",0,0,"1432",,terminal_output +2905,5109569,"TERMINAL",0,0,"2543",,terminal_output +2906,5110624,"TERMINAL",0,0,"3654",,terminal_output +2907,5111651,"TERMINAL",0,0,"4765",,terminal_output +2908,5112659,"TERMINAL",0,0,"5876",,terminal_output +2909,5113830,"TERMINAL",0,0,"6987",,terminal_output +2910,5114760,"TERMINAL",0,0,"73098",,terminal_output +2911,5115776,"TERMINAL",0,0,"81309",,terminal_output +2912,5116595,"TERMINAL",0,0,"watch",,terminal_focus +2913,5116824,"TERMINAL",0,0,"9327:01",,terminal_output +2914,5117779,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670",,terminal_output +2915,5119551,"TERMINAL",0,0,"scancel 3536945",,terminal_command +2916,5119557,"TERMINAL",0,0,"]633;C]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670",,terminal_output +2917,5120755,"TERMINAL",0,0,"dev",,terminal_command +2918,5122132,"TERMINAL",0,0,"queue",,terminal_command +2919,5122184,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1991.localdomain: Thu Oct 2 13:25:55 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3532383 accelerat train_to tum_cte0 R 1:46:38\t 1 hkn04253536945 dev_accel train_dy tum_cte0 CG\t8:34\t 1 hkn04013536954 dev_accel interact tum_cte0 PD\t0:00\t 1 (QOSMaxJobsPerUserLimit)3533705 large generate tum_cte0 R 1-21:27:06\t 1 hkn1901",,terminal_output +2920,5123226,"TERMINAL",0,0,"697",,terminal_output +2921,5124258,"TERMINAL",0,0,"7408",,terminal_output +2922,5125298,"TERMINAL",0,0,"819",,terminal_output +2923,5126337,"TERMINAL",0,0,"9210",,terminal_output +2924,5127396,"TERMINAL",0,0,"6:0031",,terminal_output +2925,5128746,"TERMINAL",0,0,"142",,terminal_output +2926,5129822,"TERMINAL",0,0,"253",,terminal_output +2927,5130903,"TERMINAL",0,0,"375",,terminal_output +2928,5131015,"TERMINAL",0,0,"salloc",,terminal_focus +2929,5131861,"TERMINAL",0,0,"586",,terminal_output +2930,5132207,"TERMINAL",0,0,"so",,terminal_output +2931,5132313,"TERMINAL",0,0,"u",,terminal_output +2932,5132409,"TERMINAL",0,0,"r",,terminal_output +2933,5132622,"TERMINAL",0,0,"c",,terminal_output +2934,5132713,"TERMINAL",0,0,"e",,terminal_output +2935,5132947,"TERMINAL",0,0,"697",,terminal_output +2936,5132962,"TERMINAL",0,0," .",,terminal_output +2937,5133080,"TERMINAL",0,0,"v",,terminal_output +2938,5133123,"TERMINAL",0,0,"\t",,terminal_output +2939,5133405,"TERMINAL",0,0,"b",,terminal_output +2940,5133562,"TERMINAL",0,0,"\t",,terminal_output +2941,5133867,"TERMINAL",0,0,"a",,terminal_output +2942,5133949,"TERMINAL",0,0,"c",,terminal_output +2943,5133978,"TERMINAL",0,0,"7508",,terminal_output +2944,5134239,"TERMINAL",0,0,"\t",,terminal_output +2945,5135003,"TERMINAL",0,0,"819",,terminal_output +2946,5135590,"TERMINAL",0,0,"\r\n",,terminal_output +2947,5136133,"TERMINAL",0,0,"9220",,terminal_output +2948,5137003,"TERMINAL",0,0,"s",,terminal_output +2949,5137068,"TERMINAL",0,0,"1031",,terminal_output +2950,5137126,"TERMINAL",0,0,"h",,terminal_output +2951,5137183,"TERMINAL",0,0," ",,terminal_output +2952,5137680,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +2953,5138269,"TERMINAL",0,0,"142",,terminal_output +2954,5138321,"TERMINAL",0,0,"\r\n",,terminal_output +2955,5139149,"TERMINAL",0,0,"253",,terminal_output +2956,5140222,"TERMINAL",0,0,"364",,terminal_output +2957,5141221,"TERMINAL",0,0,"475",,terminal_output +2958,5142250,"TERMINAL",0,0,"586",,terminal_output +2959,5143290,"TERMINAL",0,0,"697",,terminal_output +2960,5144345,"TERMINAL",0,0,"77:00Resources)8",,terminal_output +2961,5145376,"TERMINAL",0,0,"819",,terminal_output +2962,5146411,"TERMINAL",0,0,"9230",,terminal_output +2963,5147450,"TERMINAL",0,0,"2031",,terminal_output +2964,5148513,"TERMINAL",0,0,"142",,terminal_output +2965,5149538,"TERMINAL",0,0,"253",,terminal_output +2966,5150608,"TERMINAL",0,0,"364",,terminal_output +2967,5151624,"TERMINAL",0,0,"475",,terminal_output +2968,5152650,"TERMINAL",0,0,"\r586",,terminal_output +2969,5153716,"TERMINAL",0,0,"697",,terminal_output +2970,5154725,"TERMINAL",0,0,"7108",,terminal_output +2971,5155784,"TERMINAL",0,0,"819",,terminal_output +2972,5156839,"TERMINAL",0,0,"9240",,terminal_output +2973,5157859,"TERMINAL",0,0,"3042",,terminal_output +2974,5159063,"TERMINAL",0,0,"253",,terminal_output +2975,5159928,"TERMINAL",0,0,"364",,terminal_output +2976,5161005,"TERMINAL",0,0,"475",,terminal_output +2977,5162030,"TERMINAL",0,0,"586",,terminal_output +2978,5163052,"TERMINAL",0,0,"697",,terminal_output +2979,5164091,"TERMINAL",0,0,"watch",,terminal_focus +2980,5164121,"TERMINAL",0,0,"7208",,terminal_output +2981,5164428,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +2982,5165331,"TERMINAL",0,0,"idling",,terminal_command +2983,5165407,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1991.localdomain: Thu Oct 2 13:26:38 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 132 nodes idle\rPartition dev_accelerated:\t 1 nodes idle\rPartition accelerated:\t 9 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 5 nodes idle\rPartition accelerated-h200:\t 2 nodes idle",,terminal_output +2984,5166439,"TERMINAL",0,0,"9",,terminal_output +2985,5167506,"TERMINAL",0,0,"40",,terminal_output +2986,5168507,"TERMINAL",0,0,"1",,terminal_output +2987,5169554,"TERMINAL",0,0,"2",,terminal_output +2988,5170627,"TERMINAL",0,0,"3",,terminal_output +2989,5171683,"TERMINAL",0,0,"4",,terminal_output +2990,5172652,"TERMINAL",0,0,"5",,terminal_output +2991,5173239,"TERMINAL",0,0,"salloc",,terminal_focus +2992,5173751,"TERMINAL",0,0,"6",,terminal_output +2993,5173942,"TERMINAL",0,0,"^Csalloc: Job allocation 3536954 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +2994,5174736,"TERMINAL",0,0,"7",,terminal_output +2995,5175234,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +2996,5175311,"TERMINAL",0,0,"]633;Csalloc: Pending job allocation 3536975\r\nsalloc: job 3536975 queued and waiting for resources\r\n",,terminal_output +2997,5175763,"TERMINAL",0,0,"8",,terminal_output +2998,5176423,"TERMINAL",0,0,"^Csalloc: Job allocation 3536975 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +2999,5176800,"TERMINAL",0,0,"9",,terminal_output +3000,5177904,"TERMINAL",0,0,"50",,terminal_output +3001,5178926,"TERMINAL",0,0,"2",,terminal_output +3002,5179923,"TERMINAL",0,0,"3",,terminal_output +3003,5180965,"TERMINAL",0,0,"4",,terminal_output +3004,5182073,"TERMINAL",0,0,"5",,terminal_output +3005,5182494,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +3006,5182547,"TERMINAL",0,0,"]633;Csalloc: Pending job allocation 3536976\r\nsalloc: job 3536976 queued and waiting for resources\r\n",,terminal_output +3007,5183166,"TERMINAL",0,0,"6",,terminal_output +3008,5189427,"TERMINAL",0,0,"7",,terminal_output +3009,5189486,"TERMINAL",0,0,"^Csalloc: Job allocation 3536976 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +3010,5189502,"TERMINAL",0,0,"897:0012",,terminal_output +3011,5190332,"TERMINAL",0,0,"3",,terminal_output +3012,5191366,"TERMINAL",0,0,"4",,terminal_output +3013,5192403,"TERMINAL",0,0,"5",,terminal_output +3014,5193445,"TERMINAL",0,0,"6",,terminal_output +3015,5194512,"TERMINAL",0,0,"7",,terminal_output +3016,5195174,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +3017,5195236,"TERMINAL",0,0,"]633;Csalloc: Pending job allocation 3536977\r\nsalloc: job 3536977 queued and waiting for resources\r\n",,terminal_output +3018,5195524,"TERMINAL",0,0,"8",,terminal_output +3019,5196254,"TERMINAL",0,0,"^Csalloc: Job allocation 3536977 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +3020,5196577,"TERMINAL",0,0,"9",,terminal_output +3021,5197657,"TERMINAL",0,0,"10",,terminal_output +3022,5198664,"TERMINAL",0,0,"1",,terminal_output +3023,5199688,"TERMINAL",0,0,"2",,terminal_output +3024,5200741,"TERMINAL",0,0,"3",,terminal_output +3025,5201744,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_command +3026,5201783,"TERMINAL",0,0,"]633;CSubmitted batch job 3536978\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +3027,5201798,"TERMINAL",0,0,"4",,terminal_output +3028,5202807,"TERMINAL",0,0,"5",,terminal_output +3029,5203785,"TERMINAL",0,0,"Every 1.0s: sinfo_t_idlehkn1991.localdomain: Thu Oct 2 13:27:16 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 132 nodes idle\rPartition dev_accelerated:\t 1 nodes idle\rPartition accelerated:\t 9 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 5 nodes idle\rPartition accelerated-h200:\t 2 nodes idle",,terminal_output +3030,5204800,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/train_dynamics.py",0,0,"",python,tab +3031,5204963,"TERMINAL",0,0,"7",,terminal_output +3032,5205907,"TERMINAL",0,0,"8",,terminal_output +3033,5206705,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",0,0,"",python,tab +3034,5206891,"TERMINAL",0,0,"20",,terminal_output +3035,5207935,"TERMINAL",0,0,"1",,terminal_output +3036,5208322,"data/jasmine_data/coinrun/generate_coinrun_dataset.py",0,0,"",python,tab +3037,5208990,"TERMINAL",0,0,"2",,terminal_output +3038,5209994,"TERMINAL",0,0,"3",,terminal_output +3039,5211061,"TERMINAL",0,0,"4",,terminal_output +3040,5212141,"TERMINAL",0,0,"5",,terminal_output +3041,5213098,"TERMINAL",0,0,"6",,terminal_output +3042,5214162,"TERMINAL",0,0,"7",,terminal_output +3043,5215241,"TERMINAL",0,0,"8",,terminal_output +3044,5216220,"TERMINAL",0,0,"9",,terminal_output +3045,5217268,"TERMINAL",0,0,"30",,terminal_output +3046,5218371,"jasmine/train_dynamics.py",0,0,"",python,tab +3047,5218372,"jasmine/train_dynamics.py",11744,0,"",python,selection_command +3048,5218406,"TERMINAL",0,0,"1",,terminal_output +3049,5219387,"TERMINAL",0,0,"2",,terminal_output +3050,5220446,"TERMINAL",0,0,"3",,terminal_output +3051,5221461,"TERMINAL",0,0,"4",,terminal_output +3052,5222511,"TERMINAL",0,0,"5",,terminal_output +3053,5223575,"TERMINAL",0,0,"6",,terminal_output +3054,5224559,"TERMINAL",0,0,"7",,terminal_output +3055,5225625,"TERMINAL",0,0,"8",,terminal_output +3056,5226645,"TERMINAL",0,0,"9",,terminal_output +3057,5227679,"TERMINAL",0,0,"40",,terminal_output +3058,5228785,"TERMINAL",0,0,"1",,terminal_output +3059,5229813,"TERMINAL",0,0,"2",,terminal_output +3060,5230796,"TERMINAL",0,0,"3",,terminal_output +3061,5231833,"TERMINAL",0,0,"4",,terminal_output +3062,5232894,"TERMINAL",0,0,"6",,terminal_output +3063,5233928,"TERMINAL",0,0,"7",,terminal_output +3064,5234105,"jasmine/train_dynamics.py",20271,0,"",python,selection_mouse +3065,5235146,"TERMINAL",0,0,"8",,terminal_output +3066,5236072,"TERMINAL",0,0,"9",,terminal_output +3067,5237050,"TERMINAL",0,0,"50",,terminal_output +3068,5238036,"jasmine/train_dynamics.py",20288,0,"",python,selection_mouse +3069,5238036,"jasmine/train_dynamics.py",20287,0,"",python,selection_command +3070,5238100,"TERMINAL",0,0,"12",,terminal_output +3071,5238614,"jasmine/train_dynamics.py",20288,0,"",python,selection_mouse +3072,5238616,"jasmine/train_dynamics.py",20287,0,"",python,selection_command +3073,5239140,"TERMINAL",0,0,"2",,terminal_output +3074,5240168,"TERMINAL",0,0,"3",,terminal_output +3075,5240312,"jasmine/train_dynamics.py",20257,32,"",python,content +3076,5241203,"TERMINAL",0,0,"4",,terminal_output +3077,5242249,"TERMINAL",0,0,"5",,terminal_output +3078,5243292,"TERMINAL",0,0,"6",,terminal_output +3079,5244335,"TERMINAL",0,0,"7",,terminal_output +3080,5245373,"TERMINAL",0,0,"8",,terminal_output +3081,5246445,"TERMINAL",0,0,"9",,terminal_output +3082,5246502,"jasmine/train_dynamics.py",20342,0,"",python,selection_mouse +3083,5247480,"TERMINAL",0,0,"8:00",,terminal_output +3084,5248509,"TERMINAL",0,0,"1",,terminal_output +3085,5249574,"TERMINAL",0,0,"2",,terminal_output +3086,5250607,"TERMINAL",0,0,"3",,terminal_output +3087,5251482,"TERMINAL",0,0,"watch",,terminal_focus +3088,5251649,"TERMINAL",0,0,"4",,terminal_output +3089,5252745,"TERMINAL",0,0,"5",,terminal_output +3090,5253474,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",0,0,"",python,tab +3091,5253755,"TERMINAL",0,0,"6",,terminal_output +3092,5254784,"TERMINAL",0,0,"7",,terminal_output +3093,5255927,"TERMINAL",0,0,"8",,terminal_output +3094,5256825,"TERMINAL",0,0,"9",,terminal_output +3095,5257169,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +3096,5259544,"TERMINAL",0,0,"git status",,terminal_command +3097,5259603,"TERMINAL",0,0,"]633;COn branch ablation/use-pytorch-dataloader\r\nLast commands done (2 commands done):\r\n pick ba37453 feat: generate coinrun dataset with val split\r\n pick faadd10 feat: implemented validation loss for all three models\r\nNext commands to do (26 remaining commands):\r\n pick 9a17dbb fix: pass val data path to dataloader\r\n pick 6e69cdb fix typo in image logging\r\n (use ""git rebase --edit-todo"" to view and edit)\r\nYou are currently editing a commit while rebasing branch 'gt-actions' on 'c7522f2'.\r\n (use ""git commit --amend"" to amend the current commit)\r\n (use ""git rebase --continue"" once you are satisfied with your changes)\r\n\r\nChanges to be committed:\r\n (use ""git restore --staged ..."" to unstage)\r\n\tmodified: jasmine/train_dynamics.py\r\n\r\nChanges not staged for commit:\r\n (use ""git add ..."" to update what will be committed)\r\n (use ""git restore ..."" to discard changes in working directory)\r\n\tmodified: jasmine/utils/dataloader_torch.py\r\n\r\nUntracked files:\r\n (use ""git add ..."" to include in what will be committed)\r\n\tali-old-branch.diff\r\n\tdata/_vizdoom.ini\r\n\tdata/data/\r\n\tdata/jasmine_data/vizdoom/\r\n\tdata/uv.lock\r\n\tdiff.diff\r\n\tdiff2.diff\r\n\tinput_pipeline/\r\n\tkiller.sh\r\n\tkiller_partition.sh\r\n\tlog.log\r\n\tmessage.md\r\n\toverfit_dir.zip\r\n\trequirements-franz.txt\r\n\tsamples/\r\n\tscripts_cremers/\r\n\tslurm/\r\n\ttest.py\r\n\tutils/\r\n\tuv.lock\r\n\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +3098,5263245,"input_pipeline/generate_breakout_dataset_agent.py",0,0,"""""""\nTrain a DQN agent on MinAtar-Breakout and collect episodes to build a dataset.\nEpisodes are chunked and stored using save_chunks.\n""""""\n\nfrom dataclasses import dataclass\nimport numpy as np\nimport os\nimport json\nimport tyro\nfrom typing import List, Tuple\n\nfrom minatar import Environment # pip install minatar\nfrom gym import spaces\nimport gym\nfrom stable_baselines3 import DQN # pip install stable-baselines3[extra]\n\nfrom utils import save_chunks # reuse your existing utility\n\n\n# ---- Wrap MinAtar in a Gym API ----\nclass MinAtarBreakout(gym.Env):\n def __init__(self):\n super().__init__()\n self.env = Environment(""breakout"", sticky_action_prob=0.1)\n self.action_space = spaces.Discrete(self.env.num_actions())\n self.observation_space = spaces.Box(\n low=0, high=1, shape=self.env.state().shape, dtype=np.float32\n )\n\n def reset(self):\n self.env.reset()\n return self.env.state()\n\n def step(self, action):\n reward, done = self.env.act(action)\n return self.env.state(), reward, done, {}\n\n\n@dataclass\nclass Args:\n output_dir: str = ""data/breakout_episodes""\n num_episodes: int = 2000\n min_episode_length: int = 100\n max_episode_length: int = 500\n chunk_size: int = 50\n chunks_per_file: int = 100\n train_timesteps: int = 200_000 # DQN training timesteps\n collection_interval: int = 10_000 # Collect episodes every N training steps\n seed: int = 0\n\n\nargs = tyro.cli(Args)\n\n\ndef collect_episodes(model: DQN, num_eps: int) -> Tuple[List[np.ndarray], List[np.ndarray], List[dict]]:\n obs_chunks, act_chunks, ep_metadata = [], [], []\n for ep in range(num_eps):\n env = MinAtarBreakout()\n obs = env.reset()\n done = False\n ep_obs, ep_act = [], []\n steps = 0\n while not done and steps < args.max_episode_length:\n action, _ = model.predict(obs, deterministic=False)\n ep_obs.append(obs.astype(np.uint8))\n ep_act.append(action)\n obs, reward, done, _ = env.step(action)\n steps += 1\n\n if steps >= args.min_episode_length:\n # Chunk episode\n episode_obs_chunks, episode_act_chunks = [], []\n for i in range(0, len(ep_obs), args.chunk_size):\n o_chunk = np.stack(ep_obs[i:i + args.chunk_size])\n a_chunk = np.array(ep_act[i:i + args.chunk_size])\n episode_obs_chunks.append(o_chunk)\n episode_act_chunks.append(a_chunk)\n\n obs_chunks.extend(episode_obs_chunks)\n act_chunks.extend(episode_act_chunks)\n\n avg_len = np.mean([len(ep_obs)])\n ep_metadata.append({""avg_seq_len"": avg_len})\n else:\n print(f""Episode too short ({steps}), discarded."")\n\n return obs_chunks, act_chunks, ep_metadata\n\n\ndef main():\n np.random.seed(args.seed)\n os.makedirs(args.output_dir, exist_ok=True)\n\n # --- Create training environment and model ---\n env = MinAtarBreakout()\n model = DQN(\n ""MlpPolicy"",\n env,\n learning_starts=1_000,\n buffer_size=50_000,\n train_freq=4,\n verbose=1,\n seed=args.seed,\n )\n\n episode_metadata_all = []\n obs_chunks, act_chunks = [], []\n file_idx = 0\n\n # --- Train and periodically collect ---\n timesteps = 0\n while timesteps < args.train_timesteps:\n model.learn(total_timesteps=args.collection_interval, reset_num_timesteps=False)\n timesteps += args.collection_interval\n\n # Collect episodes from the current policy\n o_chunks, a_chunks, ep_meta = collect_episodes(model, num_eps=20)\n obs_chunks.extend(o_chunks)\n act_chunks.extend(a_chunks)\n episode_metadata_all.extend(ep_meta)\n\n # Save chunks to disk\n obs_chunks, act_chunks_meta, file_idx, act_chunks = save_chunks(\n obs_chunks, file_idx, args.chunks_per_file, args.output_dir, act_chunks\n )\n episode_metadata_all.extend(act_chunks_meta)\n\n print(f""Collected and saved dataset at {timesteps} timesteps."")\n\n # --- Save metadata ---\n metadata = {\n ""env"": ""MinAtar-Breakout"",\n ""num_actions"": env.action_space.n,\n ""train_timesteps"": args.train_timesteps,\n ""num_episodes"": args.num_episodes,\n ""avg_episode_len"": float(np.mean([ep[""avg_seq_len""] for ep in episode_metadata_all])),\n ""episode_metadata"": episode_metadata_all,\n }\n\n with open(os.path.join(args.output_dir, ""metadata.json""), ""w"") as f:\n json.dump(metadata, f)\n\n print(""Dataset generation complete."")\n\n\nif __name__ == ""__main__"":\n main()\n",python,tab +3099,5263791,"jasmine/utils/dataloader_torch.py",0,0,"",python,tab +3100,5263792,"jasmine/utils/dataloader_torch.py",851,0,"",python,selection_command +3101,5267810,"TERMINAL",0,0,"git status",,terminal_command +3102,5267872,"TERMINAL",0,0,"]633;C",,terminal_output +3103,5267972,"TERMINAL",0,0,"On branch ablation/use-pytorch-dataloader\r\nLast commands done (2 commands done):\r\n pick ba37453 feat: generate coinrun dataset with val split\r\n pick faadd10 feat: implemented validation loss for all three models\r\nNext commands to do (26 remaining commands):\r\n pick 9a17dbb fix: pass val data path to dataloader\r\n pick 6e69cdb fix typo in image logging\r\n (use ""git rebase --edit-todo"" to view and edit)\r\nYou are currently editing a commit while rebasing branch 'gt-actions' on 'c7522f2'.\r\n (use ""git commit --amend"" to amend the current commit)\r\n (use ""git rebase --continue"" once you are satisfied with your changes)\r\n\r\nChanges to be committed:\r\n (use ""git restore --staged ..."" to unstage)\r\n\tmodified: jasmine/train_dynamics.py\r\n\tmodified: jasmine/utils/dataloader_torch.py\r\n\r\nUntracked files:\r\n (use ""git add ..."" to include in what will be committed)\r\n\tali-old-branch.diff\r\n\tdata/_vizdoom.ini\r\n\tdata/data/\r\n\tdata/jasmine_data/vizdoom/\r\n\tdata/uv.lock\r\n\tdiff.diff\r\n\tdiff2.diff\r\n\tinput_pipeline/\r\n\tkiller.sh\r\n\tkiller_partition.sh\r\n\tlog.log\r\n\tmessage.md\r\n\toverfit_dir.zip\r\n\trequirements-franz.txt\r\n\tsamples/\r\n\tscripts_cremers/\r\n\tslurm/\r\n\ttest.py\r\n\tutils/\r\n\tuv.lock\r\n\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +3104,5288349,"TERMINAL",0,0,"git commit -m ""implemented dynamics model training with pytorch dataloader""",,terminal_command +3105,5288368,"TERMINAL",0,0,"]633;C",,terminal_output +3106,5289538,"TERMINAL",0,0,"black....................................................................",,terminal_output +3107,5290867,"TERMINAL",0,0,"Passed\r\n",,terminal_output +3108,5291014,"TERMINAL",0,0,"[ablation/use-pytorch-dataloader 174c42b] implemented dynamics model training with pytorch dataloader\r\n 2 files changed, 7 insertions(+), 39 deletions(-)\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +3109,5307169,"TERMINAL",0,0,"git push",,terminal_command +3110,5307204,"TERMINAL",0,0,"]633;Cfatal: The current branch ablation/use-pytorch-dataloader has no upstream branch.\r\nTo push the current branch and set the remote as upstream, use\r\n\r\n git push --set-upstream origin ablation/use-pytorch-dataloader\r\n\r\nTo have this happen automatically for branches without a tracking\r\nupstream, see 'push.autoSetupRemote' in 'git help config'.\r\n\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +3111,5311598,"TERMINAL",0,0,"git push --set-upstream origin ablation/use-pytorch-dataloader",,terminal_command +3112,5311672,"TERMINAL",0,0,"]633;C",,terminal_output +3113,5313210,"TERMINAL",0,0,"Enumerating objects: 22, done.\r\nCounting objects: 4% (1/22)\rCounting objects: 9% (2/22)\rCounting objects: 13% (3/22)\rCounting objects: 18% (4/22)\rCounting objects: 22% (5/22)\rCounting objects: 27% (6/22)\rCounting objects: 31% (7/22)\rCounting objects: 36% (8/22)\rCounting objects: 40% (9/22)\rCounting objects: 45% (10/22)\rCounting objects: 50% (11/22)\rCounting objects: 54% (12/22)\rCounting objects: 59% (13/22)\rCounting objects: 63% (14/22)\rCounting objects: 68% (15/22)\rCounting objects: 72% (16/22)\rCounting objects: 77% (17/22)\rCounting objects: 81% (18/22)\rCounting objects: 86% (19/22)\rCounting objects: 90% (20/22)\rCounting objects: 95% (21/22)\rCounting objects: 100% (22/22)\rCounting objects: 100% (22/22), done.\r\nDelta compression using up to 152 threads\r\nCompressing objects: 5% (1/17)\rCompressing objects: 11% (2/17)\rCompressing objects: 17% (3/17)\rCompressing objects: 23% (4/17)\rCompressing objects: 29% (5/17)\rCompressing objects: 35% (6/17)\rCompressing objects: 41% (7/17)\rCompressing objects: 47% (8/17)\rCompressing objects: 52% (9/17)\rCompressing objects: 58% (10/17)\rCompressing objects: 64% (11/17)\rCompressing objects: 70% (12/17)\rCompressing objects: 76% (13/17)\rCompressing objects: 82% (14/17)\rCompressing objects: 88% (15/17)\rCompressing objects: 94% (16/17)\rCompressing objects: 100% (17/17)\rCompressing objects: 100% (17/17), done.\r\nWriting objects: 5% (1/17)\rWriting objects: 11% (2/17)\rWriting objects: 17% (3/17)\rWriting objects: 23% (4/17)\rWriting objects: 29% (5/17)\rWriting objects: 47% (8/17)\rWriting objects: 52% (9/17)\rWriting objects: 58% (10/17)\rWriting objects: 64% (11/17)\rWriting objects: 70% (12/17)\rWriting objects: 76% (13/17)\rWriting objects: 82% (14/17)\rWriting objects: 88% (15/17)\rWriting objects: 94% (16/17)\rWriting objects: 100% (17/17)\rWriting objects: 100% (17/17), 2.28 KiB | 583.00 KiB/s, done.\r\nTotal 17 (delta 13), reused 0 (delta 0), pack-reused 0\r\n",,terminal_output +3114,5313273,"TERMINAL",0,0,"remote: Resolving deltas: 0% (0/13)\rremote: Resolving deltas: 7% (1/13)\rremote: Resolving deltas: 15% (2/13)\rremote: Resolving deltas: 23% (3/13)\rremote: Resolving deltas: 30% (4/13)\rremote: Resolving deltas: 38% (5/13)\rremote: Resolving deltas: 46% (6/13)\rremote: Resolving deltas: 53% (7/13)\rremote: Resolving deltas: 61% (8/13)\rremote: Resolving deltas: 69% (9/13)\rremote: Resolving deltas: 76% (10/13)\rremote: Resolving deltas: 84% (11/13)\rremote: Resolving deltas: 92% (12/13)\rremote: Resolving deltas: 100% (13/13)\rremote: Resolving deltas: 100% (13/13), completed with 5 local objects.\r\n",,terminal_output +3115,5313601,"TERMINAL",0,0,"remote: \r\nremote: Create a pull request for 'ablation/use-pytorch-dataloader' on GitHub by visiting:\r\nremote: https://github.com/p-doom/jasmine/pull/new/ablation/use-pytorch-dataloader\r\nremote: \r\nTo github.com:p-doom/jasmine.git\r\n * [new branch] ablation/use-pytorch-dataloader -> ablation/use-pytorch-dataloader\r\nbranch 'ablation/use-pytorch-dataloader' set up to track 'origin/ablation/use-pytorch-dataloader'.\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +3116,5322642,"TERMINAL",0,0,"queue",,terminal_command +3117,5322692,"TERMINAL",0,0,"]633;C",,terminal_output +3118,5322931,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: squeue --mehkn1991.localdomain: Thu Oct 2 13:29:15 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3532383 accelerat train_to tum_cte0 R 1:49:59\t 1 hkn04253536978 dev_accel train_dy tum_cte0 R\t1:26\t 1 hkn04013533705 large generate tum_cte0 R 1-21:30:27\t 1 hkn1901",,terminal_output +3119,5323939,"TERMINAL",0,0,"750:0078",,terminal_output +3120,5325067,"TERMINAL",0,0,"8189",,terminal_output +3121,5326028,"TERMINAL",0,0,"92930",,terminal_output +3122,5327041,"TERMINAL",0,0,"203301",,terminal_output +3123,5328185,"TERMINAL",0,0,"1412",,terminal_output +3124,5329198,"TERMINAL",0,0,"2523",,terminal_output +3125,5330412,"TERMINAL",0,0,"3634",,terminal_output +3126,5331218,"TERMINAL",0,0,"4745",,terminal_output +3127,5332244,"TERMINAL",0,0,"5856",,terminal_output +3128,5333285,"TERMINAL",0,0,"6967",,terminal_output +3129,5334318,"TERMINAL",0,0,"71078",,terminal_output +3130,5335418,"TERMINAL",0,0,"8189",,terminal_output +3131,5336427,"TERMINAL",0,0,"92940",,terminal_output +3132,5337455,"TERMINAL",0,0,"303401",,terminal_output +3133,5338480,"TERMINAL",0,0,"1412",,terminal_output +3134,5339513,"TERMINAL",0,0,"2523",,terminal_output +3135,5340552,"TERMINAL",0,0,"3634",,terminal_output +3136,5341624,"TERMINAL",0,0,"4745",,terminal_output +3137,5342675,"TERMINAL",0,0,"5856",,terminal_output +3138,5343666,"TERMINAL",0,0,"6967",,terminal_output +3139,5344743,"TERMINAL",0,0,"72078",,terminal_output +3140,5345743,"TERMINAL",0,0,"8189",,terminal_output +3141,5346827,"TERMINAL",0,0,"92950",,terminal_output +3142,5347853,"TERMINAL",0,0,"404512",,terminal_output +3143,5348860,"TERMINAL",0,0,"2523",,terminal_output +3144,5349898,"TERMINAL",0,0,"3634",,terminal_output +3145,5350941,"TERMINAL",0,0,"4745",,terminal_output +3146,5352004,"TERMINAL",0,0,"5856",,terminal_output +3147,5353019,"TERMINAL",0,0,"6967",,terminal_output +3148,5354055,"TERMINAL",0,0,"73078",,terminal_output +3149,5355243,"TERMINAL",0,0,"8189",,terminal_output +3150,5356211,"TERMINAL",0,0,"9291:00",,terminal_output +3151,5357207,"TERMINAL",0,0,"5032:001",,terminal_output +3152,5358260,"TERMINAL",0,0,"1412",,terminal_output +3153,5359278,"TERMINAL",0,0,"2523",,terminal_output +3154,5360287,"TERMINAL",0,0,"3634",,terminal_output +3155,5361329,"TERMINAL",0,0,"4745",,terminal_output +3156,5362360,"TERMINAL",0,0,"5856",,terminal_output +3157,5363486,"TERMINAL",0,0,"6967",,terminal_output +3158,5364437,"TERMINAL",0,0,"74078",,terminal_output +3159,5365496,"TERMINAL",0,0,"8189",,terminal_output +3160,5366506,"TERMINAL",0,0,"92910",,terminal_output +3161,5367552,"TERMINAL",0,0,"30:003101",,terminal_output +3162,5368585,"TERMINAL",0,0,"1412",,terminal_output +3163,5369649,"TERMINAL",0,0,"2523",,terminal_output +3164,5370659,"TERMINAL",0,0,"3634",,terminal_output +3165,5371710,"TERMINAL",0,0,"4745",,terminal_output +3166,5372753,"TERMINAL",0,0,"5856",,terminal_output +3167,5373783,"TERMINAL",0,0,"6967",,terminal_output +3168,5374833,"TERMINAL",0,0,"75189",,terminal_output +3169,5375866,"TERMINAL",0,0,"92920",,terminal_output +3170,5376985,"TERMINAL",0,0,"103201",,terminal_output +3171,5378112,"TERMINAL",0,0,"1412",,terminal_output +3172,5378976,"TERMINAL",0,0,"2523",,terminal_output +3173,5380075,"TERMINAL",0,0,"3634",,terminal_output +3174,5381079,"TERMINAL",0,0,"4745",,terminal_output +3175,5382210,"TERMINAL",0,0,"5856",,terminal_output +3176,5383332,"TERMINAL",0,0,"6967",,terminal_output +3177,5384255,"TERMINAL",0,0,"71:0078",,terminal_output +3178,5385279,"TERMINAL",0,0,"8189",,terminal_output +3179,5386265,"TERMINAL",0,0,"92930",,terminal_output +3180,5387315,"TERMINAL",0,0,"203301",,terminal_output +3181,5388360,"TERMINAL",0,0,"1412",,terminal_output +3182,5389389,"TERMINAL",0,0,"2523",,terminal_output +3183,5390432,"TERMINAL",0,0,"3634",,terminal_output +3184,5391468,"TERMINAL",0,0,"4745",,terminal_output +3185,5392510,"TERMINAL",0,0,"5856",,terminal_output +3186,5393572,"TERMINAL",0,0,"6967",,terminal_output +3187,5394598,"TERMINAL",0,0,"71078",,terminal_output +3188,5395755,"TERMINAL",0,0,"8189",,terminal_output +3189,5396913,"TERMINAL",0,0,"92940",,terminal_output +3190,5397706,"TERMINAL",0,0,"303401",,terminal_output +3191,5398865,"TERMINAL",0,0,"1412",,terminal_output +3192,5399780,"TERMINAL",0,0,"2523",,terminal_output +3193,5400844,"TERMINAL",0,0,"3745",,terminal_output +3194,5401855,"TERMINAL",0,0,"5856",,terminal_output +3195,5402992,"TERMINAL",0,0,"6967",,terminal_output +3196,5403944,"TERMINAL",0,0,"72078",,terminal_output +3197,5405044,"TERMINAL",0,0,"8189",,terminal_output +3198,5406064,"TERMINAL",0,0,"92950",,terminal_output +3199,5407090,"TERMINAL",0,0,"403501",,terminal_output +3200,5408119,"TERMINAL",0,0,"1412",,terminal_output +3201,5409347,"TERMINAL",0,0,"2523",,terminal_output +3202,5410183,"TERMINAL",0,0,"3634",,terminal_output +3203,5411291,"TERMINAL",0,0,"4745",,terminal_output +3204,5412257,"TERMINAL",0,0,"5856",,terminal_output +3205,5413300,"TERMINAL",0,0,"6967",,terminal_output +3206,5414467,"TERMINAL",0,0,"73078",,terminal_output +3207,5415372,"TERMINAL",0,0,"8189",,terminal_output +3208,5416405,"TERMINAL",0,0,"9292:00",,terminal_output +3209,5417445,"TERMINAL",0,0,"5033:001",,terminal_output +3210,5418562,"TERMINAL",0,0,"1412",,terminal_output +3211,5419617,"TERMINAL",0,0,"2523",,terminal_output +3212,5420616,"TERMINAL",0,0,"3634",,terminal_output +3213,5421631,"TERMINAL",0,0,"4745",,terminal_output +3214,5422646,"TERMINAL",0,0,"5856",,terminal_output +3215,5423783,"TERMINAL",0,0,"6967",,terminal_output +3216,5424807,"TERMINAL",0,0,"74078",,terminal_output +3217,5425756,"TERMINAL",0,0,"8189",,terminal_output +3218,5426787,"TERMINAL",0,0,"92910",,terminal_output +3219,5427889,"TERMINAL",0,0,"1:004112",,terminal_output +3220,5429121,"TERMINAL",0,0,"2523",,terminal_output +3221,5429916,"TERMINAL",0,0,"3634",,terminal_output +3222,5430937,"TERMINAL",0,0,"4745",,terminal_output +3223,5432079,"TERMINAL",0,0,"5856",,terminal_output +3224,5433108,"TERMINAL",0,0,"6967",,terminal_output +3225,5434224,"TERMINAL",0,0,"75078",,terminal_output +3226,5435112,"TERMINAL",0,0,"8189",,terminal_output +3227,5436157,"TERMINAL",0,0,"92920",,terminal_output +3228,5437196,"TERMINAL",0,0,"103201",,terminal_output +3229,5438386,"TERMINAL",0,0,"1412",,terminal_output +3230,5439274,"TERMINAL",0,0,"2523",,terminal_output +3231,5440307,"TERMINAL",0,0,"3634",,terminal_output +3232,5441461,"TERMINAL",0,0,"4745",,terminal_output +3233,5442383,"TERMINAL",0,0,"5856",,terminal_output +3234,5443440,"TERMINAL",0,0,"6967",,terminal_output +3235,5444570,"TERMINAL",0,0,"72:0078",,terminal_output +3236,5445498,"TERMINAL",0,0,"8189",,terminal_output +3237,5446534,"TERMINAL",0,0,"92930",,terminal_output +3238,5447576,"TERMINAL",0,0,"203301",,terminal_output +3239,5448665,"TERMINAL",0,0,"1412",,terminal_output +3240,5449687,"TERMINAL",0,0,"2523",,terminal_output +3241,5450684,"TERMINAL",0,0,"3634",,terminal_output +3242,5451726,"TERMINAL",0,0,"4745",,terminal_output +3243,5452762,"TERMINAL",0,0,"5856",,terminal_output +3244,5453799,"TERMINAL",0,0,"6967",,terminal_output +3245,5454838,"TERMINAL",0,0,"71189",,terminal_output +3246,5455895,"TERMINAL",0,0,"92940",,terminal_output +3247,5456956,"TERMINAL",0,0,"303401",,terminal_output +3248,5457973,"TERMINAL",0,0,"1412",,terminal_output +3249,5459047,"TERMINAL",0,0,"2523",,terminal_output +3250,5460061,"TERMINAL",0,0,"3634",,terminal_output +3251,5461084,"TERMINAL",0,0,"4745",,terminal_output +3252,5462157,"TERMINAL",0,0,"5856",,terminal_output +3253,5463205,"TERMINAL",0,0,"6967",,terminal_output +3254,5464234,"TERMINAL",0,0,"72078",,terminal_output +3255,5465252,"TERMINAL",0,0,"8189",,terminal_output +3256,5466274,"TERMINAL",0,0,"92950",,terminal_output +3257,5467320,"TERMINAL",0,0,"403501",,terminal_output +3258,5468379,"TERMINAL",0,0,"1412",,terminal_output +3259,5469393,"TERMINAL",0,0,"2523",,terminal_output +3260,5470433,"TERMINAL",0,0,"3634",,terminal_output +3261,5471474,"TERMINAL",0,0,"4745",,terminal_output +3262,5472524,"TERMINAL",0,0,"5856",,terminal_output +3263,5473653,"TERMINAL",0,0,"6967",,terminal_output +3264,5474621,"TERMINAL",0,0,"73078",,terminal_output +3265,5475622,"TERMINAL",0,0,"8189",,terminal_output +3266,5476662,"TERMINAL",0,0,"9293:00",,terminal_output +3267,5477698,"TERMINAL",0,0,"5034:001",,terminal_output +3268,5478772,"TERMINAL",0,0,"1412",,terminal_output +3269,5479798,"TERMINAL",0,0,"2523",,terminal_output +3270,5480918,"TERMINAL",0,0,"3745",,terminal_output +3271,5481887,"TERMINAL",0,0,"5856",,terminal_output +3272,5482970,"TERMINAL",0,0,"6967",,terminal_output +3273,5483990,"TERMINAL",0,0,"74078",,terminal_output +3274,5484975,"TERMINAL",0,0,"8189",,terminal_output +3275,5486016,"TERMINAL",0,0,"92910",,terminal_output +3276,5487227,"TERMINAL",0,0,"2:003101",,terminal_output +3277,5488095,"TERMINAL",0,0,"1412",,terminal_output +3278,5489126,"TERMINAL",0,0,"2523",,terminal_output +3279,5490165,"TERMINAL",0,0,"3634",,terminal_output +3280,5491263,"TERMINAL",0,0,"4745",,terminal_output +3281,5492352,"TERMINAL",0,0,"5856",,terminal_output +3282,5493374,"TERMINAL",0,0,"6967",,terminal_output +3283,5494335,"TERMINAL",0,0,"75078",,terminal_output +3284,5495363,"TERMINAL",0,0,"8189",,terminal_output +3285,5496398,"TERMINAL",0,0,"92920",,terminal_output +3286,5497507,"TERMINAL",0,0,"103201",,terminal_output +3287,5498476,"TERMINAL",0,0,"1412",,terminal_output +3288,5499557,"TERMINAL",0,0,"2523",,terminal_output +3289,5500587,"TERMINAL",0,0,"3634",,terminal_output +3290,5501601,"TERMINAL",0,0,"4745",,terminal_output +3291,5502639,"TERMINAL",0,0,"5856",,terminal_output +3292,5503763,"TERMINAL",0,0,"6967",,terminal_output +3293,5504704,"TERMINAL",0,0,"73:0078",,terminal_output +3294,5505807,"TERMINAL",0,0,"8189",,terminal_output +3295,5506823,"TERMINAL",0,0,"92930",,terminal_output +3296,5507818,"TERMINAL",0,0,"204312",,terminal_output +3297,5508878,"TERMINAL",0,0,"2523",,terminal_output +3298,5509895,"TERMINAL",0,0,"3634",,terminal_output +3299,5511025,"TERMINAL",0,0,"4745",,terminal_output +3300,5512051,"TERMINAL",0,0,"5856",,terminal_output +3301,5513076,"TERMINAL",0,0,"6967",,terminal_output +3302,5514054,"TERMINAL",0,0,"71078",,terminal_output +3303,5515121,"TERMINAL",0,0,"8189",,terminal_output +3304,5516146,"TERMINAL",0,0,"92940",,terminal_output +3305,5517272,"TERMINAL",0,0,"303401",,terminal_output +3306,5518296,"TERMINAL",0,0,"1412",,terminal_output +3307,5519321,"TERMINAL",0,0,"2523",,terminal_output +3308,5520300,"TERMINAL",0,0,"3634",,terminal_output +3309,5521357,"TERMINAL",0,0,"4745",,terminal_output +3310,5522418,"TERMINAL",0,0,"5856",,terminal_output +3311,5523438,"TERMINAL",0,0,"6967",,terminal_output +3312,5524577,"TERMINAL",0,0,"72078",,terminal_output +3313,5525601,"TERMINAL",0,0,"8189",,terminal_output +3314,5526559,"TERMINAL",0,0,"92950",,terminal_output +3315,5527625,"TERMINAL",0,0,"403501",,terminal_output +3316,5528642,"TERMINAL",0,0,"1412",,terminal_output +3317,5529764,"TERMINAL",0,0,"2523",,terminal_output +3318,5530794,"TERMINAL",0,0,"3634",,terminal_output +3319,5531773,"TERMINAL",0,0,"4745",,terminal_output +3320,5532816,"TERMINAL",0,0,"5856",,terminal_output +3321,5533967,"TERMINAL",0,0,"73078",,terminal_output +3322,5534900,"TERMINAL",0,0,"8189",,terminal_output +3323,5536015,"TERMINAL",0,0,"9294:00",,terminal_output +3324,5537038,"TERMINAL",0,0,"5035:001",,terminal_output +3325,5538058,"TERMINAL",0,0,"1412",,terminal_output +3326,5539084,"TERMINAL",0,0,"2523",,terminal_output +3327,5540106,"TERMINAL",0,0,"3634",,terminal_output +3328,5541262,"TERMINAL",0,0,"4745",,terminal_output +3329,5542162,"TERMINAL",0,0,"5856",,terminal_output +3330,5543280,"TERMINAL",0,0,"6967",,terminal_output +3331,5544240,"TERMINAL",0,0,"74078",,terminal_output +3332,5545332,"TERMINAL",0,0,"8189",,terminal_output +3333,5546357,"TERMINAL",0,0,"92910",,terminal_output +3334,5547383,"TERMINAL",0,0,"3:003101",,terminal_output +3335,5548437,"TERMINAL",0,0,"1412",,terminal_output +3336,5549477,"TERMINAL",0,0,"2523",,terminal_output +3337,5550552,"TERMINAL",0,0,"3634",,terminal_output +3338,5551601,"TERMINAL",0,0,"4745",,terminal_output +3339,5552600,"TERMINAL",0,0,"5856",,terminal_output +3340,5553623,"TERMINAL",0,0,"6967",,terminal_output +3341,5554749,"TERMINAL",0,0,"75078",,terminal_output +3342,5555700,"TERMINAL",0,0,"8189",,terminal_output +3343,5556811,"TERMINAL",0,0,"92920",,terminal_output +3344,5557880,"TERMINAL",0,0,"103201",,terminal_output +3345,5558836,"TERMINAL",0,0,"1523",,terminal_output +3346,5559876,"TERMINAL",0,0,"3634",,terminal_output +3347,5560998,"TERMINAL",0,0,"4745",,terminal_output +3348,5562020,"TERMINAL",0,0,"5856",,terminal_output +3349,5563047,"TERMINAL",0,0,"6967",,terminal_output +3350,5564068,"TERMINAL",0,0,"74:0078",,terminal_output +3351,5565101,"TERMINAL",0,0,"8189",,terminal_output +3352,5566218,"TERMINAL",0,0,"92930",,terminal_output +3353,5567274,"TERMINAL",0,0,"203301",,terminal_output +3354,5568267,"TERMINAL",0,0,"1412",,terminal_output +3355,5569283,"TERMINAL",0,0,"2523",,terminal_output +3356,5570323,"TERMINAL",0,0,"3634",,terminal_output +3357,5571362,"TERMINAL",0,0,"4745",,terminal_output +3358,5572408,"TERMINAL",0,0,"5856",,terminal_output +3359,5573467,"TERMINAL",0,0,"6967",,terminal_output +3360,5574495,"TERMINAL",0,0,"71078",,terminal_output +3361,5575648,"TERMINAL",0,0,"8189",,terminal_output +3362,5576586,"TERMINAL",0,0,"92940",,terminal_output +3363,5577627,"TERMINAL",0,0,"303401",,terminal_output +3364,5578717,"TERMINAL",0,0,"1412",,terminal_output +3365,5579719,"TERMINAL",0,0,"2523",,terminal_output +3366,5580865,"TERMINAL",0,0,"3634",,terminal_output +3367,5581812,"TERMINAL",0,0,"4745",,terminal_output +3368,5582912,"TERMINAL",0,0,"5967",,terminal_output +3369,5583941,"TERMINAL",0,0,"72078",,terminal_output +3370,5584939,"TERMINAL",0,0,"8189",,terminal_output +3371,5586082,"TERMINAL",0,0,"92950",,terminal_output +3372,5587078,"TERMINAL",0,0,"403501",,terminal_output +3373,5588100,"TERMINAL",0,0,"1412",,terminal_output +3374,5589201,"TERMINAL",0,0,"2523",,terminal_output +3375,5590241,"TERMINAL",0,0,"3634",,terminal_output +3376,5591289,"TERMINAL",0,0,"4745",,terminal_output +3377,5592285,"TERMINAL",0,0,"5856",,terminal_output +3378,5593359,"TERMINAL",0,0,"6967",,terminal_output +3379,5593509,"/home/hk-project-p0023960/tum_cte0515/Projects/jafar/generate_dataset.py",0,0,"",python,tab +3380,5594359,"TERMINAL",0,0,"73078",,terminal_output +3381,5594950,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",0,0,"",python,tab +3382,5595442,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/genie.py",0,0,"",python,tab +3383,5595537,"TERMINAL",0,0,"8189",,terminal_output +3384,5596482,"TERMINAL",0,0,"9295:00",,terminal_output +3385,5597529,"TERMINAL",0,0,"5036:001",,terminal_output +3386,5598150,"jasmine/train_dynamics.py",0,0,"",python,tab +3387,5598613,"TERMINAL",0,0,"1412",,terminal_output +3388,5599615,"TERMINAL",0,0,"2523",,terminal_output +3389,5599725,"jasmine/train_dynamics.py",20678,0,"",python,selection_mouse +3390,5601142,"TERMINAL",0,0,"3634",,terminal_output +3391,5601684,"TERMINAL",0,0,"4745",,terminal_output +3392,5602724,"TERMINAL",0,0,"5856",,terminal_output +3393,5603769,"TERMINAL",0,0,"6967",,terminal_output +3394,5604823,"TERMINAL",0,0,"74189",,terminal_output +3395,5605863,"TERMINAL",0,0,"92910",,terminal_output +3396,5606975,"TERMINAL",0,0,"4:003101",,terminal_output +3397,5607967,"TERMINAL",0,0,"1412",,terminal_output +3398,5609000,"TERMINAL",0,0,"2523",,terminal_output +3399,5610086,"TERMINAL",0,0,"3634",,terminal_output +3400,5611403,"TERMINAL",0,0,"4745",,terminal_output +3401,5612209,"TERMINAL",0,0,"5856",,terminal_output +3402,5613196,"TERMINAL",0,0,"6967",,terminal_output +3403,5614286,"TERMINAL",0,0,"75078",,terminal_output +3404,5615374,"TERMINAL",0,0,"8189",,terminal_output +3405,5616319,"TERMINAL",0,0,"92920",,terminal_output +3406,5617369,"TERMINAL",0,0,"103201",,terminal_output +3407,5618409,"TERMINAL",0,0,"1412",,terminal_output +3408,5619464,"TERMINAL",0,0,"2523",,terminal_output +3409,5620701,"TERMINAL",0,0,"3634",,terminal_output +3410,5621415,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",0,0,"",shellscript,tab +3411,5621643,"TERMINAL",0,0,"4745",,terminal_output +3412,5622674,"TERMINAL",0,0,"5856",,terminal_output +3413,5623680,"TERMINAL",0,0,"6967",,terminal_output +3414,5624695,"TERMINAL",0,0,"75:0078",,terminal_output +3415,5625785,"TERMINAL",0,0,"8189",,terminal_output +3416,5626784,"TERMINAL",0,0,"92930",,terminal_output +3417,5627825,"TERMINAL",0,0,"204312",,terminal_output +3418,5628396,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +3419,5628881,"TERMINAL",0,0,"2523",,terminal_output +3420,5629951,"TERMINAL",0,0,"3634",,terminal_output +3421,5630694,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1196,0,"",shellscript,selection_mouse +3422,5630968,"TERMINAL",0,0,"4745",,terminal_output +3423,5631892,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1183,26,"",shellscript,content +3424,5632012,"TERMINAL",0,0,"5856",,terminal_output +3425,5632230,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1184,0,"",shellscript,selection_command +3426,5632709,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1224,0,"",shellscript,selection_command +3427,5632731,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1242,0,"",shellscript,selection_command +3428,5632766,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1266,0,"",shellscript,selection_command +3429,5632797,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1289,0,"",shellscript,selection_command +3430,5632833,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1322,0,"",shellscript,selection_command +3431,5632865,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1345,0,"",shellscript,selection_command +3432,5632899,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1366,0,"",shellscript,selection_command +3433,5632932,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1384,0,"",shellscript,selection_command +3434,5632966,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1404,0,"",shellscript,selection_command +3435,5632991,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1434,0,"",shellscript,selection_command +3436,5633023,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1468,0,"",shellscript,selection_command +3437,5633053,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1494,0,"",shellscript,selection_command +3438,5633082,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1506,0,"",shellscript,selection_command +3439,5633083,"TERMINAL",0,0,"6967",,terminal_output +3440,5633123,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1549,0,"",shellscript,selection_command +3441,5633157,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1578,0,"",shellscript,selection_command +3442,5633157,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1604,0,"",shellscript,selection_command +3443,5633198,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1626,0,"",shellscript,selection_command +3444,5633227,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1649,0,"",shellscript,selection_command +3445,5633263,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1675,0,"",shellscript,selection_command +3446,5633404,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1696,0,"",shellscript,selection_command +3447,5633547,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1738,0,"",shellscript,selection_command +3448,5633682,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1789,0,"",shellscript,selection_command +3449,5633865,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1833,0,"",shellscript,selection_command +3450,5633993,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1856,0,"",shellscript,selection_command +3451,5634140,"TERMINAL",0,0,"71078",,terminal_output +3452,5634171,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1880,0,"",shellscript,selection_command +3453,5634320,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1856,0,"",shellscript,selection_command +3454,5635163,"TERMINAL",0,0,"8189",,terminal_output +3455,5635717,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1879,0,"",shellscript,selection_command +3456,5636140,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",630,0,"",shellscript,selection_keyboard +3457,5636198,"TERMINAL",0,0,"92940",,terminal_output +3458,5636899,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1879,0,"",shellscript,selection_keyboard +3459,5637249,"TERMINAL",0,0,"303401",,terminal_output +3460,5637460,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1878,0,"",shellscript,selection_command +3461,5637634,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1877,0,"",shellscript,selection_command +3462,5637795,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1876,0,"",shellscript,selection_command +3463,5638337,"TERMINAL",0,0,"1412",,terminal_output +3464,5638438,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1875,1,"",shellscript,content +3465,5638724,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1875,0,"5",shellscript,content +3466,5638725,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1876,0,"",shellscript,selection_keyboard +3467,5639243,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1875,0,"",shellscript,selection_command +3468,5639360,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1852,0,"",shellscript,selection_command +3469,5639371,"TERMINAL",0,0,"2523",,terminal_output +3470,5639617,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1875,0,"",shellscript,selection_command +3471,5639817,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1899,0,"",shellscript,selection_command +3472,5640373,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1920,0,"",shellscript,selection_command +3473,5640410,"TERMINAL",0,0,"3634",,terminal_output +3474,5641031,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1899,0,"",shellscript,selection_command +3475,5641447,"TERMINAL",0,0,"4745",,terminal_output +3476,5641572,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1875,0,"",shellscript,selection_command +3477,5641577,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1852,0,"",shellscript,selection_command +3478,5641608,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1808,0,"",shellscript,selection_command +3479,5641646,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1757,0,"",shellscript,selection_command +3480,5641678,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1715,0,"",shellscript,selection_command +3481,5641714,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1694,0,"",shellscript,selection_command +3482,5641715,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1668,0,"",shellscript,selection_command +3483,5641746,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1645,0,"",shellscript,selection_command +3484,5641776,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1623,0,"",shellscript,selection_command +3485,5641821,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1597,0,"",shellscript,selection_command +3486,5641830,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1568,0,"",shellscript,selection_command +3487,5641872,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1525,0,"",shellscript,selection_command +3488,5641916,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1504,0,"",shellscript,selection_command +3489,5641990,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1487,0,"",shellscript,selection_command +3490,5641991,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1453,0,"",shellscript,selection_command +3491,5641991,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1423,0,"",shellscript,selection_command +3492,5642005,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1402,0,"",shellscript,selection_command +3493,5642046,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1382,0,"",shellscript,selection_command +3494,5642079,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1364,0,"",shellscript,selection_command +3495,5642110,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1341,0,"",shellscript,selection_command +3496,5642303,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1308,0,"",shellscript,selection_command +3497,5642474,"TERMINAL",0,0,"5856",,terminal_output +3498,5642596,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1341,0,"",shellscript,selection_command +3499,5642767,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1364,0,"",shellscript,selection_command +3500,5643051,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1382,0,"",shellscript,selection_command +3501,5643477,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1402,0,"",shellscript,selection_command +3502,5643570,"TERMINAL",0,0,"6967",,terminal_output +3503,5643673,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1423,0,"",shellscript,selection_command +3504,5643836,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1453,0,"",shellscript,selection_command +3505,5643986,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1487,0,"",shellscript,selection_command +3506,5644135,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1504,0,"",shellscript,selection_command +3507,5644253,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1525,0,"",shellscript,selection_command +3508,5644565,"TERMINAL",0,0,"72078",,terminal_output +3509,5644955,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1504,0,"",shellscript,selection_command +3510,5645158,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1487,0,"",shellscript,selection_command +3511,5645294,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1453,0,"",shellscript,selection_command +3512,5645466,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1423,0,"",shellscript,selection_command +3513,5645594,"TERMINAL",0,0,"8189",,terminal_output +3514,5645759,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1402,0,"",shellscript,selection_command +3515,5646125,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1382,0,"",shellscript,selection_command +3516,5646288,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1364,0,"",shellscript,selection_command +3517,5646462,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1341,0,"",shellscript,selection_command +3518,5646644,"TERMINAL",0,0,"92950",,terminal_output +3519,5647077,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1340,0,"",shellscript,selection_command +3520,5647207,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1339,0,"",shellscript,selection_command +3521,5647553,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1339,4,"",shellscript,content +3522,5647769,"TERMINAL",0,0,"403501",,terminal_output +3523,5648734,"TERMINAL",0,0,"1412",,terminal_output +3524,5649308,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1223,0,"",shellscript,selection_mouse +3525,5649344,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1223,0,"3",shellscript,content +3526,5649345,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1224,0,"",shellscript,selection_keyboard +3527,5649386,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1224,0,"6",shellscript,content +3528,5649387,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1225,0,"",shellscript,selection_keyboard +3529,5649456,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1225,0," ",shellscript,content +3530,5649457,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1226,0,"",shellscript,selection_keyboard +3531,5649764,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1225,0,"",shellscript,selection_command +3532,5649801,"TERMINAL",0,0,"2523",,terminal_output +3533,5650826,"TERMINAL",0,0,"3745",,terminal_output +3534,5651101,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1225,1,"",shellscript,content +3535,5651567,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1223,2,"",shellscript,content +3536,5651881,"TERMINAL",0,0,"5856",,terminal_output +3537,5652944,"TERMINAL",0,0,"6967",,terminal_output +3538,5653976,"TERMINAL",0,0,"73078",,terminal_output +3539,5654011,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1339,0,"110 ",shellscript,content +3540,5654600,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1338,0,"",shellscript,selection_command +3541,5654801,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1361,0,"",shellscript,selection_command +3542,5655037,"TERMINAL",0,0,"8189",,terminal_output +3543,5655349,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1382,0,"",shellscript,selection_command +3544,5655494,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1400,0,"",shellscript,selection_command +3545,5656112,"TERMINAL",0,0,"9296:00",,terminal_output +3546,5657018,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1420,0,"",shellscript,selection_command +3547,5657224,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1450,0,"",shellscript,selection_command +3548,5657312,"TERMINAL",0,0,"5037:001",,terminal_output +3549,5657435,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1467,0,"",shellscript,selection_command +3550,5657855,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1466,0,"",shellscript,selection_command +3551,5658081,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1465,0,"",shellscript,selection_command +3552,5658144,"TERMINAL",0,0,"1412",,terminal_output +3553,5658682,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1464,1,"",shellscript,content +3554,5658838,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1464,0,"5",shellscript,content +3555,5658839,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1465,0,"",shellscript,selection_keyboard +3556,5658913,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1465,0,"0",shellscript,content +3557,5658914,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1466,0,"",shellscript,selection_keyboard +3558,5659080,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1465,0,"",shellscript,selection_command +3559,5659295,"TERMINAL",0,0,"2523",,terminal_output +3560,5660185,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1487,0,"",shellscript,selection_mouse +3561,5660332,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1485,7,"maskgit",shellscript,selection_mouse +3562,5660336,"TERMINAL",0,0,"3634",,terminal_output +3563,5660864,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1623,0,"",shellscript,selection_mouse +3564,5661020,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",1619,5,"jafar",shellscript,selection_mouse +3565,5661531,"TERMINAL",0,0,"4745",,terminal_output +3566,5662482,"TERMINAL",0,0,"5856",,terminal_output +3567,5663450,"TERMINAL",0,0,"6967",,terminal_output +3568,5664490,"TERMINAL",0,0,"74078",,terminal_output +3569,5665537,"TERMINAL",0,0,"8189",,terminal_output +3570,5666601,"TERMINAL",0,0,"92910",,terminal_output +3571,5667655,"TERMINAL",0,0,"5:003101",,terminal_output +3572,5668727,"TERMINAL",0,0,"1412",,terminal_output +3573,5669703,"TERMINAL",0,0,"2523",,terminal_output +3574,5670771,"TERMINAL",0,0,"3634",,terminal_output +3575,5671809,"TERMINAL",0,0,"4745",,terminal_output +3576,5672842,"TERMINAL",0,0,"5967",,terminal_output +3577,5673955,"TERMINAL",0,0,"75078",,terminal_output +3578,5674945,"TERMINAL",0,0,"8189",,terminal_output +3579,5676011,"TERMINAL",0,0,"92920",,terminal_output +3580,5677023,"TERMINAL",0,0,"103201",,terminal_output +3581,5678145,"TERMINAL",0,0,"1412",,terminal_output +3582,5679106,"TERMINAL",0,0,"2523",,terminal_output +3583,5680151,"TERMINAL",0,0,"3634",,terminal_output +3584,5681223,"TERMINAL",0,0,"4745",,terminal_output +3585,5682244,"TERMINAL",0,0,"5856",,terminal_output +3586,5683470,"TERMINAL",0,0,"6967",,terminal_output +3587,5684399,"TERMINAL",0,0,"76:0078",,terminal_output +3588,5685414,"TERMINAL",0,0,"8189",,terminal_output +3589,5686430,"TERMINAL",0,0,"92930",,terminal_output +3590,5687439,"TERMINAL",0,0,"203301",,terminal_output +3591,5688483,"TERMINAL",0,0,"1412",,terminal_output +3592,5689538,"TERMINAL",0,0,"2523",,terminal_output +3593,5690568,"TERMINAL",0,0,"3634",,terminal_output +3594,5691766,"TERMINAL",0,0,"4745",,terminal_output +3595,5692663,"TERMINAL",0,0,"5856",,terminal_output +3596,5693701,"TERMINAL",0,0,"6967",,terminal_output +3597,5694759,"TERMINAL",0,0,"71078",,terminal_output +3598,5695860,"TERMINAL",0,0,"8189",,terminal_output +3599,5696836,"TERMINAL",0,0,"934041",,terminal_output +3600,5697907,"TERMINAL",0,0,"31412",,terminal_output +3601,5698924,"TERMINAL",0,0,"2523",,terminal_output +3602,5700395,"TERMINAL",0,0,"3634",,terminal_output +3603,5701087,"TERMINAL",0,0,"4745",,terminal_output +3604,5702104,"TERMINAL",0,0,"5856",,terminal_output +3605,5703085,"TERMINAL",0,0,"6967",,terminal_output +3606,5704132,"TERMINAL",0,0,"72078",,terminal_output +3607,5705171,"TERMINAL",0,0,"8189",,terminal_output +3608,5706242,"TERMINAL",0,0,"92950",,terminal_output +3609,5707265,"TERMINAL",0,0,"403501",,terminal_output +3610,5708385,"TERMINAL",0,0,"1412",,terminal_output +3611,5709428,"TERMINAL",0,0,"2523",,terminal_output +3612,5710398,"TERMINAL",0,0,"3634",,terminal_output +3613,5711431,"TERMINAL",0,0,"4745",,terminal_output +3614,5712475,"TERMINAL",0,0,"5856",,terminal_output +3615,5713521,"TERMINAL",0,0,"6967",,terminal_output +3616,5714592,"TERMINAL",0,0,"73078",,terminal_output +3617,5715623,"TERMINAL",0,0,"8189",,terminal_output +3618,5716658,"TERMINAL",0,0,"9297:00",,terminal_output +3619,5717738,"TERMINAL",0,0,"5038:001",,terminal_output +3620,5718731,"TERMINAL",0,0,"bash",,terminal_focus +3621,5718863,"TERMINAL",0,0,"1412",,terminal_output +3622,5719805,"TERMINAL",0,0,"2523",,terminal_output +3623,5720821,"TERMINAL",0,0,"3745",,terminal_output +3624,5721903,"TERMINAL",0,0,"5856",,terminal_output +3625,5723002,"TERMINAL",0,0,"6967",,terminal_output +3626,5723195,"TERMINAL",0,0,"watch",,terminal_focus +3627,5723950,"TERMINAL",0,0,"74078",,terminal_output +3628,5724504,"TERMINAL",0,0,"bash",,terminal_focus +3629,5725028,"TERMINAL",0,0,"8189",,terminal_output +3630,5726173,"TERMINAL",0,0,"92910",,terminal_output +3631,5727125,"TERMINAL",0,0,"6:003101",,terminal_output +3632,5727459,"TERMINAL",0,0,"scancel 3536978",,terminal_command +3633,5727523,"TERMINAL",0,0,"]633;C]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +3634,5728119,"TERMINAL",0,0,"14CG2",,terminal_output +3635,5729171,"TERMINAL",0,0,"253",,terminal_output +3636,5729232,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +3637,5729263,"TERMINAL",0,0,"]633;Csalloc: Pending job allocation 3536989\r\nsalloc: job 3536989 queued and waiting for resources\r\n",,terminal_output +3638,5730266,"TERMINAL",0,0,"366989 dev_accel interactPD 0:00(Resources)3533705 large generate tum_cte0 R 1-21:37:14\t 1 hkn1901",,terminal_output +3639,5731441,"TERMINAL",0,0,"475",,terminal_output +3640,5732043,"TERMINAL",0,0,"^Csalloc: Job allocation 3536989 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +3641,5732371,"TERMINAL",0,0,"\r583705 large generate R 1-21:37:16hkn1901",,terminal_output +3642,5733452,"TERMINAL",0,0,"697",,terminal_output +3643,5734396,"TERMINAL",0,0,"7508",,terminal_output +3644,5735446,"TERMINAL",0,0,"819",,terminal_output +3645,5736489,"TERMINAL",0,0,"9220",,terminal_output +3646,5737619,"TERMINAL",0,0,"1031",,terminal_output +3647,5739861,"TERMINAL",0,0,"153",,terminal_output +3648,5739925,"TERMINAL",0,0,"364",,terminal_output +3649,5740979,"TERMINAL",0,0,"475",,terminal_output +3650,5742039,"TERMINAL",0,0,"586",,terminal_output +3651,5743059,"TERMINAL",0,0,"697",,terminal_output +3652,5744100,"TERMINAL",0,0,"77:008",,terminal_output +3653,5745147,"TERMINAL",0,0,"819",,terminal_output +3654,5746342,"TERMINAL",0,0,"9230",,terminal_output +3655,5747265,"TERMINAL",0,0,"2031",,terminal_output +3656,5748277,"TERMINAL",0,0,"142",,terminal_output +3657,5749381,"TERMINAL",0,0,"253",,terminal_output +3658,5750441,"TERMINAL",0,0,"364",,terminal_output +3659,5751446,"TERMINAL",0,0,"475",,terminal_output +3660,5752464,"TERMINAL",0,0,"586",,terminal_output +3661,5753515,"TERMINAL",0,0,"697",,terminal_output +3662,5754551,"TERMINAL",0,0,"7108",,terminal_output +3663,5755593,"TERMINAL",0,0,"819",,terminal_output +3664,5756650,"TERMINAL",0,0,"9240",,terminal_output +3665,5757687,"TERMINAL",0,0,"3031",,terminal_output +3666,5758731,"TERMINAL",0,0,"142",,terminal_output +3667,5759770,"TERMINAL",0,0,"253",,terminal_output +3668,5760890,"TERMINAL",0,0,"375",,terminal_output +3669,5761864,"TERMINAL",0,0,"\r583705 large generate R 1-21:37:4619",,terminal_output +3670,5762932,"TERMINAL",0,0,"697",,terminal_output +3671,5763988,"TERMINAL",0,0,"7208",,terminal_output +3672,5765084,"TERMINAL",0,0,"819",,terminal_output +3673,5766104,"TERMINAL",0,0,"9250",,terminal_output +3674,5767234,"TERMINAL",0,0,"4031",,terminal_output +3675,5768151,"TERMINAL",0,0,"142",,terminal_output +3676,5769285,"TERMINAL",0,0,"253",,terminal_output +3677,5770222,"TERMINAL",0,0,"364",,terminal_output +3678,5771365,"TERMINAL",0,0,"475",,terminal_output +3679,5772404,"TERMINAL",0,0,"586",,terminal_output +3680,5773385,"TERMINAL",0,0,"697",,terminal_output +3681,5774431,"TERMINAL",0,0,"7308",,terminal_output +3682,5775473,"TERMINAL",0,0,"819",,terminal_output +3683,5776533,"TERMINAL",0,0,"928:00",,terminal_output +3684,5777567,"TERMINAL",0,0,"5031",,terminal_output +3685,5778700,"TERMINAL",0,0,"142",,terminal_output +3686,5779665,"TERMINAL",0,0,"253",,terminal_output +3687,5780706,"TERMINAL",0,0,"364",,terminal_output +3688,5781747,"TERMINAL",0,0,"475",,terminal_output +3689,5782793,"TERMINAL",0,0,"586",,terminal_output +3690,5783931,"TERMINAL",0,0,"6408",,terminal_output +3691,5784905,"TERMINAL",0,0,"819",,terminal_output +3692,5785982,"TERMINAL",0,0,"9210",,terminal_output +3693,5786986,"TERMINAL",0,0,"7:0031",,terminal_output +3694,5788123,"TERMINAL",0,0,"142",,terminal_output +3695,5789149,"TERMINAL",0,0,"253",,terminal_output +3696,5790107,"TERMINAL",0,0,"364",,terminal_output +3697,5791190,"TERMINAL",0,0,"475",,terminal_output +3698,5792218,"TERMINAL",0,0,"586",,terminal_output +3699,5793241,"TERMINAL",0,0,"697",,terminal_output +3700,5794287,"TERMINAL",0,0,"7508",,terminal_output +3701,5795351,"TERMINAL",0,0,"819",,terminal_output +3702,5795413,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +3703,5795480,"TERMINAL",0,0,"]633;Csalloc: Pending job allocation 3536992\r\nsalloc: job 3536992 queued and waiting for resources\r\n",,terminal_output +3704,5796448,"TERMINAL",0,0,"926992 dev_accel interactPD 0:00(Resources)3533705 large generate tum_cte0 R 1-21:38:20\t 1 hkn1901",,terminal_output +3705,5797480,"TERMINAL",0,0,"1031",,terminal_output +3706,5798479,"TERMINAL",0,0,"142",,terminal_output +3707,5799513,"TERMINAL",0,0,"253",,terminal_output +3708,5800558,"TERMINAL",0,0,"364",,terminal_output +3709,5801337,"TERMINAL",0,0,"^Csalloc: Job allocation 3536992 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1991:~/Projects/jasmine",,terminal_output +3710,5801615,"TERMINAL",0,0,"\r473705 large generate R 1-21:38:25hkn1901",,terminal_output +3711,5802644,"TERMINAL",0,0,"586",,terminal_output +3712,5803706,"TERMINAL",0,0,"697",,terminal_output +3713,5804734,"TERMINAL",0,0,"78:008",,terminal_output +3714,5804905,"TERMINAL",0,0,"salloc --time=30:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +3715,5805778,"TERMINAL",0,0,"819",,terminal_output +3716,5806858,"TERMINAL",0,0,"9331",,terminal_output +3717,5807910,"TERMINAL",0,0,"2142",,terminal_output +3718,5808915,"TERMINAL",0,0,"253",,terminal_output +3719,5809958,"TERMINAL",0,0,"364",,terminal_output +3720,5811005,"TERMINAL",0,0,"475",,terminal_output +3721,5812054,"TERMINAL",0,0,"salloc --time=00:30:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8",,terminal_command +3722,5812096,"TERMINAL",0,0,"586",,terminal_output +3723,5812108,"TERMINAL",0,0,"]633;Csalloc: Pending job allocation 3536993\r\nsalloc: job 3536993 queued and waiting for resources\r\n",,terminal_output +3724,5813107,"TERMINAL",0,0,"696993 dev_accel interactPD 0:00(Resources)3533705 large generate tum_cte0 R 1-21:38:37\t 1 hkn1901",,terminal_output +3725,5814135,"TERMINAL",0,0,"7108",,terminal_output +3726,5815208,"TERMINAL",0,0,"819",,terminal_output +3727,5816238,"TERMINAL",0,0,"9240",,terminal_output +3728,5816921,"TERMINAL",0,0,"s",,terminal_output +3729,5817136,"TERMINAL",0,0,"o",,terminal_output +3730,5817319,"TERMINAL",0,0,"u",,terminal_output +3731,5817320,"TERMINAL",0,0,"3031",,terminal_output +3732,5817432,"TERMINAL",0,0,"r",,terminal_output +3733,5817612,"TERMINAL",0,0,"c",,terminal_output +3734,5817767,"TERMINAL",0,0,"e",,terminal_output +3735,5817820,"TERMINAL",0,0," ",,terminal_output +3736,5817966,"TERMINAL",0,0,".",,terminal_output +3737,5818216,"TERMINAL",0,0,"\t",,terminal_output +3738,5818378,"TERMINAL",0,0,"142",,terminal_output +3739,5819109,"TERMINAL",0,0,"",,terminal_output +3740,5819391,"TERMINAL",0,0,"253",,terminal_output +3741,5819544,"TERMINAL",0,0," ",,terminal_output +3742,5820500,"TERMINAL",0,0,"364",,terminal_output +3743,5820552,"TERMINAL",0,0,".",,terminal_output +3744,5821003,"TERMINAL",0,0,"ve",,terminal_output +3745,5821065,"TERMINAL",0,0,"n",,terminal_output +3746,5821458,"TERMINAL",0,0,"\t",,terminal_output +3747,5821467,"TERMINAL",0,0,"475",,terminal_output +3748,5822061,"TERMINAL",0,0,"b",,terminal_output +3749,5822259,"TERMINAL",0,0,"i",,terminal_output +3750,5822324,"TERMINAL",0,0,"n",,terminal_output +3751,5822489,"TERMINAL",0,0,"586",,terminal_output +3752,5822748,"TERMINAL",0,0,"\t",,terminal_output +3753,5823147,"TERMINAL",0,0,"a",,terminal_output +3754,5823250,"TERMINAL",0,0,"c",,terminal_output +3755,5823532,"TERMINAL",0,0,"697",,terminal_output +3756,5823937,"TERMINAL",0,0,"t",,terminal_output +3757,5824423,"TERMINAL",0,0,"\t",,terminal_output +3758,5824568,"TERMINAL",0,0,"7208",,terminal_output +3759,5825218,"TERMINAL",0,0,"\r\n",,terminal_output +3760,5825605,"TERMINAL",0,0,"819",,terminal_output +3761,5826643,"TERMINAL",0,0,"9250",,terminal_output +3762,5827687,"TERMINAL",0,0,"4031",,terminal_output +3763,5828714,"TERMINAL",0,0,"142",,terminal_output +3764,5829814,"TERMINAL",0,0,"253",,terminal_output +3765,5830824,"TERMINAL",0,0,"364",,terminal_output +3766,5831953,"TERMINAL",0,0,"486",,terminal_output +3767,5832895,"TERMINAL",0,0,"697",,terminal_output +3768,5833934,"TERMINAL",0,0,"7308",,terminal_output +3769,5834969,"TERMINAL",0,0,"819",,terminal_output +3770,5836027,"TERMINAL",0,0,"929:00",,terminal_output +3771,5837056,"TERMINAL",0,0,"5031",,terminal_output +3772,5838640,"TERMINAL",0,0,"14 Rhkn04012",,terminal_output +3773,5838641,"TERMINAL",0,0,"salloc: job 3536993 has been allocated resources\r\nsalloc: Granted job allocation 3536993\r\nsalloc: Waiting for resource configuration\r\n",,terminal_output +3774,5839246,"TERMINAL",0,0,"2513",,terminal_output +3775,5840174,"TERMINAL",0,0,"3624",,terminal_output +3776,5841276,"TERMINAL",0,0,"4735",,terminal_output +3777,5842566,"TERMINAL",0,0,"5846",,terminal_output +3778,5843309,"TERMINAL",0,0,"6957",,terminal_output +3779,5844442,"TERMINAL",0,0,"74068",,terminal_output +3780,5845470,"TERMINAL",0,0,"8179",,terminal_output +3781,5846445,"TERMINAL",0,0,"92810",,terminal_output +3782,5847502,"TERMINAL",0,0,"8:00391",,terminal_output +3783,5848525,"TERMINAL",0,0,"14102",,terminal_output +3784,5849562,"TERMINAL",0,0,"2513",,terminal_output +3785,5850608,"TERMINAL",0,0,"3624",,terminal_output +3786,5851650,"TERMINAL",0,0,"4735",,terminal_output +3787,5852693,"TERMINAL",0,0,"5846",,terminal_output +3788,5853767,"TERMINAL",0,0,"6957",,terminal_output +3789,5854887,"TERMINAL",0,0,"75068",,terminal_output +3790,5855840,"TERMINAL",0,0,"82820",,terminal_output +3791,5857036,"TERMINAL",0,0,"10391",,terminal_output +3792,5857961,"TERMINAL",0,0,"14202",,terminal_output +3793,5858984,"TERMINAL",0,0,"2513",,terminal_output +3794,5859993,"TERMINAL",0,0,"3624",,terminal_output +3795,5861097,"TERMINAL",0,0,"4735",,terminal_output +3796,5862083,"TERMINAL",0,0,"5846",,terminal_output +3797,5863183,"TERMINAL",0,0,"6957",,terminal_output +3798,5864181,"TERMINAL",0,0,"79:0068",,terminal_output +3799,5865242,"TERMINAL",0,0,"8179",,terminal_output +3800,5865270,"TERMINAL",0,0,"salloc: Nodes hkn0401 are ready for job\r\n",,terminal_output +3801,5865412,"TERMINAL",0,0,"source .ven\tbin\tact\t\r\n",,terminal_output +3802,5866035,"TERMINAL",0,0,"]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h[tum_cte0515@hkn0401 jasmine]$ source .venv/bin/activate\r\n[?2004l\r]0;tum_cte0515@hkn0401:~/Projects/jasmine[?2004h(jasmine) [tum_cte0515@hkn0401 jasmine]$ ",,terminal_output +3803,5866267,"TERMINAL",0,0,"92830",,terminal_output +3804,5867420,"TERMINAL",0,0,"20391",,terminal_output +3805,5868412,"TERMINAL",0,0,"14302",,terminal_output +3806,5869437,"TERMINAL",0,0,"2513",,terminal_output +3807,5870459,"TERMINAL",0,0,"3624",,terminal_output +3808,5871499,"TERMINAL",0,0,"4735",,terminal_output +3809,5872543,"TERMINAL",0,0,"5846",,terminal_output +3810,5873623,"TERMINAL",0,0,"6957",,terminal_output +3811,5874634,"TERMINAL",0,0,"71068",,terminal_output +3812,5875989,"TERMINAL",0,0,"8179",,terminal_output +3813,5876732,"TERMINAL",0,0,"92840",,terminal_output +3814,5878056,"TERMINAL",0,0,"30391",,terminal_output +3815,5878819,"TERMINAL",0,0,"15413",,terminal_output +3816,5879860,"TERMINAL",0,0,"3624",,terminal_output +3817,5881103,"TERMINAL",0,0,"4735",,terminal_output +3818,5882134,"TERMINAL",0,0,"5846",,terminal_output +3819,5883450,"TERMINAL",0,0,"6957",,terminal_output +3820,5884172,"TERMINAL",0,0,"72068",,terminal_output +3821,5885206,"TERMINAL",0,0,"8179",,terminal_output +3822,5886150,"TERMINAL",0,0,"92850",,terminal_output +3823,5887268,"TERMINAL",0,0,"40391",,terminal_output +3824,5888232,"TERMINAL",0,0,"14502",,terminal_output +3825,5889276,"TERMINAL",0,0,"2513",,terminal_output +3826,5890421,"TERMINAL",0,0,"3624",,terminal_output +3827,5891364,"TERMINAL",0,0,"4735",,terminal_output +3828,5892448,"TERMINAL",0,0,"5846",,terminal_output +3829,5893501,"TERMINAL",0,0,"6957",,terminal_output +3830,5894574,"TERMINAL",0,0,"73068",,terminal_output +3831,5895549,"TERMINAL",0,0,"8179",,terminal_output +3832,5896666,"TERMINAL",0,0,"92840:00",,terminal_output +3833,5897649,"TERMINAL",0,0,"50391",,terminal_output +3834,5898600,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",0,0,"",shellscript,tab +3835,5898732,"TERMINAL",0,0,"141:002",,terminal_output +3836,5899722,"TERMINAL",0,0,"2513",,terminal_output +3837,5900855,"TERMINAL",0,0,"3624",,terminal_output +3838,5901868,"TERMINAL",0,0,"4735",,terminal_output +3839,5902766,"TERMINAL",0,0,"s",,terminal_output +3840,5902871,"TERMINAL",0,0,"h",,terminal_output +3841,5902902,"TERMINAL",0,0,"6957",,terminal_output +3842,5902965,"TERMINAL",0,0," ",,terminal_output +3843,5903490,"TERMINAL",0,0,"^M",,terminal_output +3844,5903941,"TERMINAL",0,0,"74068",,terminal_output +3845,5904280,"TERMINAL",0,0,"",,terminal_output +3846,5904362,"TERMINAL",0,0,"",,terminal_output +3847,5904909,"TERMINAL",0,0," ",,terminal_output +3848,5904937,"TERMINAL",0,0,"8179",,terminal_output +3849,5905220,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh",,terminal_output +3850,5905993,"TERMINAL",0,0,"92810",,terminal_output +3851,5907113,"TERMINAL",0,0,"9:00391",,terminal_output +3852,5908134,"TERMINAL",0,0,"14102",,terminal_output +3853,5908804,"TERMINAL",0,0,"\rslurm/jobs/mihir/horeka/coinrun/train_dyn_single_gpu.sh\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=1\r\n#SBATCH --ntasks-per-node=1\r\n#SBATCH --time=00:40:00\r\n#SBATCH --partition=dev_accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:1\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/dynamics/%x_%j.log\r\n#SBATCH --job-name=train_dyn_single_gpu\r\n#SBATCH --requeue\r\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/npy_test\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dyn/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\ntokenizer_checkpoint=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/interactive/3536670\r\n\r\nenv | grep SLURM\r\n\r\n\r\nsrun python jasmine/train_dynamics.py \\r\n --save_ckpt \\r\n --image_height=64 \\r\n --image_width=64 \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=110 \\r\n --patch_size=4 \\r\n --init_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=50 \\r\n --log_checkpoint_interval=50 \\r\n --dyna_type=maskgit \\r\n --log \\r\n --name=coinrun-dyn-dev-$slurm_job_id \\r\n --tags dyn coinrun dev \\r\n --entity instant-uv \\r\n --project jafar \\r\n --warmup_steps 0 \\r\n --wsd_decay_steps 0 \\r\n --num_steps 10 \\r\n --data_dir $array_records_dir_train \\r\n --tokenizer_checkpoint $tokenizer_checkpoint \\r\n --val_data_dir $array_records_dir_val \\r\n --log_interval 1 \\r\n --val_interval 50 \\r\n --eval_full_frame \\r\n --val_steps 5\r\n",,terminal_output +3854,5908949,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1517629\r\nSLURM_JOB_GPUS=0\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine\r\nSLURMD_NODENAME=hkn0401\r\nSLURM_JOB_START_TIME=1759405071\r\nSLURM_STEP_NODELIST=hkn0401\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1759406871\r\nSLURM_PMI2_SRUN_PORT=33699\r\nSLURM_CPUS_ON_NODE=8\r\nSLURM_JOB_CPUS_PER_NODE=8\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=dev_accelerated\r\nSLURM_TRES_PER_TASK=cpu=8\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=1\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3536993\r\nSLURM_PTY_PORT=38337\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.199\r\nSLURM_PTY_WIN_ROW=40\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0401\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.199\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=182\r\nSLURM_NODELIST=hkn0401\r\nSLURM_SRUN_COMM_PORT=42793\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NNODES=1\r\nSLURM_SUBMIT_HOST=hkn1991.localdomain\r\nSLURM_JOB_ID=3536993\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0401\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_STEP_LAUNCHER_PORT=42793\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0401\r\n",,terminal_output +3855,5909078,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +3856,5909103,"TERMINAL",0,0,"2513",,terminal_output +3857,5910193,"TERMINAL",0,0,"3624",,terminal_output +3858,5911213,"TERMINAL",0,0,"4735",,terminal_output +3859,5912234,"TERMINAL",0,0,"5846",,terminal_output +3860,5913259,"TERMINAL",0,0,"6957",,terminal_output +3861,5914386,"TERMINAL",0,0,"75068",,terminal_output +3862,5915463,"TERMINAL",0,0,"8179",,terminal_output +3863,5916439,"TERMINAL",0,0,"92820",,terminal_output +3864,5917503,"TERMINAL",0,0,"10391",,terminal_output +3865,5918441,"TERMINAL",0,0,"14202",,terminal_output +3866,5919504,"TERMINAL",0,0,"2513",,terminal_output +3867,5919922,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +3868,5920529,"TERMINAL",0,0,"3624",,terminal_output +3869,5920537,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.22.0\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/wandb/run-20251002_133912-juyysibs\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run coinrun-dyn-dev-3536993\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/juyysibs\r\n",,terminal_output +3870,5921552,"TERMINAL",0,0,"4735",,terminal_output +3871,5922589,"TERMINAL",0,0,"5846",,terminal_output +3872,5922890,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/.venv/lib64/python3.12/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py:1269: UserWarning: Sharding info not provided when restoring. Populating sharding info from sharding file. Please note restoration time will be slightly increased due to reading from file. Note also that this option is unsafe when restoring on a different topology than the checkpoint was saved with.\r\n warnings.warn(\r\n",,terminal_output +3873,5923628,"TERMINAL",0,0,"6957",,terminal_output +3874,5924960,"TERMINAL",0,0,"72:00:0068",,terminal_output +3875,5925971,"TERMINAL",0,0,"8179",,terminal_output +3876,5926789,"TERMINAL",0,0,"92830",,terminal_output +3877,5927780,"TERMINAL",0,0,"20391",,terminal_output +3878,5928922,"TERMINAL",0,0,"15313",,terminal_output +3879,5929940,"TERMINAL",0,0,"3624",,terminal_output +3880,5930972,"TERMINAL",0,0,"4735",,terminal_output +3881,5931999,"TERMINAL",0,0,"5846",,terminal_output +3882,5933021,"TERMINAL",0,0,"6957",,terminal_output +3883,5934023,"TERMINAL",0,0,"71068",,terminal_output +3884,5935047,"TERMINAL",0,0,"8179",,terminal_output +3885,5936197,"TERMINAL",0,0,"92840",,terminal_output +3886,5937124,"TERMINAL",0,0,"30391",,terminal_output +3887,5938245,"TERMINAL",0,0,"14402",,terminal_output +3888,5939265,"TERMINAL",0,0,"2513",,terminal_output +3889,5940237,"TERMINAL",0,0,"3624",,terminal_output +3890,5941316,"TERMINAL",0,0,"4735",,terminal_output +3891,5942344,"TERMINAL",0,0,"5846",,terminal_output +3892,5943364,"TERMINAL",0,0,"6957",,terminal_output +3893,5944489,"TERMINAL",0,0,"72068",,terminal_output +3894,5945552,"TERMINAL",0,0,"8179",,terminal_output +3895,5946461,"TERMINAL",0,0,"92850",,terminal_output +3896,5947517,"TERMINAL",0,0,"40391",,terminal_output +3897,5948565,"TERMINAL",0,0,"14502",,terminal_output +3898,5949602,"TERMINAL",0,0,"2513",,terminal_output +3899,5950638,"TERMINAL",0,0,"3624",,terminal_output +3900,5951675,"TERMINAL",0,0,"4735",,terminal_output +3901,5952714,"TERMINAL",0,0,"5846",,terminal_output +3902,5953814,"TERMINAL",0,0,"6957",,terminal_output +3903,5954789,"TERMINAL",0,0,"73068",,terminal_output +3904,5955855,"TERMINAL",0,0,"8281:00",,terminal_output +3905,5956878,"TERMINAL",0,0,"50391",,terminal_output +3906,5958009,"TERMINAL",0,0,"142:002",,terminal_output +3907,5959045,"TERMINAL",0,0,"2513",,terminal_output +3908,5959985,"TERMINAL",0,0,"3624",,terminal_output +3909,5961075,"TERMINAL",0,0,"4735",,terminal_output +3910,5962105,"TERMINAL",0,0,"5846",,terminal_output +3911,5963115,"TERMINAL",0,0,"6957",,terminal_output +3912,5964173,"TERMINAL",0,0,"74068",,terminal_output +3913,5965224,"TERMINAL",0,0,"8179",,terminal_output +3914,5966267,"TERMINAL",0,0,"92810",,terminal_output +3915,5967312,"TERMINAL",0,0,"40:00391",,terminal_output +3916,5968488,"TERMINAL",0,0,"14102",,terminal_output +3917,5969475,"TERMINAL",0,0,"2513",,terminal_output +3918,5970519,"TERMINAL",0,0,"3624",,terminal_output +3919,5971496,"TERMINAL",0,0,"4735",,terminal_output +3920,5972575,"TERMINAL",0,0,"5846",,terminal_output +3921,5973583,"TERMINAL",0,0,"6957",,terminal_output +3922,5974623,"TERMINAL",0,0,"75068",,terminal_output +3923,5975683,"TERMINAL",0,0,"8179",,terminal_output +3924,5976717,"TERMINAL",0,0,"92820",,terminal_output +3925,5977757,"TERMINAL",0,0,"10391",,terminal_output +3926,5978797,"TERMINAL",0,0,"14202",,terminal_output +3927,5979899,"TERMINAL",0,0,"2624",,terminal_output +3928,5981046,"TERMINAL",0,0,"4735",,terminal_output +3929,5981969,"TERMINAL",0,0,"5846",,terminal_output +3930,5983031,"TERMINAL",0,0,"6957",,terminal_output +3931,5984119,"TERMINAL",0,0,"71:0068",,terminal_output +3932,5985071,"TERMINAL",0,0,"8179",,terminal_output +3933,5986462,"TERMINAL",0,0,"92830",,terminal_output +3934,5987187,"TERMINAL",0,0,"20391",,terminal_output +3935,5988207,"TERMINAL",0,0,"14302",,terminal_output +3936,5989339,"TERMINAL",0,0,"2513",,terminal_output +3937,5990366,"TERMINAL",0,0,"3624",,terminal_output +3938,5991490,"TERMINAL",0,0,"4735",,terminal_output +3939,5992414,"TERMINAL",0,0,"5846",,terminal_output +3940,5993470,"TERMINAL",0,0,"6957",,terminal_output +3941,5994561,"TERMINAL",0,0,"71068",,terminal_output +3942,5995801,"TERMINAL",0,0,"8179",,terminal_output +3943,5996548,"TERMINAL",0,0,"92840",,terminal_output +3944,5997614,"TERMINAL",0,0,"30391",,terminal_output +3945,5998636,"TERMINAL",0,0,"14402",,terminal_output +3946,5999677,"TERMINAL",0,0,"2513",,terminal_output +3947,6000727,"TERMINAL",0,0,"3624",,terminal_output +3948,6001835,"TERMINAL",0,0,"4735",,terminal_output +3949,6002861,"TERMINAL",0,0,"5846",,terminal_output +3950,6003983,"TERMINAL",0,0,"72068",,terminal_output +3951,6004941,"TERMINAL",0,0,"8179",,terminal_output +3952,6005944,"TERMINAL",0,0,"92850",,terminal_output +3953,6007159,"TERMINAL",0,0,"40391",,terminal_output +3954,6008179,"TERMINAL",0,0,"14502",,terminal_output +3955,6009111,"TERMINAL",0,0,"2513",,terminal_output +3956,6010126,"TERMINAL",0,0,"3624",,terminal_output +3957,6011253,"TERMINAL",0,0,"4735",,terminal_output +3958,6012383,"TERMINAL",0,0,"5846",,terminal_output +3959,6013305,"TERMINAL",0,0,"6957",,terminal_output +3960,6014327,"TERMINAL",0,0,"73068",,terminal_output +3961,6015356,"TERMINAL",0,0,"8179",,terminal_output +3962,6016474,"TERMINAL",0,0,"9282:00",,terminal_output +3963,6017507,"TERMINAL",0,0,"50391",,terminal_output +3964,6018536,"TERMINAL",0,0,"143:002",,terminal_output +3965,6019557,"TERMINAL",0,0,"2513",,terminal_output +3966,6020627,"TERMINAL",0,0,"3624",,terminal_output +3967,6021594,"TERMINAL",0,0,"4735",,terminal_output +3968,6022640,"TERMINAL",0,0,"5846",,terminal_output +3969,6023683,"TERMINAL",0,0,"6957",,terminal_output +3970,6024717,"TERMINAL",0,0,"74068",,terminal_output +3971,6025912,"TERMINAL",0,0,"8179",,terminal_output +3972,6026810,"TERMINAL",0,0,"92810",,terminal_output +3973,6027851,"TERMINAL",0,0,"1:004102",,terminal_output +3974,6028973,"TERMINAL",0,0,"2513",,terminal_output +3975,6029948,"TERMINAL",0,0,"3624",,terminal_output +3976,6031120,"TERMINAL",0,0,"4735",,terminal_output +3977,6032057,"TERMINAL",0,0,"5846",,terminal_output +3978,6033385,"TERMINAL",0,0,"6957",,terminal_output +3979,6034209,"TERMINAL",0,0,"75068",,terminal_output +3980,6035218,"TERMINAL",0,0,"8179",,terminal_output +3981,6036243,"TERMINAL",0,0,"92820",,terminal_output +3982,6037266,"TERMINAL",0,0,"10391",,terminal_output +3983,6038396,"TERMINAL",0,0,"14202",,terminal_output +3984,6039414,"TERMINAL",0,0,"2513",,terminal_output +3985,6040443,"TERMINAL",0,0,"3624",,terminal_output +3986,6041485,"TERMINAL",0,0,"4735",,terminal_output +3987,6042501,"TERMINAL",0,0,"5846",,terminal_output +3988,6043515,"TERMINAL",0,0,"6957",,terminal_output +3989,6044566,"TERMINAL",0,0,"72:0068",,terminal_output +3990,6045611,"TERMINAL",0,0,"8179",,terminal_output +3991,6046652,"TERMINAL",0,0,"92830",,terminal_output +3992,6047694,"TERMINAL",0,0,"20391",,terminal_output +3993,6048836,"TERMINAL",0,0,"14302",,terminal_output +3994,6049805,"TERMINAL",0,0,"2513",,terminal_output +3995,6050849,"TERMINAL",0,0,"3735",,terminal_output +3996,6051906,"TERMINAL",0,0,"5846",,terminal_output +3997,6053139,"TERMINAL",0,0,"6957",,terminal_output +3998,6054056,"TERMINAL",0,0,"71068",,terminal_output +3999,6055031,"TERMINAL",0,0,"8179",,terminal_output +4000,6056106,"TERMINAL",0,0,"92840",,terminal_output +4001,6057127,"TERMINAL",0,0,"30391",,terminal_output +4002,6058165,"TERMINAL",0,0,"14402",,terminal_output +4003,6059214,"TERMINAL",0,0,"2513",,terminal_output +4004,6060310,"TERMINAL",0,0,"3624",,terminal_output +4005,6061328,"TERMINAL",0,0,"4735",,terminal_output +4006,6062376,"TERMINAL",0,0,"5846",,terminal_output +4007,6063482,"TERMINAL",0,0,"6957",,terminal_output +4008,6064509,"TERMINAL",0,0,"72068",,terminal_output +4009,6065542,"TERMINAL",0,0,"8179",,terminal_output +4010,6066551,"TERMINAL",0,0,"92850",,terminal_output +4011,6067579,"TERMINAL",0,0,"40391",,terminal_output +4012,6068630,"TERMINAL",0,0,"14502",,terminal_output +4013,6069673,"TERMINAL",0,0,"2513",,terminal_output +4014,6070718,"TERMINAL",0,0,"3624",,terminal_output +4015,6071760,"TERMINAL",0,0,"4735",,terminal_output +4016,6072903,"TERMINAL",0,0,"5846",,terminal_output +4017,6073923,"TERMINAL",0,0,"63068",,terminal_output +4018,6074947,"TERMINAL",0,0,"8179",,terminal_output +4019,6075970,"TERMINAL",0,0,"9283:00",,terminal_output +4020,6076998,"TERMINAL",0,0,"50391",,terminal_output +4021,6078226,"TERMINAL",0,0,"144:002",,terminal_output +4022,6079144,"TERMINAL",0,0,"2513",,terminal_output +4023,6080170,"TERMINAL",0,0,"3624",,terminal_output +4024,6081191,"TERMINAL",0,0,"4735",,terminal_output +4025,6082324,"TERMINAL",0,0,"5846",,terminal_output +4026,6083343,"TERMINAL",0,0,"6957",,terminal_output +4027,6084295,"TERMINAL",0,0,"74068",,terminal_output +4028,6085524,"TERMINAL",0,0,"8179",,terminal_output +4029,6086444,"TERMINAL",0,0,"92810",,terminal_output +4030,6087439,"TERMINAL",0,0,"2:00391",,terminal_output +4031,6088461,"TERMINAL",0,0,"14102",,terminal_output +4032,6089547,"TERMINAL",0,0,"2513",,terminal_output +4033,6090620,"TERMINAL",0,0,"3624",,terminal_output +4034,6091586,"TERMINAL",0,0,"4735",,terminal_output +4035,6092643,"TERMINAL",0,0,"5846",,terminal_output +4036,6093674,"TERMINAL",0,0,"6957",,terminal_output +4037,6094921,"TERMINAL",0,0,"75068",,terminal_output +4038,6095759,"TERMINAL",0,0,"8179",,terminal_output +4039,6096859,"TERMINAL",0,0,"92820",,terminal_output +4040,6097843,"TERMINAL",0,0,"104202",,terminal_output +4041,6098910,"TERMINAL",0,0,"2513",,terminal_output +4042,6099909,"TERMINAL",0,0,"3624",,terminal_output +4043,6100961,"TERMINAL",0,0,"4735",,terminal_output +4044,6102087,"TERMINAL",0,0,"5846",,terminal_output +4045,6103042,"TERMINAL",0,0,"6957",,terminal_output +4046,6104268,"TERMINAL",0,0,"73:0068",,terminal_output +4047,6105155,"TERMINAL",0,0,"8179",,terminal_output +4048,6106140,"TERMINAL",0,0,"92830",,terminal_output +4049,6107006,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jasmine/jasmine/models/tokenizer.py",0,0,"",python,tab +4050,6107192,"TERMINAL",0,0,"20391",,terminal_output +4051,6108278,"TERMINAL",0,0,"14302",,terminal_output +4052,6109077,"jasmine/train_dynamics.py",0,0,"",python,tab +4053,6109321,"TERMINAL",0,0,"2513",,terminal_output +4054,6110334,"TERMINAL",0,0,"3624",,terminal_output +4055,6111366,"TERMINAL",0,0,"4735",,terminal_output +4056,6112391,"TERMINAL",0,0,"5846",,terminal_output +4057,6113463,"TERMINAL",0,0,"6957",,terminal_output +4058,6114480,"TERMINAL",0,0,"71068",,terminal_output +4059,6115505,"TERMINAL",0,0,"8179",,terminal_output +4060,6116537,"TERMINAL",0,0,"92840",,terminal_output +4061,6117572,"TERMINAL",0,0,"30391",,terminal_output +4062,6118605,"TERMINAL",0,0,"14402",,terminal_output +4063,6119666,"TERMINAL",0,0,"2513",,terminal_output +4064,6120681,"TERMINAL",0,0,"3624",,terminal_output +4065,6121714,"TERMINAL",0,0,"4735",,terminal_output +4066,6122749,"TERMINAL",0,0,"5846",,terminal_output +4067,6124214,"TERMINAL",0,0,"6957",,terminal_output +4068,6124851,"TERMINAL",0,0,"72179",,terminal_output +4069,6125293,"jasmine/train_dynamics.py",19409,0,"",python,selection_mouse +4070,6125437,"jasmine/train_dynamics.py",19409,4,"next",python,selection_mouse +4071,6125593,"jasmine/train_dynamics.py",19379,34,") == 0:\n first_batch = next",python,selection_mouse +4072,6125626,"jasmine/train_dynamics.py",19380,33," == 0:\n first_batch = next",python,selection_mouse +4073,6125657,"jasmine/train_dynamics.py",19381,32,"== 0:\n first_batch = next",python,selection_mouse +4074,6125687,"jasmine/train_dynamics.py",19382,31,"= 0:\n first_batch = next",python,selection_mouse +4075,6125687,"jasmine/train_dynamics.py",19383,30," 0:\n first_batch = next",python,selection_mouse +4076,6125717,"jasmine/train_dynamics.py",19384,29,"0:\n first_batch = next",python,selection_mouse +4077,6125718,"jasmine/train_dynamics.py",19385,28,":\n first_batch = next",python,selection_mouse +4078,6125748,"jasmine/train_dynamics.py",19386,27,"\n first_batch = next",python,selection_mouse +4079,6125865,"TERMINAL",0,0,"92850",,terminal_output +4080,6125939,"jasmine/train_dynamics.py",19409,24,"next(iter(train_iterator",python,selection_mouse +4081,6126305,"jasmine/train_dynamics.py",19409,25,"next(iter(train_iterator)",python,selection_mouse +4082,6126597,"jasmine/train_dynamics.py",19434,0,"",python,selection_mouse +4083,6126624,"jasmine/train_dynamics.py",19433,2,"))",python,selection_mouse +4084,6126757,"jasmine/train_dynamics.py",19433,2,"))",python,selection_mouse +4085,6126831,"jasmine/train_dynamics.py",19419,16,"train_iterator))",python,selection_mouse +4086,6126874,"jasmine/train_dynamics.py",19433,39,"))\n first_batch[""rng""] = rng # ",python,selection_mouse +4087,6126906,"jasmine/train_dynamics.py",19433,36,"))\n first_batch[""rng""] = rng ",python,selection_mouse +4088,6126907,"jasmine/train_dynamics.py",19433,35,"))\n first_batch[""rng""] = rng",python,selection_mouse +4089,6126907,"TERMINAL",0,0,"40391",,terminal_output +4090,6126937,"jasmine/train_dynamics.py",19433,32,"))\n first_batch[""rng""] = ",python,selection_mouse +4091,6126968,"jasmine/train_dynamics.py",19433,30,"))\n first_batch[""rng""] ",python,selection_mouse +4092,6126969,"jasmine/train_dynamics.py",19433,29,"))\n first_batch[""rng""]",python,selection_mouse +4093,6127257,"jasmine/train_dynamics.py",19433,28,"))\n first_batch[""rng""",python,selection_mouse +4094,6127355,"jasmine/train_dynamics.py",19433,27,"))\n first_batch[""rng",python,selection_mouse +4095,6127839,"jasmine/train_dynamics.py",19410,0,"",python,selection_mouse +4096,6127957,"TERMINAL",0,0,"14502",,terminal_output +4097,6127989,"jasmine/train_dynamics.py",19409,4,"next",python,selection_mouse +4098,6128198,"jasmine/train_dynamics.py",19409,9,"next(iter",python,selection_mouse +4099,6128219,"jasmine/train_dynamics.py",19409,24,"next(iter(train_iterator",python,selection_mouse +4100,6128413,"jasmine/train_dynamics.py",19409,25,"next(iter(train_iterator)",python,selection_mouse +4101,6128441,"jasmine/train_dynamics.py",19409,26,"next(iter(train_iterator))",python,selection_mouse +4102,6128822,"jasmine/train_dynamics.py",19435,0,"",python,selection_mouse +4103,6128829,"jasmine/train_dynamics.py",19434,0,"",python,selection_command +4104,6128983,"TERMINAL",0,0,"2513",,terminal_output +4105,6129107,"jasmine/train_dynamics.py",19433,1,")",python,selection_mouse +4106,6129108,"jasmine/train_dynamics.py",19432,2,"r)",python,selection_mouse +4107,6129108,"jasmine/train_dynamics.py",19433,2,"))",python,selection_command +4108,6129121,"jasmine/train_dynamics.py",19430,5,"tor))",python,selection_mouse +4109,6129136,"jasmine/train_dynamics.py",19428,7,"rator))",python,selection_mouse +4110,6129165,"jasmine/train_dynamics.py",19425,10,"iterator))",python,selection_mouse +4111,6129197,"jasmine/train_dynamics.py",19423,12,"n_iterator))",python,selection_mouse +4112,6129231,"jasmine/train_dynamics.py",19422,13,"in_iterator))",python,selection_mouse +4113,6129231,"jasmine/train_dynamics.py",19421,14,"ain_iterator))",python,selection_mouse +4114,6129263,"jasmine/train_dynamics.py",19420,15,"rain_iterator))",python,selection_mouse +4115,6129292,"jasmine/train_dynamics.py",19419,16,"train_iterator))",python,selection_mouse +4116,6129327,"jasmine/train_dynamics.py",19418,17,"(train_iterator))",python,selection_mouse +4117,6129358,"jasmine/train_dynamics.py",19417,18,"r(train_iterator))",python,selection_mouse +4118,6129440,"jasmine/train_dynamics.py",19416,19,"er(train_iterator))",python,selection_mouse +4119,6129474,"jasmine/train_dynamics.py",19415,20,"ter(train_iterator))",python,selection_mouse +4120,6129494,"jasmine/train_dynamics.py",19435,29,"\n first_batch[""rng""] =",python,selection_mouse +4121,6129539,"jasmine/train_dynamics.py",19435,28,"\n first_batch[""rng""] ",python,selection_mouse +4122,6129808,"jasmine/train_dynamics.py",19463,0,"",python,selection_mouse +4123,6130015,"TERMINAL",0,0,"3624",,terminal_output +4124,6130066,"jasmine/train_dynamics.py",19415,0,"",python,selection_mouse +4125,6130204,"jasmine/train_dynamics.py",19414,4,"iter",python,selection_mouse +4126,6130397,"jasmine/train_dynamics.py",19414,5,"iter(",python,selection_mouse +4127,6130430,"jasmine/train_dynamics.py",19414,19,"iter(train_iterator",python,selection_mouse +4128,6131067,"TERMINAL",0,0,"4735",,terminal_output +4129,6131521,"jasmine/train_dynamics.py",19419,14,"train_iterator",python,selection_mouse +4130,6131977,"jasmine/train_dynamics.py",19418,15,"(train_iterator",python,selection_mouse +4131,6132093,"TERMINAL",0,0,"5846",,terminal_output +4132,6132259,"jasmine/train_dynamics.py",19414,19,"iter(train_iterator",python,selection_mouse +4133,6132752,"jasmine/train_dynamics.py",19416,0,"",python,selection_mouse +4134,6132823,"jasmine/train_dynamics.py",19414,4,"iter",python,selection_mouse +4135,6133031,"jasmine/train_dynamics.py",19414,19,"iter(train_iterator",python,selection_mouse +4136,6133173,"TERMINAL",0,0,"6957",,terminal_output +4137,6134236,"TERMINAL",0,0,"73068",,terminal_output +4138,6135201,"TERMINAL",0,0,"8179",,terminal_output +4139,6136263,"TERMINAL",0,0,"9284:00",,terminal_output +4140,6137330,"TERMINAL",0,0,"50391",,terminal_output +4141,6138378,"TERMINAL",0,0,"145:002",,terminal_output +4142,6139395,"TERMINAL",0,0,"2513",,terminal_output +4143,6140352,"TERMINAL",0,0,"\r\n",,terminal_output +4144,6140501,"TERMINAL",0,0,"3624",,terminal_output +4145,6141432,"TERMINAL",0,0,"4735",,terminal_output +4146,6142512,"TERMINAL",0,0,"5846",,terminal_output +4147,6143504,"TERMINAL",0,0,"6957",,terminal_output +4148,6144589,"TERMINAL",0,0,"74068",,terminal_output +4149,6145704,"TERMINAL",0,0,"8179",,terminal_output +4150,6146649,"TERMINAL",0,0,"92810",,terminal_output +4151,6147688,"TERMINAL",0,0,"3:00391",,terminal_output +4152,6148780,"TERMINAL",0,0,"14102",,terminal_output +4153,6149724,"TERMINAL",0,0,"2513",,terminal_output +4154,6150823,"TERMINAL",0,0,"3624",,terminal_output +4155,6151847,"TERMINAL",0,0,"4735",,terminal_output +4156,6152835,"TERMINAL",0,0,"5957",,terminal_output +4157,6153902,"TERMINAL",0,0,"75068",,terminal_output +4158,6154910,"TERMINAL",0,0,"8179",,terminal_output +4159,6156052,"TERMINAL",0,0,"92820",,terminal_output +4160,6157074,"TERMINAL",0,0,"10391",,terminal_output +4161,6158033,"TERMINAL",0,0,"14202",,terminal_output +4162,6159129,"TERMINAL",0,0,"2513",,terminal_output +4163,6160095,"TERMINAL",0,0,"3624",,terminal_output +4164,6161158,"TERMINAL",0,0,"4735",,terminal_output +4165,6162192,"TERMINAL",0,0,"5846",,terminal_output +4166,6163316,"TERMINAL",0,0,"6957",,terminal_output +4167,6164249,"TERMINAL",0,0,"74:0068",,terminal_output +4168,6165377,"TERMINAL",0,0,"8179",,terminal_output +4169,6166316,"TERMINAL",0,0,"92830",,terminal_output +4170,6167414,"TERMINAL",0,0,"20391",,terminal_output +4171,6168446,"TERMINAL",0,0,"14302",,terminal_output +4172,6169427,"TERMINAL",0,0,"2513",,terminal_output +4173,6170483,"TERMINAL",0,0,"3624",,terminal_output +4174,6171498,"TERMINAL",0,0,"4735",,terminal_output +4175,6172562,"TERMINAL",0,0,"5846",,terminal_output +4176,6173601,"TERMINAL",0,0,"6957",,terminal_output +4177,6174615,"TERMINAL",0,0,"71068",,terminal_output +4178,6175652,"TERMINAL",0,0,"8179",,terminal_output +4179,6176750,"TERMINAL",0,0,"92840",,terminal_output +4180,6177730,"TERMINAL",0,0,"30391",,terminal_output +4181,6178807,"TERMINAL",0,0,"14402",,terminal_output +4182,6179807,"TERMINAL",0,0,"2513",,terminal_output +4183,6180933,"TERMINAL",0,0,"3735",,terminal_output +4184,6181968,"TERMINAL",0,0,"5846",,terminal_output +4185,6182980,"TERMINAL",0,0,"6957",,terminal_output +4186,6184000,"TERMINAL",0,0,"72068",,terminal_output +4187,6185013,"TERMINAL",0,0,"8179",,terminal_output +4188,6186087,"TERMINAL",0,0,"92850",,terminal_output +4189,6187093,"TERMINAL",0,0,"40391",,terminal_output +4190,6188310,"TERMINAL",0,0,"14502",,terminal_output +4191,6189225,"TERMINAL",0,0,"2513",,terminal_output +4192,6190252,"TERMINAL",0,0,"3624",,terminal_output +4193,6191280,"TERMINAL",0,0,"4735",,terminal_output +4194,6192324,"TERMINAL",0,0,"5846",,terminal_output +4195,6193325,"TERMINAL",0,0,"6957",,terminal_output +4196,6194466,"TERMINAL",0,0,"73068",,terminal_output +4197,6195482,"TERMINAL",0,0,"8179",,terminal_output +4198,6196447,"TERMINAL",0,0,"9285:00",,terminal_output +4199,6197503,"TERMINAL",0,0,"50391",,terminal_output +4200,6198574,"TERMINAL",0,0,"146:002",,terminal_output +4201,6199571,"TERMINAL",0,0,"2513",,terminal_output +4202,6200599,"TERMINAL",0,0,"3624",,terminal_output +4203,6201650,"TERMINAL",0,0,"4735",,terminal_output +4204,6202680,"TERMINAL",0,0,"5846",,terminal_output +4205,6203709,"TERMINAL",0,0,"6957",,terminal_output +4206,6204754,"TERMINAL",0,0,"74068",,terminal_output +4207,6205811,"TERMINAL",0,0,"8179",,terminal_output +4208,6206835,"TERMINAL",0,0,"93911",,terminal_output +4209,6207962,"TERMINAL",0,0,"4:014102",,terminal_output +4210,6209000,"TERMINAL",0,0,"2513",,terminal_output +4211,6209953,"TERMINAL",0,0,"3624",,terminal_output +4212,6211040,"TERMINAL",0,0,"4735",,terminal_output +4213,6212059,"TERMINAL",0,0,"5846",,terminal_output +4214,6213062,"TERMINAL",0,0,"6957",,terminal_output +4215,6214107,"TERMINAL",0,0,"75068",,terminal_output +4216,6215236,"TERMINAL",0,0,"8179",,terminal_output +4217,6216259,"TERMINAL",0,0,"92820",,terminal_output +4218,6217282,"TERMINAL",0,0,"10391",,terminal_output +4219,6218307,"TERMINAL",0,0,"14202",,terminal_output +4220,6219332,"TERMINAL",0,0,"2513",,terminal_output +4221,6220331,"TERMINAL",0,0,"3624",,terminal_output +4222,6221379,"TERMINAL",0,0,"4735",,terminal_output +4223,6222408,"TERMINAL",0,0,"5846",,terminal_output +4224,6223443,"TERMINAL",0,0,"6957",,terminal_output +4225,6224588,"TERMINAL",0,0,"75:0068",,terminal_output +4226,6225542,"TERMINAL",0,0,"8179",,terminal_output +4227,6226554,"TERMINAL",0,0,"92830",,terminal_output +4228,6227642,"TERMINAL",0,0,"20391",,terminal_output +4229,6228665,"TERMINAL",0,0,"14302",,terminal_output +4230,6229662,"TERMINAL",0,0,"2513",,terminal_output +4231,6230698,"TERMINAL",0,0,"3624",,terminal_output +4232,6231732,"TERMINAL",0,0,"4735",,terminal_output +4233,6232851,"TERMINAL",0,0,"5846",,terminal_output +4234,6233806,"TERMINAL",0,0,"6957",,terminal_output +4235,6234897,"TERMINAL",0,0,"71179",,terminal_output +4236,6235919,"TERMINAL",0,0,"92840",,terminal_output +4237,6236918,"TERMINAL",0,0,"30391",,terminal_output +4238,6237969,"TERMINAL",0,0,"14402",,terminal_output +4239,6239095,"TERMINAL",0,0,"2513",,terminal_output +4240,6240028,"TERMINAL",0,0,"3624",,terminal_output +4241,6241141,"TERMINAL",0,0,"4735",,terminal_output +4242,6242163,"TERMINAL",0,0,"5846",,terminal_output +4243,6243188,"TERMINAL",0,0,"6957",,terminal_output +4244,6244214,"TERMINAL",0,0,"72068",,terminal_output +4245,6245236,"TERMINAL",0,0,"8179",,terminal_output +4246,6246260,"TERMINAL",0,0,"92850",,terminal_output +4247,6247289,"TERMINAL",0,0,"40391",,terminal_output +4248,6248354,"TERMINAL",0,0,"14502",,terminal_output +4249,6249388,"TERMINAL",0,0,"2513",,terminal_output +4250,6250459,"TERMINAL",0,0,"3624",,terminal_output +4251,6251440,"TERMINAL",0,0,"4735",,terminal_output +4252,6252500,"TERMINAL",0,0,"5846",,terminal_output +4253,6253534,"TERMINAL",0,0,"6957",,terminal_output +4254,6254587,"TERMINAL",0,0,"73068",,terminal_output +4255,6255687,"TERMINAL",0,0,"8179",,terminal_output +4256,6256660,"TERMINAL",0,0,"9286:00",,terminal_output +4257,6257665,"TERMINAL",0,0,"50391",,terminal_output +4258,6258704,"TERMINAL",0,0,"147:002",,terminal_output +4259,6259741,"TERMINAL",0,0,"2513",,terminal_output +4260,6260805,"TERMINAL",0,0,"3624",,terminal_output +4261,6261815,"TERMINAL",0,0,"4735",,terminal_output +4262,6262845,"TERMINAL",0,0,"5957",,terminal_output +4263,6263889,"TERMINAL",0,0,"74068",,terminal_output +4264,6264925,"TERMINAL",0,0,"8179",,terminal_output +4265,6266131,"TERMINAL",0,0,"92810",,terminal_output +4266,6266997,"TERMINAL",0,0,"5:00391",,terminal_output +4267,6268079,"TERMINAL",0,0,"14102",,terminal_output +4268,6269100,"TERMINAL",0,0,"2513",,terminal_output +4269,6270111,"TERMINAL",0,0,"3624",,terminal_output +4270,6271249,"TERMINAL",0,0,"4735",,terminal_output +4271,6272274,"TERMINAL",0,0,"5846",,terminal_output +4272,6273217,"TERMINAL",0,0,"6957",,terminal_output +4273,6274324,"TERMINAL",0,0,"75068",,terminal_output +4274,6275296,"TERMINAL",0,0,"8179",,terminal_output +4275,6276367,"TERMINAL",0,0,"92820",,terminal_output +4276,6277392,"TERMINAL",0,0,"10391",,terminal_output +4277,6278417,"TERMINAL",0,0,"14202",,terminal_output +4278,6279554,"TERMINAL",0,0,"2513",,terminal_output +4279,6280567,"TERMINAL",0,0,"3624",,terminal_output +4280,6281522,"TERMINAL",0,0,"4735",,terminal_output +4281,6282572,"TERMINAL",0,0,"5846",,terminal_output +4282,6283625,"TERMINAL",0,0,"6957",,terminal_output +4283,6284658,"TERMINAL",0,0,"76:0068",,terminal_output +4284,6285700,"TERMINAL",0,0,"8179",,terminal_output +4285,6286742,"TERMINAL",0,0,"92830",,terminal_output +4286,6287780,"TERMINAL",0,0,"20391",,terminal_output +4287,6288819,"TERMINAL",0,0,"15313",,terminal_output +4288,6289857,"TERMINAL",0,0,"3624",,terminal_output +4289,6290896,"TERMINAL",0,0,"4735",,terminal_output +4290,6292038,"TERMINAL",0,0,"5846",,terminal_output +4291,6293061,"TERMINAL",0,0,"6957",,terminal_output +4292,6294018,"TERMINAL",0,0,"71068",,terminal_output +4293,6295061,"TERMINAL",0,0,"8179",,terminal_output +4294,6296139,"TERMINAL",0,0,"92840",,terminal_output +4295,6297155,"TERMINAL",0,0,"30391",,terminal_output +4296,6298182,"TERMINAL",0,0,"14402",,terminal_output +4297,6299311,"TERMINAL",0,0,"2513",,terminal_output +4298,6300257,"TERMINAL",0,0,"3624",,terminal_output +4299,6301293,"TERMINAL",0,0,"4735",,terminal_output +4300,6302378,"TERMINAL",0,0,"5846",,terminal_output +4301,6303504,"TERMINAL",0,0,"6957",,terminal_output +4302,6304425,"TERMINAL",0,0,"72068",,terminal_output +4303,6305449,"TERMINAL",0,0,"8179",,terminal_output +4304,6306473,"TERMINAL",0,0,"92850",,terminal_output +4305,6307511,"TERMINAL",0,0,"40391",,terminal_output +4306,6308620,"TERMINAL",0,0,"14502",,terminal_output +4307,6309651,"TERMINAL",0,0,"2513",,terminal_output +4308,6310673,"TERMINAL",0,0,"3624",,terminal_output +4309,6311660,"TERMINAL",0,0,"4735",,terminal_output +4310,6312695,"TERMINAL",0,0,"5846",,terminal_output +4311,6313732,"TERMINAL",0,0,"6957",,terminal_output +4312,6314767,"TERMINAL",0,0,"73068",,terminal_output +4313,6315895,"TERMINAL",0,0,"8179",,terminal_output +4314,6316918,"TERMINAL",0,0,"9397:01",,terminal_output +4315,6317941,"TERMINAL",0,0,"5148:002",,terminal_output +4316,6318965,"TERMINAL",0,0,"2513",,terminal_output +4317,6319957,"TERMINAL",0,0,"3624",,terminal_output +4318,6321017,"TERMINAL",0,0,"4735",,terminal_output +4319,6322141,"TERMINAL",0,0,"5846",,terminal_output +4320,6323170,"TERMINAL",0,0,"6957",,terminal_output +4321,6324105,"TERMINAL",0,0,"74068",,terminal_output +4322,6325210,"TERMINAL",0,0,"8179",,terminal_output +4323,6326237,"TERMINAL",0,0,"92810",,terminal_output +4324,6327216,"TERMINAL",0,0,"6:00391",,terminal_output +4325,6328286,"TERMINAL",0,0,"14102",,terminal_output +4326,6329312,"TERMINAL",0,0,"2513",,terminal_output +4327,6330332,"TERMINAL",0,0,"3624",,terminal_output +4328,6331364,"TERMINAL",0,0,"4735",,terminal_output +4329,6332483,"TERMINAL",0,0,"5846",,terminal_output +4330,6333510,"TERMINAL",0,0,"6957",,terminal_output +4331,6334532,"TERMINAL",0,0,"75068",,terminal_output +4332,6335554,"TERMINAL",0,0,"8179",,terminal_output +4333,6336554,"TERMINAL",0,0,"92820",,terminal_output +4334,6337606,"TERMINAL",0,0,"10391",,terminal_output +4335,6338627,"TERMINAL",0,0,"14202",,terminal_output +4336,6339662,"TERMINAL",0,0,"2513",,terminal_output +4337,6340703,"TERMINAL",0,0,"3624",,terminal_output +4338,6341740,"TERMINAL",0,0,"4735",,terminal_output +4339,6342773,"TERMINAL",0,0,"5846",,terminal_output +4340,6343814,"TERMINAL",0,0,"6957",,terminal_output +4341,6344849,"TERMINAL",0,0,"87:0179",,terminal_output +4342,6345900,"TERMINAL",0,0,"92830",,terminal_output +4343,6346922,"TERMINAL",0,0,"20391",,terminal_output +4344,6348007,"TERMINAL",0,0,"14302",,terminal_output +4345,6349078,"TERMINAL",0,0,"2513",,terminal_output +4346,6350062,"TERMINAL",0,0,"3624",,terminal_output +4347,6351120,"TERMINAL",0,0,"4735",,terminal_output +4348,6352245,"TERMINAL",0,0,"5846",,terminal_output +4349,6353271,"TERMINAL",0,0,"6957",,terminal_output +4350,6354216,"TERMINAL",0,0,"71068",,terminal_output +4351,6355317,"TERMINAL",0,0,"8179",,terminal_output +4352,6356346,"TERMINAL",0,0,"92840",,terminal_output +4353,6357365,"TERMINAL",0,0,"30391",,terminal_output +4354,6358364,"TERMINAL",0,0,"14402",,terminal_output +4355,6359445,"TERMINAL",0,0,"2513",,terminal_output +4356,6360542,"TERMINAL",0,0,"3624",,terminal_output +4357,6361469,"TERMINAL",0,0,"4735",,terminal_output +4358,6362509,"TERMINAL",0,0,"5846",,terminal_output +4359,6363641,"TERMINAL",0,0,"6957",,terminal_output +4360,6364638,"TERMINAL",0,0,"72068",,terminal_output +4361,6365623,"TERMINAL",0,0,"8179",,terminal_output +4362,6366669,"TERMINAL",0,0,"92850",,terminal_output +4363,6367706,"TERMINAL",0,0,"40391",,terminal_output +4364,6368742,"TERMINAL",0,0,"14502",,terminal_output +4365,6369779,"TERMINAL",0,0,"2513",,terminal_output +4366,6370814,"TERMINAL",0,0,"3735",,terminal_output +4367,6371910,"TERMINAL",0,0,"5846",,terminal_output +4368,6372933,"TERMINAL",0,0,"6957",,terminal_output +4369,6373958,"TERMINAL",0,0,"73068",,terminal_output +4370,6374982,"TERMINAL",0,0,"8179",,terminal_output +4371,6376107,"TERMINAL",0,0,"9288:00",,terminal_output +4372,6377133,"TERMINAL",0,0,"50391",,terminal_output +4373,6378156,"TERMINAL",0,0,"149:002",,terminal_output +4374,6379179,"TERMINAL",0,0,"2513",,terminal_output +4375,6380146,"TERMINAL",0,0,"3624",,terminal_output +4376,6381201,"TERMINAL",0,0,"4735",,terminal_output +4377,6382252,"TERMINAL",0,0,"5846",,terminal_output +4378,6383375,"TERMINAL",0,0,"6957",,terminal_output +4379,6384401,"TERMINAL",0,0,"74068",,terminal_output +4380,6385352,"TERMINAL",0,0,"8179",,terminal_output +4381,6386441,"TERMINAL",0,0,"92810",,terminal_output +4382,6387470,"TERMINAL",0,0,"7:00391",,terminal_output +4383,6388476,"TERMINAL",0,0,"14102",,terminal_output +4384,6389521,"TERMINAL",0,0,"2513",,terminal_output +4385,6390545,"TERMINAL",0,0,"3624",,terminal_output +4386,6391565,"TERMINAL",0,0,"4735",,terminal_output +4387,6392629,"TERMINAL",0,0,"5846",,terminal_output +4388,6393719,"TERMINAL",0,0,"6957",,terminal_output +4389,6394679,"TERMINAL",0,0,"75068",,terminal_output +4390,6395724,"TERMINAL",0,0,"8179",,terminal_output +4391,6396752,"TERMINAL",0,0,"92820",,terminal_output +4392,6397794,"TERMINAL",0,0,"10391",,terminal_output +4393,6398837,"TERMINAL",0,0,"15213",,terminal_output +4394,6399966,"TERMINAL",0,0,"3624",,terminal_output +4395,6400990,"TERMINAL",0,0,"4735",,terminal_output +4396,6402016,"TERMINAL",0,0,"5846",,terminal_output +4397,6403039,"TERMINAL",0,0,"6957",,terminal_output +4398,6404064,"TERMINAL",0,0,"78:0068",,terminal_output +4399,6405050,"TERMINAL",0,0,"8179",,terminal_output +4400,6406092,"TERMINAL",0,0,"92830",,terminal_output +4401,6407135,"TERMINAL",0,0,"20391",,terminal_output +4402,6408261,"TERMINAL",0,0,"14302",,terminal_output +4403,6409287,"TERMINAL",0,0,"2513",,terminal_output +4404,6410234,"TERMINAL",0,0,"3624",,terminal_output +4405,6411295,"TERMINAL",0,0,"4735",,terminal_output +4406,6412357,"TERMINAL",0,0,"5846",,terminal_output +4407,6413380,"TERMINAL",0,0,"6957",,terminal_output +4408,6414416,"TERMINAL",0,0,"71068",,terminal_output +4409,6415451,"TERMINAL",0,0,"8179",,terminal_output +4410,6416488,"TERMINAL",0,0,"92840",,terminal_output +4411,6417523,"TERMINAL",0,0,"30391",,terminal_output +4412,6418560,"TERMINAL",0,0,"14402",,terminal_output +4413,6419596,"TERMINAL",0,0,"2513",,terminal_output +4414,6420677,"TERMINAL",0,0,"3624",,terminal_output +4415,6421665,"TERMINAL",0,0,"4735",,terminal_output +4416,6422702,"TERMINAL",0,0,"5846",,terminal_output +4417,6423740,"TERMINAL",0,0,"6957",,terminal_output +4418,6424779,"TERMINAL",0,0,"72068",,terminal_output +4419,6425870,"TERMINAL",0,0,"82850",,terminal_output +4420,6426850,"TERMINAL",0,0,"40391",,terminal_output +4421,6427887,"TERMINAL",0,0,"14502",,terminal_output +4422,6428946,"TERMINAL",0,0,"2513",,terminal_output +4423,6429962,"TERMINAL",0,0,"3624",,terminal_output +4424,6430999,"TERMINAL",0,0,"4735",,terminal_output +4425,6432117,"TERMINAL",0,0,"5846",,terminal_output +4426,6433142,"TERMINAL",0,0,"6957",,terminal_output +4427,6434168,"TERMINAL",0,0,"73068",,terminal_output +4428,6435145,"TERMINAL",0,0,"8179",,terminal_output +4429,6436213,"TERMINAL",0,0,"9289:00",,terminal_output +4430,6437246,"TERMINAL",0,0,"50391",,terminal_output +4431,6438366,"TERMINAL",0,0,"1410:002",,terminal_output +4432,6439392,"TERMINAL",0,0,"2513",,terminal_output +4433,6440325,"TERMINAL",0,0,"3624",,terminal_output +4434,6441439,"TERMINAL",0,0,"4735",,terminal_output +4435,6442464,"TERMINAL",0,0,"5846",,terminal_output +4436,6443487,"TERMINAL",0,0,"6957",,terminal_output +4437,6444508,"TERMINAL",0,0,"74068",,terminal_output +4438,6445533,"TERMINAL",0,0,"8179",,terminal_output +4439,6446550,"TERMINAL",0,0,"92810",,terminal_output +4440,6447589,"TERMINAL",0,0,"8:00391",,terminal_output +4441,6448709,"TERMINAL",0,0,"14102",,terminal_output +4442,6449671,"TERMINAL",0,0,"2513",,terminal_output +4443,6450708,"TERMINAL",0,0,"3624",,terminal_output +4444,6451746,"TERMINAL",0,0,"4735",,terminal_output +4445,6452782,"TERMINAL",0,0,"5846",,terminal_output +4446,6453822,"TERMINAL",0,0,"65068",,terminal_output +4447,6454853,"TERMINAL",0,0,"8179",,terminal_output +4448,6455978,"TERMINAL",0,0,"92820",,terminal_output +4449,6457002,"TERMINAL",0,0,"10391",,terminal_output +4450,6457961,"TERMINAL",0,0,"14202",,terminal_output +4451,6459051,"TERMINAL",0,0,"2513",,terminal_output +4452,6460040,"TERMINAL",0,0,"3624",,terminal_output +4453,6461079,"TERMINAL",0,0,"4735",,terminal_output +4454,6462116,"TERMINAL",0,0,"5846",,terminal_output +4455,6463252,"TERMINAL",0,0,"6957",,terminal_output +4456,6464273,"TERMINAL",0,0,"79:0068",,terminal_output +4457,6465227,"TERMINAL",0,0,"8179",,terminal_output +4458,6466320,"TERMINAL",0,0,"92830",,terminal_output +4459,6467302,"TERMINAL",0,0,"20391",,terminal_output +4460,6468367,"TERMINAL",0,0,"14302",,terminal_output +4461,6469393,"TERMINAL",0,0,"2513",,terminal_output +4462,6470419,"TERMINAL",0,0,"3624",,terminal_output +4463,6471452,"TERMINAL",0,0,"4735",,terminal_output +4464,6472490,"TERMINAL",0,0,"5846",,terminal_output +4465,6473601,"TERMINAL",0,0,"6957",,terminal_output +4466,6474621,"TERMINAL",0,0,"71068",,terminal_output +4467,6475670,"TERMINAL",0,0,"8179",,terminal_output +4468,6476666,"TERMINAL",0,0,"92840",,terminal_output +4469,6477687,"TERMINAL",0,0,"30391",,terminal_output +4470,6478723,"TERMINAL",0,0,"14402",,terminal_output +4471,6479762,"TERMINAL",0,0,"2513",,terminal_output +4472,6480803,"TERMINAL",0,0,"3624",,terminal_output +4473,6481884,"TERMINAL",0,0,"4846",,terminal_output +4474,6482881,"TERMINAL",0,0,"6957",,terminal_output +4475,6483937,"TERMINAL",0,0,"72068",,terminal_output +4476,6485046,"TERMINAL",0,0,"8179",,terminal_output +4477,6486088,"TERMINAL",0,0,"92850",,terminal_output +4478,6487110,"TERMINAL",0,0,"40391",,terminal_output +4479,6488069,"TERMINAL",0,0,"14502",,terminal_output +4480,6489158,"TERMINAL",0,0,"2513",,terminal_output +4481,6490143,"TERMINAL",0,0,"3624",,terminal_output +4482,6491206,"TERMINAL",0,0,"4735",,terminal_output +4483,6492228,"TERMINAL",0,0,"5846",,terminal_output +4484,6493353,"TERMINAL",0,0,"6957",,terminal_output +4485,6494287,"TERMINAL",0,0,"73068",,terminal_output +4486,6495329,"TERMINAL",0,0,"8179",,terminal_output +4487,6496427,"TERMINAL",0,0,"92850:00",,terminal_output +4488,6497450,"TERMINAL",0,0,"50391",,terminal_output +4489,6498475,"TERMINAL",0,0,"141:002",,terminal_output +4490,6499480,"TERMINAL",0,0,"2513",,terminal_output +4491,6500655,"TERMINAL",0,0,"3624",,terminal_output +4492,6501555,"TERMINAL",0,0,"4735",,terminal_output +4493,6502589,"TERMINAL",0,0,"5846",,terminal_output +4494,6503671,"TERMINAL",0,0,"6957",,terminal_output +4495,6504676,"TERMINAL",0,0,"74068",,terminal_output +4496,6505745,"TERMINAL",0,0,"8179",,terminal_output +4497,6506749,"TERMINAL",0,0,"92810",,terminal_output +4498,6507786,"TERMINAL",0,0,"9:00391",,terminal_output +4499,6508823,"TERMINAL",0,0,"15113",,terminal_output +4500,6509944,"TERMINAL",0,0,"3624",,terminal_output +4501,6510903,"TERMINAL",0,0,"4735",,terminal_output +4502,6512098,"TERMINAL",0,0,"5846",,terminal_output +4503,6513017,"TERMINAL",0,0,"6957",,terminal_output +4504,6514041,"TERMINAL",0,0,"75068",,terminal_output +4505,6515046,"TERMINAL",0,0,"8179",,terminal_output +4506,6516193,"TERMINAL",0,0,"92820",,terminal_output +4507,6517214,"TERMINAL",0,0,"10391",,terminal_output +4508,6518158,"TERMINAL",0,0,"14202",,terminal_output +4509,6519262,"TERMINAL",0,0,"2513",,terminal_output +4510,6520245,"TERMINAL",0,0,"3624",,terminal_output +4511,6521309,"TERMINAL",0,0,"4735",,terminal_output +4512,6522440,"TERMINAL",0,0,"5846",,terminal_output +4513,6523458,"TERMINAL",0,0,"6957",,terminal_output +4514,6524441,"TERMINAL",0,0,"710:0068",,terminal_output +4515,6525512,"TERMINAL",0,0,"8179",,terminal_output +4516,6526513,"TERMINAL",0,0,"92830",,terminal_output +4517,6527559,"TERMINAL",0,0,"20391",,terminal_output +4518,6528682,"TERMINAL",0,0,"14302",,terminal_output +4519,6529668,"TERMINAL",0,0,"2513",,terminal_output +4520,6530680,"TERMINAL",0,0,"3624",,terminal_output +4521,6531768,"TERMINAL",0,0,"4735",,terminal_output +4522,6532852,"TERMINAL",0,0,"5846",,terminal_output +4523,6533819,"TERMINAL",0,0,"61068",,terminal_output +4524,6534931,"TERMINAL",0,0,"8179",,terminal_output +4525,6535899,"TERMINAL",0,0,"92840",,terminal_output +4526,6536976,"TERMINAL",0,0,"30391",,terminal_output +4527,6538003,"TERMINAL",0,0,"14402",,terminal_output +4528,6539128,"TERMINAL",0,0,"2513",,terminal_output +4529,6540078,"TERMINAL",0,0,"3624",,terminal_output +4530,6541181,"TERMINAL",0,0,"4735",,terminal_output +4531,6542205,"TERMINAL",0,0,"5846",,terminal_output +4532,6543200,"TERMINAL",0,0,"6957",,terminal_output +4533,6544253,"TERMINAL",0,0,"72068",,terminal_output +4534,6545278,"TERMINAL",0,0,"8179",,terminal_output +4535,6546433,"TERMINAL",0,0,"92850",,terminal_output +4536,6547426,"TERMINAL",0,0,"40391",,terminal_output +4537,6548449,"TERMINAL",0,0,"14502",,terminal_output +4538,6549472,"TERMINAL",0,0,"2513",,terminal_output +4539,6550495,"TERMINAL",0,0,"3624",,terminal_output +4540,6551527,"TERMINAL",0,0,"4735",,terminal_output +4541,6552563,"TERMINAL",0,0,"5846",,terminal_output +4542,6553688,"TERMINAL",0,0,"6957",,terminal_output +4543,6554671,"TERMINAL",0,0,"73068",,terminal_output +4544,6555718,"TERMINAL",0,0,"8179",,terminal_output +4545,6556724,"TERMINAL",0,0,"9281:00",,terminal_output +4546,6557762,"TERMINAL",0,0,"50391",,terminal_output +4547,6558810,"TERMINAL",0,0,"142:002",,terminal_output +4548,6559857,"TERMINAL",0,0,"2624",,terminal_output +4549,6560937,"TERMINAL",0,0,"4735",,terminal_output +4550,6561964,"TERMINAL",0,0,"5846",,terminal_output +4551,6562969,"TERMINAL",0,0,"6957",,terminal_output +4552,6564007,"TERMINAL",0,0,"74068",,terminal_output +4553,6565038,"TERMINAL",0,0,"8179",,terminal_output +4554,6566075,"TERMINAL",0,0,"92810",,terminal_output +4555,6567184,"TERMINAL",0,0,"50:00391",,terminal_output +4556,6568151,"TERMINAL",0,0,"14102",,terminal_output +4557,6569240,"TERMINAL",0,0,"2513",,terminal_output +4558,6570258,"TERMINAL",0,0,"3624",,terminal_output +4559,6571284,"TERMINAL",0,0,"4735",,terminal_output +4560,6572409,"TERMINAL",0,0,"5846",,terminal_output +4561,6573349,"TERMINAL",0,0,"6957",,terminal_output +4562,6574557,"TERMINAL",0,0,"75068",,terminal_output +4563,6575582,"TERMINAL",0,0,"8179",,terminal_output +4564,6576549,"TERMINAL",0,0,"92820",,terminal_output +4565,6577588,"TERMINAL",0,0,"10391",,terminal_output +4566,6578626,"TERMINAL",0,0,"14202",,terminal_output +4567,6579704,"TERMINAL",0,0,"2513",,terminal_output +4568,6580703,"TERMINAL",0,0,"3624",,terminal_output +4569,6581736,"TERMINAL",0,0,"4735",,terminal_output +4570,6582774,"TERMINAL",0,0,"5846",,terminal_output +4571,6583813,"TERMINAL",0,0,"6957",,terminal_output +4572,6584848,"TERMINAL",0,0,"81:0179",,terminal_output +4573,6585888,"TERMINAL",0,0,"92830",,terminal_output +4574,6586953,"TERMINAL",0,0,"20391",,terminal_output +4575,6587958,"TERMINAL",0,0,"14302",,terminal_output +4576,6588996,"TERMINAL",0,0,"2513",,terminal_output +4577,6590064,"TERMINAL",0,0,"3624",,terminal_output +4578,6591080,"TERMINAL",0,0,"4735",,terminal_output +4579,6592175,"TERMINAL",0,0,"5846",,terminal_output +4580,6593200,"TERMINAL",0,0,"6957",,terminal_output +4581,6594224,"TERMINAL",0,0,"71068",,terminal_output +4582,6595247,"TERMINAL",0,0,"8179",,terminal_output +4583,6596282,"TERMINAL",0,0,"92840",,terminal_output +4584,6597397,"TERMINAL",0,0,"30391",,terminal_output +4585,6598362,"TERMINAL",0,0,"14402",,terminal_output +4586,6599444,"TERMINAL",0,0,"2513",,terminal_output +4587,6600432,"TERMINAL",0,0,"3624",,terminal_output +4588,6601479,"TERMINAL",0,0,"4735",,terminal_output +4589,6602506,"TERMINAL",0,0,"5846",,terminal_output +4590,6603642,"TERMINAL",0,0,"6957",,terminal_output +4591,6604671,"TERMINAL",0,0,"72068",,terminal_output +4592,6605615,"TERMINAL",0,0,"8179",,terminal_output +4593,6606711,"TERMINAL",0,0,"92850",,terminal_output +4594,6607693,"TERMINAL",0,0,"40391",,terminal_output +4595,6608734,"TERMINAL",0,0,"14502",,terminal_output +4596,6609790,"TERMINAL",0,0,"2513",,terminal_output +4597,6610806,"TERMINAL",0,0,"3624",,terminal_output +4598,6611934,"TERMINAL",0,0,"4846",,terminal_output +4599,6612960,"TERMINAL",0,0,"6957",,terminal_output +4600,6613924,"TERMINAL",0,0,"73068",,terminal_output +4601,6615009,"TERMINAL",0,0,"8179",,terminal_output +4602,6616030,"TERMINAL",0,0,"9282:00",,terminal_output +4603,6617055,"TERMINAL",0,0,"50391",,terminal_output +4604,6618179,"TERMINAL",0,0,"143:002",,terminal_output +4605,6619207,"TERMINAL",0,0,"2513",,terminal_output +4606,6620150,"TERMINAL",0,0,"3624",,terminal_output +4607,6621252,"TERMINAL",0,0,"4735",,terminal_output +4608,6622281,"TERMINAL",0,0,"5846",,terminal_output +4609,6623266,"TERMINAL",0,0,"6957",,terminal_output +4610,6624299,"TERMINAL",0,0,"74068",,terminal_output +4611,6625351,"TERMINAL",0,0,"8179",,terminal_output +4612,6626452,"TERMINAL",0,0,"92810",,terminal_output +4613,6627511,"TERMINAL",0,0,"1:00391",,terminal_output +4614,6628448,"TERMINAL",0,0,"14102",,terminal_output +4615,6629549,"TERMINAL",0,0,"2513",,terminal_output +4616,6630580,"TERMINAL",0,0,"3624",,terminal_output +4617,6631563,"TERMINAL",0,0,"4735",,terminal_output +4618,6632600,"TERMINAL",0,0,"5846",,terminal_output +4619,6633748,"TERMINAL",0,0,"6957",,terminal_output +4620,6634772,"TERMINAL",0,0,"75068",,terminal_output +4621,6635716,"TERMINAL",0,0,"8179",,terminal_output +4622,6636751,"TERMINAL",0,0,"92820",,terminal_output +4623,6637791,"TERMINAL",0,0,"10391",,terminal_output +4624,6638865,"TERMINAL",0,0,"15213",,terminal_output +4625,6639872,"TERMINAL",0,0,"3624",,terminal_output +4626,6640918,"TERMINAL",0,0,"4735",,terminal_output +4627,6641960,"TERMINAL",0,0,"5846",,terminal_output +4628,6643067,"TERMINAL",0,0,"6957",,terminal_output +4629,6644063,"TERMINAL",0,0,"72:0068",,terminal_output +4630,6645064,"TERMINAL",0,0,"8179",,terminal_output +4631,6646135,"TERMINAL",0,0,"92830",,terminal_output +4632,6647135,"TERMINAL",0,0,"20391",,terminal_output +4633,6648170,"TERMINAL",0,0,"14302",,terminal_output +4634,6649317,"TERMINAL",0,0,"2513",,terminal_output +4635,6650333,"TERMINAL",0,0,"3624",,terminal_output +4636,6651357,"TERMINAL",0,0,"4735",,terminal_output +4637,6652320,"TERMINAL",0,0,"5846",,terminal_output +4638,6653408,"TERMINAL",0,0,"6957",,terminal_output +4639,6654401,"TERMINAL",0,0,"71068",,terminal_output +4640,6655440,"TERMINAL",0,0,"8179",,terminal_output +4641,6656480,"TERMINAL",0,0,"92840",,terminal_output +4642,6657519,"TERMINAL",0,0,"30391",,terminal_output +4643,6658630,"TERMINAL",0,0,"14402",,terminal_output +4644,6659598,"TERMINAL",0,0,"2513",,terminal_output +4645,6660708,"TERMINAL",0,0,"3624",,terminal_output +4646,6661702,"TERMINAL",0,0,"4735",,terminal_output +4647,6662706,"TERMINAL",0,0,"5846",,terminal_output +4648,6663744,"TERMINAL",0,0,"6957",,terminal_output +4649,6664780,"TERMINAL",0,0,"72068",,terminal_output +4650,6665817,"TERMINAL",0,0,"82850",,terminal_output +4651,6666859,"TERMINAL",0,0,"40391",,terminal_output +4652,6667946,"TERMINAL",0,0,"14502",,terminal_output +4653,6668974,"TERMINAL",0,0,"2513",,terminal_output +4654,6669998,"TERMINAL",0,0,"3624",,terminal_output +4655,6671021,"TERMINAL",0,0,"4735",,terminal_output +4656,6672146,"TERMINAL",0,0,"5846",,terminal_output +4657,6673172,"TERMINAL",0,0,"6957",,terminal_output +4658,6674199,"TERMINAL",0,0,"73068",,terminal_output +4659,6675151,"TERMINAL",0,0,"8179",,terminal_output +4660,6676242,"TERMINAL",0,0,"9283:00",,terminal_output +4661,6677268,"TERMINAL",0,0,"50391",,terminal_output +4662,6678291,"TERMINAL",0,0,"144:002",,terminal_output +4663,6680028,"TERMINAL",0,0,"2624",,terminal_output +4664,6680962,"TERMINAL",0,0,"4735",,terminal_output +4665,6682080,"TERMINAL",0,0,"5846",,terminal_output +4666,6683052,"TERMINAL",0,0,"6957",,terminal_output +4667,6684103,"TERMINAL",0,0,"74068",,terminal_output +4668,6685142,"TERMINAL",0,0,"8179",,terminal_output +4669,6686276,"TERMINAL",0,0,"92810",,terminal_output +4670,6687299,"TERMINAL",0,0,"2:00391",,terminal_output +4671,6688276,"TERMINAL",0,0,"14102",,terminal_output +4672,6689347,"TERMINAL",0,0,"2513",,terminal_output +4673,6690474,"TERMINAL",0,0,"3624",,terminal_output +4674,6691441,"TERMINAL",0,0,"4735",,terminal_output +4675,6692499,"TERMINAL",0,0,"5846",,terminal_output +4676,6693493,"TERMINAL",0,0,"6957",,terminal_output +4677,6694572,"TERMINAL",0,0,"75068",,terminal_output +4678,6695612,"TERMINAL",0,0,"8179",,terminal_output +4679,6696720,"TERMINAL",0,0,"92820",,terminal_output +4680,6697678,"TERMINAL",0,0,"10391",,terminal_output +4681,6698723,"TERMINAL",0,0,"14202",,terminal_output +4682,6699775,"TERMINAL",0,0,"2513",,terminal_output +4683,6700821,"TERMINAL",0,0,"3735",,terminal_output +4684,6701908,"TERMINAL",0,0,"5846",,terminal_output +4685,6702920,"TERMINAL",0,0,"6957",,terminal_output +4686,6703994,"TERMINAL",0,0,"73:0068",,terminal_output +4687,6705120,"TERMINAL",0,0,"8179",,terminal_output +4688,6706146,"TERMINAL",0,0,"92830",,terminal_output +4689,6707168,"TERMINAL",0,0,"20391",,terminal_output +4690,6708152,"TERMINAL",0,0,"14302",,terminal_output +4691,6709219,"TERMINAL",0,0,"2513",,terminal_output +4692,6710345,"TERMINAL",0,0,"3624",,terminal_output +4693,6711364,"TERMINAL",0,0,"4735",,terminal_output +4694,6712332,"TERMINAL",0,0,"5846",,terminal_output +4695,6713412,"TERMINAL",0,0,"6957",,terminal_output +4696,6714428,"TERMINAL",0,0,"71068",,terminal_output diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-94286bd5-ed1d-4a7b-9c23-3104eeb712871757340853772-2025_09_08-16.14.51.932/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-94286bd5-ed1d-4a7b-9c23-3104eeb712871757340853772-2025_09_08-16.14.51.932/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..5496f0cf53bd60e9e40c343d11f3ab4a009e029e --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-94286bd5-ed1d-4a7b-9c23-3104eeb712871757340853772-2025_09_08-16.14.51.932/source.csv @@ -0,0 +1,2047 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +1,6,"utils/evaluate_lam.py",0,0,"",python,tab +2,914,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"4:14:51 PM [info] Activating crowd-code\n4:14:51 PM [info] Recording started\n4:14:51 PM [info] Initializing git provider using file system watchers...\n",Log,tab +3,1470,"extension-output-pdoom-org.crowd-code-#1-crowd-code",150,0,"4:14:52 PM [info] Git repository found\n4:14:52 PM [info] Git provider initialized successfully\n4:14:52 PM [info] Initial git state: [object Object]\n",Log,content +4,2481,"utils/evaluate_lam.py",0,0,"",python,tab +5,11956,"TERMINAL",0,0,"source .venv/bin/activate",,terminal_command +6,11981,"TERMINAL",0,0,"]633;C]0;tum_cte0515@hkn1990:~/Projects/jasmine",,terminal_output +7,17727,"TERMINAL",0,0,"cd $ws_dir",,terminal_command +8,17802,"TERMINAL",0,0,"]633;C]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared",,terminal_output +9,22499,"TERMINAL",0,0,"cd data_coinrun/",,terminal_command +10,22886,"TERMINAL",0,0,"ls",,terminal_command +11,22925,"TERMINAL",0,0,"]633;Carray_records coinrun_episodes coinrun_episodes_10m coinrun_episodes_10m_chunked coinrun_episodes_10m_chunked_2 coinrun_episodes_10m_gt_actions coinrun_episodes_10m_gt_actions_split dev\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun",,terminal_output +12,27732,"TERMINAL",0,0,"cd coinrun_episodes_10m_chunked",,terminal_command +13,33842,"TERMINAL",0,0,"ls -l | wc -l",,terminal_command +14,33889,"TERMINAL",0,0,"]633;C",,terminal_output +15,43502,"TERMINAL",0,0,"6073\r\n]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked",,terminal_output +16,53276,"TERMINAL",0,0,"queue",,terminal_command +17,53341,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1990.localdomain: Mon Sep 8 16:15:45 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3476346 cpuonly preproce tum_cte0 R 1:11:06\t 1 hkn03383476347 cpuonly preproce tum_cte0 R 1:11:06\t 1 hkn03543476960 large preproce tum_cte0 R 1:07:03\t 1 hkn1901",,terminal_output +18,54381,"TERMINAL",0,0,"6774",,terminal_output +19,54671,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1990:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_chunked",,terminal_output +20,237872,"TERMINAL",0,0,"bash",,terminal_focus +21,239323,"utils/evaluate_lam.py",0,0,"",python,tab +22,244772,"TERMINAL",0,0,"bash",,terminal_focus +23,271466,"TERMINAL",0,0,"bash",,terminal_focus +24,354433,"TERMINAL",0,0,"git status",,terminal_command +25,354544,"TERMINAL",0,0,"]633;COn branch input_pipeline/add-npy2array_record\r\nYour branch is up to date with 'origin/input_pipeline/add-npy2array_record'.\r\n\r\nUntracked files:\r\n (use ""git add ..."" to include in what will be committed)\r\n\tdata/\r\n\tdiff.diff\r\n\tgenerate_dataset_chunk.py\r\n\tkiller.sh\r\n\tkiller_partition.sh\r\n\tlog.log\r\n\tlogs/\r\n\toverfit_dir.zip\r\n\tread_tf_record.py\r\n\trequirements-franz.txt\r\n\tsamples/\r\n\tscripts_cremers/\r\n\tslurm/\r\n\tutils/visualizer.py\r\n\r\nnothing added to commit but untracked files present (use ""git add"" to track)\r\n]0;tum_cte0515@hkn1990:~/Projects/jasmine",,terminal_output +26,359265,"TERMINAL",0,0,"git add generate_dataset_chunk.py",,terminal_command +27,359321,"TERMINAL",0,0,"]633;C",,terminal_output +28,359510,"TERMINAL",0,0,"]0;tum_cte0515@hkn1990:~/Projects/jasmine",,terminal_output +29,370845,"TERMINAL",0,0,"git commit -am ""added cpinrun generation in chunked format""",,terminal_command +30,370906,"TERMINAL",0,0,"]633;C",,terminal_output +31,371002,"TERMINAL",0,0,"[input_pipeline/add-npy2array_record 32692f9] added cpinrun generation in chunked format\r\n 1 file changed, 89 insertions(+)\r\n create mode 100644 generate_dataset_chunk.py\r\n]0;tum_cte0515@hkn1990:~/Projects/jasmine",,terminal_output +32,372331,"TERMINAL",0,0,"git push",,terminal_command +33,372393,"TERMINAL",0,0,"]633;C",,terminal_output +34,373776,"TERMINAL",0,0,"Enumerating objects: 4, done.\r\nCounting objects: 25% (1/4)\rCounting objects: 50% (2/4)\rCounting objects: 75% (3/4)\rCounting objects: 100% (4/4)\rCounting objects: 100% (4/4), done.\r\nDelta compression using up to 152 threads\r\nCompressing objects: 33% (1/3)\rCompressing objects: 66% (2/3)\rCompressing objects: 100% (3/3)\rCompressing objects: 100% (3/3), done.\r\nWriting objects: 33% (1/3)\rWriting objects: 66% (2/3)\rWriting objects: 100% (3/3)\rWriting objects: 100% (3/3), 1.45 KiB | 743.00 KiB/s, done.\r\nTotal 3 (delta 1), reused 0 (delta 0), pack-reused 0\r\nremote: Resolving deltas: 0% (0/1)\rremote: Resolving deltas: 100% (1/1)\rremote: Resolving deltas: 100% (1/1), completed with 1 local object.\r\n",,terminal_output +35,374195,"TERMINAL",0,0,"To github.com:p-doom/jasmine.git\r\n 1c3f025..32692f9 input_pipeline/add-npy2array_record -> input_pipeline/add-npy2array_record\r\n]0;tum_cte0515@hkn1990:~/Projects/jasmine",,terminal_output +36,376865,"TERMINAL",0,0,"git branch",,terminal_command +37,377048,"TERMINAL",0,0,"]633;C[?1h=\r add-wandb-name-and-tags\r\n before-nnx\r\n causal-mem-reduce\r\n causal-spatiotemporal-kv-cache\r\n causal-st-transformer\r\n causal-transformer-dynamics-model\r\n causal-transformer-nnx-no-kv-cache\r\n coinrun-gt-actions\r\n convert-to-jax-array-in-iter\r\n correct-batched-sampling\r\n dev\r\n dont-let-tf-see-gpu\r\n feat/darkness-filter\r\n feat/explicit-image-dims\r\n fix-action-padding-lam-future-information-access\r\n fix-sampling\r\n fix-transformer-forwardpass\r\n fix/spatiotemporal-pe-once-in-STTransformer\r\n grad-norm-log-and-clip\r\n grain-dataloader\r\n* input_pipeline/add-npy2array_record\r\n logging-variants\r\n lr-schedules\r\n main\r\n maskgit-different-maskprob-per-sample\r\n maskgit-sampling-iterative-unmasking-fix\r\n metrics-logging-for-dynamics-model\r\n monkey-patch\r\n new-arch-sampling\r\n preprocess_video\r\n refactor-tmp\r\n revised-dataloader\r\n runner\r\n runner-grain\r\n sample-ali-branch\r\n sample-from-different-topologies\r\n sampling-startframe-indexing-fix\r\n speedup-tfrecord-preprocessing\r\n tmp\r\n train_lam_coinrun_ablation_wsd_3e-6_28747\r\n validation-loss\r\n\r[?1l>]0;tum_cte0515@hkn1990:~/Projects/jasmine",,terminal_output +38,407037,"TERMINAL",0,0,"git checkout action-mapper",,terminal_command +39,407117,"TERMINAL",0,0,"]633;C",,terminal_output +40,407832,"TERMINAL",0,0,"branch 'action-mapper' set up to track 'origin/action-mapper'.\r\nSwitched to a new branch 'action-mapper'\r\n]0;tum_cte0515@hkn1990:~/Projects/jasmine",,terminal_output +41,409795,"TERMINAL",0,0,"git pull",,terminal_command +42,409863,"TERMINAL",0,0,"]633;C",,terminal_output +43,410490,"",0,0,"Switched from branch 'input_pipeline/add-npy2array_record' to 'action-mapper'",,git_branch_checkout +44,411682,"TERMINAL",0,0,"remote: Enumerating objects: 4, done.\r\nremote: Counting objects: 25% (1/4)\rremote: Counting objects: 50% (2/4)\rremote: Counting objects: 75% (3/4)\rremote: Counting objects: 100% (4/4)\rremote: Counting objects: 100% (4/4), done.\r\nremote: Compressing objects: 50% (1/2)\rremote: Compressing objects: 100% (2/2)\rremote: Compressing objects: 100% (2/2), done.\r\nremote: Total 4 (delta 2), reused 3 (delta 2), pack-reused 0 (from 0)\r\nUnpacking objects: 25% (1/4)\rUnpacking objects: 50% (2/4)\rUnpacking objects: 75% (3/4)\rUnpacking objects: 100% (4/4)\rUnpacking objects: 100% (4/4), 1.28 KiB | 12.00 KiB/s, done.\r\n",,terminal_output +45,411881,"TERMINAL",0,0,"From github.com:p-doom/jasmine\r\n * [new branch] demo-notebook -> origin/demo-notebook\r\n 49fb80d..975ead7 main -> origin/main\r\n",,terminal_output +46,412050,"TERMINAL",0,0,"Already up to date.\r\n]0;tum_cte0515@hkn1990:~/Projects/jasmine",,terminal_output +47,442859,"TERMINAL",0,0,"git merge validation-loss",,terminal_command +48,442969,"TERMINAL",0,0,"]633;C",,terminal_output +49,443244,"TERMINAL",0,0,"hint: Waiting for your editor to close the file... ",,terminal_output +50,443411,"TERMINAL",0,0,"[?1049h[>4;2m[?1h=[?2004h[?1004h[?12h[?12l[?25l""~/Projects/jasmine/.git/MERGE_MSG"" 6L, 276B▽ Pzz\[0%m [>c]10;?]11;?Merge branch 'validation-loss' into action-mapper\r\n# Please enter a commit message to explain why this merge is necessary,# especially if it merges an updated upstream into a topic branch.#\r\n# Lines starting with '#' will be ignored, and an empty message aborts\r\n# the commit.\r\n~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ 1,1All[?25h",,terminal_output +51,443456,"TERMINAL",0,0,"P+q436f\P+q6b75\P+q6b64\P+q6b72\P+q6b6c\P+q2332\P+q2334\P+q2569\P+q2a37\P+q6b31\[?12$p[?25l/3333/3333 [?25h[?25l/f6f6/e3e3 [?25h",,terminal_output +52,444973,"TERMINAL",0,0,"[?25l^[",,terminal_output +53,445109,"TERMINAL",0,0," ^[ [?25h",,terminal_output +54,445232,"TERMINAL",0,0,"[?25l::[?25h",,terminal_output +55,445361,"TERMINAL",0,0,"w",,terminal_output +56,445528,"TERMINAL",0,0,"q",,terminal_output +57,446046,"TERMINAL",0,0,"\r[?25l[?2004l[>4;m"".git/MERGE_MSG"" 6L, 276B written\r\r\r\n[?1004l[?2004l[?1l>[?25h[>4;m[?1049l\rMerge made by the 'ort' strategy.\r\n",,terminal_output +58,446207,"TERMINAL",0,0," README.md | 117 +++++--\r\n generate_dataset.py | 114 +++++--\r\n genie.py | 599 +++++++++++++++++++++-------------\r\n input_pipeline/download/download_array_records.sh | 2 +-\r\n input_pipeline/download/openai/download_actions_files.py | 93 ++++++\r\n input_pipeline/preprocess/npy_to_tfrecord.py | 125 --------\r\n input_pipeline/preprocess/pngs_to_array_records.py | 130 ++++++++\r\n input_pipeline/preprocess/video_to_array_records.py | 68 ++--\r\n input_pipeline/preprocess/video_to_npy.py | 117 -------\r\n models/dynamics.py | 176 ++++++++--\r\n models/lam.py | 146 ++++++---\r\n models/tokenizer.py | 125 +++++---\r\n requirements.txt | 12 +-\r\n sample.py | 303 +++++++++++-------\r\n tests/data/generate_dummy_data.py | 42 +--\r\n tests/test_dataloader.py | 18 +-\r\n train_dynamics.py | 352 +++++++++++++-------\r\n train_lam.py | 329 +++++++++++++------\r\n train_tokenizer.py | 310 ++++++++++++------\r\n utils/dataloader.py | 62 +++-\r\n utils/dataset_utils.py | 164 +++++-----\r\n utils/lr_utils.py | 35 +-\r\n utils/nn.py | 638 +++++++++++++++++++++++++++----------\r\n utils/parameter_utils.py | 12 +-\r\n utils/preprocess.py | 1 -\r\n 25 files changed, 2706 insertions(+), 1384 deletions(-)\r\n create mode 100644 input_pipeline/download/openai/download_actions_files.py\r\n delete mode 100644 input_pipeline/preprocess/npy_to_tfrecord.py\r\n create mode 100644 input_pipeline/preprocess/pngs_to_array_records.py\r\n delete mode 100644 input_pipeline/preprocess/video_to_npy.py\r\n]0;tum_cte0515@hkn1990:~/Projects/jasmine",,terminal_output +59,478151,"train_lamap.py",0,0,"from dataclasses import dataclass, field\nimport os\n\nfrom flax.training.train_state import TrainState\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\n\nfrom models.lamap import LatentActionMapper\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n warmup_steps: int = 5000\n lr_schedule : str = ""wsd"" # supported options: wsd, cos\n # LAM\n model_dim: int = 512\n latent_dim: int = 32\n num_latents: int = 6\n patch_size: int = 16\n num_blocks: int = 8\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.0\n param_dtype: jnp.dtype = jnp.float32\n dtype: jnp.dtype = jnp.bfloat16\n # LAMAP\n # FIXME (f.srambical): this assumes that the number of actions is the same as the number of latent actions\n action_dim: int = 32\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_lamap""\n tags: list[str] = field(default_factory=lambda: [""lamap""])\n log_interval: int = 5\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\ndef lamap_loss_fn(params, state, inputs):\n # --- Compute loss ---\n inputs[""videos""] = inputs[""videos""].astype(args.dtype) / 255.0\n outputs = state.apply_fn(\n params, inputs, training=True, rngs={""dropout"": inputs[""rng""]}\n )\n loss = optax.softmax_cross_entropy(outputs[""action_predictions""], inputs[""actions""]).mean()\n return loss\n\n@jax.jit\ndef train_step(state, inputs):\n grad_fn = jax.value_and_grad(lamap_loss_fn, allow_int=True)\n loss, grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n return state, loss\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n\n # --- Initialize model ---\n lamap = LatentActionMapper(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n latent_dim=args.latent_dim,\n action_dim=args.action_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n )\n\n image_shape = (args.image_height, args.image_width, args.image_channels)\n rng, _rng = jax.random.split(rng)\n inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=args.dtype,\n ),\n rng=_rng,\n )\n rng, _rng = jax.random.split(rng)\n init_params = lamap.init(_rng, inputs)\n\n param_counts = count_parameters_by_component(init_params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(args.lr_schedule, \n args.init_lr, \n args.max_lr, \n args.decay_end, \n args.num_steps, \n args.warmup_steps, \n args.wsd_decay_steps)\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4, mu_dtype=args.dtype)\n train_state = TrainState.create(apply_fn=lamap.apply, params=init_params, tx=tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler\n )\n handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler) # type: ignore\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_train_state = jax.tree_util.tree_map(\n ocp.utils.to_shape_dtype_struct, train_state\n )\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_train_state),\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator),\n ),\n )\n train_state = restored[""model_state""]\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader = (jax.make_array_from_process_local_data(videos_sharding, elem) for elem in grain_iterator) # type: ignore\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng = jax.random.split(rng)\n\n # FIXME (f.srambical): use mock actions for now\n rng, _rng_actions = jax.random.split(rng)\n actions = jax.random.uniform(\n _rng_actions,\n (per_device_batch_size_for_init, args.action_dim),\n dtype=args.dtype,\n )\n\n inputs = dict(\n videos=videos,\n actions=actions,\n rng=_rng\n )\n train_state, loss = train_step(\n train_state, inputs\n )\n lr = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n ""lr"": lr,\n }\n )\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.StandardSave(train_state),\n dataloader_state=grain.checkpoint.CheckpointSave(\n grain_iterator\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +60,485581,"TERMINAL",0,0,"bash",,terminal_focus +61,490457,"train_lam.py",0,0,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom models.lam import LatentActionModel\nfrom utils.dataloader import create_dataloader_iterator\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n vq_beta: float = 0.25\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n vq_reset_thresh: int = 50\n # LAM\n model_dim: int = 512\n ffn_dim: int = 2048\n latent_dim: int = 32\n num_latents: int = 6\n patch_size: int = 16\n num_blocks: int = 4\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.0\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_lam""\n tags: list[str] = field(default_factory=lambda: [""lam""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\n\ndef lam_loss_fn(\n model: LatentActionModel, inputs: dict, training: bool = True\n) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n # --- Compute loss ---\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=(index_counts != 0).mean(),\n )\n return loss, (outputs[""recon""], index_counts, metrics)\n\n\n@nnx.jit\ndef train_step(\n lam: LatentActionModel,\n optimizer: nnx.Optimizer,\n inputs: dict,\n action_last_active: jax.Array,\n rng: jax.Array,\n) -> tuple[jax.Array, jax.Array, jax.Array, dict]:\n def loss_fn(\n model: LatentActionModel,\n ) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n model.train()\n return lam_loss_fn(model, inputs, training=True)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)\n\n # --- Reset inactive latent actions ---\n codebook = lam.vq.codebook\n num_codes = len(codebook)\n active_codes = idx_counts != 0.0\n action_last_active = jnp.where(active_codes, 0, action_last_active + 1)\n p_code = active_codes / active_codes.sum()\n reset_idxs = jax.random.choice(rng, num_codes, shape=(num_codes,), p=p_code)\n do_reset = action_last_active >= args.vq_reset_thresh\n new_codebook = jnp.where(\n jnp.expand_dims(do_reset, -1), codebook[reset_idxs], codebook.value\n )\n lam.vq.codebook.value = new_codebook\n action_last_active = jnp.where(do_reset, 0, action_last_active)\n return loss, recon, action_last_active, metrics\n\n@nnx.jit\ndef val_step(lam: LatentActionModel, inputs: dict) -> tuple[jax.Array, jax.Array, dict]:\n lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n ffn_dim=args.ffn_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n rngs=rngs,\n )\n\n # Count parameters\n _, params, _ = nnx.split(lam, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n optimizer = nnx.Optimizer(lam, tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n train_iterator = create_dataloader_iterator(args.data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n if args.val_data_dir:\n val_iterator = create_dataloader_iterator(args.val_data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n if args.val_data_dir:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n val_dataloader_state=grain.checkpoint.CheckpointRestore(val_iterator), # type: ignore\n )\n else: \n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n )\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=restore_args\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n train_iterator = restored[""train_dataloader_state""]\n if args.val_data_dir:\n val_iterator = restored[""val_dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader_train = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in train_iterator\n )\n\n if args.val_data_dir:\n dataloader_val = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in val_iterator\n )\n print(f""Starting training from step {step}..."")\n action_last_active = jnp.zeros(args.num_latents, dtype=jnp.int32)\n while step < args.num_steps:\n for videos in dataloader_train:\n # --- Train step ---\n rng, _rng = jax.random.split(rng)\n\n inputs = dict(videos=videos, rng=_rng)\n rng, _rng = jax.random.split(rng)\n loss, recon, action_last_active, metrics = train_step(\n lam, optimizer, inputs, action_last_active, _rng\n )\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Validation loss ---\n if args.val_data_dir and step % args.val_interval == 0:\n print(f""Calculating validation metrics..."")\n val_loss, val_metrics, val_gt_batch, val_recon = calculate_validation_metrics(dataloader_val, lam)\n print(f""Step {step}, validation loss: {val_loss}"")\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n log_dict = {\n ""loss"": loss,\n ""step"": step,\n **metrics\n }\n if args.val_data_dir and step % args.val_interval == 0:\n log_dict.update({\n ""val_loss"": val_loss,\n **val_metrics\n })\n wandb.log(log_dict)\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0, 1:].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if args.val_data_dir and step % args.val_interval == 0:\n gt_seq_val = val_gt_batch[""videos""][0, 1:].astype(jnp.float32) / 255.0\n recon_seq_val = val_recon[0].clip(0, 1)\n val_comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n val_comparison_seq = einops.rearrange(\n val_comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n if args.val_data_dir and step % args.val_interval == 0:\n log_images.update(\n dict(\n val_image=wandb.Image(np.asarray(gt_seq_val[0])),\n val_recon=wandb.Image(np.asarray(recon_seq_val[0])),\n val_true_vs_recon=wandb.Image(\n np.asarray(val_comparison_seq.astype(np.uint8))\n )\n )\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n if args.val_data_dir:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n val_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n val_iterator # type: ignore\n )\n )\n else: \n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n )\n )\n checkpoint_manager.save(\n step,\n args=ckpt_manager_args\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +62,493091,"train_lam.py",0,0,"",python,tab +63,494819,"train_lamap.py",0,0,"",python,tab +64,494822,"train_lamap.py",5816,0,"",python,selection_mouse +65,497513,"train_lam.py",0,0,"",python,tab +66,497514,"train_lam.py",1561,0,"",python,selection_mouse +67,497571,"train_lam.py",1560,0,"",python,selection_command +68,503535,"train_lam.py",1953,0,"",python,selection_mouse +69,504298,"train_lam.py",1942,23," val_steps: int = 50",python,selection_command +70,504508,"train_lam.py",1911,54," val_interval: int = 20_000\n val_steps: int = 50",python,selection_command +71,504630,"train_lam.py",1884,81," val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50",python,selection_command +72,504990,"train_lam.py",1884,0,"",python,selection_command +73,506602,"train_lamap.py",0,0,"",python,tab +74,506603,"train_lamap.py",1900,0,"",python,selection_mouse +75,507410,"train_lamap.py",1878,0,"",python,selection_mouse +76,509138,"train_lamap.py",1866,0," val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n",python,content +77,509215,"train_lamap.py",1870,0,"",python,selection_command +78,557003,"train_lamap.py",2049,0,"",python,selection_mouse +79,566664,"train_lamap.py",2617,0,"",python,selection_mouse +80,566682,"train_lamap.py",2616,0,"",python,selection_command +81,622308,"train_lam.py",0,0,"",python,tab +82,622309,"train_lam.py",2051,0,"",python,selection_mouse +83,623899,"train_lam.py",2052,1,"d",python,selection_command +84,624024,"train_lam.py",2809,2,"da",python,selection_command +85,624088,"train_lam.py",2809,3,"dat",python,selection_command +86,624239,"train_lam.py",4993,4,"data",python,selection_command +87,624623,"train_lam.py",9650,5,"data_",python,selection_command +88,624851,"train_lam.py",9650,6,"data_d",python,selection_command +89,624981,"train_lam.py",9650,7,"data_di",python,selection_command +90,625092,"train_lam.py",9650,8,"data_dir",python,selection_command +91,626432,"train_lamap.py",0,0,"",python,tab +92,626433,"train_lamap.py",2676,0,"",python,selection_mouse +93,628274,"train_lamap.py",2688,1,"d",python,selection_command +94,628486,"train_lamap.py",4485,2,"da",python,selection_command +95,628544,"train_lamap.py",4485,3,"dat",python,selection_command +96,628635,"train_lamap.py",5564,4,"data",python,selection_command +97,629055,"train_lamap.py",6959,5,"data_",python,selection_command +98,629239,"train_lamap.py",6959,6,"data_d",python,selection_command +99,629390,"train_lamap.py",6959,7,"data_di",python,selection_command +100,629538,"train_lamap.py",6959,8,"data_dir",python,selection_command +101,632596,"train_lamap.py",7005,8,"data_dir",python,selection_command +102,637148,"train_lam.py",0,0,"",python,tab +103,637149,"train_lam.py",10076,0,"",python,selection_mouse +104,637150,"train_lam.py",9992,84," cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )",python,selection_mouse +105,637150,"train_lam.py",9862,214," )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )",python,selection_mouse +106,637151,"train_lam.py",9729,347," grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )",python,selection_mouse +107,637151,"train_lam.py",9692,384," ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )",python,selection_mouse +108,637151,"train_lam.py",9661,415," handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )",python,selection_mouse +109,637152,"train_lam.py",9635,441," if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )",python,selection_mouse +110,637152,"train_lam.py",9628,448," )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )",python,selection_mouse +111,637241,"train_lam.py",9442,634," handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )",python,selection_mouse +112,637242,"train_lam.py",10075,0,"",python,selection_command +113,637243,"train_lam.py",9992,84," cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )",python,selection_command +114,637287,"train_lam.py",9253,823," handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )",python,selection_mouse +115,642282,"train_lamap.py",0,0,"",python,tab +116,642283,"train_lamap.py",6558,0,"",python,selection_mouse +117,642353,"train_lamap.py",6557,0,"",python,selection_command +118,643460,"train_lamap.py",6422,0,"",python,selection_mouse +119,644138,"train_lamap.py",6313,0,"",python,selection_mouse +120,644254,"train_lamap.py",6312,1,"d",python,selection_mouse +121,644366,"train_lamap.py",6300,13,"_registry.add",python,selection_mouse +122,644367,"train_lamap.py",6166,147,"dler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add",python,selection_mouse +123,644367,"train_lamap.py",6163,150,"handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add",python,selection_mouse +124,644368,"train_lamap.py",6161,152," handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add",python,selection_mouse +125,644441,"train_lamap.py",6154,159," )\n handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add",python,selection_mouse +126,644474,"train_lamap.py",6153,160," )\n handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add",python,selection_mouse +127,644702,"train_lamap.py",6065,248," ""model_state"", ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler\n )\n handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add",python,selection_mouse +128,645099,"train_lamap.py",6153,160," )\n handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add",python,selection_mouse +129,645961,"train_lamap.py",6422,0,"",python,selection_mouse +130,646402,"train_lamap.py",6292,130," handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler) # type: ignore\n",python,selection_mouse +131,646570,"train_lamap.py",6291,131," handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler) # type: ignore\n",python,selection_mouse +132,646571,"train_lamap.py",6290,132," handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler) # type: ignore\n",python,selection_mouse +133,646632,"train_lamap.py",6160,262," handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler) # type: ignore\n",python,selection_mouse +134,649763,"train_lamap.py",6159,263," handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler) # type: ignore\n",python,selection_mouse +135,659330,"train_lamap.py",6159,264,"",python,content +136,660424,"train_lamap.py",6158,0,"\n ",python,content +137,661026,"train_lamap.py",6163,0," handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )",python,content +138,662186,"train_lamap.py",6167,0,"",python,selection_mouse +139,662926,"train_lamap.py",6163,4,"",python,content +140,673087,"train_lam.py",0,0,"",python,tab +141,673088,"train_lam.py",9252,0,"",python,selection_mouse +142,673089,"train_lam.py",9165,87," ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,selection_mouse +143,673089,"train_lam.py",9164,88," ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,selection_mouse +144,673090,"train_lam.py",9163,89," ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,selection_mouse +145,673198,"train_lam.py",9137,115," handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,selection_mouse +146,673199,"train_lam.py",9251,0,"",python,selection_command +147,673199,"train_lam.py",9137,114," handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n ",python,selection_mouse +148,673200,"train_lam.py",9165,87," ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,selection_command +149,673253,"train_lam.py",9137,115," handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,selection_mouse +150,673304,"train_lam.py",9131,121," )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,selection_mouse +151,673364,"train_lam.py",9050,202," ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,selection_mouse +152,673458,"train_lam.py",9024,228," handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,selection_mouse +153,679322,"train_lam.py",9252,0,"",python,selection_mouse +154,679373,"train_lam.py",9251,0,"",python,selection_command +155,679810,"train_lam.py",9252,0,"",python,selection_mouse +156,679827,"train_lam.py",9251,0,"",python,selection_command +157,680106,"train_lam.py",9251,1,")",python,selection_mouse +158,680107,"train_lam.py",9169,82," ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n ",python,selection_mouse +159,680108,"train_lam.py",9138,113," handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n ",python,selection_mouse +160,680108,"train_lam.py",9131,120," )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n ",python,selection_mouse +161,680108,"train_lam.py",9050,201," ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n ",python,selection_mouse +162,680113,"train_lam.py",9252,0,"",python,selection_command +163,680187,"train_lam.py",9024,228," handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,selection_mouse +164,680270,"train_lam.py",8953,299," handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,selection_mouse +165,680314,"train_lam.py",8940,312," step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,selection_mouse +166,680814,"train_lam.py",8940,0,"",python,selection_command +167,682360,"train_lamap.py",0,0,"",python,tab +168,682361,"train_lamap.py",6158,0,"",python,selection_mouse +169,683400,"train_lamap.py",6157,0,"",python,selection_command +170,683668,"train_lamap.py",6158,0," step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )",python,content +171,683716,"train_lamap.py",6158,0,"",python,selection_command +172,684981,"train_lamap.py",6158,0,"\n ",python,content +173,686320,"train_lamap.py",6163,1,"",python,content +174,686479,"train_lamap.py",6163,1,"",python,content +175,686632,"train_lamap.py",6163,1,"",python,content +176,686782,"train_lamap.py",6163,1,"",python,content +177,687083,"train_lamap.py",6162,0,"",python,selection_command +178,687266,"train_lamap.py",6156,0,"",python,selection_command +179,688080,"train_lamap.py",6153,5," )",python,selection_command +180,688294,"train_lamap.py",6065,93," ""model_state"", ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler\n )",python,selection_command +181,688412,"train_lamap.py",6039,119," handler_registry.add(\n ""model_state"", ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler\n )",python,selection_command +182,688691,"train_lamap.py",6033,125," )\n handler_registry.add(\n ""model_state"", ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler\n )",python,selection_command +183,688820,"train_lamap.py",5948,210," ""model_state"", ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler\n )",python,selection_command +184,688863,"train_lamap.py",5922,236," handler_registry.add(\n ""model_state"", ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler\n )",python,selection_command +185,688966,"train_lamap.py",5851,307," handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler\n )",python,selection_command +186,689143,"train_lamap.py",5838,320," step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler\n )",python,selection_command +187,689548,"train_lamap.py",5838,321,"",python,content +188,689594,"train_lamap.py",5842,0,"",python,selection_command +189,692721,"train_lam.py",0,0,"",python,tab +190,692722,"train_lam.py",9866,0,"",python,selection_mouse +191,692795,"train_lam.py",9865,0,"",python,selection_command +192,693430,"train_lam.py",9786,0,"",python,selection_mouse +193,693578,"train_lam.py",9783,4,"cast",python,selection_mouse +194,697986,"train_lam.py",58,0,"",python,selection_mouse +195,700395,"train_lamap.py",0,0,"",python,tab +196,700396,"train_lamap.py",694,0,"",python,selection_mouse +197,701952,"train_lamap.py",50,0,"",python,selection_mouse +198,701959,"train_lamap.py",49,0,"",python,selection_command +199,702792,"train_lamap.py",50,0,"\nfrom typing import cast",python,content +200,702874,"train_lamap.py",51,0,"",python,selection_command +201,706671,"train_lamap.py",70,0,"",python,selection_command +202,707148,"train_lamap.py",6284,0,"",python,selection_command +203,708985,"train_lamap.py",6362,0,"",python,selection_command +204,709180,"train_lamap.py",6372,0,"",python,selection_command +205,709324,"train_lamap.py",6398,0,"",python,selection_command +206,709825,"train_lamap.py",6432,0,"",python,selection_command +207,709881,"train_lamap.py",6476,0,"",python,selection_command +208,709939,"train_lamap.py",6554,0,"",python,selection_command +209,710044,"train_lamap.py",6564,0,"",python,selection_command +210,710045,"train_lamap.py",6590,0,"",python,selection_command +211,710157,"train_lamap.py",6620,0,"",python,selection_command +212,710158,"train_lamap.py",6656,0,"",python,selection_command +213,710158,"train_lamap.py",6701,0,"",python,selection_command +214,710210,"train_lamap.py",6787,0,"",python,selection_command +215,710211,"train_lamap.py",6797,0,"",python,selection_command +216,710212,"train_lamap.py",6827,0,"",python,selection_command +217,710212,"train_lamap.py",6863,0,"",python,selection_command +218,710667,"train_lamap.py",6911,0,"",python,selection_command +219,710995,"train_lamap.py",6915,0,"",python,selection_command +220,713774,"train_lamap.py",7535,8,"data_dir",python,selection_command +221,737932,"train_lam.py",0,0,"",python,tab +222,737933,"train_lam.py",10791,0,"",python,selection_mouse +223,738354,"train_lam.py",10906,0,"",python,selection_mouse +224,740805,"train_lam.py",10906,1,"\n",python,selection_command +225,741375,"train_lam.py",10782,124," val_iterator = create_dataloader_iterator(args.val_data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n",python,selection_command +226,741525,"train_lam.py",10756,150," if args.val_data_dir:\n val_iterator = create_dataloader_iterator(args.val_data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n",python,selection_command +227,741637,"train_lam.py",10638,268," train_iterator = create_dataloader_iterator(args.data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n if args.val_data_dir:\n val_iterator = create_dataloader_iterator(args.val_data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n",python,selection_command +228,741945,"train_lam.py",10561,345," image_shape = (args.image_height, args.image_width, args.image_channels)\n train_iterator = create_dataloader_iterator(args.data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n if args.val_data_dir:\n val_iterator = create_dataloader_iterator(args.val_data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n",python,selection_command +229,742167,"train_lam.py",10561,0,"",python,selection_command +230,744002,"train_lamap.py",0,0,"",python,tab +231,744003,"train_lamap.py",7439,0,"",python,selection_mouse +232,745744,"train_lamap.py",7440,0," image_shape = (args.image_height, args.image_width, args.image_channels)\n train_iterator = create_dataloader_iterator(args.data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n if args.val_data_dir:\n val_iterator = create_dataloader_iterator(args.val_data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n\n",python,content +233,745815,"train_lamap.py",7440,0,"",python,selection_command +234,747286,"train_lamap.py",7440,346,"",python,content +235,747349,"train_lamap.py",7439,0,"",python,selection_command +236,747615,"train_lamap.py",7481,0,"\n ",python,content +237,748311,"train_lamap.py",7482,4,"",python,content +238,749429,"train_lamap.py",7482,0," image_shape = (args.image_height, args.image_width, args.image_channels)\n train_iterator = create_dataloader_iterator(args.data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n if args.val_data_dir:\n val_iterator = create_dataloader_iterator(args.val_data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n\n",python,content +239,749446,"train_lamap.py",7482,0,"",python,selection_command +240,751092,"train_lamap.py",7559,0,"",python,selection_command +241,751650,"train_lamap.py",7677,0,"",python,selection_command +242,751947,"train_lamap.py",7703,0,"",python,selection_command +243,752233,"train_lamap.py",7827,0,"",python,selection_command +244,752543,"train_lamap.py",7828,0,"",python,selection_command +245,753009,"train_lamap.py",7829,0,"",python,selection_command +246,753969,"train_lamap.py",7829,26," array_record_files = [",python,selection_command +247,754293,"train_lamap.py",7829,65," array_record_files = [\n os.path.join(args.data_dir, x)",python,selection_command +248,754750,"train_lamap.py",7829,108," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)",python,selection_command +249,754833,"train_lamap.py",7829,147," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")",python,selection_command +250,755036,"train_lamap.py",7829,153," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]",python,selection_command +251,755037,"train_lamap.py",7829,192," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(",python,selection_command +252,755037,"train_lamap.py",7829,220," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,",python,selection_command +253,755149,"train_lamap.py",7829,242," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,",python,selection_command +254,755149,"train_lamap.py",7829,301," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size",python,selection_command +255,755150,"train_lamap.py",7829,366," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes",python,selection_command +256,755150,"train_lamap.py",7829,391," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,",python,selection_command +257,755150,"train_lamap.py",7829,413," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,",python,selection_command +258,755151,"train_lamap.py",7829,436," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,",python,selection_command +259,755196,"train_lamap.py",7829,468," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,",python,selection_command +260,755257,"train_lamap.py",7829,492," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,",python,selection_command +261,755410,"train_lamap.py",7829,498," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )",python,selection_command +262,755556,"train_lamap.py",7829,559," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()",python,selection_command +263,755743,"train_lamap.py",7829,638," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)",python,selection_command +264,756073,"train_lamap.py",7829,639,"",python,content +265,756482,"train_lamap.py",7828,0,"",python,selection_command +266,756933,"train_lamap.py",7828,1,"",python,content +267,757425,"train_lamap.py",7828,1,"",python,content +268,757450,"train_lamap.py",7832,0,"",python,selection_command +269,787826,"train_lamap.py",8750,0,"",python,selection_command +270,790489,"train_lamap.py",8138,0,"",python,selection_command +271,795489,"train_lamap.py",7827,0,"",python,selection_command +272,798749,"train_lamap.py",7736,0,"",python,selection_mouse +273,804016,"slurm/jobs/mihir/horeka/coinrun/train_tokenizer_single_gpu.sh",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=00:20:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/coinrun/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_1e-4\n#SBATCH --requeue\n#SBATCH --signal=b:usr1@300 # 5 min before timeout\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir_train=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_gt_actions_split/train\narray_records_dir_val=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m_gt_actions_split/val\n\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n --image_height=64 \\n --image_width=64 \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=125 \\n --init_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=250 \\n --log_checkpoint_interval=250 \\n --log \\n --name=coinrun-tokenizer-dev-$slurm_job_id \\n --tags tokenizer coinrun dev \\n --entity instant-uv \\n --project jafar \\n --warmup_steps 0 \\n --wsd_decay_steps 0 \\n --num_steps 250 \\n --data_dir $array_records_dir_train\n",shellscript,tab +274,806158,"train_lamap.py",0,0,"",python,tab +275,810828,"train_lamap.py",7772,0,"",python,selection_command +276,812475,"train_lamap.py",7770,0,"",python,selection_command +277,812953,"train_lamap.py",7758,0,"",python,selection_command +278,813405,"train_lamap.py",7757,0,"",python,selection_command +279,813768,"train_lamap.py",7753,0,"",python,selection_command +280,814056,"train_lamap.py",7752,0,"",python,selection_command +281,820059,"train_lam.py",0,0,"",python,tab +282,820060,"train_lam.py",550,0,"",python,selection_command +283,821628,"train_lam.py",543,0,"",python,selection_mouse +284,821735,"train_lam.py",543,0," array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)",python,content +285,825090,"train_lam.py",543,638,"",python,content +286,825937,"train_lam.py",542,0,"",python,selection_command +287,830514,"train_lamap.py",0,0,"",python,tab +288,830515,"train_lamap.py",529,0,"",python,selection_command +289,834880,"train_lamap.py",569,0,"",python,selection_mouse +290,834886,"train_lamap.py",568,0,"",python,selection_command +291,835548,"train_lamap.py",558,0,"",python,selection_command +292,835711,"train_lamap.py",557,0,"",python,selection_command +293,835862,"train_lamap.py",502,0,"",python,selection_command +294,835986,"train_lamap.py",459,0,"",python,selection_command +295,836202,"train_lamap.py",415,0,"",python,selection_command +296,836476,"train_lamap.py",449,0,"",python,selection_command +297,836826,"train_lamap.py",435,14,"",python,content +298,837474,"train_lamap.py",435,0,"c",python,content +299,837475,"train_lamap.py",436,0,"",python,selection_keyboard +300,837666,"train_lamap.py",436,0,"r",python,content +301,837668,"train_lamap.py",437,0,"",python,selection_keyboard +302,837856,"train_lamap.py",437,0,"e",python,content +303,837858,"train_lamap.py",438,0,"",python,selection_keyboard +304,838067,"train_lamap.py",438,0,"a",python,content +305,838068,"train_lamap.py",439,0,"",python,selection_keyboard +306,839744,"train_lamap.py",439,0,"t",python,content +307,839746,"train_lamap.py",440,0,"",python,selection_keyboard +308,839974,"train_lamap.py",440,0,"e",python,content +309,839976,"train_lamap.py",441,0,"",python,selection_keyboard +310,841090,"train_lamap.py",435,6,"create_dataloader_iterator",python,content +311,841723,"train_lamap.py",460,0,"",python,selection_command +312,843412,"train_lamap.py",7592,0,"",python,selection_command +313,845762,"train_lamap.py",7710,0,"",python,selection_command +314,845952,"train_lamap.py",7736,0,"",python,selection_command +315,846109,"train_lamap.py",7839,0,"",python,selection_command +316,846262,"train_lamap.py",7861,0,"",python,selection_command +317,846399,"train_lamap.py",7894,0,"",python,selection_command +318,846542,"train_lamap.py",7920,0,"",python,selection_command +319,846682,"train_lamap.py",7975,0,"",python,selection_command +320,846836,"train_lamap.py",8019,0,"",python,selection_command +321,846996,"train_lamap.py",8042,0,"",python,selection_command +322,847150,"train_lamap.py",8089,0,"",python,selection_command +323,847272,"train_lamap.py",8135,0,"",python,selection_command +324,847420,"train_lamap.py",8172,0,"",python,selection_command +325,848054,"train_lamap.py",8248,0,"",python,selection_command +326,857115,"train_lamap.py",8260,1,"g",python,selection_command +327,857217,"train_lamap.py",8260,2,"gr",python,selection_command +328,857404,"train_lamap.py",8260,3,"gra",python,selection_command +329,857464,"train_lamap.py",8260,4,"grai",python,selection_command +330,857529,"train_lamap.py",8260,5,"grain",python,selection_command +331,857918,"train_lamap.py",8295,6,"grain_",python,selection_command +332,858329,"train_lamap.py",8295,7,"grain_i",python,selection_command +333,858489,"train_lamap.py",8295,8,"grain_it",python,selection_command +334,858678,"train_lamap.py",8295,9,"grain_ite",python,selection_command +335,858824,"train_lamap.py",8295,10,"grain_iter",python,selection_command +336,859139,"train_lamap.py",8295,11,"grain_itera",python,selection_command +337,859321,"train_lamap.py",8295,12,"grain_iterat",python,selection_command +338,859439,"train_lamap.py",8295,13,"grain_iterato",python,selection_command +339,859565,"train_lamap.py",8295,14,"grain_iterator",python,selection_command +340,873837,"train_lamap.py",8295,14,"train_iterator",python,content +341,873843,"train_lamap.py",8391,14,"grain_iterator",python,selection_command +342,875242,"train_lamap.py",8391,14,"train_iterator",python,content +343,875247,"train_lamap.py",8679,14,"grain_iterator",python,selection_command +344,877233,"train_lamap.py",8679,14,"train_iterator",python,content +345,877237,"train_lamap.py",10305,14,"grain_iterator",python,selection_command +346,878275,"train_lamap.py",10305,14,"train_iterator",python,content +347,934854,"train_lamap.py",2423,0,"",python,selection_mouse +348,934879,"train_lamap.py",2422,0,"",python,selection_command +349,935850,"train_lamap.py",2424,0,"",python,selection_command +350,936054,"train_lamap.py",2432,0,"",python,selection_command +351,936683,"train_lamap.py",2448,0,"",python,selection_command +352,936684,"train_lamap.py",2479,0,"",python,selection_command +353,936744,"train_lamap.py",2543,0,"",python,selection_command +354,936819,"train_lamap.py",2598,0,"",python,selection_command +355,936982,"train_lamap.py",2645,0,"",python,selection_command +356,937223,"train_lamap.py",2598,0,"",python,selection_command +357,937382,"train_lamap.py",2543,0,"",python,selection_command +358,937543,"train_lamap.py",2598,0,"",python,selection_command +359,937716,"train_lamap.py",2645,0,"",python,selection_command +360,937871,"train_lamap.py",2654,0,"",python,selection_command +361,938001,"train_lamap.py",2655,0,"",python,selection_command +362,938365,"train_lamap.py",2654,0,"",python,selection_command +363,938639,"train_lamap.py",2654,0,"\nfrom utils.parameter_utils import count_parameters_by_component",python,content +364,938665,"train_lamap.py",2655,0,"",python,selection_command +365,940470,"train_lamap.py",2655,64,"",python,content +366,940519,"train_lamap.py",10305,1,"g",python,content +367,940534,"train_lamap.py",8679,1,"g",python,content +368,940547,"train_lamap.py",8391,1,"g",python,content +369,940561,"train_lamap.py",8295,1,"g",python,content +370,940574,"train_lamap.py",2422,0,"",python,selection_command +371,940721,"train_lamap.py",2423,0,"\n ",python,content +372,941733,"train_lamap.py",2424,4,"",python,content +373,941931,"train_lamap.py",2425,1,"",python,content +374,941984,"train_lamap.py",2422,0,"",python,selection_command +375,942342,"train_lamap.py",2424,0,"",python,selection_command +376,942484,"train_lamap.py",2432,0,"",python,selection_command +377,942645,"train_lamap.py",2448,0,"",python,selection_command +378,942822,"train_lamap.py",2479,0,"",python,selection_command +379,942924,"train_lamap.py",2543,0,"",python,selection_command +380,943091,"train_lamap.py",2598,0,"",python,selection_command +381,943228,"train_lamap.py",2645,0,"",python,selection_command +382,943443,"train_lamap.py",2654,0,"",python,selection_command +383,943659,"train_lamap.py",2654,0,"\n",python,content +384,951546,"train_lam.py",0,0,"",python,tab +385,951547,"train_lam.py",5738,0,"",python,selection_mouse +386,953014,"train_lam.py",5690,48," return val_loss, val_metrics, inputs, recon\n",python,selection_command +387,953517,"train_lam.py",5684,54," }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +388,953576,"train_lam.py",5638,100," for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +389,953726,"train_lam.py",5572,166," f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +390,953727,"train_lam.py",5552,186," val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +391,953881,"train_lam.py",5514,224," val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +392,953882,"train_lam.py",5513,225,"\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +393,953882,"train_lam.py",5374,364," print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +394,953882,"train_lam.py",5344,394," if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +395,954115,"train_lam.py",5343,395,"\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +396,954115,"train_lam.py",5325,413," break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +397,954116,"train_lam.py",5291,447," if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +398,954116,"train_lam.py",5273,465," step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +399,954116,"train_lam.py",5232,506," metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +400,954319,"train_lam.py",5197,541," loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +401,954320,"train_lam.py",5144,594," loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +402,954321,"train_lam.py",5107,631," inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +403,954321,"train_lam.py",5073,665," for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +404,954321,"train_lam.py",5047,691," metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +405,954322,"train_lam.py",5024,714," loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +406,954322,"train_lam.py",5011,727," step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +407,954322,"train_lam.py",4956,782,"def calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +408,954432,"train_lam.py",4955,783,"\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +409,954433,"train_lam.py",4923,815," return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +410,954433,"train_lam.py",4848,890," (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +411,954434,"train_lam.py",4833,905," lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +412,954434,"train_lam.py",4744,994,"def val_step(lam: LatentActionModel, inputs: dict) -> tuple[jax.Array, jax.Array, dict]:\n lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +413,954453,"train_lam.py",4735,1003,"@nnx.jit\ndef val_step(lam: LatentActionModel, inputs: dict) -> tuple[jax.Array, jax.Array, dict]:\n lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +414,954542,"train_lam.py",4734,1004,"\n@nnx.jit\ndef val_step(lam: LatentActionModel, inputs: dict) -> tuple[jax.Array, jax.Array, dict]:\n lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +415,954695,"train_lam.py",4682,1056," return loss, recon, action_last_active, metrics\n\n@nnx.jit\ndef val_step(lam: LatentActionModel, inputs: dict) -> tuple[jax.Array, jax.Array, dict]:\n lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +416,954938,"train_lam.py",4734,1004,"\n@nnx.jit\ndef val_step(lam: LatentActionModel, inputs: dict) -> tuple[jax.Array, jax.Array, dict]:\n lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +417,955090,"train_lam.py",4735,1003,"@nnx.jit\ndef val_step(lam: LatentActionModel, inputs: dict) -> tuple[jax.Array, jax.Array, dict]:\n lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +418,955317,"train_lam.py",4744,994,"def val_step(lam: LatentActionModel, inputs: dict) -> tuple[jax.Array, jax.Array, dict]:\n lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +419,955613,"train_lam.py",4735,1003,"@nnx.jit\ndef val_step(lam: LatentActionModel, inputs: dict) -> tuple[jax.Array, jax.Array, dict]:\n lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,selection_command +420,955850,"train_lam.py",4735,0,"",python,selection_command +421,958395,"train_lamap.py",0,0,"",python,tab +422,960422,"train_lamap.py",2655,0,"\n@nnx.jit\ndef val_step(lam: LatentActionModel, inputs: dict) -> tuple[jax.Array, jax.Array, dict]:\n lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n",python,content +423,960533,"train_lamap.py",2656,0,"",python,selection_command +424,961101,"train_lamap.py",2655,0,"",python,selection_command +425,961540,"train_lamap.py",2655,1,"",python,content +426,961894,"train_lamap.py",2664,0,"",python,selection_command +427,962124,"train_lamap.py",2753,0,"",python,selection_command +428,963116,"train_lamap.py",2768,0,"",python,selection_command +429,963309,"train_lamap.py",2753,0,"",python,selection_command +430,963466,"train_lamap.py",2664,0,"",python,selection_command +431,963629,"train_lamap.py",2655,0,"",python,selection_command +432,963736,"train_lamap.py",2654,0,"",python,selection_command +433,963884,"train_lamap.py",2631,0,"",python,selection_command +434,964053,"train_lamap.py",2584,0,"",python,selection_command +435,964188,"train_lamap.py",2529,0,"",python,selection_command +436,964328,"train_lamap.py",2465,0,"",python,selection_command +437,964464,"train_lamap.py",2434,0,"",python,selection_command +438,964622,"train_lamap.py",2425,0,"",python,selection_command +439,964891,"train_lamap.py",2426,0,"",python,selection_command +440,965135,"train_lamap.py",2426,3,"",python,content +441,965457,"train_lamap.py",2426,0,"n",python,content +442,965458,"train_lamap.py",2427,0,"",python,selection_keyboard +443,965632,"train_lamap.py",2427,0,"n",python,content +444,965634,"train_lamap.py",2428,0,"",python,selection_keyboard +445,965906,"train_lamap.py",2428,0,"x",python,content +446,965908,"train_lamap.py",2429,0,"",python,selection_keyboard +447,966849,"train_lamap.py",2428,0,"",python,selection_command +448,983739,"train_lam.py",0,0,"",python,tab +449,983740,"train_lam.py",415,0,"",python,selection_mouse +450,987672,"train_lam.py",403,0,"",python,selection_mouse +451,990626,"train_lamap.py",0,0,"",python,tab +452,990626,"train_lamap.py",2682,0,"",python,selection_mouse +453,991206,"train_lamap.py",2682,17,"",python,content +454,992177,"train_lamap.py",2681,0,"",python,selection_command +455,992362,"train_lamap.py",2680,0,"",python,selection_command +456,992564,"train_lamap.py",2680,0,"a",python,content +457,992565,"train_lamap.py",2681,0,"",python,selection_keyboard +458,992873,"train_lamap.py",2681,0,"p",python,content +459,992874,"train_lamap.py",2682,0,"",python,selection_keyboard +460,993224,"train_lamap.py",2683,0,"",python,selection_command +461,993461,"train_lamap.py",2684,0,"",python,selection_command +462,995010,"train_lamap.py",2684,0,"L",python,content +463,995012,"train_lamap.py",2685,0,"",python,selection_keyboard +464,995199,"train_lamap.py",2685,0,"a",python,content +465,995201,"train_lamap.py",2686,0,"",python,selection_keyboard +466,995318,"train_lamap.py",2686,0,"n",python,content +467,995322,"train_lamap.py",2687,0,"",python,selection_keyboard +468,995665,"train_lamap.py",2686,1,"",python,content +469,995865,"train_lamap.py",2686,0,"t",python,content +470,995867,"train_lamap.py",2687,0,"",python,selection_keyboard +471,996060,"train_lamap.py",2687,0,"e",python,content +472,996062,"train_lamap.py",2688,0,"",python,selection_keyboard +473,996142,"train_lamap.py",2688,0,"n",python,content +474,996144,"train_lamap.py",2689,0,"",python,selection_keyboard +475,996313,"train_lamap.py",2689,0,"t",python,content +476,996315,"train_lamap.py",2690,0,"",python,selection_keyboard +477,1001016,"train_lamap.py",2684,6,"LatentActionMapper",python,content +478,1002693,"train_lamap.py",2701,0,"",python,selection_command +479,1003032,"train_lamap.py",2769,0,"",python,selection_command +480,1003425,"train_lamap.py",2808,0,"",python,selection_command +481,1004424,"train_lamap.py",2809,0,"",python,selection_command +482,1004662,"train_lamap.py",2810,0,"",python,selection_command +483,1005007,"train_lamap.py",2809,0,"",python,selection_command +484,1005162,"train_lamap.py",2808,0,"",python,selection_command +485,1005755,"train_lamap.py",2808,0,"a",python,content +486,1005756,"train_lamap.py",2809,0,"",python,selection_keyboard +487,1006079,"train_lamap.py",2809,0,"p",python,content +488,1006080,"train_lamap.py",2810,0,"",python,selection_keyboard +489,1006437,"train_lamap.py",2809,0,"",python,selection_command +490,1006686,"train_lamap.py",2769,0,"",python,selection_command +491,1007124,"train_lamap.py",2768,0,"",python,selection_command +492,1007236,"train_lamap.py",2767,0,"",python,selection_command +493,1007409,"train_lamap.py",2766,0,"",python,selection_command +494,1007533,"train_lamap.py",2765,0,"",python,selection_command +495,1007726,"train_lamap.py",2764,0,"",python,selection_command +496,1007819,"train_lamap.py",2763,0,"",python,selection_command +497,1008290,"train_lamap.py",2763,0,"a",python,content +498,1008291,"train_lamap.py",2764,0,"",python,selection_keyboard +499,1008620,"train_lamap.py",2764,0,"p",python,content +500,1008622,"train_lamap.py",2765,0,"",python,selection_keyboard +501,1008885,"train_lamap.py",2764,0,"",python,selection_command +502,1009146,"train_lamap.py",2672,0,"",python,selection_command +503,1009350,"train_lamap.py",2662,0,"",python,selection_command +504,1009468,"train_lamap.py",2654,0,"",python,selection_command +505,1009663,"train_lamap.py",2639,0,"",python,selection_command +506,1009794,"train_lamap.py",2592,0,"",python,selection_command +507,1009954,"train_lamap.py",2537,0,"",python,selection_command +508,1010067,"train_lamap.py",2473,0,"",python,selection_command +509,1010284,"train_lamap.py",2442,0,"",python,selection_command +510,1013149,"train_lamap.py",2464,0,"\n ",python,content +511,1013632,"train_lamap.py",2469,0,"l",python,content +512,1013633,"train_lamap.py",2470,0,"",python,selection_keyboard +513,1013777,"train_lamap.py",2470,0,"a",python,content +514,1013779,"train_lamap.py",2471,0,"",python,selection_keyboard +515,1013952,"train_lamap.py",2471,0,"m",python,content +516,1013954,"train_lamap.py",2472,0,"",python,selection_keyboard +517,1014145,"train_lamap.py",2472,0,"a",python,content +518,1014147,"train_lamap.py",2473,0,"",python,selection_keyboard +519,1014248,"train_lamap.py",2473,0,"p",python,content +520,1014250,"train_lamap.py",2474,0,"",python,selection_keyboard +521,1014581,"train_lamap.py",2474,0,".",python,content +522,1014582,"train_lamap.py",2475,0,"",python,selection_keyboard +523,1015021,"train_lamap.py",2475,0,"t",python,content +524,1015023,"train_lamap.py",2476,0,"",python,selection_keyboard +525,1015237,"train_lamap.py",2476,0,"r",python,content +526,1015238,"train_lamap.py",2477,0,"",python,selection_keyboard +527,1015402,"train_lamap.py",2477,0,"a",python,content +528,1015404,"train_lamap.py",2478,0,"",python,selection_keyboard +529,1015437,"train_lamap.py",2478,0,"i",python,content +530,1015438,"train_lamap.py",2479,0,"",python,selection_keyboard +531,1015521,"train_lamap.py",2479,0,"n",python,content +532,1015522,"train_lamap.py",2480,0,"",python,selection_keyboard +533,1016267,"train_lamap.py",2480,0,"()",python,content +534,1016268,"train_lamap.py",2481,0,"",python,selection_keyboard +535,1016370,"train_lamap.py",2481,1,")",python,content +536,1016371,"train_lamap.py",2482,0,"",python,selection_keyboard +537,1016423,"train_lamap.py",2481,0,"",python,selection_command +538,1017727,"train_lamap.py",2499,0,"",python,selection_command +539,1017936,"train_lamap.py",2563,0,"",python,selection_command +540,1018416,"train_lamap.py",2618,0,"",python,selection_command +541,1018467,"train_lamap.py",2665,0,"",python,selection_command +542,1018520,"train_lamap.py",2672,0,"",python,selection_command +543,1018526,"train_lamap.py",2680,0,"",python,selection_command +544,1018607,"train_lamap.py",2698,0,"",python,selection_command +545,1018608,"train_lamap.py",2789,0,"",python,selection_command +546,1019013,"train_lamap.py",2807,0,"",python,selection_command +547,1019213,"train_lamap.py",2808,0,"",python,selection_command +548,1019783,"train_lamap.py",2809,0,"",python,selection_command +549,1019783,"train_lamap.py",2810,0,"",python,selection_command +550,1019827,"train_lamap.py",2811,0,"",python,selection_command +551,1019879,"train_lamap.py",2812,0,"",python,selection_command +552,1019936,"train_lamap.py",2813,0,"",python,selection_command +553,1019937,"train_lamap.py",2814,0,"",python,selection_command +554,1020020,"train_lamap.py",2815,0,"",python,selection_command +555,1020021,"train_lamap.py",2816,0,"",python,selection_command +556,1020091,"train_lamap.py",2817,0,"",python,selection_command +557,1020092,"train_lamap.py",2818,0,"",python,selection_command +558,1020093,"train_lamap.py",2819,0,"",python,selection_command +559,1020142,"train_lamap.py",2820,0,"",python,selection_command +560,1020143,"train_lamap.py",2821,0,"",python,selection_command +561,1020199,"train_lamap.py",2822,0,"",python,selection_command +562,1020200,"train_lamap.py",2823,0,"",python,selection_command +563,1020216,"train_lamap.py",2824,0,"",python,selection_command +564,1020217,"train_lamap.py",2825,0,"",python,selection_command +565,1020370,"train_lamap.py",2826,0,"",python,selection_command +566,1020371,"train_lamap.py",2827,0,"",python,selection_command +567,1020419,"train_lamap.py",2828,0,"",python,selection_command +568,1020420,"train_lamap.py",2829,0,"",python,selection_command +569,1020420,"train_lamap.py",2830,0,"",python,selection_command +570,1020489,"train_lamap.py",2831,0,"",python,selection_command +571,1020490,"train_lamap.py",2832,0,"",python,selection_command +572,1020490,"train_lamap.py",2833,0,"",python,selection_command +573,1020536,"train_lamap.py",2834,0,"",python,selection_command +574,1020551,"train_lamap.py",2835,0,"",python,selection_command +575,1020616,"train_lamap.py",2836,0,"",python,selection_command +576,1020803,"train_lamap.py",2837,0,"",python,selection_command +577,1020920,"train_lamap.py",2838,0,"",python,selection_command +578,1021071,"train_lamap.py",2839,0,"",python,selection_command +579,1021222,"train_lamap.py",2840,0,"",python,selection_command +580,1021362,"train_lamap.py",2841,0,"",python,selection_command +581,1021540,"train_lamap.py",2842,0,"",python,selection_command +582,1021864,"train_lamap.py",2842,0,"a",python,content +583,1021865,"train_lamap.py",2843,0,"",python,selection_keyboard +584,1021998,"train_lamap.py",2843,0,"p",python,content +585,1021999,"train_lamap.py",2844,0,"",python,selection_keyboard +586,1022343,"train_lamap.py",2843,0,"",python,selection_command +587,1022581,"train_lamap.py",2844,0,"",python,selection_command +588,1022677,"train_lamap.py",2845,0,"",python,selection_command +589,1023117,"train_lamap.py",2846,0,"",python,selection_command +590,1023159,"train_lamap.py",2847,0,"",python,selection_command +591,1023329,"train_lamap.py",2848,0,"",python,selection_command +592,1023351,"train_lamap.py",2849,0,"",python,selection_command +593,1023352,"train_lamap.py",2850,0,"",python,selection_command +594,1023353,"train_lamap.py",2851,0,"",python,selection_command +595,1023457,"train_lamap.py",2852,0,"",python,selection_command +596,1023609,"train_lamap.py",2853,0,"",python,selection_command +597,1023792,"train_lamap.py",2854,0,"",python,selection_command +598,1024109,"train_lamap.py",2854,8,"",python,content +599,1024347,"train_lamap.py",2854,1,"",python,content +600,1024842,"train_lamap.py",2854,5,"",python,content +601,1025307,"train_lamap.py",2854,0,"a",python,content +602,1025309,"train_lamap.py",2855,0,"",python,selection_keyboard +603,1025708,"train_lamap.py",2855,0,"l",python,content +604,1025710,"train_lamap.py",2856,0,"",python,selection_keyboard +605,1025833,"train_lamap.py",2856,0,"l",python,content +606,1025835,"train_lamap.py",2857,0,"",python,selection_keyboard +607,1025975,"train_lamap.py",2857,0,"o",python,content +608,1025977,"train_lamap.py",2858,0,"",python,selection_keyboard +609,1026103,"train_lamap.py",2858,0,"w",python,content +610,1026104,"train_lamap.py",2859,0,"",python,selection_keyboard +611,1026303,"train_lamap.py",2859,0,"_",python,content +612,1026304,"train_lamap.py",2860,0,"",python,selection_keyboard +613,1026640,"train_lamap.py",2860,0,"i",python,content +614,1026642,"train_lamap.py",2861,0,"",python,selection_keyboard +615,1026732,"train_lamap.py",2861,0,"n",python,content +616,1026733,"train_lamap.py",2862,0,"",python,selection_keyboard +617,1026856,"train_lamap.py",2862,0,"t",python,content +618,1026857,"train_lamap.py",2863,0,"",python,selection_keyboard +619,1027565,"train_lamap.py",2863,0,"=",python,content +620,1027567,"train_lamap.py",2864,0,"",python,selection_keyboard +621,1027883,"train_lamap.py",2864,0,"F",python,content +622,1027884,"train_lamap.py",2865,0,"",python,selection_keyboard +623,1028072,"train_lamap.py",2865,0,"a",python,content +624,1028073,"train_lamap.py",2866,0,"",python,selection_keyboard +625,1028344,"train_lamap.py",2866,0,"l",python,content +626,1028345,"train_lamap.py",2867,0,"",python,selection_keyboard +627,1028617,"train_lamap.py",2867,0,"s",python,content +628,1028618,"train_lamap.py",2868,0,"",python,selection_keyboard +629,1028808,"train_lamap.py",2868,0,"e",python,content +630,1028810,"train_lamap.py",2869,0,"",python,selection_keyboard +631,1029053,"train_lamap.py",2868,0,"",python,selection_command +632,1029265,"train_lamap.py",2789,0,"",python,selection_command +633,1029544,"train_lamap.py",2788,0,"",python,selection_command +634,1030218,"train_lamap.py",2805,0,"",python,selection_command +635,1030701,"train_lamap.py",2885,0,"",python,selection_command +636,1031003,"train_lamap.py",2805,0,"",python,selection_command +637,1032349,"train_lamap.py",2852,0,"",python,selection_mouse +638,1032506,"train_lamap.py",2846,6,"inputs",python,selection_mouse +639,1033582,"train_lamap.py",2902,0,"",python,selection_mouse +640,1033633,"train_lamap.py",2901,0,"",python,selection_command +641,1034262,"train_lamap.py",2845,0,"",python,selection_mouse +642,1035221,"train_lamap.py",2842,0,"",python,selection_mouse +643,1036133,"train_lamap.py",2840,0,"",python,selection_mouse +644,1036292,"train_lamap.py",2839,5,"lamap",python,selection_mouse +645,1037094,"train_lamap.py",2830,0,"",python,selection_mouse +646,1039579,"train_lamap.py",2790,0,"",python,selection_mouse +647,1039580,"train_lamap.py",2789,0,"",python,selection_command +648,1040166,"train_lamap.py",2902,0,"",python,selection_mouse +649,1040171,"train_lamap.py",2901,0,"",python,selection_command +650,1040778,"train_lamap.py",2840,0,"",python,selection_mouse +651,1040915,"train_lamap.py",2839,5,"lamap",python,selection_mouse +652,1041712,"train_lamap.py",2672,0,"",python,selection_mouse +653,1042586,"train_lamap.py",2573,0,"",python,selection_mouse +654,1056364,"train_lam.py",0,0,"",python,tab +655,1056365,"train_lam.py",387,0,"",python,selection_mouse +656,1064519,"train_lam.py",49,0,"",python,selection_mouse +657,1066093,"train_lamap.py",0,0,"",python,tab +658,1066094,"train_lamap.py",360,0,"",python,selection_mouse +659,1066142,"train_lamap.py",359,0,"",python,selection_command +660,1066939,"train_lamap.py",360,0,"\n",python,content +661,1068509,"train_lamap.py",361,0,"i",python,content +662,1068509,"train_lamap.py",362,0,"",python,selection_keyboard +663,1068618,"train_lamap.py",362,0,"m",python,content +664,1068619,"train_lamap.py",363,0,"",python,selection_keyboard +665,1068795,"train_lamap.py",363,0,"p",python,content +666,1068796,"train_lamap.py",364,0,"",python,selection_keyboard +667,1069000,"train_lamap.py",364,0,"o",python,content +668,1069001,"train_lamap.py",365,0,"",python,selection_keyboard +669,1069213,"train_lamap.py",365,0,"r",python,content +670,1069214,"train_lamap.py",366,0,"",python,selection_keyboard +671,1069299,"train_lamap.py",366,0,"t",python,content +672,1069300,"train_lamap.py",367,0,"",python,selection_keyboard +673,1069419,"train_lamap.py",367,0," ",python,content +674,1069420,"train_lamap.py",368,0,"",python,selection_keyboard +675,1069950,"train_lamap.py",368,0,"f",python,content +676,1069951,"train_lamap.py",369,0,"",python,selection_keyboard +677,1070033,"train_lamap.py",369,0,"l",python,content +678,1070034,"train_lamap.py",370,0,"",python,selection_keyboard +679,1070143,"train_lamap.py",370,0,"a",python,content +680,1070144,"train_lamap.py",371,0,"",python,selection_keyboard +681,1070339,"train_lamap.py",371,0,"x",python,content +682,1070341,"train_lamap.py",372,0,"",python,selection_keyboard +683,1070500,"train_lamap.py",372,0,".",python,content +684,1070501,"train_lamap.py",373,0,"",python,selection_keyboard +685,1070759,"train_lamap.py",373,0,"n",python,content +686,1070761,"train_lamap.py",374,0,"",python,selection_keyboard +687,1070898,"train_lamap.py",374,0,"n",python,content +688,1070899,"train_lamap.py",375,0,"",python,selection_keyboard +689,1071006,"train_lamap.py",375,0,"x",python,content +690,1071007,"train_lamap.py",376,0,"",python,selection_keyboard +691,1071304,"train_lamap.py",376,0," ",python,content +692,1071305,"train_lamap.py",377,0,"",python,selection_keyboard +693,1071449,"train_lamap.py",377,0,"a",python,content +694,1071450,"train_lamap.py",378,0,"",python,selection_keyboard +695,1071602,"train_lamap.py",378,0,"s",python,content +696,1071603,"train_lamap.py",379,0,"",python,selection_keyboard +697,1071730,"train_lamap.py",379,0," ",python,content +698,1071731,"train_lamap.py",380,0,"",python,selection_keyboard +699,1071847,"train_lamap.py",380,0,"n",python,content +700,1071848,"train_lamap.py",381,0,"",python,selection_keyboard +701,1072014,"train_lamap.py",381,0,"n",python,content +702,1072015,"train_lamap.py",382,0,"",python,selection_keyboard +703,1072130,"train_lamap.py",382,0,"x",python,content +704,1072131,"train_lamap.py",383,0,"",python,selection_keyboard +705,1072618,"train_lamap.py",382,0,"",python,selection_command +706,1088497,"train_lamap.py",90,0,"",python,selection_mouse +707,1090660,"train_lamap.py",76,49,"",python,content +708,1132433,"train_lam.py",0,0,"",python,tab +709,1132434,"train_lam.py",2223,0,"",python,selection_mouse +710,1133286,"train_lam.py",2224,0,"",python,selection_mouse +711,1133430,"train_lam.py",2224,2,"gt",python,selection_mouse +712,1168321,"train_lamap.py",0,0,"",python,tab +713,1168322,"train_lamap.py",2398,0,"",python,selection_mouse +714,1170181,"train_lamap.py",2314,0,"",python,selection_mouse +715,1221342,"train_lamap.py",2368,0,"",python,selection_mouse +716,1221492,"train_lamap.py",2364,7,"actions",python,selection_mouse +717,1226374,"train_lamap.py",2407,0,"",python,selection_mouse +718,1226422,"train_lamap.py",2406,0,"",python,selection_command +719,1228676,"train_lamap.py",2407,0,"",python,selection_mouse +720,1228692,"train_lamap.py",2406,0,"",python,selection_command +721,1229333,"train_lamap.py",2478,0,"",python,selection_mouse +722,1232460,"train_lamap.py",3451,0,"",python,selection_mouse +723,1236916,"train_lam.py",0,0,"",python,tab +724,1236917,"train_lam.py",271,0,"",python,selection_command +725,1239562,"train_lam.py",270,0,"",python,selection_command +726,1244433,"train_lamap.py",0,0,"",python,tab +727,1244434,"train_lamap.py",273,0,"",python,selection_mouse +728,1244483,"train_lamap.py",272,0,"",python,selection_command +729,1245254,"train_lamap.py",248,0,"",python,selection_command +730,1245421,"train_lamap.py",230,0,"",python,selection_command +731,1245943,"train_lamap.py",238,0,"\nimport numpy as np",python,content +732,1246012,"train_lamap.py",239,0,"",python,selection_command +733,1247161,"train_lamap.py",258,0,"",python,selection_command +734,1247670,"train_lamap.py",269,0,"",python,selection_command +735,1247824,"train_lamap.py",293,0,"",python,selection_command +736,1247825,"train_lamap.py",305,0,"",python,selection_command +737,1247825,"train_lamap.py",318,0,"",python,selection_command +738,1247939,"train_lamap.py",331,0,"",python,selection_command +739,1247939,"train_lamap.py",354,0,"",python,selection_command +740,1247940,"train_lamap.py",355,0,"",python,selection_command +741,1248008,"train_lamap.py",399,0,"",python,selection_command +742,1248009,"train_lamap.py",455,0,"",python,selection_command +743,1248009,"train_lamap.py",498,0,"",python,selection_command +744,1248100,"train_lamap.py",562,0,"",python,selection_command +745,1248101,"train_lamap.py",563,0,"",python,selection_command +746,1248154,"train_lamap.py",564,0,"",python,selection_command +747,1252680,"train_lamap.py",6368,0,"",python,selection_mouse +748,1274139,"train_lam.py",0,0,"",python,tab +749,1274139,"train_lam.py",6233,0,"",python,selection_mouse +750,1276142,"train_lam.py",6230,35," rng = jax.random.key(args.seed)",python,selection_command +751,1276358,"train_lam.py",6230,36," rng = jax.random.key(args.seed)\n",python,selection_command +752,1276535,"train_lam.py",6230,67," rng = jax.random.key(args.seed)\n\n # --- Initialize model ---",python,selection_command +753,1276700,"train_lam.py",6230,105," rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)",python,selection_command +754,1276828,"train_lam.py",6230,131," rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)",python,selection_command +755,1277070,"train_lam.py",6230,0,"",python,selection_command +756,1278443,"train_lamap.py",0,0,"",python,tab +757,1278444,"train_lamap.py",4212,0,"",python,selection_mouse +758,1280181,"train_lamap.py",4212,0,"\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)",python,content +759,1280271,"train_lamap.py",4217,0,"",python,selection_command +760,1280761,"train_lamap.py",4212,0,"",python,selection_command +761,1281260,"train_lamap.py",4176,0,"",python,selection_command +762,1281792,"train_lamap.py",4172,40,"",python,content +763,1282865,"train_lamap.py",4173,0,"",python,selection_command +764,1283163,"train_lamap.py",4209,0,"",python,selection_command +765,1284924,"train_lamap.py",4173,0,"",python,selection_command +766,1285149,"train_lamap.py",4172,0,"",python,selection_command +767,1285655,"train_lamap.py",4172,1,"",python,content +768,1285689,"train_lamap.py",4176,0,"",python,selection_command +769,1285865,"train_lamap.py",4208,0,"",python,selection_command +770,1286022,"train_lamap.py",4213,0,"",python,selection_command +771,1286188,"train_lamap.py",4244,0,"",python,selection_command +772,1286371,"train_lamap.py",4282,0,"",python,selection_command +773,1286563,"train_lamap.py",4308,0,"",python,selection_command +774,1286670,"train_lamap.py",4339,0,"",python,selection_command +775,1287842,"train_lamap.py",4371,0,"",python,selection_command +776,1288146,"train_lamap.py",4339,0,"",python,selection_command +777,1288641,"train_lamap.py",4371,0,"",python,selection_command +778,1290759,"train_lamap.py",4407,0,"",python,selection_command +779,1292907,"train_lamap.py",4441,0,"",python,selection_command +780,1300567,"train_lamap.py",4359,0,"",python,selection_mouse +781,1301038,"models/lamap.py",0,0,"from typing import Dict, Any\n\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\n\nfrom models.lam import LatentActionModel\n\n\nclass LatentActionMapper(nn.Module):\n """"""Latent Action Mapper""""""\n\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n # --- LAM ---\n in_dim: int\n model_dim: int\n latent_dim: int\n num_latents: int\n patch_size: int\n num_blocks: int\n num_heads: int\n dropout: float\n codebook_dropout: float\n use_flash_attention: bool\n\n # --- Mapper ---\n action_dim: int\n\n def setup(self):\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n latent_dim=self.latent_dim,\n num_latents=self.num_latents,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=self.lam_dropout,\n codebook_dropout=self.lam_codebook_dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n )\n self.action_map = nn.Dense(\n self.action_dim,\n use_bias=False,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\n latent_actions = jax.lax.stop_gradient(lam_outputs[""z_q""])\n outputs = dict(\n action_predictions=self.action_map(latent_actions),\n )\n return outputs",python,tab +782,1301040,"models/lamap.py",139,0,"",python,selection_command +783,1309774,"models/lamap.py",1626,0,"",python,selection_mouse +784,1309796,"models/lamap.py",1625,0,"",python,selection_command +785,1309918,"models/lamap.py",1625,1,"s",python,selection_mouse +786,1309934,"models/lamap.py",1626,0,"",python,selection_command +787,1310053,"models/lamap.py",1270,356," )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\n latent_actions = jax.lax.stop_gradient(lam_outputs[""z_q""])\n outputs = dict(\n action_predictions=self.action_map(latent_actions),\n )\n return outputs",python,selection_mouse +788,1310054,"models/lamap.py",524,1102," def setup(self):\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n latent_dim=self.latent_dim,\n num_latents=self.num_latents,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=self.lam_dropout,\n codebook_dropout=self.lam_codebook_dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n )\n self.action_map = nn.Dense(\n self.action_dim,\n use_bias=False,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\n latent_actions = jax.lax.stop_gradient(lam_outputs[""z_q""])\n outputs = dict(\n action_predictions=self.action_map(latent_actions),\n )\n return outputs",python,selection_mouse +789,1310136,"models/lamap.py",304,1322," latent_dim: int\n num_latents: int\n patch_size: int\n num_blocks: int\n num_heads: int\n dropout: float\n codebook_dropout: float\n use_flash_attention: bool\n\n # --- Mapper ---\n action_dim: int\n\n def setup(self):\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n latent_dim=self.latent_dim,\n num_latents=self.num_latents,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=self.lam_dropout,\n codebook_dropout=self.lam_codebook_dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n )\n self.action_map = nn.Dense(\n self.action_dim,\n use_bias=False,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\n latent_actions = jax.lax.stop_gradient(lam_outputs[""z_q""])\n outputs = dict(\n action_predictions=self.action_map(latent_actions),\n )\n return outputs",python,selection_mouse +790,1310243,"models/lamap.py",0,1626,"from typing import Dict, Any\n\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\n\nfrom models.lam import LatentActionModel\n\n\nclass LatentActionMapper(nn.Module):\n """"""Latent Action Mapper""""""\n\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n # --- LAM ---\n in_dim: int\n model_dim: int\n latent_dim: int\n num_latents: int\n patch_size: int\n num_blocks: int\n num_heads: int\n dropout: float\n codebook_dropout: float\n use_flash_attention: bool\n\n # --- Mapper ---\n action_dim: int\n\n def setup(self):\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n latent_dim=self.latent_dim,\n num_latents=self.num_latents,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=self.lam_dropout,\n codebook_dropout=self.lam_codebook_dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n use_flash_attention=self.use_flash_attention,\n )\n self.action_map = nn.Dense(\n self.action_dim,\n use_bias=False,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\n latent_actions = jax.lax.stop_gradient(lam_outputs[""z_q""])\n outputs = dict(\n action_predictions=self.action_map(latent_actions),\n )\n return outputs",python,selection_mouse +791,1311660,"models/lamap.py",0,0,"",python,selection_command +792,1333365,"models/lamap.py",65,0,"import flax.nnx as nnx\n",python,content +793,1333367,"models/lamap.py",88,24,"",python,content +794,1334177,"models/lamap.py",132,0,"class LatentActionMapper(nnx.Module):\n",python,content +795,1334681,"models/lamap.py",170,37,"",python,content +796,1339268,"models/lamap.py",524,0," def __init__(self, rngs: nnx.Rngs):\n",python,content +797,1339895,"models/lamap.py",564,0," self.rngs = rngs\n",python,content +798,1340170,"models/lamap.py",589,21,"",python,content +799,1340395,"models/lamap.py",659,0," model_dim=self.model_dim,\n",python,content +800,1341077,"models/lamap.py",697,0," ffn_dim=self.model_dim * 4, # Assuming ffn_dim is 4 times model_dim\n",python,content +801,1341265,"models/lamap.py",778,36,"",python,content +802,1341606,"models/lamap.py",860,0," patch_size=self.patch_size,\n",python,content +803,1341801,"models/lamap.py",900,0," num_blocks=self.num_blocks,\n",python,content +804,1341996,"models/lamap.py",940,0," num_heads=self.num_heads,\n",python,content +805,1342161,"models/lamap.py",978,0," dropout=self.dropout,\n",python,content +806,1342653,"models/lamap.py",1012,0," codebook_dropout=self.codebook_dropout,\n",python,content +807,1343066,"models/lamap.py",1064,0," param_dtype=self.param_dtype,\n",python,content +808,1343524,"models/lamap.py",1106,0," dtype=self.dtype,\n",python,content +809,1343914,"models/lamap.py",1136,296,"",python,content +810,1344227,"models/lamap.py",1194,0," rngs=self.rngs,\n",python,content +811,1344430,"models/lamap.py",1232,0," self.action_map = nnx.Linear(\n",python,content +812,1344623,"models/lamap.py",1270,0," self.latent_dim,\n",python,content +813,1344765,"models/lamap.py",1299,36,"",python,content +814,1345973,"models/lamap.py",1428,0," rngs=self.rngs,\n",python,content +815,1346519,"models/lamap.py",1456,13," )\n )",python,content +816,1346520,"models/lamap.py",1466,14,"",python,content +817,1347905,"models/lamap.py",1555,0," lam_outputs = self.lam.vq_encode(batch[""videos""], training=training)\n",python,content +818,1348400,"models/lamap.py",1632,74,"",python,content +819,1348581,"models/lamap.py",1699,0," action_predictions = self.action_map(latent_actions)\n",python,content +820,1349663,"models/lamap.py",1784,0," action_predictions=action_predictions,\n",python,content +821,1350050,"models/lamap.py",1835,64,"",python,content +822,1382181,"models/lamap.py",1624,0,"",python,selection_mouse +823,1382399,"models/lamap.py",1622,8,"training",python,selection_mouse +824,1384326,"models/lamap.py",1622,8,"",python,content +825,1384804,"models/lamap.py",1622,0,"F",python,content +826,1384805,"models/lamap.py",1623,0,"",python,selection_keyboard +827,1385010,"models/lamap.py",1623,0,"a",python,content +828,1385012,"models/lamap.py",1624,0,"",python,selection_keyboard +829,1385069,"models/lamap.py",1624,0,"l",python,content +830,1385070,"models/lamap.py",1625,0,"",python,selection_keyboard +831,1385158,"models/lamap.py",1625,0,"s",python,content +832,1385159,"models/lamap.py",1626,0,"",python,selection_keyboard +833,1385340,"models/lamap.py",1626,0,"e",python,content +834,1385342,"models/lamap.py",1627,0,"",python,selection_keyboard +835,1385864,"models/lamap.py",1626,0,"",python,selection_command +836,1401133,"models/lamap.py",315,0,"",python,selection_mouse +837,1403452,"models/lamap.py",323,0,"\n latent_dim: int",python,content +838,1403478,"models/lamap.py",328,0,"",python,selection_command +839,1403845,"models/lamap.py",329,0,"",python,selection_command +840,1404360,"models/lamap.py",328,0,"",python,selection_command +841,1404645,"models/lamap.py",328,1,"",python,content +842,1404813,"models/lamap.py",328,1,"",python,content +843,1405016,"models/lamap.py",328,1,"",python,content +844,1405164,"models/lamap.py",328,1,"",python,content +845,1405360,"models/lamap.py",328,1,"",python,content +846,1405699,"models/lamap.py",328,1,"",python,content +847,1406156,"models/lamap.py",328,0,"f",python,content +848,1406157,"models/lamap.py",329,0,"",python,selection_keyboard +849,1406300,"models/lamap.py",329,0,"f",python,content +850,1406302,"models/lamap.py",330,0,"",python,selection_keyboard +851,1406368,"models/lamap.py",330,0,"n",python,content +852,1406369,"models/lamap.py",331,0,"",python,selection_keyboard +853,1406850,"models/lamap.py",330,0,"",python,selection_command +854,1407035,"models/lamap.py",347,0,"",python,selection_command +855,1407544,"models/lamap.py",368,0,"",python,selection_command +856,1407611,"models/lamap.py",388,0,"",python,selection_command +857,1407660,"models/lamap.py",408,0,"",python,selection_command +858,1407705,"models/lamap.py",427,0,"",python,selection_command +859,1407706,"models/lamap.py",446,0,"",python,selection_command +860,1407732,"models/lamap.py",474,0,"",python,selection_command +861,1407804,"models/lamap.py",498,0,"",python,selection_command +862,1407864,"models/lamap.py",505,0,"",python,selection_command +863,1407865,"models/lamap.py",526,0,"",python,selection_command +864,1407986,"models/lamap.py",540,0,"",python,selection_command +865,1407987,"models/lamap.py",547,0,"",python,selection_command +866,1407987,"models/lamap.py",587,0,"",python,selection_command +867,1408046,"models/lamap.py",612,0,"",python,selection_command +868,1408047,"models/lamap.py",650,0,"",python,selection_command +869,1408048,"models/lamap.py",682,0,"",python,selection_command +870,1408161,"models/lamap.py",720,0,"",python,selection_command +871,1408486,"models/lamap.py",721,0,"",python,selection_command +872,1408996,"models/lamap.py",722,0,"",python,selection_command +873,1409071,"models/lamap.py",723,0,"",python,selection_command +874,1409084,"models/lamap.py",724,0,"",python,selection_command +875,1409122,"models/lamap.py",725,0,"",python,selection_command +876,1409147,"models/lamap.py",726,0,"",python,selection_command +877,1409187,"models/lamap.py",727,0,"",python,selection_command +878,1409300,"models/lamap.py",728,0,"",python,selection_command +879,1409325,"models/lamap.py",729,0,"",python,selection_command +880,1409326,"models/lamap.py",730,0,"",python,selection_command +881,1409326,"models/lamap.py",731,0,"",python,selection_command +882,1409452,"models/lamap.py",732,0,"",python,selection_command +883,1409453,"models/lamap.py",733,0,"",python,selection_command +884,1409453,"models/lamap.py",734,0,"",python,selection_command +885,1409489,"models/lamap.py",735,0,"",python,selection_command +886,1409490,"models/lamap.py",736,0,"",python,selection_command +887,1409490,"models/lamap.py",737,0,"",python,selection_command +888,1409528,"models/lamap.py",738,0,"",python,selection_command +889,1409565,"models/lamap.py",739,0,"",python,selection_command +890,1409587,"models/lamap.py",740,0,"",python,selection_command +891,1409960,"models/lamap.py",739,0,"",python,selection_command +892,1410073,"models/lamap.py",738,0,"",python,selection_command +893,1410398,"models/lamap.py",738,1,"",python,content +894,1410910,"models/lamap.py",738,10,"",python,content +895,1411338,"models/lamap.py",738,2,"",python,content +896,1411738,"models/lamap.py",738,1,"",python,content +897,1413478,"models/lamap.py",738,0,".",python,content +898,1413479,"models/lamap.py",739,0,"",python,selection_keyboard +899,1413844,"models/lamap.py",739,0,"f",python,content +900,1413844,"models/lamap.py",740,0,"",python,selection_keyboard +901,1414057,"models/lamap.py",740,0,"f",python,content +902,1414058,"models/lamap.py",741,0,"",python,selection_keyboard +903,1414538,"models/lamap.py",739,2,"ffn_dim",python,content +904,1414936,"models/lamap.py",745,0,"",python,selection_command +905,1415124,"models/lamap.py",746,0,"",python,selection_command +906,1415269,"models/lamap.py",747,0,"",python,selection_command +907,1415412,"models/lamap.py",748,0,"",python,selection_command +908,1415579,"models/lamap.py",749,0,"",python,selection_command +909,1416294,"models/lamap.py",749,39,"",python,content +910,1416329,"models/lamap.py",748,0,"",python,selection_command +911,1417720,"models/lamap.py",746,2,"",python,content +912,1418819,"models/lamap.py",747,0,"",python,selection_command +913,1418904,"models/lamap.py",747,0,",",python,content +914,1418905,"models/lamap.py",748,0,"",python,selection_keyboard +915,1419547,"models/lamap.py",747,1,"",python,content +916,1419694,"models/lamap.py",746,1,"",python,content +917,1419924,"models/lamap.py",746,0,",",python,content +918,1419925,"models/lamap.py",747,0,"",python,selection_keyboard +919,1420085,"models/lamap.py",746,0,"",python,selection_command +920,1425534,"models/lamap.py",332,0,"",python,selection_mouse +921,1427049,"models/lamap.py",324,17,"",python,content +922,1427072,"models/lamap.py",328,0,"",python,selection_command +923,1427072,"models/lamap.py",308,0,"",python,selection_command +924,1427307,"models/lamap.py",289,0,"",python,selection_command +925,1427552,"models/lamap.py",303,0,"\n ffn_dim: int",python,content +926,1427570,"models/lamap.py",308,0,"",python,selection_command +927,1428142,"models/lamap.py",325,0,"",python,selection_command +928,1441172,"train_lamap.py",0,0,"",python,tab +929,1443051,"train_lamap.py",0,0,"",python,selection_command +930,1443506,"train_lamap.py",0,40,"from dataclasses import dataclass, field",python,selection_command +931,1443906,"train_lamap.py",0,11655,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom models.lamap import LatentActionMapper\nfrom utils.dataloader import create_dataloader_iterator\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n warmup_steps: int = 5000\n lr_schedule : str = ""wsd"" # supported options: wsd, cos\n # LAM\n model_dim: int = 512\n latent_dim: int = 32\n num_latents: int = 6\n patch_size: int = 16\n num_blocks: int = 8\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.0\n param_dtype: jnp.dtype = jnp.float32\n dtype: jnp.dtype = jnp.bfloat16\n # LAMAP\n # FIXME (f.srambical): this assumes that the number of actions is the same as the number of latent actions\n action_dim: int = 32\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_lamap""\n tags: list[str] = field(default_factory=lambda: [""lamap""])\n log_interval: int = 5\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n wandb_id: str = """"\n use_flash_attention: bool = True\n\n\nargs = tyro.cli(Args)\n\ndef lamap_loss_fn(params, state, inputs):\n # --- Compute loss ---\n inputs[""videos""] = inputs[""videos""].astype(args.dtype) / 255.0\n outputs = state.apply_fn(\n params, inputs, training=True, rngs={""dropout"": inputs[""rng""]}\n )\n loss = optax.softmax_cross_entropy(outputs[""action_predictions""], inputs[""actions""]).mean()\n return loss\n\n@nnx.jit\ndef train_step(state, inputs):\n lamap.train()\n grad_fn = jax.value_and_grad(lamap_loss_fn, allow_int=True)\n loss, grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n return state, loss\n\n@nnx.jit\ndef val_step(lamap: LatentActionMapper, inputs: dict) -> tuple[jax.Array, jax.Array, dict]:\n lamap.eval()\n (loss, (recon, _, metrics)) = lamap_loss_fn(lamap, inputs, allow_int=False)\n return loss, recon, metrics\n\ndef calculate_validation_metrics(val_dataloader, lam):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n inputs = dict(videos=videos)\n loss, recon, metrics = val_step(lam, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n # --- Initialize model ---\n lamap = LatentActionMapper(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n latent_dim=args.latent_dim,\n action_dim=args.action_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n )\n\n image_shape = (args.image_height, args.image_width, args.image_channels)\n rng, _rng = jax.random.split(rng)\n inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=args.dtype,\n ),\n rng=_rng,\n )\n rng, _rng = jax.random.split(rng)\n init_params = lamap.init(_rng, inputs)\n\n param_counts = count_parameters_by_component(init_params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(args.lr_schedule, \n args.init_lr, \n args.max_lr, \n args.decay_end, \n args.num_steps, \n args.warmup_steps, \n args.wsd_decay_steps)\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4, mu_dtype=args.dtype)\n train_state = TrainState.create(apply_fn=lamap.apply, params=init_params, tx=tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n train_iterator = create_dataloader_iterator(args.data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n if args.val_data_dir:\n val_iterator = create_dataloader_iterator(args.val_data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_train_state = jax.tree_util.tree_map(\n ocp.utils.to_shape_dtype_struct, train_state\n )\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_train_state),\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator),\n ),\n )\n train_state = restored[""model_state""]\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader = (jax.make_array_from_process_local_data(videos_sharding, elem) for elem in grain_iterator) # type: ignore\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng = jax.random.split(rng)\n\n # FIXME (f.srambical): use mock actions for now\n rng, _rng_actions = jax.random.split(rng)\n actions = jax.random.uniform(\n _rng_actions,\n (per_device_batch_size_for_init, args.action_dim),\n dtype=args.dtype,\n )\n\n inputs = dict(\n videos=videos,\n actions=actions,\n rng=_rng\n )\n train_state, loss = train_step(\n train_state, inputs\n )\n lr = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n ""lr"": lr,\n }\n )\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.StandardSave(train_state),\n dataloader_state=grain.checkpoint.CheckpointSave(\n grain_iterator\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,selection_command +932,1444836,"train_lamap.py",0,0,"",python,selection_command +933,1457182,"train_lamap.py",41,10,"",python,content +934,1457383,"train_lamap.py",66,0,"import jax\n",python,content +935,1457618,"train_lamap.py",77,0,"import jax.numpy as jnp\n",python,content +936,1457619,"train_lamap.py",101,0,"import numpy as np\n",python,content +937,1457835,"train_lamap.py",120,119,"",python,content +938,1457836,"train_lamap.py",133,85,"",python,content +939,1458770,"train_lamap.py",194,0,"from jax.sharding import Mesh, PartitionSpec, NamedSharding\n",python,content +940,1459340,"train_lamap.py",254,0,"from jax.experimental.mesh_utils import create_device_mesh\n",python,content +941,1459952,"train_lamap.py",313,0,"import orbax.checkpoint as ocp\n",python,content +942,1471080,"train_lamap.py",948,0," wsd_decay_steps: int = 10000\n",python,content +943,1471734,"train_lamap.py",981,101,"",python,content +944,1471945,"train_lamap.py",1010,0," lr_schedule: str = ""wsd""\n",python,content +945,1472119,"train_lamap.py",1039,60,"",python,content +946,1476147,"train_lamap.py",1344,111,"",python,content +947,1481530,"train_lamap.py",1842,0,"\n",python,content +948,1482333,"train_lamap.py",1843,0,"def lamap_loss_fn(model: LatentActionMapper, inputs: dict, training: bool = True) -> jax.Array:\n",python,content +949,1483011,"train_lamap.py",1939,69,"",python,content +950,1483273,"train_lamap.py",2006,0," outputs = model(inputs, training=training)\n",python,content +951,1484071,"train_lamap.py",2053,0," loss = optax.softmax_cross_entropy(outputs[""action_predictions""], inputs[""actions""]).mean()\n",python,content +952,1484333,"train_lamap.py",2149,203,"",python,content +953,1484334,"train_lamap.py",2166,0,"\n",python,content +954,1486592,"train_lamap.py",2207,18,"",python,content +955,1490248,"train_lamap.py",2397,0,"\n",python,content +956,1492184,"train_lamap.py",2499,17,"",python,content +957,1492643,"train_lamap.py",2612,0,"\n",python,content +958,1510609,"train_lam.py",0,0,"",python,tab +959,1515833,"train_lamap.py",0,0,"",python,tab +960,1519131,"train_lamap.py",4538,0," rngs=rngs,\n",python,content +961,1529447,"train_lamap.py",5679,0," lr_schedule = get_lr_schedule(\n",python,content +962,1529689,"train_lamap.py",5714,0," args.lr_schedule,\n",python,content +963,1529972,"train_lamap.py",5740,0," args.init_lr,\n",python,content +964,1530111,"train_lamap.py",5762,0," args.max_lr,\n",python,content +965,1530340,"train_lamap.py",5783,0," args.decay_end,\n",python,content +966,1530585,"train_lamap.py",5807,0," args.num_steps,\n",python,content +967,1535469,"train_lamap.py",5831,0," args.warmup_steps,\n",python,content +968,1535731,"train_lamap.py",5858,0," args.wsd_decay_steps,\n",python,content +969,1535985,"train_lamap.py",5888,0," )\n",python,content +970,1535986,"train_lamap.py",5894,0," tx = optax.adamw(\n",python,content +971,1536961,"train_lamap.py",5916,0," learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4, mu_dtype=args.dtype\n",python,content +972,1536962,"train_lamap.py",6006,0," )\n",python,content +973,1536963,"train_lamap.py",6012,0," train_state = nnx.TrainState.create(apply_fn=lamap.apply, params=init_params, tx=tx)\n",python,content +974,1536964,"train_lamap.py",6101,0,"\n",python,content +975,1536965,"train_lamap.py",6102,0," # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n",python,content +976,1536966,"train_lamap.py",6184,0," device_mesh_arr = create_device_mesh((num_devices,))\n",python,content +977,1537349,"train_lamap.py",6102,691,"",python,content +978,1553749,"train_lamap.py",8266,0," train_iterator = create_dataloader_iterator(\n",python,content +979,1554401,"train_lamap.py",8315,0," args.data_dir, image_shape, args.seq_len, args.batch_size, args.seed\n",python,content +980,1554559,"train_lamap.py",8392,0," )\n",python,content +981,1554875,"train_lamap.py",8398,118,"",python,content +982,1555121,"train_lamap.py",8424,0," val_iterator = create_dataloader_iterator(\n",python,content +983,1555861,"train_lamap.py",8475,0," args.val_data_dir, image_shape, args.seq_len, args.batch_size, args.seed\n",python,content +984,1556121,"train_lamap.py",8560,0," )\n",python,content +985,1556122,"train_lamap.py",8570,124,"",python,content +986,1558122,"train_lamap.py",8424,145," val_iterator = create_dataloader_iterator(args.val_data_dir, image_shape, args.seq_len, args.batch_size, args.seed)",python,content +987,1558123,"train_lamap.py",8266,131," train_iterator = create_dataloader_iterator(args.data_dir, image_shape, args.seq_len, args.batch_size, args.seed)",python,content +988,1558123,"train_lamap.py",5679,421," lr_schedule = get_lr_schedule(args.lr_schedule, \n args.init_lr, \n args.max_lr, \n args.decay_end, \n args.num_steps, \n args.warmup_steps, \n args.wsd_decay_steps)\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4, mu_dtype=args.dtype)\n train_state = TrainState.create(apply_fn=lamap.apply, params=init_params, tx=tx)",python,content +989,1558123,"train_lamap.py",4538,19,"",python,content +990,1558123,"train_lamap.py",4024,0," # --- Initialize model ---\n",python,content +991,1558123,"train_lamap.py",3888,39," rng = jax.random.key(args.seed)",python,content +992,1558123,"train_lamap.py",2611,1,"",python,content +993,1558123,"train_lamap.py",2499,0," lamap.eval()\n",python,content +994,1558123,"train_lamap.py",2396,1,"",python,content +995,1558123,"train_lamap.py",2207,0," lamap.train()\n",python,content +996,1558126,"train_lamap.py",2165,1,"",python,content +997,1558126,"train_lamap.py",2006,46," outputs = state.apply_fn(\n params, inputs, training=True, rngs={""dropout"": inputs[""rng""]}\n )",python,content +998,1558126,"train_lamap.py",1843,95,"def lamap_loss_fn(params, state, inputs):\n # --- Compute loss ---",python,content +999,1558126,"train_lamap.py",1841,1,"",python,content +1000,1558126,"train_lamap.py",1344,0," # FIXME (f.srambical): this assumes that the number of actions is the same as the number of latent actions\n",python,content +1001,1558126,"train_lamap.py",1010,28," lr_schedule : str = ""wsd"" # supported options: wsd, cos",python,content +1002,1558126,"train_lamap.py",948,32," wsd_decay_steps: int = 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule",python,content +1003,1558126,"train_lamap.py",194,150,"",python,content +1004,1558126,"train_lamap.py",101,32,"",python,content +1005,1558126,"train_lamap.py",66,0,"from jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\n",python,content +1006,1558127,"train_lamap.py",41,0,"import os\n",python,content +1007,1559389,"train_lamap.py",50,0,"",python,selection_mouse +1008,1559441,"train_lamap.py",49,0,"",python,selection_command +1009,1560142,"train_lamap.py",41,10,"",python,content +1010,1560294,"train_lamap.py",65,0,"",python,selection_command +1011,1560461,"train_lamap.py",66,0,"",python,selection_command +1012,1560992,"train_lamap.py",126,0,"",python,selection_command +1013,1561070,"train_lamap.py",185,0,"",python,selection_command +1014,1561126,"train_lamap.py",198,0,"",python,selection_command +1015,1561127,"train_lamap.py",229,0,"",python,selection_command +1016,1561292,"train_lamap.py",248,0,"",python,selection_command +1017,1561293,"train_lamap.py",259,0,"",python,selection_command +1018,1561294,"train_lamap.py",283,0,"",python,selection_command +1019,1561498,"train_lamap.py",295,0,"",python,selection_command +1020,1561498,"train_lamap.py",308,0,"",python,selection_command +1021,1561499,"train_lamap.py",321,0,"",python,selection_command +1022,1561499,"train_lamap.py",344,0,"",python,selection_command +1023,1561499,"train_lamap.py",345,0,"",python,selection_command +1024,1561677,"train_lamap.py",389,0,"",python,selection_command +1025,1561678,"train_lamap.py",445,0,"",python,selection_command +1026,1561678,"train_lamap.py",488,0,"",python,selection_command +1027,1561678,"train_lamap.py",552,0,"",python,selection_command +1028,1561679,"train_lamap.py",553,0,"",python,selection_command +1029,1561679,"train_lamap.py",554,0,"",python,selection_command +1030,1561845,"train_lamap.py",565,0,"",python,selection_command +1031,1561846,"train_lamap.py",577,0,"",python,selection_command +1032,1561846,"train_lamap.py",594,0,"",python,selection_command +1033,1561846,"train_lamap.py",623,0,"",python,selection_command +1034,1561847,"train_lamap.py",641,0,"",python,selection_command +1035,1561847,"train_lamap.py",663,0,"",python,selection_command +1036,1561893,"train_lamap.py",691,0,"",python,selection_command +1037,1561894,"train_lamap.py",718,0,"",python,selection_command +1038,1561895,"train_lamap.py",745,0,"",python,selection_command +1039,1561895,"train_lamap.py",768,0,"",python,selection_command +1040,1561895,"train_lamap.py",796,0,"",python,selection_command +1041,1561938,"train_lamap.py",827,0,"",python,selection_command +1042,1562024,"train_lamap.py",846,0,"",python,selection_command +1043,1562233,"train_lamap.py",871,0,"",python,selection_command +1044,1562358,"train_lamap.py",896,0,"",python,selection_command +1045,1562546,"train_lamap.py",921,0,"",python,selection_command +1046,1563031,"train_lamap.py",948,0,"",python,selection_command +1047,1563088,"train_lamap.py",1049,0,"",python,selection_command +1048,1563110,"train_lamap.py",1078,0,"",python,selection_command +1049,1563144,"train_lamap.py",1138,0,"",python,selection_command +1050,1563199,"train_lamap.py",1148,0,"",python,selection_command +1051,1563200,"train_lamap.py",1173,0,"",python,selection_command +1052,1563345,"train_lamap.py",1198,0,"",python,selection_command +1053,1563591,"train_lamap.py",1223,0,"",python,selection_command +1054,1563817,"train_lamap.py",1248,0,"",python,selection_command +1055,1564080,"train_lamap.py",1272,0,"",python,selection_command +1056,1578266,"train_lamap.py",1210,0,"",python,selection_mouse +1057,1579963,"train_lam.py",0,0,"",python,tab +1058,1580474,"train_lam.py",1339,0,"",python,selection_mouse +1059,1580947,"train_lam.py",1293,0,"",python,selection_mouse +1060,1582966,"train_lamap.py",0,0,"",python,tab +1061,1582967,"train_lamap.py",1155,0,"",python,selection_mouse +1062,1583344,"train_lamap.py",1172,0,"\n ffn_dim: int = 2048",python,content +1063,1583413,"train_lamap.py",1177,0,"",python,selection_command +1064,1584223,"train_lamap.py",1201,0,"",python,selection_command +1065,1584768,"train_lamap.py",1226,0,"",python,selection_command +1066,1584894,"train_lamap.py",1251,0,"",python,selection_command +1067,1584895,"train_lamap.py",1276,0,"",python,selection_command +1068,1584895,"train_lamap.py",1300,0,"",python,selection_command +1069,1585094,"train_lamap.py",1323,0,"",python,selection_command +1070,1585095,"train_lamap.py",1348,0,"",python,selection_command +1071,1585245,"train_lamap.py",1382,0,"",python,selection_command +1072,1585245,"train_lamap.py",1423,0,"",python,selection_command +1073,1585246,"train_lamap.py",1459,0,"",python,selection_command +1074,1585246,"train_lamap.py",1471,0,"",python,selection_command +1075,1585246,"train_lamap.py",1582,0,"",python,selection_command +1076,1585246,"train_lamap.py",1607,0,"",python,selection_command +1077,1585247,"train_lamap.py",1621,0,"",python,selection_command +1078,1585285,"train_lamap.py",1643,0,"",python,selection_command +1079,1585286,"train_lamap.py",1664,0,"",python,selection_command +1080,1585286,"train_lamap.py",1686,0,"",python,selection_command +1081,1585286,"train_lamap.py",1716,0,"",python,selection_command +1082,1585364,"train_lamap.py",1779,0,"",python,selection_command +1083,1585546,"train_lamap.py",1805,0,"",python,selection_command +1084,1586469,"train_lamap.py",1828,0,"",python,selection_command +1085,1586974,"train_lamap.py",1869,0,"",python,selection_command +1086,1587425,"train_lamap.py",1913,0,"",python,selection_command +1087,1587425,"train_lamap.py",1940,0,"",python,selection_command +1088,1587498,"train_lamap.py",1971,0,"",python,selection_command +1089,1587499,"train_lamap.py",1995,0,"",python,selection_command +1090,1587499,"train_lamap.py",2018,0,"",python,selection_command +1091,1587499,"train_lamap.py",2051,0,"",python,selection_command +1092,1587499,"train_lamap.py",2052,0,"",python,selection_command +1093,1587500,"train_lamap.py",2057,0,"",python,selection_command +1094,1587676,"train_lamap.py",2075,0,"",python,selection_command +1095,1587974,"train_lamap.py",2080,0,"",python,selection_command +1096,1588645,"train_lamap.py",2122,0,"",python,selection_command +1097,1588995,"train_lamap.py",2149,0,"",python,selection_command +1098,1589232,"train_lamap.py",2216,0,"",python,selection_command +1099,1589408,"train_lamap.py",2246,0,"",python,selection_command +1100,1589562,"train_lamap.py",2317,0,"",python,selection_command +1101,1589736,"train_lamap.py",2323,0,"",python,selection_command +1102,1589924,"train_lamap.py",2419,0,"",python,selection_command +1103,1590097,"train_lamap.py",2431,0,"",python,selection_command +1104,1590223,"train_lamap.py",2436,0,"",python,selection_command +1105,1590391,"train_lamap.py",2445,0,"",python,selection_command +1106,1590533,"train_lamap.py",2476,0,"",python,selection_command +1107,1590683,"train_lamap.py",2494,0,"",python,selection_command +1108,1590816,"train_lamap.py",2558,0,"",python,selection_command +1109,1590990,"train_lamap.py",2613,0,"",python,selection_command +1110,1591124,"train_lamap.py",2660,0,"",python,selection_command +1111,1591229,"train_lamap.py",2679,0,"",python,selection_command +1112,1597132,"train_lamap.py",2135,0,"",python,selection_mouse +1113,1598802,"train_lamap.py",2166,0,"",python,selection_mouse +1114,1617047,"train_lam.py",0,0,"",python,tab +1115,1617048,"train_lam.py",2165,0,"",python,selection_mouse +1116,1619677,"train_lam.py",2135,57,") -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:",python,selection_command +1117,1619840,"train_lam.py",2069,123," model: LatentActionModel, inputs: dict, training: bool = True\n) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:",python,selection_command +1118,1620052,"train_lam.py",2052,140,"def lam_loss_fn(\n model: LatentActionModel, inputs: dict, training: bool = True\n) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:",python,selection_command +1119,1621414,"train_lam.py",2052,0,"",python,selection_command +1120,1623419,"train_lamap.py",0,0,"",python,tab +1121,1623420,"train_lamap.py",2075,0,"",python,selection_mouse +1122,1624059,"train_lamap.py",2075,0,"\ndef lam_loss_fn(\n model: LatentActionModel, inputs: dict, training: bool = True\n) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:",python,content +1123,1624132,"train_lamap.py",2076,0,"",python,selection_command +1124,1624751,"train_lamap.py",2075,0,"",python,selection_command +1125,1625279,"train_lamap.py",2076,0,"",python,selection_command +1126,1625585,"train_lamap.py",2077,0,"",python,selection_command +1127,1625735,"train_lamap.py",2078,0,"",python,selection_command +1128,1625908,"train_lamap.py",2079,0,"",python,selection_command +1129,1626047,"train_lamap.py",2080,0,"",python,selection_command +1130,1626161,"train_lamap.py",2081,0,"",python,selection_command +1131,1626331,"train_lamap.py",2082,0,"",python,selection_command +1132,1626610,"train_lamap.py",2083,0,"",python,selection_command +1133,1626909,"train_lamap.py",2083,0,"a",python,content +1134,1626910,"train_lamap.py",2084,0,"",python,selection_keyboard +1135,1627233,"train_lamap.py",2084,0,"p",python,content +1136,1627234,"train_lamap.py",2085,0,"",python,selection_keyboard +1137,1627806,"train_lamap.py",2084,0,"",python,selection_command +1138,1628150,"train_lamap.py",2103,0,"",python,selection_command +1139,1628788,"train_lamap.py",2104,0,"",python,selection_command +1140,1629317,"train_lamap.py",2105,0,"",python,selection_command +1141,1629355,"train_lamap.py",2106,0,"",python,selection_command +1142,1629355,"train_lamap.py",2107,0,"",python,selection_command +1143,1629475,"train_lamap.py",2108,0,"",python,selection_command +1144,1629475,"train_lamap.py",2109,0,"",python,selection_command +1145,1629560,"train_lamap.py",2110,0,"",python,selection_command +1146,1629561,"train_lamap.py",2111,0,"",python,selection_command +1147,1629561,"train_lamap.py",2112,0,"",python,selection_command +1148,1629659,"train_lamap.py",2113,0,"",python,selection_command +1149,1629660,"train_lamap.py",2114,0,"",python,selection_command +1150,1629660,"train_lamap.py",2115,0,"",python,selection_command +1151,1629727,"train_lamap.py",2116,0,"",python,selection_command +1152,1629728,"train_lamap.py",2117,0,"",python,selection_command +1153,1629785,"train_lamap.py",2118,0,"",python,selection_command +1154,1630472,"train_lamap.py",2118,5,"",python,content +1155,1631191,"train_lamap.py",2118,0,"M",python,content +1156,1631192,"train_lamap.py",2119,0,"",python,selection_keyboard +1157,1632055,"train_lamap.py",2119,0,"a",python,content +1158,1632056,"train_lamap.py",2120,0,"",python,selection_keyboard +1159,1632154,"train_lamap.py",2120,0,"p",python,content +1160,1632156,"train_lamap.py",2121,0,"",python,selection_keyboard +1161,1632360,"train_lamap.py",2121,0,"p",python,content +1162,1632363,"train_lamap.py",2122,0,"",python,selection_keyboard +1163,1632437,"train_lamap.py",2122,0,"e",python,content +1164,1632438,"train_lamap.py",2123,0,"",python,selection_keyboard +1165,1632564,"train_lamap.py",2123,0,"r",python,content +1166,1632564,"train_lamap.py",2124,0,"",python,selection_keyboard +1167,1634093,"train_lamap.py",2123,0,"",python,selection_command +1168,1635530,"train_lamap.py",2190,0,"",python,selection_command +1169,1636940,"train_lamap.py",2219,0,"\n",python,content +1170,1638532,"train_lamap.py",2220,0,"\n",python,content +1171,1638670,"train_lamap.py",2221,0,"\n",python,content +1172,1639123,"train_lamap.py",2221,0,"",python,selection_command +1173,1639410,"train_lamap.py",2220,0,"",python,selection_command +1174,1640118,"train_lamap.py",2220,0," ",python,content +1175,1641360,"train_lam.py",0,0,"",python,tab +1176,1641360,"train_lam.py",2302,0,"",python,selection_mouse +1177,1641731,"train_lam.py",2239,0,"",python,selection_mouse +1178,1642574,"train_lam.py",2220,65," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0",python,selection_command +1179,1642850,"train_lam.py",2220,110," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)",python,selection_command +1180,1643441,"train_lam.py",2220,157," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)",python,selection_command +1181,1643592,"train_lam.py",2220,217," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)",python,selection_command +1182,1643593,"train_lam.py",2220,250," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]",python,selection_command +1183,1643593,"train_lam.py",2220,315," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()",python,selection_command +1184,1643593,"train_lam.py",2220,400," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()",python,selection_command +1185,1643971,"train_lam.py",2220,434," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(",python,selection_command +1186,1643971,"train_lam.py",2220,495," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])",python,selection_command +1187,1643972,"train_lam.py",2220,508," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()",python,selection_command +1188,1643972,"train_lam.py",2220,565," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss",python,selection_command +1189,1643972,"train_lam.py",2220,566," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n",python,selection_command +1190,1643972,"train_lam.py",2220,607," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---",python,selection_command +1191,1643973,"train_lam.py",2220,685," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])",python,selection_command +1192,1644114,"train_lam.py",2220,766," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])",python,selection_command +1193,1644115,"train_lam.py",2220,817," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()",python,selection_command +1194,1644115,"train_lam.py",2220,868," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()",python,selection_command +1195,1644115,"train_lam.py",2220,935," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())",python,selection_command +1196,1644116,"train_lam.py",2220,993," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))",python,selection_command +1197,1644116,"train_lam.py",2220,1013," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(",python,selection_command +1198,1644116,"train_lam.py",2220,1032," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,",python,selection_command +1199,1644116,"train_lam.py",2220,1049," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,",python,selection_command +1200,1644117,"train_lam.py",2220,1072," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,",python,selection_command +1201,1644176,"train_lam.py",2220,1113," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,",python,selection_command +1202,1644291,"train_lam.py",2220,1132," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,",python,selection_command +1203,1644397,"train_lam.py",2220,1151," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,",python,selection_command +1204,1644533,"train_lam.py",2220,1202," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=(index_counts != 0).mean(),",python,selection_command +1205,1644674,"train_lam.py",2220,1208," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=(index_counts != 0).mean(),\n )",python,selection_command +1206,1644940,"train_lam.py",2220,1267," gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=(index_counts != 0).mean(),\n )\n return loss, (outputs[""recon""], index_counts, metrics)",python,selection_command +1207,1645368,"train_lam.py",2220,0,"",python,selection_command +1208,1647003,"train_lamap.py",0,0,"",python,tab +1209,1647875,"train_lamap.py",2223,0,"",python,selection_command +1210,1648952,"train_lamap.py",2224,0,"\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=training)\n outputs[""recon""] = outputs[""recon""].astype(jnp.float32)\n gt_future_frames = gt[:, 1:]\n mse = jnp.square(gt_future_frames - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = gt_future_frames.clip(0, 1).reshape(-1, *gt_future_frames.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n count_fn = jax.vmap(lambda i: (outputs[""indices""] == i).sum())\n index_counts = count_fn(jnp.arange(args.num_latents))\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=(index_counts != 0).mean(),\n )\n return loss, (outputs[""recon""], index_counts, metrics)",python,content +1211,1648952,"train_lamap.py",2220,4,"",python,content +1212,1649029,"train_lamap.py",2225,0,"",python,selection_command +1213,1649888,"train_lamap.py",2220,0,"",python,selection_command +1214,1650644,"train_lamap.py",2220,1,"",python,content +1215,1650680,"train_lamap.py",2224,0,"",python,selection_command +1216,1650804,"train_lamap.py",2220,0,"\n",python,content +1217,1650873,"train_lamap.py",2220,0,"",python,selection_command +1218,1651237,"train_lamap.py",2225,1263,"",python,content +1219,1651292,"train_lamap.py",2220,1,"",python,content +1220,1651344,"train_lamap.py",2223,0,"",python,selection_command +1221,1654366,"train_lam.py",0,0,"",python,tab +1222,1654367,"train_lam.py",2301,0,"",python,selection_mouse +1223,1654980,"train_lam.py",2239,0,"",python,selection_mouse +1224,1657653,"train_lam.py",2302,0,"",python,selection_mouse +1225,1661166,"train_lam.py",2236,0,"",python,selection_command +1226,1665582,"train_lamap.py",0,0,"",python,tab +1227,1665582,"train_lamap.py",2224,0,"",python,selection_mouse +1228,1666437,"train_lamap.py",2223,0,"",python,selection_command +1229,1666803,"train_lamap.py",2224,0,"\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)",python,content +1230,1666841,"train_lamap.py",2229,0,"",python,selection_command +1231,1667482,"train_lamap.py",2223,0,"",python,selection_command +1232,1667834,"train_lamap.py",2220,5,"",python,content +1233,1667903,"train_lamap.py",2224,0,"",python,selection_command +1234,1668191,"train_lamap.py",2290,0,"",python,selection_command +1235,1669781,"train_lam.py",0,0,"",python,tab +1236,1669782,"train_lam.py",2300,0,"",python,selection_mouse +1237,1670209,"train_lam.py",2392,0,"",python,selection_mouse +1238,1671047,"train_lam.py",2345,0,"",python,selection_mouse +1239,1678112,"train_lamap.py",0,0,"",python,tab +1240,1678113,"train_lamap.py",2311,0,"",python,selection_mouse +1241,1678712,"train_lamap.py",2330,0,"\n outputs = model(inputs, training=training)",python,content +1242,1678771,"train_lamap.py",2335,0,"",python,selection_command +1243,1682117,"models/lamap.py",0,0,"",python,tab +1244,1685747,"train_lamap.py",0,0,"",python,tab +1245,1688744,"train_lam.py",0,0,"",python,tab +1246,1688745,"train_lam.py",2452,0,"",python,selection_mouse +1247,1690655,"train_lam.py",2476,0,"",python,selection_mouse +1248,1694648,"train_lamap.py",0,0,"",python,tab +1249,1694649,"train_lamap.py",2636,0,"",python,selection_mouse +1250,1695145,"train_lamap.py",2734,0,"",python,selection_mouse +1251,1695177,"train_lamap.py",2733,0,"",python,selection_command +1252,1695906,"train_lamap.py",2648,0,"",python,selection_mouse +1253,1696952,"train_lamap.py",2621,0,"",python,selection_command +1254,1697094,"train_lamap.py",2571,0,"",python,selection_command +1255,1697261,"train_lamap.py",2541,0,"",python,selection_command +1256,1697400,"train_lamap.py",2474,0,"",python,selection_command +1257,1697574,"train_lamap.py",2447,0,"",python,selection_command +1258,1697718,"train_lamap.py",2405,0,"",python,selection_command +1259,1697838,"train_lamap.py",2379,0,"",python,selection_command +1260,1698006,"train_lamap.py",2378,0,"",python,selection_command +1261,1698137,"train_lamap.py",2356,0,"",python,selection_command +1262,1698355,"train_lamap.py",2377,0,"\n loss = optax.softmax_cross_entropy(outputs[""action_predictions""], inputs[""actions""]).mean()",python,content +1263,1698404,"train_lamap.py",2382,0,"",python,selection_command +1264,1698946,"train_lamap.py",2474,0,"",python,selection_command +1265,1699268,"train_lamap.py",2382,0,"",python,selection_command +1266,1699497,"train_lamap.py",2473,0,"\n ",python,content +1267,1699961,"train_lamap.py",2478,0,"r",python,content +1268,1699961,"train_lamap.py",2479,0,"",python,selection_keyboard +1269,1700144,"train_lamap.py",2479,0,"e",python,content +1270,1700145,"train_lamap.py",2480,0,"",python,selection_keyboard +1271,1700344,"train_lamap.py",2480,0,"t",python,content +1272,1700346,"train_lamap.py",2481,0,"",python,selection_keyboard +1273,1700946,"train_lamap.py",2481,0,"u",python,content +1274,1700947,"train_lamap.py",2482,0,"",python,selection_keyboard +1275,1701123,"train_lamap.py",2482,0,"r",python,content +1276,1701125,"train_lamap.py",2483,0,"",python,selection_keyboard +1277,1701240,"train_lamap.py",2483,0,"n",python,content +1278,1701241,"train_lamap.py",2484,0,"",python,selection_keyboard +1279,1701293,"train_lamap.py",2484,0," ",python,content +1280,1701294,"train_lamap.py",2485,0,"",python,selection_keyboard +1281,1701408,"train_lamap.py",2485,0,"l",python,content +1282,1701408,"train_lamap.py",2486,0,"",python,selection_keyboard +1283,1701581,"train_lamap.py",2486,0,"o",python,content +1284,1701582,"train_lamap.py",2487,0,"",python,selection_keyboard +1285,1701679,"train_lamap.py",2487,0,"s",python,content +1286,1701680,"train_lamap.py",2488,0,"",python,selection_keyboard +1287,1701789,"train_lamap.py",2488,0,"s",python,content +1288,1701790,"train_lamap.py",2489,0,"",python,selection_keyboard +1289,1702149,"train_lamap.py",2488,0,"",python,selection_command +1290,1702309,"train_lamap.py",2490,0,"",python,selection_command +1291,1702467,"train_lamap.py",2491,0,"",python,selection_command +1292,1702856,"train_lamap.py",2506,0,"",python,selection_command +1293,1713469,"train_lamap.py",2491,0,"",python,selection_command +1294,1713699,"train_lamap.py",2490,0,"",python,selection_command +1295,1713846,"train_lamap.py",2488,0,"",python,selection_command +1296,1713997,"train_lamap.py",2392,0,"",python,selection_command +1297,1714382,"train_lamap.py",2345,0,"",python,selection_command +1298,1732884,"train_dynamics.py",0,0,"from dataclasses import dataclass, field\nimport os\nfrom typing import cast\n\nimport einops\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\nimport flax.nnx as nnx\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import create_dataloader_iterator\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = (\n 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n )\n warmup_steps: int = 5000\n lr_schedule: str = ""wsd"" # supported options: wsd, cos\n # Tokenizer\n tokenizer_dim: int = 512\n tokenizer_ffn_dim: int = 2048\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 4\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n lam_ffn_dim: int = 2048\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 4\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_type: str = ""maskgit"" # supported options: maskgit, causal\n dyna_dim: int = 512\n dyna_ffn_dim: int = 2048\n dyna_num_blocks: int = 6\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n param_dtype = jnp.float32\n dtype = jnp.bfloat16\n use_flash_attention: bool = True\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n val_data_dir: str = """"\n val_interval: int = 20_000\n val_steps: int = 50\n wandb_id: str = """"\n\n\nargs = tyro.cli(Args)\n\n\ndef dynamics_loss_fn(\n model: Genie, inputs: dict\n) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n """"""Compute masked dynamics loss""""""\n gt = jnp.asarray(inputs[""videos""], dtype=jnp.float32) / 255.0\n inputs[""videos""] = gt.astype(args.dtype)\n outputs = model(inputs, training=True)\n mask = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n gt = gt.clip(0, 1).reshape(-1, *gt.shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = jnp.asarray(pix.psnr(gt, recon)).mean()\n ssim = jnp.asarray(pix.ssim(gt, recon)).mean()\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]), size=args.num_latent_actions, fill_value=0\n )\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]), size=args.num_patch_latents, fill_value=0\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n psnr=psnr,\n ssim=ssim,\n codebook_usage_lam=codebook_usage_lam,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n\n@nnx.jit\ndef train_step(\n model: Genie, optimizer: nnx.Optimizer, inputs: dict\n) -> tuple[jax.Array, jax.Array, dict]:\n """"""Update state and compute metrics""""""\n\n def loss_fn(model: Genie) -> tuple[jax.Array, tuple[jax.Array, dict]]:\n model.train()\n return dynamics_loss_fn(model, inputs)\n\n (loss, (recon, metrics)), grads = nnx.value_and_grad(loss_fn, has_aux=True)(model)\n optimizer.update(grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return loss, recon, metrics\n\n@nnx.jit\ndef val_step(genie: Genie, inputs: dict) -> tuple[jax.Array, jax.Array, dict]:\n """"""Evaluate model and compute metrics""""""\n genie.eval()\n (loss, (recon, metrics)) = dynamics_loss_fn(genie, inputs)\n return loss, recon, metrics\n\n\ndef calculate_validation_metrics(val_dataloader, genie, rng):\n step = 0\n loss_per_step = []\n metrics_per_step = []\n for videos in val_dataloader:\n rng, _rng_mask = jax.random.split(rng, 2)\n inputs = dict(videos=videos, mask_rng=_rng_mask)\n loss, recon, metrics = val_step(genie, inputs)\n loss_per_step.append(loss)\n metrics_per_step.append(metrics)\n step += 1\n if step > args.val_steps:\n break\n\n if step < args.val_steps:\n print(f""Warning: Your validation dataset is too small to make val_steps many steps. Made {step} steps, expected {args.val_steps}"")\n\n val_loss = np.mean(loss_per_step)\n val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }\n return val_loss, val_metrics, inputs, recon\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.key(args.seed)\n\n # --- Initialize model ---\n rng, _rng = jax.random.split(rng)\n rngs = nnx.Rngs(_rng)\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n tokenizer_ffn_dim=args.tokenizer_ffn_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n lam_ffn_dim=args.lam_ffn_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n # Dynamics\n dyna_type=args.dyna_type,\n dyna_dim=args.dyna_dim,\n dyna_ffn_dim=args.dyna_ffn_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n use_flash_attention=args.use_flash_attention,\n decode=False,\n rngs=rngs,\n )\n\n _, params, _ = nnx.split(genie, nnx.Param, ...)\n param_counts = count_parameters_by_component(params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(\n args.lr_schedule,\n args.init_lr,\n args.max_lr,\n args.decay_end,\n args.num_steps,\n args.warmup_steps,\n args.wsd_decay_steps,\n )\n tx = optax.adamw(\n learning_rate=lr_schedule,\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n mu_dtype=args.dtype,\n )\n optimizer = nnx.Optimizer(genie, tx)\n del genie\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n\n model_state = nnx.state(optimizer.model)\n model_sharded_state = jax.lax.with_sharding_constraint(\n model_state, replicated_sharding\n )\n nnx.update(optimizer.model, model_sharded_state)\n optimizer_state = nnx.state(optimizer, nnx.optimizer.OptState)\n optimizer_sharded_state = jax.lax.with_sharding_constraint(\n optimizer_state, replicated_sharding\n )\n nnx.update(optimizer, optimizer_sharded_state)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeSave, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.PyTreeRestore, ocp.handlers.PyTreeCheckpointHandler\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""train_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n if args.val_data_dir:\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointSave,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n handler_registry.add(\n ""val_dataloader_state"",\n grain.checkpoint.CheckpointRestore,\n cast(ocp.handlers.CheckpointHandler, grain.checkpoint.CheckpointHandler),\n )\n\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n image_shape = (args.image_height, args.image_width, args.image_channels)\n train_iterator = create_dataloader_iterator(args.data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n if args.val_data_dir:\n val_iterator = create_dataloader_iterator(args.val_data_dir, image_shape, args.seq_len, args.batch_size, args.seed)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_optimizer = nnx.eval_shape(lambda: optimizer)\n abstract_optimizer_state = nnx.state(abstract_optimizer)\n if args.val_data_dir:\n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n val_dataloader_state=grain.checkpoint.CheckpointRestore(val_iterator), # type: ignore\n )\n else: \n restore_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeRestore(abstract_optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointRestore(train_iterator), # type: ignore\n )\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=restore_args\n )\n restored_optimizer_state = restored[""model_state""]\n nnx.update(optimizer, restored_optimizer_state)\n train_iterator = restored[""train_dataloader_state""]\n if args.val_data_dir:\n val_iterator = restored[""val_dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n optimizer = restore_genie_components(optimizer, replicated_sharding, rng, args)\n # NOTE: We have to remove the (unused) tokenizer vq dropout due flax.nnx lazily initializing modules.\n # Specifically, the first dynamics model checkpoint will contain the vq dropout module,\n # but the first full restore will fail due to nnx not initializing the module when\n # dropout is set to 0.0.\n del optimizer.model.tokenizer.vq.drop\n\n # --- TRAIN LOOP ---\n dataloader_train = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in train_iterator\n )\n if args.val_data_dir:\n dataloader_val = (\n jax.make_array_from_process_local_data(videos_sharding, elem)\n for elem in val_iterator\n )\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader_train:\n # --- Train step ---\n rng, _rng_mask = jax.random.split(rng, 2)\n inputs = dict(videos=videos, mask_rng=_rng_mask)\n loss, recon, metrics = train_step(optimizer.model, optimizer, inputs)\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Validation loss ---\n if args.val_data_dir and step % args.val_interval == 0:\n rng, _rng_mask_val = jax.random.split(rng, 2)\n print(f""Calculating validation metrics..."")\n val_loss, val_metrics, val_gt_batch, val_recon = calculate_validation_metrics(dataloader_val, optimizer.model, _rng_mask_val)\n print(f""Step {step}, validation loss: {val_loss}"")\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n log_dict = {\n ""loss"": loss,\n ""step"": step,\n **metrics\n }\n if args.val_data_dir and step % args.val_interval == 0:\n log_dict.update({\n ""val_loss"": val_loss,\n **val_metrics\n })\n wandb.log(log_dict)\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if args.val_data_dir and step % args.val_interval == 0:\n gt_seq_val = val_gt_batch[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq_val = val_recon[0].clip(0, 1)\n val_comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n val_comparison_seq = einops.rearrange(\n val_comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n if args.val_data_dir and step % args.val_interval == 0:\n log_images.update(\n dict(\n val_image=wandb.Image(np.asarray(gt_seq_val[0])),\n val_recon=wandb.Image(np.asarray(recon_seq_val[0])),\n val_true_vs_recon=wandb.Image(\n np.asarray(val_comparison_seq.astype(np.uint8))\n )\n )\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n optimizer_state = nnx.state(optimizer)\n if args.val_data_dir:\n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n ),\n val_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n val_iterator # type: ignore\n )\n )\n else: \n ckpt_manager_args = ocp.args.Composite(\n model_state=ocp.args.PyTreeSave(optimizer_state), # type: ignore\n train_dataloader_state=grain.checkpoint.CheckpointSave( # type: ignore\n train_iterator # type: ignore\n )\n )\n checkpoint_manager.save(\n step,\n args=ckpt_manager_args\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +1299,1737148,"train_dynamics.py",3555,0,"",python,selection_mouse +1300,1744873,"train_dynamics.py",3311,0,"",python,selection_mouse +1301,1749765,"train_lamap.py",0,0,"",python,tab +1302,1758117,"train_lamap.py",3695,0,"",python,selection_mouse +1303,1758158,"train_lamap.py",3694,0,"",python,selection_command +1304,1760079,"train_lamap.py",2407,0,"",python,selection_mouse +1305,1761407,"train_lamap.py",2473,0,"\n ",python,content +1306,1762362,"train_lamap.py",2478,0,"#",python,content +1307,1762363,"train_lamap.py",2479,0,"",python,selection_keyboard +1308,1762631,"train_lamap.py",2479,0," ",python,content +1309,1762631,"train_lamap.py",2480,0,"",python,selection_keyboard +1310,1762786,"train_lamap.py",2480,0,"T",python,content +1311,1762788,"train_lamap.py",2481,0,"",python,selection_keyboard +1312,1763047,"train_lamap.py",2481,0,"O",python,content +1313,1763048,"train_lamap.py",2482,0,"",python,selection_keyboard +1314,1763414,"train_lamap.py",2482,0,"D",python,content +1315,1763415,"train_lamap.py",2483,0,"",python,selection_keyboard +1316,1763417,"train_lamap.py",2483,0,"O",python,content +1317,1763417,"train_lamap.py",2484,0,"",python,selection_keyboard +1318,1763571,"train_lamap.py",2484,0," ",python,content +1319,1763572,"train_lamap.py",2485,0,"",python,selection_keyboard +1320,1763573,"train_lamap.py",2485,0,"m",python,content +1321,1763573,"train_lamap.py",2486,0,"",python,selection_keyboard +1322,1763631,"train_lamap.py",2486,0,"i",python,content +1323,1763631,"train_lamap.py",2487,0,"",python,selection_keyboard +1324,1763674,"train_lamap.py",2487,0,"h",python,content +1325,1763675,"train_lamap.py",2488,0,"",python,selection_keyboard +1326,1763740,"train_lamap.py",2488,0,"i",python,content +1327,1763741,"train_lamap.py",2489,0,"",python,selection_keyboard +1328,1763884,"train_lamap.py",2489,0,"r",python,content +1329,1763885,"train_lamap.py",2490,0,"",python,selection_keyboard +1330,1764027,"train_lamap.py",2490,0," ",python,content +1331,1764028,"train_lamap.py",2491,0,"",python,selection_keyboard +1332,1764110,"train_lamap.py",2491,0,"a",python,content +1333,1764111,"train_lamap.py",2492,0,"",python,selection_keyboard +1334,1764210,"train_lamap.py",2492,0,"l",python,content +1335,1764212,"train_lamap.py",2493,0,"",python,selection_keyboard +1336,1764327,"train_lamap.py",2493,0,"s",python,content +1337,1764329,"train_lamap.py",2494,0,"",python,selection_keyboard +1338,1764378,"train_lamap.py",2494,0,"o",python,content +1339,1764379,"train_lamap.py",2495,0,"",python,selection_keyboard +1340,1764478,"train_lamap.py",2495,0," ",python,content +1341,1764480,"train_lamap.py",2496,0,"",python,selection_keyboard +1342,1764625,"train_lamap.py",2496,0,"c",python,content +1343,1764626,"train_lamap.py",2497,0,"",python,selection_keyboard +1344,1764850,"train_lamap.py",2497,0,"a",python,content +1345,1764851,"train_lamap.py",2498,0,"",python,selection_keyboard +1346,1764944,"train_lamap.py",2498,0,"l",python,content +1347,1764945,"train_lamap.py",2499,0,"",python,selection_keyboard +1348,1765050,"train_lamap.py",2499,0,"c",python,content +1349,1765052,"train_lamap.py",2500,0,"",python,selection_keyboard +1350,1765143,"train_lamap.py",2500,0,"u",python,content +1351,1765144,"train_lamap.py",2501,0,"",python,selection_keyboard +1352,1765266,"train_lamap.py",2501,0,"l",python,content +1353,1765268,"train_lamap.py",2502,0,"",python,selection_keyboard +1354,1765417,"train_lamap.py",2502,0,"a",python,content +1355,1765419,"train_lamap.py",2503,0,"",python,selection_keyboard +1356,1765640,"train_lamap.py",2503,0,"t",python,content +1357,1765641,"train_lamap.py",2504,0,"",python,selection_keyboard +1358,1766209,"train_lamap.py",2504,0,"e",python,content +1359,1766211,"train_lamap.py",2505,0,"",python,selection_keyboard +1360,1766294,"train_lamap.py",2505,0," ",python,content +1361,1766296,"train_lamap.py",2506,0,"",python,selection_keyboard +1362,1766433,"train_lamap.py",2506,0,"a",python,content +1363,1766434,"train_lamap.py",2507,0,"",python,selection_keyboard +1364,1766627,"train_lamap.py",2507,0,"c",python,content +1365,1766628,"train_lamap.py",2508,0,"",python,selection_keyboard +1366,1766762,"train_lamap.py",2508,0,"c",python,content +1367,1766763,"train_lamap.py",2509,0,"",python,selection_keyboard +1368,1766837,"train_lamap.py",2509,0,"t",python,content +1369,1766838,"train_lamap.py",2510,0,"",python,selection_keyboard +1370,1766961,"train_lamap.py",2510,0,"u",python,content +1371,1766962,"train_lamap.py",2511,0,"",python,selection_keyboard +1372,1767184,"train_lamap.py",2511,0,"r",python,content +1373,1767184,"train_lamap.py",2512,0,"",python,selection_keyboard +1374,1767705,"train_lamap.py",2511,1,"",python,content +1375,1767741,"train_lamap.py",2510,1,"",python,content +1376,1767855,"train_lamap.py",2509,1,"",python,content +1377,1768424,"train_lamap.py",2509,0,"u",python,content +1378,1768426,"train_lamap.py",2510,0,"",python,selection_keyboard +1379,1769335,"train_lamap.py",2510,0,"u",python,content +1380,1769336,"train_lamap.py",2511,0,"",python,selection_keyboard +1381,1769681,"train_lamap.py",2510,1,"",python,content +1382,1769796,"train_lamap.py",2510,0,"r",python,content +1383,1769797,"train_lamap.py",2511,0,"",python,selection_keyboard +1384,1770011,"train_lamap.py",2511,0,"a",python,content +1385,1770011,"train_lamap.py",2512,0,"",python,selection_keyboard +1386,1770173,"train_lamap.py",2512,0,"c",python,content +1387,1770175,"train_lamap.py",2513,0,"",python,selection_keyboard +1388,1770556,"train_lamap.py",2513,0,"y",python,content +1389,1770558,"train_lamap.py",2514,0,"",python,selection_keyboard +1390,1770674,"train_lamap.py",2514,0,".",python,content +1391,1770675,"train_lamap.py",2515,0,"",python,selection_keyboard +1392,1771155,"train_lamap.py",2514,0,"",python,selection_command +1393,1771664,"train_lamap.py",2530,0,"",python,selection_command +1394,1771892,"train_lamap.py",2532,0,"",python,selection_command +1395,1772146,"train_lamap.py",2533,0,"",python,selection_command +1396,1772322,"train_lamap.py",2574,0,"",python,selection_command +1397,1772477,"train_lamap.py",2601,0,"",python,selection_command +1398,1774945,"train_lamap.py",2574,0,"",python,selection_command +1399,1776861,"train_lamap.py",2534,41,"def lamap_loss_fn(params, state, inputs):",python,selection_command +1400,1777088,"train_lamap.py",2534,68,"def lamap_loss_fn(params, state, inputs):\n # --- Compute loss ---",python,selection_command +1401,1777197,"train_lamap.py",2534,135,"def lamap_loss_fn(params, state, inputs):\n # --- Compute loss ---\n inputs[""videos""] = inputs[""videos""].astype(args.dtype) / 255.0",python,selection_command +1402,1777347,"train_lamap.py",2534,165,"def lamap_loss_fn(params, state, inputs):\n # --- Compute loss ---\n inputs[""videos""] = inputs[""videos""].astype(args.dtype) / 255.0\n outputs = state.apply_fn(",python,selection_command +1403,1777508,"train_lamap.py",2534,236,"def lamap_loss_fn(params, state, inputs):\n # --- Compute loss ---\n inputs[""videos""] = inputs[""videos""].astype(args.dtype) / 255.0\n outputs = state.apply_fn(\n params, inputs, training=True, rngs={""dropout"": inputs[""rng""]}",python,selection_command +1404,1777756,"train_lamap.py",2534,242,"def lamap_loss_fn(params, state, inputs):\n # --- Compute loss ---\n inputs[""videos""] = inputs[""videos""].astype(args.dtype) / 255.0\n outputs = state.apply_fn(\n params, inputs, training=True, rngs={""dropout"": inputs[""rng""]}\n )",python,selection_command +1405,1777841,"train_lamap.py",2534,338,"def lamap_loss_fn(params, state, inputs):\n # --- Compute loss ---\n inputs[""videos""] = inputs[""videos""].astype(args.dtype) / 255.0\n outputs = state.apply_fn(\n params, inputs, training=True, rngs={""dropout"": inputs[""rng""]}\n )\n loss = optax.softmax_cross_entropy(outputs[""action_predictions""], inputs[""actions""]).mean()",python,selection_command +1406,1777955,"train_lamap.py",2534,354,"def lamap_loss_fn(params, state, inputs):\n # --- Compute loss ---\n inputs[""videos""] = inputs[""videos""].astype(args.dtype) / 255.0\n outputs = state.apply_fn(\n params, inputs, training=True, rngs={""dropout"": inputs[""rng""]}\n )\n loss = optax.softmax_cross_entropy(outputs[""action_predictions""], inputs[""actions""]).mean()\n return loss",python,selection_command +1407,1778302,"train_lamap.py",2534,355,"",python,content +1408,1778641,"train_lamap.py",2533,0,"",python,selection_command +1409,1779122,"train_lamap.py",2533,1,"",python,content +1410,1779683,"train_lamap.py",2533,1,"",python,content +1411,1785176,"train_dynamics.py",0,0,"",python,tab +1412,1787945,"train_dynamics.py",3272,0,"",python,selection_mouse +1413,1791540,"train_dynamics.py",3230,0,"",python,selection_command +1414,1791973,"train_dynamics.py",3272,0,"",python,selection_command +1415,1792108,"train_dynamics.py",3230,0,"",python,selection_command +1416,1792306,"train_dynamics.py",3158,0,"",python,selection_command +1417,1793444,"train_dynamics.py",3230,0,"",python,selection_command +1418,1797273,"train_dynamics.py",3158,0,"",python,selection_command +1419,1797534,"train_dynamics.py",3108,0,"",python,selection_command +1420,1797770,"train_dynamics.py",3080,0,"",python,selection_command +1421,1799247,"models/lamap.py",0,0,"",python,tab +1422,1801338,"train_lamap.py",0,0,"",python,tab +1423,1804450,"train_dynamics.py",0,0,"",python,tab +1424,1812043,"train_lamap.py",0,0,"",python,tab +1425,1813712,"train_lamap.py",2404,0,"",python,selection_mouse +1426,1815642,".venv/lib/python3.10/site-packages/optax/__init__.py",0,0,"# Copyright 2019 DeepMind Technologies Limited. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the ""License"");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an ""AS IS"" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n""""""Optax: composable gradient processing and optimization, in JAX.""""""\n\n# pylint: disable=wrong-import-position\n# pylint: disable=g-importing-member\n\nimport typing as _typing\n\nfrom optax import assignment\nfrom optax import contrib\nfrom optax import losses\nfrom optax import monte_carlo\nfrom optax import perturbations\nfrom optax import projections\nfrom optax import schedules\nfrom optax import second_order\nfrom optax import transforms\nfrom optax import tree\nfrom optax import tree_utils\nfrom optax._src.alias import adabelief\nfrom optax._src.alias import adadelta\nfrom optax._src.alias import adafactor\nfrom optax._src.alias import adagrad\nfrom optax._src.alias import adam\nfrom optax._src.alias import adamax\nfrom optax._src.alias import adamaxw\nfrom optax._src.alias import adamw\nfrom optax._src.alias import adan\nfrom optax._src.alias import amsgrad\nfrom optax._src.alias import fromage\nfrom optax._src.alias import lamb\nfrom optax._src.alias import lars\nfrom optax._src.alias import lbfgs\nfrom optax._src.alias import lion\nfrom optax._src.alias import MaskOrFn\nfrom optax._src.alias import nadam\nfrom optax._src.alias import nadamw\nfrom optax._src.alias import noisy_sgd\nfrom optax._src.alias import novograd\nfrom optax._src.alias import optimistic_adam\nfrom optax._src.alias import optimistic_gradient_descent\nfrom optax._src.alias import polyak_sgd\nfrom optax._src.alias import radam\nfrom optax._src.alias import rmsprop\nfrom optax._src.alias import rprop\nfrom optax._src.alias import sgd\nfrom optax._src.alias import sign_sgd\nfrom optax._src.alias import sm3\nfrom optax._src.alias import yogi\nfrom optax._src.base import EmptyState\nfrom optax._src.base import GradientTransformation\nfrom optax._src.base import GradientTransformationExtraArgs\nfrom optax._src.base import identity\nfrom optax._src.base import OptState\nfrom optax._src.base import Params\nfrom optax._src.base import ScalarOrSchedule\nfrom optax._src.base import Schedule\nfrom optax._src.base import set_to_zero\nfrom optax._src.base import stateless\nfrom optax._src.base import stateless_with_tree_map\nfrom optax._src.base import TransformInitFn\nfrom optax._src.base import TransformUpdateExtraArgsFn\nfrom optax._src.base import TransformUpdateFn\nfrom optax._src.base import Updates\nfrom optax._src.base import with_extra_args_support\nfrom optax._src.factorized import FactoredState\nfrom optax._src.factorized import scale_by_factored_rms\nfrom optax._src.linear_algebra import global_norm\nfrom optax._src.linear_algebra import matrix_inverse_pth_root\nfrom optax._src.linear_algebra import nnls\nfrom optax._src.linear_algebra import power_iteration\nfrom optax._src.linesearch import scale_by_backtracking_linesearch\nfrom optax._src.linesearch import scale_by_zoom_linesearch\nfrom optax._src.linesearch import ScaleByBacktrackingLinesearchState\nfrom optax._src.linesearch import ScaleByZoomLinesearchState\nfrom optax._src.linesearch import ZoomLinesearchInfo\nfrom optax._src.lookahead import lookahead\nfrom optax._src.lookahead import LookaheadParams\nfrom optax._src.lookahead import LookaheadState\nfrom optax._src.numerics import safe_increment\nfrom optax._src.numerics import safe_int32_increment\nfrom optax._src.numerics import safe_norm\nfrom optax._src.numerics import safe_root_mean_squares\nfrom optax._src.transform import apply_every\nfrom optax._src.transform import ApplyEvery\nfrom optax._src.transform import centralize\nfrom optax._src.transform import normalize_by_update_norm\nfrom optax._src.transform import scale\nfrom optax._src.transform import scale_by_adadelta\nfrom optax._src.transform import scale_by_adam\nfrom optax._src.transform import scale_by_adamax\nfrom optax._src.transform import scale_by_adan\nfrom optax._src.transform import scale_by_amsgrad\nfrom optax._src.transform import scale_by_belief\nfrom optax._src.transform import scale_by_distance_over_gradients\nfrom optax._src.transform import scale_by_lbfgs\nfrom optax._src.transform import scale_by_learning_rate\nfrom optax._src.transform import scale_by_lion\nfrom optax._src.transform import scale_by_novograd\nfrom optax._src.transform import scale_by_optimistic_gradient\nfrom optax._src.transform import scale_by_param_block_norm\nfrom optax._src.transform import scale_by_param_block_rms\nfrom optax._src.transform import scale_by_polyak\nfrom optax._src.transform import scale_by_radam\nfrom optax._src.transform import scale_by_rms\nfrom optax._src.transform import scale_by_rprop\nfrom optax._src.transform import scale_by_rss\nfrom optax._src.transform import scale_by_schedule\nfrom optax._src.transform import scale_by_sign\nfrom optax._src.transform import scale_by_sm3\nfrom optax._src.transform import scale_by_stddev\nfrom optax._src.transform import scale_by_trust_ratio\nfrom optax._src.transform import scale_by_yogi\nfrom optax._src.transform import ScaleByAdaDeltaState\nfrom optax._src.transform import ScaleByAdamState\nfrom optax._src.transform import ScaleByAdanState\nfrom optax._src.transform import ScaleByAmsgradState\nfrom optax._src.transform import ScaleByBeliefState\nfrom optax._src.transform import ScaleByLBFGSState\nfrom optax._src.transform import ScaleByLionState\nfrom optax._src.transform import ScaleByNovogradState\nfrom optax._src.transform import ScaleByRmsState\nfrom optax._src.transform import ScaleByRpropState\nfrom optax._src.transform import ScaleByRssState\nfrom optax._src.transform import ScaleByRStdDevState\nfrom optax._src.transform import ScaleByScheduleState\nfrom optax._src.transform import ScaleBySM3State\nfrom optax._src.update import apply_updates\nfrom optax._src.update import incremental_update\nfrom optax._src.update import periodic_update\nfrom optax._src.utils import multi_normal\nfrom optax._src.utils import scale_gradient\nfrom optax._src.utils import value_and_grad_from_state\n\n# TODO(mtthss): remove contrib aliases from flat namespace once users updated.\n# Deprecated modules\nfrom optax.contrib import differentially_private_aggregate as _deprecated_differentially_private_aggregate\nfrom optax.contrib import DifferentiallyPrivateAggregateState as _deprecated_DifferentiallyPrivateAggregateState\nfrom optax.contrib import dpsgd as _deprecated_dpsgd\n\n\n# TODO(mtthss): remove aliases after updates.\nadaptive_grad_clip = transforms.adaptive_grad_clip\nAdaptiveGradClipState = EmptyState\nclip = transforms.clip\nclip_by_block_rms = transforms.clip_by_block_rms\nclip_by_global_norm = transforms.clip_by_global_norm\nClipByGlobalNormState = EmptyState\nClipState = EmptyState\nper_example_global_norm_clip = transforms.per_example_global_norm_clip\nper_example_layer_norm_clip = transforms.per_example_layer_norm_clip\nkeep_params_nonnegative = transforms.keep_params_nonnegative\nNonNegativeParamsState = transforms.NonNegativeParamsState\nzero_nans = transforms.zero_nans\nZeroNansState = transforms.ZeroNansState\nchain = transforms.chain\npartition = transforms.partition\nPartitionState = transforms.PartitionState\nmulti_transform = transforms.partition # for backwards compatibility\nMultiTransformState = transforms.PartitionState # for backwards compatibility\nnamed_chain = transforms.named_chain\ntrace = transforms.trace\nTraceState = transforms.TraceState\nema = transforms.ema\nEmaState = transforms.EmaState\nadd_noise = transforms.add_noise\nAddNoiseState = transforms.AddNoiseState\nadd_decayed_weights = transforms.add_decayed_weights\nAddDecayedWeightsState = EmptyState\nScaleByTrustRatioState = EmptyState\nScaleState = EmptyState\napply_if_finite = transforms.apply_if_finite\nApplyIfFiniteState = transforms.ApplyIfFiniteState\nconditionally_mask = transforms.conditionally_mask\nconditionally_transform = transforms.conditionally_transform\nConditionallyMaskState = transforms.ConditionallyMaskState\nConditionallyTransformState = transforms.ConditionallyTransformState\nflatten = transforms.flatten\nmasked = transforms.masked\nMaskedNode = transforms.MaskedNode\nMaskedState = transforms.MaskedState\nMultiSteps = transforms.MultiSteps\nMultiStepsState = transforms.MultiStepsState\nShouldSkipUpdateFunction = transforms.ShouldSkipUpdateFunction\nskip_large_updates = transforms.skip_large_updates\nskip_not_finite = transforms.skip_not_finite\n\n# TODO(mtthss): remove tree_utils aliases after updates.\ntree_map_params = tree_utils.tree_map_params\nbias_correction = tree_utils.tree_bias_correction\nupdate_infinity_moment = tree_utils.tree_update_infinity_moment\nupdate_moment = tree_utils.tree_update_moment\nupdate_moment_per_elem_norm = tree_utils.tree_update_moment_per_elem_norm\n\n# TODO(mtthss): remove schedules aliases from flat namespaces after user updates\nconstant_schedule = schedules.constant_schedule\ncosine_decay_schedule = schedules.cosine_decay_schedule\ncosine_onecycle_schedule = schedules.cosine_onecycle_schedule\nexponential_decay = schedules.exponential_decay\ninject_hyperparams = schedules.inject_hyperparams\nInjectHyperparamsState = schedules.InjectHyperparamsState\njoin_schedules = schedules.join_schedules\nlinear_onecycle_schedule = schedules.linear_onecycle_schedule\nlinear_schedule = schedules.linear_schedule\npiecewise_constant_schedule = schedules.piecewise_constant_schedule\npiecewise_interpolate_schedule = schedules.piecewise_interpolate_schedule\npolynomial_schedule = schedules.polynomial_schedule\nsgdr_schedule = schedules.sgdr_schedule\nwarmup_constant_schedule = schedules.warmup_constant_schedule\nwarmup_cosine_decay_schedule = schedules.warmup_cosine_decay_schedule\nwarmup_exponential_decay_schedule = schedules.warmup_exponential_decay_schedule\ninject_stateful_hyperparams = schedules.inject_stateful_hyperparams\nInjectStatefulHyperparamsState = schedules.InjectStatefulHyperparamsState\nWrappedSchedule = schedules.WrappedSchedule\n\n# TODO(mtthss): remove loss aliases from flat namespace once users have updated.\nconvex_kl_divergence = losses.convex_kl_divergence\ncosine_distance = losses.cosine_distance\ncosine_similarity = losses.cosine_similarity\nctc_loss = losses.ctc_loss\nctc_loss_with_forward_probs = losses.ctc_loss_with_forward_probs\nhinge_loss = losses.hinge_loss\nhuber_loss = losses.huber_loss\nkl_divergence = losses.kl_divergence\nl2_loss = losses.l2_loss\nlog_cosh = losses.log_cosh\nntxent = losses.ntxent\nsigmoid_binary_cross_entropy = losses.sigmoid_binary_cross_entropy\nsmooth_labels = losses.smooth_labels\nsafe_softmax_cross_entropy = losses.safe_softmax_cross_entropy\nsoftmax_cross_entropy = losses.softmax_cross_entropy\nsoftmax_cross_entropy_with_integer_labels = (\n losses.softmax_cross_entropy_with_integer_labels\n)\nsquared_error = losses.squared_error\nsigmoid_focal_loss = losses.sigmoid_focal_loss\n\n_deprecations = {\n # Added Apr 2024\n ""differentially_private_aggregate"": (\n (\n ""optax.differentially_private_aggregate is deprecated: use""\n "" optax.contrib.differentially_private_aggregate (optax v0.1.8 or""\n "" newer).""\n ),\n _deprecated_differentially_private_aggregate,\n ),\n ""DifferentiallyPrivateAggregateState"": (\n (\n ""optax.DifferentiallyPrivateAggregateState is deprecated: use""\n "" optax.contrib.DifferentiallyPrivateAggregateState (optax v0.1.8""\n "" or newer).""\n ),\n _deprecated_DifferentiallyPrivateAggregateState,\n ),\n ""dpsgd"": (\n (\n ""optax.dpsgd is deprecated: use optax.contrib.dpsgd (optax v0.1.8""\n "" or newer).""\n ),\n _deprecated_dpsgd,\n ),\n}\n# pylint: disable=g-import-not-at-top\n# pylint: disable=g-bad-import-order\nif _typing.TYPE_CHECKING:\n # pylint: disable=reimported\n from optax.contrib import differentially_private_aggregate\n from optax.contrib import DifferentiallyPrivateAggregateState\n from optax.contrib import dpsgd\n # pylint: enable=reimported\n\nelse:\n from optax._src.deprecations import deprecation_getattr as _deprecation_getattr\n\n __getattr__ = _deprecation_getattr(__name__, _deprecations)\n del _deprecation_getattr\ndel _typing\n# pylint: enable=g-bad-import-order\n# pylint: enable=g-import-not-at-top\n# pylint: enable=g-importing-member\n\n\n__version__ = ""0.2.5""\n\n__all__ = (\n ""adabelief"",\n ""adadelta"",\n ""adafactor"",\n ""adagrad"",\n ""adam"",\n ""adamax"",\n ""adamaxw"",\n ""adamw"",\n ""adan"",\n ""adaptive_grad_clip"",\n ""AdaptiveGradClipState"",\n ""add_decayed_weights"",\n ""add_noise"",\n ""AddDecayedWeightsState"",\n ""AddNoiseState"",\n ""amsgrad"",\n ""apply_every"",\n ""apply_if_finite"",\n ""apply_updates"",\n ""ApplyEvery"",\n ""ApplyIfFiniteState"",\n ""assignment"",\n ""centralize"",\n ""chain"",\n ""clip_by_block_rms"",\n ""clip_by_global_norm"",\n ""clip"",\n ""ClipByGlobalNormState"",\n ""ClipState"",\n ""conditionally_mask"",\n ""ConditionallyMaskState"",\n ""conditionally_transform"",\n ""ConditionallyTransformState"",\n ""constant_schedule"",\n ""ctc_loss"",\n ""ctc_loss_with_forward_probs"",\n ""convex_kl_divergence"",\n ""cosine_decay_schedule"",\n ""cosine_distance"",\n ""cosine_onecycle_schedule"",\n ""cosine_similarity"",\n ""differentially_private_aggregate"",\n ""DifferentiallyPrivateAggregateState"",\n ""dpsgd"",\n ""ema"",\n ""EmaState"",\n ""EmptyState"",\n ""exponential_decay"",\n ""FactoredState"",\n ""flatten"",\n ""fromage"",\n ""global_norm"",\n ""GradientTransformation"",\n ""GradientTransformationExtraArgs"",\n ""hinge_loss"",\n ""huber_loss"",\n ""identity"",\n ""incremental_update"",\n ""inject_hyperparams"",\n ""InjectHyperparamsState"",\n ""join_schedules"",\n ""keep_params_nonnegative"",\n ""kl_divergence"",\n ""l2_loss"",\n ""lamb"",\n ""lars"",\n ""lbfgs"",\n ""lion"",\n ""linear_onecycle_schedule"",\n ""linear_schedule"",\n ""log_cosh"",\n ""lookahead"",\n ""LookaheadParams"",\n ""LookaheadState"",\n ""masked"",\n ""MaskOrFn"",\n ""MaskedState"",\n ""matrix_inverse_pth_root"",\n ""multi_normal"",\n ""multi_transform"", # for backwards compatibility\n ""MultiSteps"",\n ""MultiStepsState"",\n ""MultiTransformState"", # for backwards compatibility\n ""nadam"",\n ""nadamw"",\n ""nnls"",\n ""noisy_sgd"",\n ""novograd"",\n ""NonNegativeParamsState"",\n ""ntxent"",\n ""OptState"",\n ""Params"",\n ""partition"",\n ""PartitionState"",\n ""periodic_update"",\n ""per_example_global_norm_clip"",\n ""per_example_layer_norm_clip"",\n ""piecewise_constant_schedule"",\n ""piecewise_interpolate_schedule"",\n ""polynomial_schedule"",\n ""power_iteration"",\n ""polyak_sgd"",\n ""radam"",\n ""rmsprop"",\n ""rprop"",\n ""safe_increment"",\n ""safe_int32_increment"",\n ""safe_norm"",\n ""safe_root_mean_squares"",\n ""ScalarOrSchedule"",\n ""scale_by_adadelta"",\n ""scale_by_adam"",\n ""scale_by_adamax"",\n ""scale_by_adan"",\n ""scale_by_amsgrad"",\n ""scale_by_backtracking_linesearch"",\n ""scale_by_belief"",\n ""scale_by_lbfgs"",\n ""scale_by_lion"",\n ""scale_by_factored_rms"",\n ""scale_by_novograd"",\n ""scale_by_param_block_norm"",\n ""scale_by_param_block_rms"",\n ""scale_by_polyak"",\n ""scale_by_radam"",\n ""scale_by_rms"",\n ""scale_by_rprop"",\n ""scale_by_rss"",\n ""scale_by_schedule"",\n ""scale_by_sign"",\n ""scale_by_sm3"",\n ""scale_by_stddev"",\n ""scale_by_trust_ratio"",\n ""scale_by_yogi"",\n ""scale_by_zoom_linesearch"",\n ""scale_gradient"",\n ""scale"",\n ""ScaleByAdaDeltaState"",\n ""ScaleByAdamState"",\n ""ScaleByAdanState"",\n ""ScaleByAmsgradState"",\n ""ScaleByBacktrackingLinesearchState"",\n ""ScaleByBeliefState"",\n ""ScaleByLBFGSState"",\n ""ScaleByLionState"",\n ""ScaleByNovogradState"",\n ""ScaleByRmsState"",\n ""ScaleByRpropState"",\n ""ScaleByRssState"",\n ""ScaleByRStdDevState"",\n ""ScaleByScheduleState"",\n ""ScaleBySM3State"",\n ""ScaleByTrustRatioState"",\n ""ScaleByZoomLinesearchState"",\n ""ScaleState"",\n ""Schedule"",\n ""set_to_zero"",\n ""sgd"",\n ""sgdr_schedule"",\n ""ShouldSkipUpdateFunction"",\n ""sigmoid_binary_cross_entropy"",\n ""sign_sgd"",\n ""skip_large_updates"",\n ""skip_not_finite"",\n ""sm3"",\n ""smooth_labels"",\n ""softmax_cross_entropy"",\n ""softmax_cross_entropy_with_integer_labels"",\n ""stateless"",\n ""stateless_with_tree_map"",\n ""trace"",\n ""TraceState"",\n ""TransformInitFn"",\n ""TransformUpdateFn"",\n ""TransformUpdateExtraArgsFn"",\n ""Updates"",\n ""value_and_grad_from_state"",\n ""warmup_cosine_decay_schedule"",\n ""warmup_exponential_decay_schedule"",\n ""yogi"",\n ""zero_nans"",\n ""ZeroNansState"",\n ""ZoomLinesearchInfo"",\n)\n\n# _________________________________________\n# / Please don't use symbols in `_src` they \\n# \ are not part of the Optax public API. /\n# -----------------------------------------\n# \ ^__^\n# \ (oo)\_______\n# (__)\ )\/\\n# ||----w |\n# || ||\n#\n",python,tab +1427,1815643,".venv/lib/python3.10/site-packages/optax/__init__.py",10982,0,"",python,selection_command +1428,1818301,"train_lamap.py",0,0,"",python,tab +1429,1819370,"train_lamap.py",2364,0,"",python,selection_mouse +1430,1820645,"train_lamap.py",2377,0,"\n ",python,content +1431,1821075,"train_lamap.py",2382,0,"#",python,content +1432,1821076,"train_lamap.py",2383,0,"",python,selection_keyboard +1433,1821302,"train_lamap.py",2383,0," ",python,content +1434,1821303,"train_lamap.py",2384,0,"",python,selection_keyboard +1435,1821564,"train_lamap.py",2384,0,"T",python,content +1436,1821565,"train_lamap.py",2385,0,"",python,selection_keyboard +1437,1821777,"train_lamap.py",2385,0,"O",python,content +1438,1821777,"train_lamap.py",2386,0,"",python,selection_keyboard +1439,1821961,"train_lamap.py",2386,0,"D",python,content +1440,1821962,"train_lamap.py",2387,0,"",python,selection_keyboard +1441,1822006,"train_lamap.py",2387,0,"O",python,content +1442,1822007,"train_lamap.py",2388,0,"",python,selection_keyboard +1443,1822090,"train_lamap.py",2388,0," ",python,content +1444,1822092,"train_lamap.py",2389,0,"",python,selection_keyboard +1445,1822159,"train_lamap.py",2389,0,"m",python,content +1446,1822160,"train_lamap.py",2390,0,"",python,selection_keyboard +1447,1822333,"train_lamap.py",2390,0,"i",python,content +1448,1822334,"train_lamap.py",2391,0,"",python,selection_keyboard +1449,1822446,"train_lamap.py",2391,0,"h",python,content +1450,1822447,"train_lamap.py",2392,0,"",python,selection_keyboard +1451,1822557,"train_lamap.py",2392,0,"i",python,content +1452,1822558,"train_lamap.py",2393,0,"",python,selection_keyboard +1453,1822649,"train_lamap.py",2393,0,"r",python,content +1454,1822651,"train_lamap.py",2394,0,"",python,selection_keyboard +1455,1822753,"train_lamap.py",2394,0," ",python,content +1456,1822755,"train_lamap.py",2395,0,"",python,selection_keyboard +1457,1823368,"train_lamap.py",2395,0,"v",python,content +1458,1823369,"train_lamap.py",2396,0,"",python,selection_keyboard +1459,1823511,"train_lamap.py",2396,0,"e",python,content +1460,1823512,"train_lamap.py",2397,0,"",python,selection_keyboard +1461,1823604,"train_lamap.py",2397,0,"r",python,content +1462,1823604,"train_lamap.py",2398,0,"",python,selection_keyboard +1463,1824021,"train_lamap.py",2398,0,"i",python,content +1464,1824024,"train_lamap.py",2399,0,"",python,selection_keyboard +1465,1824096,"train_lamap.py",2399,0,"f",python,content +1466,1824097,"train_lamap.py",2400,0,"",python,selection_keyboard +1467,1824276,"train_lamap.py",2400,0,"y",python,content +1468,1824277,"train_lamap.py",2401,0,"",python,selection_keyboard +1469,1824459,"train_lamap.py",2401,0," ",python,content +1470,1824461,"train_lamap.py",2402,0,"",python,selection_keyboard +1471,1824667,"train_lamap.py",2402,0,"i",python,content +1472,1824669,"train_lamap.py",2403,0,"",python,selection_keyboard +1473,1824890,"train_lamap.py",2403,0,"f",python,content +1474,1824891,"train_lamap.py",2404,0,"",python,selection_keyboard +1475,1825011,"train_lamap.py",2404,0," ",python,content +1476,1825012,"train_lamap.py",2405,0,"",python,selection_keyboard +1477,1825142,"train_lamap.py",2405,0,"t",python,content +1478,1825143,"train_lamap.py",2406,0,"",python,selection_keyboard +1479,1825224,"train_lamap.py",2406,0,"h",python,content +1480,1825225,"train_lamap.py",2407,0,"",python,selection_keyboard +1481,1825377,"train_lamap.py",2407,0,"i",python,content +1482,1825378,"train_lamap.py",2408,0,"",python,selection_keyboard +1483,1825479,"train_lamap.py",2408,0,"s",python,content +1484,1825480,"train_lamap.py",2409,0,"",python,selection_keyboard +1485,1825633,"train_lamap.py",2409,0," ",python,content +1486,1825634,"train_lamap.py",2410,0,"",python,selection_keyboard +1487,1825753,"train_lamap.py",2410,0,"i",python,content +1488,1825754,"train_lamap.py",2411,0,"",python,selection_keyboard +1489,1825834,"train_lamap.py",2411,0,"s",python,content +1490,1825835,"train_lamap.py",2412,0,"",python,selection_keyboard +1491,1825913,"train_lamap.py",2412,0," ",python,content +1492,1825915,"train_lamap.py",2413,0,"",python,selection_keyboard +1493,1826408,"train_lamap.py",2413,0,"t",python,content +1494,1826409,"train_lamap.py",2414,0,"",python,selection_keyboard +1495,1826498,"train_lamap.py",2414,0,"h",python,content +1496,1826499,"train_lamap.py",2415,0,"",python,selection_keyboard +1497,1826659,"train_lamap.py",2415,0,"e",python,content +1498,1826660,"train_lamap.py",2416,0,"",python,selection_keyboard +1499,1826678,"train_lamap.py",2416,0," ",python,content +1500,1826679,"train_lamap.py",2417,0,"",python,selection_keyboard +1501,1826813,"train_lamap.py",2417,0,"c",python,content +1502,1826814,"train_lamap.py",2418,0,"",python,selection_keyboard +1503,1826913,"train_lamap.py",2418,0,"o",python,content +1504,1826913,"train_lamap.py",2419,0,"",python,selection_keyboard +1505,1827044,"train_lamap.py",2419,0,"r",python,content +1506,1827045,"train_lamap.py",2420,0,"",python,selection_keyboard +1507,1827204,"train_lamap.py",2420,0,"r",python,content +1508,1827206,"train_lamap.py",2421,0,"",python,selection_keyboard +1509,1827347,"train_lamap.py",2421,0,"e",python,content +1510,1827349,"train_lamap.py",2422,0,"",python,selection_keyboard +1511,1827476,"train_lamap.py",2422,0,"c",python,content +1512,1827478,"train_lamap.py",2423,0,"",python,selection_keyboard +1513,1827677,"train_lamap.py",2423,0,"t",python,content +1514,1827678,"train_lamap.py",2424,0,"",python,selection_keyboard +1515,1827824,"train_lamap.py",2424,0,"l",python,content +1516,1827826,"train_lamap.py",2425,0,"",python,selection_keyboard +1517,1827927,"train_lamap.py",2425,0," ",python,content +1518,1827929,"train_lamap.py",2426,0,"",python,selection_keyboard +1519,1828048,"train_lamap.py",2426,0,"l",python,content +1520,1828049,"train_lamap.py",2427,0,"",python,selection_keyboard +1521,1828250,"train_lamap.py",2427,0,"o",python,content +1522,1828251,"train_lamap.py",2428,0,"",python,selection_keyboard +1523,1828413,"train_lamap.py",2427,1,"",python,content +1524,1828576,"train_lamap.py",2426,1,"",python,content +1525,1828684,"train_lamap.py",2425,1,"",python,content +1526,1828831,"train_lamap.py",2424,1,"",python,content +1527,1829313,"train_lamap.py",2424,0," ",python,content +1528,1829314,"train_lamap.py",2425,0,"",python,selection_keyboard +1529,1829426,"train_lamap.py",2425,0,"l",python,content +1530,1829427,"train_lamap.py",2426,0,"",python,selection_keyboard +1531,1829560,"train_lamap.py",2426,0,"o",python,content +1532,1829562,"train_lamap.py",2427,0,"",python,selection_keyboard +1533,1829681,"train_lamap.py",2427,0,"s",python,content +1534,1829681,"train_lamap.py",2428,0,"",python,selection_keyboard +1535,1829830,"train_lamap.py",2428,0,"s",python,content +1536,1829832,"train_lamap.py",2429,0,"",python,selection_keyboard +1537,1829923,"train_lamap.py",2429,0," ",python,content +1538,1829924,"train_lamap.py",2430,0,"",python,selection_keyboard +1539,1830078,"train_lamap.py",2430,0,"f",python,content +1540,1830080,"train_lamap.py",2431,0,"",python,selection_keyboard +1541,1830131,"train_lamap.py",2431,0,"n",python,content +1542,1830132,"train_lamap.py",2432,0,"",python,selection_keyboard +1543,1830818,"train_lamap.py",2431,0,"",python,selection_command +1544,1830954,"train_lamap.py",2486,0,"",python,selection_command +1545,1831297,"train_lamap.py",2569,0,"",python,selection_command +1546,1831651,"train_lamap.py",2585,0,"",python,selection_command +1547,1831953,"train_lamap.py",2569,0,"",python,selection_command +1548,1832016,"train_lamap.py",2486,0,"",python,selection_command +1549,1832097,"train_lamap.py",2431,0,"",python,selection_command +1550,1832478,"train_lamap.py",2376,0,"",python,selection_command +1551,1832764,"train_lamap.py",2375,0,"",python,selection_command +1552,1833382,"train_lamap.py",2374,0,"",python,selection_command +1553,1833382,"train_lamap.py",2373,0,"",python,selection_command +1554,1833383,"train_lamap.py",2372,0,"",python,selection_command +1555,1833383,"train_lamap.py",2371,0,"",python,selection_command +1556,1833472,"train_lamap.py",2370,0,"",python,selection_command +1557,1833472,"train_lamap.py",2369,0,"",python,selection_command +1558,1833473,"train_lamap.py",2368,0,"",python,selection_command +1559,1833507,"train_lamap.py",2367,0,"",python,selection_command +1560,1833508,"train_lamap.py",2366,0,"",python,selection_command +1561,1833566,"train_lamap.py",2365,0,"",python,selection_command +1562,1833630,"train_lamap.py",2364,0,"",python,selection_command +1563,1833630,"train_lamap.py",2363,0,"",python,selection_command +1564,1833645,"train_lamap.py",2362,0,"",python,selection_command +1565,1833777,"train_lamap.py",2361,0,"",python,selection_command +1566,1834007,"train_lamap.py",2360,0,"",python,selection_command +1567,1834180,"train_lamap.py",2359,0,"",python,selection_command +1568,1834335,"train_lamap.py",2358,0,"",python,selection_command +1569,1834484,"train_lamap.py",2357,0,"",python,selection_command +1570,1834812,"train_lamap.py",2357,2,"",python,content +1571,1835117,"train_lamap.py",2357,8,"",python,content +1572,1835360,"train_lamap.py",2357,1,"",python,content +1573,1835756,"train_lamap.py",2357,8,"",python,content +1574,1837651,"models/lamap.py",0,0,"",python,tab +1575,1839466,"models/lamap.py",1488,0,"",python,selection_mouse +1576,1840939,"models/lamap.py",1483,8,"training",python,selection_mouse +1577,1841992,"models/lamap.py",1571,0,"",python,selection_mouse +1578,1842860,"models/lamap.py",1481,0,"",python,selection_mouse +1579,1843105,"models/lamap.py",1480,2,"],",python,selection_mouse +1580,1843893,"models/lamap.py",1481,0,"",python,selection_mouse +1581,1844426,"models/lamap.py",1481,2,"",python,content +1582,1844693,"models/lamap.py",1481,8,"",python,content +1583,1844933,"models/lamap.py",1481,2,"",python,content +1584,1845247,"models/lamap.py",1481,5,"",python,content +1585,1845492,"models/lamap.py",1481,2,"",python,content +1586,1845806,"models/lamap.py",1481,4,"",python,content +1587,1850302,"train_lamap.py",0,0,"",python,tab +1588,1853530,"train_lamap.py",2807,0,"",python,selection_mouse +1589,1854447,"train_lamap.py",2620,0,"",python,selection_mouse +1590,1856328,"train_lamap.py",2650,0,"",python,selection_mouse +1591,1866863,"train_lamap.py",2593,0,"",python,selection_mouse +1592,1868152,"train_lamap.py",2593,5,"",python,content +1593,1868538,"train_lamap.py",2593,0,"l",python,content +1594,1868539,"train_lamap.py",2594,0,"",python,selection_keyboard +1595,1868654,"train_lamap.py",2594,0,"a",python,content +1596,1868655,"train_lamap.py",2595,0,"",python,selection_keyboard +1597,1868780,"train_lamap.py",2595,0,"m",python,content +1598,1868781,"train_lamap.py",2596,0,"",python,selection_keyboard +1599,1869152,"train_lamap.py",2596,0,"a",python,content +1600,1869153,"train_lamap.py",2597,0,"",python,selection_keyboard +1601,1869304,"train_lamap.py",2597,0,"p",python,content +1602,1869305,"train_lamap.py",2598,0,"",python,selection_keyboard +1603,1871371,"train_lamap.py",2598,0,"\n",python,content +1604,1872263,"train_lamap.py",2599,0," ",python,content +1605,1872690,"train_lamap.py",2603,0," ",python,content +1606,1872859,"train_lamap.py",2607,0," ",python,content +1607,1873043,"train_lamap.py",2611,0," ",python,content +1608,1874476,"train_lamap.py",2611,4,"",python,content +1609,1874637,"train_lamap.py",2607,4,"",python,content +1610,1874824,"train_lamap.py",2603,4,"",python,content +1611,1874939,"train_lamap.py",2599,4,"",python,content +1612,1875089,"train_lamap.py",2598,1,"",python,content +1613,1876332,"train_lamap.py",2593,0,"\n ",python,content +1614,1877854,"train_lamap.py",2604,0,"",python,selection_command +1615,1878739,"train_lamap.py",2603,0,"",python,selection_command +1616,1879777,"train_lamap.py",2604,0,"",python,selection_command +1617,1879964,"train_lamap.py",2605,0,"",python,selection_command +1618,1880363,"train_lamap.py",2605,0,"\n ",python,content +1619,1882292,"train_lamap.py",2598,0,"",python,selection_command +1620,1884117,"train_lam.py",0,0,"",python,tab +1621,1884118,"train_lam.py",3644,0,"",python,selection_mouse +1622,1885479,"train_lam.py",3626,19," rng: jax.Array,",python,selection_command +1623,1886789,"train_lam.py",3644,0,"",python,selection_command +1624,1886976,"train_lam.py",3664,0,"",python,selection_command +1625,1887960,"train_lam.py",3646,50,") -> tuple[jax.Array, jax.Array, jax.Array, dict]:",python,selection_command +1626,1888200,"train_lam.py",3626,70," rng: jax.Array,\n) -> tuple[jax.Array, jax.Array, jax.Array, dict]:",python,selection_command +1627,1888337,"train_lam.py",3591,105," action_last_active: jax.Array,\n rng: jax.Array,\n) -> tuple[jax.Array, jax.Array, jax.Array, dict]:",python,selection_command +1628,1888513,"train_lam.py",3573,123," inputs: dict,\n action_last_active: jax.Array,\n rng: jax.Array,\n) -> tuple[jax.Array, jax.Array, jax.Array, dict]:",python,selection_command +1629,1888646,"train_lam.py",3543,153," optimizer: nnx.Optimizer,\n inputs: dict,\n action_last_active: jax.Array,\n rng: jax.Array,\n) -> tuple[jax.Array, jax.Array, jax.Array, dict]:",python,selection_command +1630,1888764,"train_lam.py",3515,181," lam: LatentActionModel,\n optimizer: nnx.Optimizer,\n inputs: dict,\n action_last_active: jax.Array,\n rng: jax.Array,\n) -> tuple[jax.Array, jax.Array, jax.Array, dict]:",python,selection_command +1631,1888891,"train_lam.py",3499,197,"def train_step(\n lam: LatentActionModel,\n optimizer: nnx.Optimizer,\n inputs: dict,\n action_last_active: jax.Array,\n rng: jax.Array,\n) -> tuple[jax.Array, jax.Array, jax.Array, dict]:",python,selection_command +1632,1889091,"train_lam.py",3499,0,"",python,selection_command +1633,1889231,"train_lam.py",3515,0,"",python,selection_command +1634,1889415,"train_lam.py",3543,0,"",python,selection_command +1635,1892160,"train_lamap.py",0,0,"",python,tab +1636,1892161,"train_lamap.py",2568,0,"",python,selection_mouse +1637,1893610,"train_lamap.py",2568,0,"\ndef train_step(\n lam: LatentActionModel,\n optimizer: nnx.Optimizer,\n inputs: dict,\n action_last_active: jax.Array,\n rng: jax.Array,\n) -> tuple[jax.Array, jax.Array, jax.Array, dict]:",python,content +1638,1893682,"train_lamap.py",2569,0,"",python,selection_command +1639,1894453,"train_lamap.py",2585,0,"",python,selection_command +1640,1895036,"train_lamap.py",2613,0,"",python,selection_command +1641,1895240,"train_lamap.py",2585,0,"",python,selection_command +1642,1895384,"train_lamap.py",2569,0,"",python,selection_command +1643,1895625,"train_lamap.py",2585,0,"",python,selection_command +1644,1896352,"train_lamap.py",2612,0,"",python,selection_command +1645,1896661,"train_lamap.py",2611,0,"",python,selection_command +1646,1896946,"train_lamap.py",2594,17,"",python,content +1647,1897360,"train_lamap.py",2594,0,"L",python,content +1648,1897361,"train_lamap.py",2595,0,"",python,selection_keyboard +1649,1898158,"train_lamap.py",2595,0,"a",python,content +1650,1898162,"train_lamap.py",2596,0,"",python,selection_keyboard +1651,1898378,"train_lamap.py",2596,0,"t",python,content +1652,1898380,"train_lamap.py",2597,0,"",python,selection_keyboard +1653,1898691,"train_lamap.py",2597,0,"e",python,content +1654,1898692,"train_lamap.py",2598,0,"",python,selection_keyboard +1655,1898800,"train_lamap.py",2598,0,"n",python,content +1656,1898800,"train_lamap.py",2599,0,"",python,selection_keyboard +1657,1898949,"train_lamap.py",2599,0,"t",python,content +1658,1898949,"train_lamap.py",2600,0,"",python,selection_keyboard +1659,1901194,"train_lamap.py",2594,6,"LatentActionMapper",python,content +1660,1901658,"train_lamap.py",2611,0,"",python,selection_command +1661,1902013,"train_lamap.py",2640,0,"",python,selection_command +1662,1902202,"train_lamap.py",2660,0,"",python,selection_command +1663,1902323,"train_lamap.py",2688,0,"",python,selection_command +1664,1902476,"train_lamap.py",2715,0,"",python,selection_command +1665,1902646,"train_lamap.py",2743,0,"",python,selection_command +1666,1903014,"train_lamap.py",2767,0,"\n",python,content +1667,1903994,"train_lamap.py",2717,0,"",python,selection_command +1668,1904245,"train_lamap.py",2768,0,"",python,selection_command +1669,1904398,"train_lamap.py",2769,0,"",python,selection_command +1670,1904953,"train_lamap.py",2768,0,"",python,selection_command +1671,1905117,"train_lamap.py",2717,0,"",python,selection_command +1672,1905240,"train_lamap.py",2697,0,"",python,selection_command +1673,1905400,"train_lamap.py",2662,0,"",python,selection_command +1674,1905555,"train_lamap.py",2644,0,"",python,selection_command +1675,1905679,"train_lamap.py",2614,0,"",python,selection_command +1676,1905812,"train_lamap.py",2585,0,"",python,selection_command +1677,1905991,"train_lamap.py",2569,0,"",python,selection_command +1678,1906090,"train_lamap.py",2568,0,"",python,selection_command +1679,1906326,"train_lamap.py",2568,0,"\n@nnx.jit",python,content +1680,1906406,"train_lamap.py",2569,0,"",python,selection_command +1681,1906693,"train_lamap.py",2578,0,"",python,selection_command +1682,1906888,"train_lamap.py",2594,0,"",python,selection_command +1683,1907047,"train_lamap.py",2623,0,"",python,selection_command +1684,1907651,"train_lamap.py",2653,0,"",python,selection_command +1685,1907652,"train_lamap.py",2671,0,"",python,selection_command +1686,1907653,"train_lamap.py",2706,0,"",python,selection_command +1687,1907701,"train_lamap.py",2726,0,"",python,selection_command +1688,1907738,"train_lamap.py",2777,0,"",python,selection_command +1689,1908161,"train_lamap.py",2726,0,"",python,selection_command +1690,1908779,"train_lamap.py",2776,0,"\n",python,content +1691,1910814,"train_lam.py",0,0,"",python,tab +1692,1910814,"train_lam.py",3853,0,"",python,selection_mouse +1693,1912211,"train_lam.py",3819,0,"",python,selection_mouse +1694,1922157,"train_lam.py",3757,0,"",python,selection_command +1695,1922313,"train_lam.py",3723,0,"",python,selection_command +1696,1922503,"train_lam.py",3706,0,"",python,selection_command +1697,1940821,"train_lam.py",3861,0,"",python,selection_mouse +1698,1940963,"train_lam.py",3859,5,"model",python,selection_mouse +1699,1941888,"train_lam.py",3868,0,"",python,selection_mouse +1700,1942097,"train_lam.py",3866,6,"inputs",python,selection_mouse +1701,1943749,"train_lam.py",3889,0,"",python,selection_mouse +1702,1948706,"train_lam.py",4051,0,"",python,selection_mouse +1703,1949284,"train_lam.py",4055,0,"",python,selection_mouse +1704,1949299,"train_lam.py",4054,0,"",python,selection_command +1705,1953337,"train_lam.py",4028,27," optimizer.update(grads)",python,selection_command +1706,1953637,"train_lam.py",4017,38," )(lam)\n optimizer.update(grads)",python,selection_command +1707,1953741,"train_lam.py",3987,68," loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)",python,selection_command +1708,1953909,"train_lam.py",3917,138," (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)",python,selection_command +1709,1954028,"train_lam.py",3890,165," # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)",python,selection_command +1710,1954198,"train_lam.py",3889,166,"\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)",python,selection_command +1711,1954830,"train_lam.py",3832,223," return lam_loss_fn(model, inputs, training=True)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)",python,selection_command +1712,1954987,"train_lam.py",3810,245," model.train()\n return lam_loss_fn(model, inputs, training=True)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)",python,selection_command +1713,1955145,"train_lam.py",3748,307," ) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n model.train()\n return lam_loss_fn(model, inputs, training=True)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)",python,selection_command +1714,1955559,"train_lam.py",3810,245," model.train()\n return lam_loss_fn(model, inputs, training=True)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)",python,selection_command +1715,1956718,"train_lam.py",3748,307," ) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n model.train()\n return lam_loss_fn(model, inputs, training=True)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)",python,selection_command +1716,1956876,"train_lam.py",3714,341," model: LatentActionModel,\n ) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n model.train()\n return lam_loss_fn(model, inputs, training=True)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)",python,selection_command +1717,1957051,"train_lam.py",3697,358," def loss_fn(\n model: LatentActionModel,\n ) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n model.train()\n return lam_loss_fn(model, inputs, training=True)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)",python,selection_command +1718,1957428,"train_lam.py",3697,0,"",python,selection_command +1719,1959895,"train_lamap.py",0,0,"",python,tab +1720,1960952,"train_lamap.py",2777,0,"\n def loss_fn(\n model: LatentActionModel,\n ) -> tuple[jax.Array, tuple[jax.Array, jax.Array, dict]]:\n model.train()\n return lam_loss_fn(model, inputs, training=True)\n\n # --- Update model ---\n (loss, (recon, idx_counts, metrics)), grads = nnx.value_and_grad(\n loss_fn, has_aux=True\n )(lam)\n optimizer.update(grads)",python,content +1721,1961014,"train_lamap.py",2782,0,"",python,selection_command +1722,1961419,"train_lamap.py",2777,0,"",python,selection_command +1723,1961833,"train_lamap.py",2777,1,"",python,content +1724,1961965,"train_lamap.py",2781,0,"",python,selection_command +1725,1961965,"train_lamap.py",2730,0,"",python,selection_command +1726,1962043,"train_lamap.py",2710,0,"",python,selection_command +1727,1962221,"train_lamap.py",2675,0,"",python,selection_command +1728,1962397,"train_lamap.py",2657,0,"",python,selection_command +1729,1962590,"train_lamap.py",2675,0,"",python,selection_command +1730,1962946,"train_lamap.py",2671,35,"",python,content +1731,1963097,"train_lamap.py",2675,0,"",python,selection_command +1732,1963098,"train_lamap.py",2695,0,"",python,selection_command +1733,1963367,"train_lamap.py",2746,0,"",python,selection_command +1734,1963516,"train_lamap.py",2763,0,"",python,selection_command +1735,1963998,"train_lamap.py",2797,0,"",python,selection_command +1736,1964144,"train_lamap.py",2859,0,"",python,selection_command +1737,1964145,"train_lamap.py",2881,0,"",python,selection_command +1738,1964145,"train_lamap.py",2934,0,"",python,selection_command +1739,1964275,"train_lamap.py",2939,0,"",python,selection_command +1740,1964276,"train_lamap.py",2966,0,"",python,selection_command +1741,1964333,"train_lamap.py",3036,0,"",python,selection_command +1742,1964334,"train_lamap.py",3066,0,"",python,selection_command +1743,1964334,"train_lamap.py",3077,0,"",python,selection_command +1744,1964335,"train_lamap.py",3101,0,"",python,selection_command +1745,1964520,"train_lamap.py",3106,0,"",python,selection_command +1746,1964735,"train_lamap.py",3101,0,"",python,selection_command +1747,1965030,"train_lamap.py",3077,0,"",python,selection_command +1748,1965209,"train_lamap.py",3066,0,"",python,selection_command +1749,1965349,"train_lamap.py",3036,0,"",python,selection_command +1750,1965490,"train_lamap.py",2966,0,"",python,selection_command +1751,1965624,"train_lamap.py",2939,0,"",python,selection_command +1752,1965760,"train_lamap.py",2934,0,"",python,selection_command +1753,1965910,"train_lamap.py",2881,0,"",python,selection_command +1754,1966056,"train_lamap.py",2859,0,"",python,selection_command +1755,1966225,"train_lamap.py",2797,0,"",python,selection_command +1756,1966341,"train_lamap.py",2763,0,"",python,selection_command +1757,1966633,"train_lamap.py",2764,0,"",python,selection_command +1758,1966800,"train_lamap.py",2765,0,"",python,selection_command +1759,1967282,"train_lamap.py",2766,0,"",python,selection_command +1760,1967409,"train_lamap.py",2767,0,"",python,selection_command +1761,1967409,"train_lamap.py",2768,0,"",python,selection_command +1762,1967448,"train_lamap.py",2769,0,"",python,selection_command +1763,1967448,"train_lamap.py",2770,0,"",python,selection_command +1764,1967503,"train_lamap.py",2771,0,"",python,selection_command +1765,1967504,"train_lamap.py",2772,0,"",python,selection_command +1766,1967528,"train_lamap.py",2773,0,"",python,selection_command +1767,1967648,"train_lamap.py",2774,0,"",python,selection_command +1768,1968125,"train_lamap.py",2774,17,"",python,content +1769,1968561,"train_lamap.py",2774,0,"L",python,content +1770,1968562,"train_lamap.py",2775,0,"",python,selection_keyboard +1771,1968809,"train_lamap.py",2775,0,"a",python,content +1772,1968810,"train_lamap.py",2776,0,"",python,selection_keyboard +1773,1970228,"train_lamap.py",2776,0,"t",python,content +1774,1970231,"train_lamap.py",2777,0,"",python,selection_keyboard +1775,1972065,"train_lamap.py",2774,3,"LatentActionMapper",python,content +1776,1973042,"train_lamap.py",2791,0,"",python,selection_command +1777,1973135,"train_lamap.py",2826,0,"",python,selection_command +1778,1973352,"train_lamap.py",2876,0,"",python,selection_command +1779,1973480,"train_lamap.py",2910,0,"",python,selection_command +1780,1973787,"train_lamap.py",2909,0,"",python,selection_command +1781,1973980,"train_lamap.py",2908,0,"",python,selection_command +1782,1974127,"train_lamap.py",2907,0,"",python,selection_command +1783,1974262,"train_lamap.py",2906,0,"",python,selection_command +1784,1974418,"train_lamap.py",2905,0,"",python,selection_command +1785,1974557,"train_lamap.py",2904,0,"",python,selection_command +1786,1975068,"train_lamap.py",2903,0,"",python,selection_command +1787,1975176,"train_lamap.py",2902,0,"",python,selection_command +1788,1975177,"train_lamap.py",2901,0,"",python,selection_command +1789,1975240,"train_lamap.py",2900,0,"",python,selection_command +1790,1975403,"train_lamap.py",2899,0,"",python,selection_command +1791,1975572,"train_lamap.py",2898,0,"",python,selection_command +1792,1975756,"train_lamap.py",2897,0,"",python,selection_command +1793,1975929,"train_lamap.py",2896,0,"",python,selection_command +1794,1976301,"train_lamap.py",2896,0,"a",python,content +1795,1976302,"train_lamap.py",2897,0,"",python,selection_keyboard +1796,1976669,"train_lamap.py",2897,0,"p",python,content +1797,1976670,"train_lamap.py",2898,0,"",python,selection_keyboard +1798,1976880,"train_lamap.py",2897,0,"",python,selection_command +1799,1977054,"train_lamap.py",2875,0,"",python,selection_command +1800,1978053,"train_lamap.py",2897,0,"",python,selection_command +1801,1978424,"train_lamap.py",2937,0,"",python,selection_command +1802,1978540,"train_lamap.py",2957,0,"",python,selection_command +1803,1991816,"train_lamap.py",2984,0,"",python,selection_command +1804,1992023,"train_lamap.py",3054,0,"",python,selection_command +1805,1992424,"train_lamap.py",3074,0,"",python,selection_command +1806,1995228,"train_lamap.py",3054,0,"",python,selection_command +1807,1995427,"train_lamap.py",2984,0,"",python,selection_command +1808,1995992,"train_lamap.py",2983,0,"",python,selection_command +1809,1996204,"train_lamap.py",2982,0,"",python,selection_command +1810,1996690,"train_lamap.py",2981,0,"",python,selection_command +1811,1996797,"train_lamap.py",2980,0,"",python,selection_command +1812,1996798,"train_lamap.py",2979,0,"",python,selection_command +1813,1996805,"train_lamap.py",2978,0,"",python,selection_command +1814,1996875,"train_lamap.py",2977,0,"",python,selection_command +1815,1997064,"train_lamap.py",2976,0,"",python,selection_command +1816,1997217,"train_lamap.py",2975,0,"",python,selection_command +1817,1997411,"train_lamap.py",2974,0,"",python,selection_command +1818,1997600,"train_lamap.py",2973,0,"",python,selection_command +1819,1997758,"train_lamap.py",2972,0,"",python,selection_command +1820,1998117,"train_lamap.py",2971,0,"",python,selection_command +1821,1999223,"train_lamap.py",2970,0,"",python,selection_command +1822,1999402,"train_lamap.py",2969,0,"",python,selection_command +1823,1999815,"train_lamap.py",2970,0,"",python,selection_command +1824,2000347,"train_lamap.py",2971,0,"",python,selection_command +1825,2000366,"train_lamap.py",2972,0,"",python,selection_command +1826,2000392,"train_lamap.py",2973,0,"",python,selection_command +1827,2000537,"train_lamap.py",2974,0,"",python,selection_command +1828,2000809,"train_lamap.py",2973,0,"",python,selection_command +1829,2000987,"train_lamap.py",2972,0,"",python,selection_command +1830,2001137,"train_lamap.py",2971,0,"",python,selection_command +1831,2001307,"train_lamap.py",2970,0,"",python,selection_command +1832,2001446,"train_lamap.py",2969,0,"",python,selection_command +1833,2001674,"train_lamap.py",2969,1,"",python,content +1834,2001846,"train_lamap.py",2970,0,"",python,selection_command +1835,2002381,"train_lamap.py",2971,0,"",python,selection_command +1836,2002404,"train_lamap.py",2972,0,"",python,selection_command +1837,2002497,"train_lamap.py",2973,0,"",python,selection_command +1838,2002646,"train_lamap.py",2974,0,"",python,selection_command +1839,2002991,"train_lamap.py",2973,0,"",python,selection_command +1840,2003214,"train_lamap.py",2973,2,"",python,content +1841,2003504,"train_lamap.py",2973,1,"",python,content +1842,2003797,"train_lamap.py",2973,5,"",python,content +1843,2004057,"train_lamap.py",2973,2,"",python,content +1844,2004309,"train_lamap.py",2973,10,"",python,content +1845,2004600,"train_lamap.py",2973,2,"",python,content +1846,2004864,"train_lamap.py",2973,7,"",python,content +1847,2005672,"train_lamap.py",2973,1,"",python,content +1848,2005837,"train_lamap.py",2973,1,"",python,content +1849,2014006,"train_lamap.py",3011,0,"",python,selection_command +1850,2014311,"train_lamap.py",3041,0,"",python,selection_command +1851,2014666,"train_lamap.py",3052,0,"",python,selection_command +1852,2016738,"train_lam.py",0,0,"",python,tab +1853,2016738,"train_lam.py",4150,0,"",python,selection_mouse +1854,2019243,"train_lam.py",4025,0,"",python,selection_mouse +1855,2019997,"train_lam.py",4036,0,"",python,selection_mouse +1856,2032082,"train_lam.py",4049,0,"",python,selection_mouse +1857,2032400,"train_lam.py",4046,0,"",python,selection_mouse +1858,2058000,"train_lam.py",4628,0,"",python,selection_mouse +1859,2058199,"train_lam.py",4618,18,"action_last_active",python,selection_mouse +1860,2059230,"train_lam.py",4702,0,"",python,selection_mouse +1861,2059364,"train_lam.py",4699,5,"recon",python,selection_mouse +1862,2059979,"train_lam.py",4698,0,"",python,selection_mouse +1863,2060103,"train_lam.py",4698,1," ",python,selection_mouse +1864,2061011,"train_lam.py",4731,0,"",python,selection_mouse +1865,2061271,"train_lam.py",4726,7,"metrics",python,selection_mouse +1866,2065719,"train_lamap.py",0,0,"",python,tab +1867,2065719,"train_lamap.py",3043,0,"",python,selection_mouse +1868,2065903,"train_lamap.py",3042,0,"",python,selection_command +1869,2066459,"train_lamap.py",3071,0,"",python,selection_mouse +1870,2066487,"train_lamap.py",3070,0,"",python,selection_command +1871,2067465,"train_lamap.py",3071,0,"\n ",python,content +1872,2090656,"train_lamap.py",3076,0,"r",python,content +1873,2090658,"train_lamap.py",3077,0,"",python,selection_keyboard +1874,2090806,"train_lamap.py",3077,0,"e",python,content +1875,2090808,"train_lamap.py",3078,0,"",python,selection_keyboard +1876,2090919,"train_lamap.py",3078,0,"t",python,content +1877,2090921,"train_lamap.py",3079,0,"",python,selection_keyboard +1878,2091130,"train_lamap.py",3079,0,"u",python,content +1879,2091131,"train_lamap.py",3080,0,"",python,selection_keyboard +1880,2091334,"train_lamap.py",3080,0,"r",python,content +1881,2091336,"train_lamap.py",3081,0,"",python,selection_keyboard +1882,2091400,"train_lamap.py",3081,0,"n",python,content +1883,2091402,"train_lamap.py",3082,0,"",python,selection_keyboard +1884,2091475,"train_lamap.py",3082,0," ",python,content +1885,2091476,"train_lamap.py",3083,0,"",python,selection_keyboard +1886,2091635,"train_lamap.py",3083,0,"l",python,content +1887,2091636,"train_lamap.py",3084,0,"",python,selection_keyboard +1888,2091807,"train_lamap.py",3084,0,"o",python,content +1889,2091808,"train_lamap.py",3085,0,"",python,selection_keyboard +1890,2091987,"train_lamap.py",3085,0,"s",python,content +1891,2091989,"train_lamap.py",3086,0,"",python,selection_keyboard +1892,2092200,"train_lamap.py",3086,0,"s",python,content +1893,2092201,"train_lamap.py",3087,0,"",python,selection_keyboard +1894,2092683,"train_lamap.py",3086,0,"",python,selection_command +1895,2092940,"train_lamap.py",3088,0,"",python,selection_command +1896,2093323,"train_lamap.py",3096,0,"",python,selection_command +1897,2093949,"train_lamap.py",3089,8,"@nnx.jit",python,selection_command +1898,2094158,"train_lamap.py",3089,24,"@nnx.jit\ndef train_step(",python,selection_command +1899,2094434,"train_lamap.py",3089,36,"@nnx.jit\ndef train_step(\n lamap, ",python,selection_command +1900,2094643,"train_lamap.py",3089,49,"@nnx.jit\ndef train_step(\n lamap, \n inputs):",python,selection_command +1901,2094792,"train_lamap.py",3089,67,"@nnx.jit\ndef train_step(\n lamap, \n inputs):\n lamap.train()",python,selection_command +1902,2094957,"train_lamap.py",3089,131,"@nnx.jit\ndef train_step(\n lamap, \n inputs):\n lamap.train()\n grad_fn = jax.value_and_grad(lamap_loss_fn, allow_int=True)",python,selection_command +1903,2095103,"train_lamap.py",3089,186,"@nnx.jit\ndef train_step(\n lamap, \n inputs):\n lamap.train()\n grad_fn = jax.value_and_grad(lamap_loss_fn, allow_int=True)\n loss, grads = grad_fn(state.params, state, inputs)",python,selection_command +1904,2095244,"train_lamap.py",3089,233,"@nnx.jit\ndef train_step(\n lamap, \n inputs):\n lamap.train()\n grad_fn = jax.value_and_grad(lamap_loss_fn, allow_int=True)\n loss, grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)",python,selection_command +1905,2095458,"train_lamap.py",3089,256,"@nnx.jit\ndef train_step(\n lamap, \n inputs):\n lamap.train()\n grad_fn = jax.value_and_grad(lamap_loss_fn, allow_int=True)\n loss, grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n return state, loss",python,selection_command +1906,2095780,"train_lamap.py",3089,257,"",python,content +1907,2096540,"train_lamap.py",3089,1,"",python,content +1908,2097276,"train_lamap.py",3098,0,"",python,selection_command +1909,2097618,"train_lamap.py",3190,0,"",python,selection_command +1910,2097829,"train_lamap.py",3207,0,"",python,selection_command +1911,2097989,"train_lamap.py",3190,0,"",python,selection_command +1912,2098133,"train_lamap.py",3098,0,"",python,selection_command +1913,2098289,"train_lamap.py",3089,0,"",python,selection_command +1914,2098462,"train_lamap.py",3088,0,"",python,selection_command +1915,2098590,"train_lamap.py",3072,0,"",python,selection_command +1916,2098748,"train_lamap.py",3044,0,"",python,selection_command +1917,2099058,"train_lamap.py",3072,0,"",python,selection_command +1918,2104044,"train_lam.py",0,0,"",python,tab +1919,2104044,"train_lam.py",4861,0,"",python,selection_mouse +1920,2104477,"train_lam.py",4946,0,"",python,selection_mouse +1921,2106808,"train_lam.py",4923,31," return loss, recon, metrics",python,selection_command +1922,2106994,"train_lam.py",4848,106," (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics",python,selection_command +1923,2107172,"train_lam.py",4833,121," lam.eval()\n (loss, (recon, _, metrics)) = lam_loss_fn(lam, inputs, training=False)\n return loss, recon, metrics",python,selection_command +1924,2108048,"train_lam.py",4846,0,"",python,selection_command +1925,2110069,"train_lamap.py",0,0,"",python,tab +1926,2110070,"train_lamap.py",3226,0,"",python,selection_mouse +1927,2111349,"train_lamap.py",3210,0,"",python,selection_mouse +1928,2111607,"train_lamap.py",3210,2," (",python,selection_mouse +1929,2111608,"train_lamap.py",3210,5," (los",python,selection_mouse +1930,2111608,"train_lamap.py",3210,8," (loss, ",python,selection_mouse +1931,2111647,"train_lamap.py",3210,13," (loss, (reco",python,selection_mouse +1932,2111688,"train_lamap.py",3210,18," (loss, (recon, _,",python,selection_mouse +1933,2111734,"train_lamap.py",3210,20," (loss, (recon, _, m",python,selection_mouse +1934,2111778,"train_lamap.py",3210,21," (loss, (recon, _, me",python,selection_mouse +1935,2111805,"train_lamap.py",3210,23," (loss, (recon, _, metr",python,selection_mouse +1936,2111863,"train_lamap.py",3210,24," (loss, (recon, _, metri",python,selection_mouse +1937,2111896,"train_lamap.py",3210,26," (loss, (recon, _, metrics",python,selection_mouse +1938,2112077,"train_lamap.py",3210,27," (loss, (recon, _, metrics)",python,selection_mouse +1939,2112184,"train_lamap.py",3210,28," (loss, (recon, _, metrics))",python,selection_mouse +1940,2112497,"train_lamap.py",3210,28,"",python,content +1941,2112970,"train_lamap.py",3210,0,"l",python,content +1942,2112973,"train_lamap.py",3211,0,"",python,selection_keyboard +1943,2113129,"train_lamap.py",3211,0,"o",python,content +1944,2113130,"train_lamap.py",3212,0,"",python,selection_keyboard +1945,2113272,"train_lamap.py",3212,0,"s",python,content +1946,2113273,"train_lamap.py",3213,0,"",python,selection_keyboard +1947,2113423,"train_lamap.py",3213,0,"s",python,content +1948,2113424,"train_lamap.py",3214,0,"",python,selection_keyboard +1949,2114667,"train_lamap.py",3210,0," ",python,content +1950,2115913,"train_lamap.py",3210,0,"",python,selection_command +1951,2116813,"train_lamap.py",3279,0,"",python,selection_mouse +1952,2116902,"train_lamap.py",3279,1,",",python,selection_mouse +1953,2117124,"train_lamap.py",3279,5,", rec",python,selection_mouse +1954,2117125,"train_lamap.py",3279,9,", recon, ",python,selection_mouse +1955,2117125,"train_lamap.py",3279,11,", recon, me",python,selection_mouse +1956,2117125,"train_lamap.py",3279,12,", recon, met",python,selection_mouse +1957,2117125,"train_lamap.py",3279,13,", recon, metr",python,selection_mouse +1958,2117126,"train_lamap.py",3279,14,", recon, metri",python,selection_mouse +1959,2117195,"train_lamap.py",3279,15,", recon, metric",python,selection_mouse +1960,2117240,"train_lamap.py",3279,16,", recon, metrics",python,selection_mouse +1961,2117649,"train_lamap.py",3279,16,"",python,content +1962,2117694,"train_lamap.py",3278,0,"",python,selection_command +1963,2125042,"train_lamap.py",4036,0,"",python,selection_mouse +1964,2125404,"train_lamap.py",4036,1,"v",python,selection_mouse +1965,2125405,"train_lamap.py",4036,3,"val",python,selection_mouse +1966,2125405,"train_lamap.py",4036,4,"val_",python,selection_mouse +1967,2125533,"train_lamap.py",4036,5,"val_m",python,selection_mouse +1968,2125534,"train_lamap.py",4036,7,"val_met",python,selection_mouse +1969,2125630,"train_lamap.py",4036,9,"val_metri",python,selection_mouse +1970,2125738,"train_lamap.py",4036,10,"val_metric",python,selection_mouse +1971,2125913,"train_lamap.py",4036,12,"val_metrics,",python,selection_mouse +1972,2127533,"train_lamap.py",4036,12,"",python,content +1973,2127858,"train_lamap.py",4036,1,"",python,content +1974,2128046,"train_lamap.py",4036,1,"",python,content +1975,2128196,"train_lamap.py",4036,1,"",python,content +1976,2128705,"train_lamap.py",4036,1,"",python,content +1977,2128827,"train_lamap.py",4036,1,"",python,content +1978,2128893,"train_lamap.py",4036,1,"",python,content +1979,2128920,"train_lamap.py",4036,1,"",python,content +1980,2128944,"train_lamap.py",4036,1,"",python,content +1981,2128989,"train_lamap.py",4036,1,"",python,content +1982,2129067,"train_lamap.py",4036,1,"",python,content +1983,2129097,"train_lamap.py",4036,1,"",python,content +1984,2129226,"train_lamap.py",4036,1,"",python,content +1985,2129651,"train_lamap.py",4036,1,"",python,content +1986,2129825,"train_lamap.py",4036,1,"",python,content +1987,2129867,"train_lamap.py",4035,0,"",python,selection_command +1988,2130847,"train_lamap.py",4034,0,"",python,selection_command +1989,2130941,"train_lamap.py",4034,1,"",python,content +1990,2132070,"train_lamap.py",4013,0,"",python,selection_command +1991,2132238,"train_lamap.py",3982,0,"",python,selection_command +1992,2132471,"train_lamap.py",4013,0,"",python,selection_command +1993,2133684,"train_lamap.py",4009,5," }",python,selection_command +1994,2133855,"train_lamap.py",3963,51," for key in metrics_per_step[0].keys()\n }",python,selection_command +1995,2134003,"train_lamap.py",3897,117," f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }",python,selection_command +1996,2134374,"train_lamap.py",3877,137," val_metrics = {\n f""val_{key}"": np.mean([m[key] for m in metrics_per_step])\n for key in metrics_per_step[0].keys()\n }",python,selection_command +1997,2136031,"train_lamap.py",3877,138,"",python,content +1998,2136134,"train_lamap.py",3881,0,"",python,selection_command +1999,2146117,"train_lamap.py",3485,0,"",python,selection_mouse +2000,2146682,"train_lamap.py",3484,0,"",python,selection_command +2001,2146934,"train_lamap.py",3483,0,"",python,selection_command +2002,2147099,"train_lamap.py",3482,0,"",python,selection_command +2003,2147472,"train_lamap.py",3481,0,"",python,selection_command +2004,2147741,"train_lamap.py",3481,2,"",python,content +2005,2148058,"train_lamap.py",3481,5,"",python,content +2006,2148340,"train_lamap.py",3481,2,"",python,content +2007,2148782,"train_lamap.py",3481,8,"",python,content +2008,2150413,"train_lamap.py",3651,0,"",python,selection_mouse +2009,2151143,"train_lamap.py",3650,0,"",python,selection_mouse +2010,2151153,"train_lamap.py",3649,0,"",python,selection_command +2011,2152196,"train_lamap.py",3562,0,"",python,selection_mouse +2012,2152688,"train_lamap.py",3540,41,"",python,content +2013,2152733,"train_lamap.py",3548,0,"",python,selection_command +2014,2152733,"train_lamap.py",3513,0,"",python,selection_command +2015,2152849,"train_lamap.py",3477,0,"",python,selection_command +2016,2153001,"train_lamap.py",3440,0,"",python,selection_command +2017,2153104,"train_lamap.py",3406,0,"",python,selection_command +2018,2153214,"train_lamap.py",3380,0,"",python,selection_command +2019,2153561,"train_lamap.py",3372,26,"",python,content +2020,2153627,"train_lamap.py",3376,0,"",python,selection_command +2021,2153721,"train_lamap.py",3410,0,"",python,selection_command +2022,2154673,"train_lamap.py",3447,0,"",python,selection_command +2023,2154891,"train_lamap.py",3483,0,"",python,selection_command +2024,2155026,"train_lamap.py",3518,0,"",python,selection_command +2025,2155226,"train_lamap.py",3536,0,"",python,selection_command +2026,2155346,"train_lamap.py",3570,0,"",python,selection_command +2027,2155528,"train_lamap.py",3584,0,"",python,selection_command +2028,2155703,"train_lamap.py",3589,0,"",python,selection_command +2029,2156095,"train_lamap.py",3584,0,"",python,selection_command +2030,2156296,"train_lamap.py",3570,0,"",python,selection_command +2031,2156485,"train_lamap.py",3536,0,"",python,selection_command +2032,2156714,"train_lamap.py",3518,0,"",python,selection_command +2033,2156893,"train_lamap.py",3483,0,"",python,selection_command +2034,2157915,"train_lamap.py",3484,0,"",python,selection_command +2035,2158427,"train_lamap.py",3485,0,"",python,selection_command +2036,2158480,"train_lamap.py",3486,0,"",python,selection_command +2037,2158534,"train_lamap.py",3487,0,"",python,selection_command +2038,2158554,"train_lamap.py",3488,0,"",python,selection_command +2039,2158569,"train_lamap.py",3489,0,"",python,selection_command +2040,2159623,"train_lamap.py",3468,0,"",python,selection_mouse +2041,2160485,"train_lamap.py",3333,0,"",python,selection_mouse +2042,2162284,"train_lamap.py",3332,0,"",python,selection_mouse +2043,2162438,"train_lamap.py",3330,3,"lam",python,selection_mouse +2044,2168643,"train_lamap.py",3333,0,"",python,selection_command +2045,2168678,"train_lamap.py",3469,0,"ap",python,content +2046,2168679,"train_lamap.py",3333,0,"ap",python,content diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-9995c396-b28a-4817-9971-5d6e293ae32a1753180233918-2025_07_22-12.30.43.584/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-9995c396-b28a-4817-9971-5d6e293ae32a1753180233918-2025_07_22-12.30.43.584/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..4bfe70d031dc435de93f6ff3bef2cfff8c697eb3 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-9995c396-b28a-4817-9971-5d6e293ae32a1753180233918-2025_07_22-12.30.43.584/source.csv @@ -0,0 +1,2 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +2,58754,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:30:43 PM [info] Activating crowd-code\n12:30:43 PM [info] Recording started\n12:30:43 PM [info] Initializing git provider using file system watchers...\n12:30:45 PM [info] Retrying git provider initialization...\n12:30:46 PM [error] Not a git repository: EntryNotFound (FileSystemError): Error: ENOENT: no such file or directory, stat '/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4/.git'\n12:30:47 PM [error] Not a git repository: EntryNotFound (FileSystemError): Error: ENOENT: no such file or directory, stat '/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4/.git'\n",Log,tab diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-a5aad456-d021-49a4-8368-8264b401bd3e1757237946327-2025_09_07-11.39.28.68/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-a5aad456-d021-49a4-8368-8264b401bd3e1757237946327-2025_09_07-11.39.28.68/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..5ac558ded6036a0357c498cccddfcd4ac645f105 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-a5aad456-d021-49a4-8368-8264b401bd3e1757237946327-2025_09_07-11.39.28.68/source.csv @@ -0,0 +1,2352 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +1,5,"train_tokenizer.py",0,0,"from dataclasses import dataclass, field\n\nimport os\nimport time\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nimport optax\nimport orbax\nfrom orbax.checkpoint import PyTreeCheckpointer\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\n\nts = int(time.time())\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_resolution: int = 64\n data_dir: str = ""data/coinrun_episodes""\n checkpoint: str = """"\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n min_lr: float = 3e-4\n max_lr: float = 3e-4\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 8\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_gradients: bool = False\n name: str = """"\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n wandb_id: str = """"\n\n\nargs = tyro.cli(Args)\n\n\ndef tokenizer_loss_fn(params, state, inputs):\n # --- Compute loss ---\n outputs = state.apply_fn(\n params, inputs, training=True, rngs={""dropout"": inputs[""rng""]}\n )\n mse = jnp.square(inputs[""videos""] - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = inputs[""videos""].clip(0, 1).reshape(-1, *inputs[""videos""].shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = pix.psnr(gt, recon).mean()\n ssim = pix.ssim(gt, recon).mean()\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n grad_fn = jax.value_and_grad(tokenizer_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n rng = jax.random.PRNGKey(args.seed)\n if args.log:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n # --- Initialize model ---\n tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_resolution, args.image_resolution, args.image_channels)\n inputs = dict(\n videos=jnp.zeros(\n (args.batch_size, args.seq_len, *image_shape), dtype=jnp.float32\n ),\n )\n init_params = tokenizer.init(_rng, inputs)\n\n # --- Load checkpoint ---\n step = 0\n if args.checkpoint:\n init_params[""params""].update(\n PyTreeCheckpointer().restore(args.checkpoint)[""model""][""params""][""params""]\n )\n # Assume checkpoint is of the form tokenizer__\n step += int(args.checkpoint.split(""_"")[-1])\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=tokenizer.apply, params=init_params, tx=tx)\n\n # --- TRAIN LOOP ---\n dataloader = get_dataloader(args.data_dir, args.seq_len, args.batch_size)\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng = jax.random.split(rng)\n inputs = dict(videos=videos, rng=_rng)\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0:\n wandb.log({""loss"": loss, ""step"": step, **metrics})\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(\n os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""\n ),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break\n",python,tab +2,1122,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"11:39:28 AM [info] Activating crowd-code\n11:39:28 AM [info] Recording started\n11:39:28 AM [info] Initializing git provider using file system watchers...\n",Log,tab +3,1421,"extension-output-pdoom-org.crowd-code-#1-crowd-code",153,0,"11:39:28 AM [info] Git repository found\n11:39:28 AM [info] Git provider initialized successfully\n11:39:28 AM [info] Initial git state: [object Object]\n",Log,content +4,2367,"extension-output-pdoom-org.crowd-code-#1-crowd-code",33,0,"",Log,selection_mouse +5,3206,"train_tokenizer.py",0,0,"",python,tab +6,7097,"TERMINAL",0,0,"undefined[tum_cte0515@hkn1991 jafar]$ source .venv/bin/activate",,terminal_command +7,10927,"train_dynamics.py",0,0,"from dataclasses import dataclass, field\nimport os\nimport time\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nimport optax\nimport orbax\nimport numpy as np\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\n\nts = int(time.time())\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_resolution: int = 64\n data_dir: str = ""data/coinrun_episodes""\n # Optimization\n batch_size: int = 36\n min_lr: float = 3e-6\n max_lr: float = 3e-5\n warmup_steps: int = 5000\n # Tokenizer\n tokenizer_dim: int = 512\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 8\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 8\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_dim: int = 512\n dyna_num_blocks: int = 12\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_gradients: bool = False\n name: str = """"\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n wandb_id: str = """"\n\nargs = tyro.cli(Args)\n\n\ndef dynamics_loss_fn(params, state, inputs):\n """"""Compute masked dynamics loss""""""\n outputs = state.apply_fn(\n params, inputs, training=True, rngs={""dropout"": inputs[""dropout_rng""]}\n )\n mask = outputs[""mask""]\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n """"""Update state and compute metrics""""""\n grad_fn = jax.value_and_grad(dynamics_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n rng = jax.random.PRNGKey(args.seed)\n if args.log:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n # --- Initialize model ---\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_resolution, args.image_resolution, args.image_channels)\n dummy_inputs = dict(\n videos=jnp.zeros(\n (args.batch_size, args.seq_len, *image_shape), dtype=jnp.float32\n ),\n mask_rng=_rng,\n )\n rng, _rng = jax.random.split(rng)\n init_params = genie.init(_rng, dummy_inputs)\n init_params = restore_genie_components(\n init_params, args.tokenizer_checkpoint, args.lam_checkpoint\n )\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=genie.apply, params=init_params, tx=tx)\n\n # --- TRAIN LOOP ---\n dataloader = get_dataloader(args.data_dir, args.seq_len, args.batch_size)\n step = 0\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _mask_rng = jax.random.split(rng, 3)\n inputs = dict(\n videos=videos,\n action=jnp.zeros((args.batch_size, args.seq_len), dtype=jnp.float32),\n dropout_rng=_rng,\n mask_rng=_mask_rng,\n )\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0:\n wandb.log({""loss"": loss, ""step"": step, **metrics})\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[15])),\n recon=wandb.Image(np.asarray(recon_seq[15])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""genie_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break\n",python,tab +8,30606,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=2-00:00:00\n#SBATCH --cpus-per-task=8\n#SBATCH --gres=gpu:1\n#SBATCH --partition=accelerated\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics/%x_%j.log\n#SBATCH --job-name=train_dynamics_coinrun_og_reproduction\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\ntags=""coinrun_og dynanmics 10m_dataset repro_mihir""\n\nnpy_records_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m""\n\n# TODO mihir: update the tokenizer and lam checkpoints\ntokenizer_ckpt_dir=""""\nlam_ckpt_dir=""""\n\nCHECKPOINT_DIR=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/dynamics/${job_name}/${slurm_job_id}""\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --tokenizer_checkpoint=""${tokenizer_ckpt_dir}"" \\n --lam_checkpoint=""${lam_ckpt_dir}"" \\n --log_image_interval=1000 \\n --log \\n --name=""${job_name}_${slurm_job_id}"" \\n --tags ${tags} \\n --entity instant-uv \\n --project jafar \\n --data_dir $npy_records_dir \\n --wandb_id $slurm_job_id\n",shellscript,tab +9,49394,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_lam_reproduction.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=2-00:00:00\n#SBATCH --cpus-per-task=8\n#SBATCH --gres=gpu:1\n#SBATCH --partition=accelerated\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/lam/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/lam/%x_%j.log\n#SBATCH --job-name=train_lam_coinrun_og_reproduction\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\ntags=""coinrun_og reproduction 10m_dataset lam repro_mihir""\n\nnpy_records_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m""\n\nCHECKPOINT_DIR=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/lam/$job_name/$slurm_job_id""\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_lam.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --log_image_interval=1000 \\n --log \\n --name=""${job_name}_${slurm_job_id}"" \\n --tags $tags \\n --entity instant-uv \\n --project jafar \\n --data_dir $npy_records_dir \\n --wandb_id $slurm_job_id\n",shellscript,tab +10,52177,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_tokenizer_repoduction.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=2-00:00:00\n#SBATCH --cpus-per-task=8\n#SBATCH --gres=gpu:1\n#SBATCH --partition=accelerated\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/tokenizer/%x_%j.log\n#SBATCH --job-name=train_tokenizer_coinrun_og_reproduction\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\ntags=""coinrun_og reproduction 10m_dataset tokenizer repro_mihir""\n\nnpy_records_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m""\n\nCHECKPOINT_DIR=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/$job_name/$slurm_job_id""\n\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --log_image_interval=1000 \\n --patch_size=16 \\n --log \\n --name=""${job_name}_${slurm_job_id}"" \\n --tags $tags \\n --entity instant-uv \\n --project jafar \\n --data_dir $npy_records_dir \\n --wandb_id $slurm_job_id\n\n",shellscript,tab +11,55293,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_tokenizer_repoduction.sbatch",523,0,"",shellscript,selection_mouse +12,55312,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_tokenizer_repoduction.sbatch",522,0,"",shellscript,selection_command +13,56560,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_lam_reproduction.sbatch",0,0,"",shellscript,tab +14,57578,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",0,0,"",shellscript,tab +15,88102,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,0,"",shellscript,selection_mouse +16,89552,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286",shellscript,content +17,96825,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1068,0,"",shellscript,selection_mouse +18,97489,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1068,0,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/lam/train_lam_coinrun_og_reproduction/3466287",shellscript,content +19,112908,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1564,0,"",shellscript,selection_mouse +20,113837,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1523,0,"",shellscript,selection_mouse +21,114595,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1523,0,"\n ",shellscript,content +22,116280,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1528,0,"-",shellscript,content +23,116282,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1529,0,"",shellscript,selection_keyboard +24,116419,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1529,0,"-",shellscript,content +25,116420,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1530,0,"",shellscript,selection_keyboard +26,116998,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1530,0,"t",shellscript,content +27,116999,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1531,0,"",shellscript,selection_keyboard +28,117147,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1531,0,"o",shellscript,content +29,117149,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1532,0,"",shellscript,selection_keyboard +30,117185,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1532,0,"k",shellscript,content +31,117187,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1533,0,"",shellscript,selection_keyboard +32,117315,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1533,0,"a",shellscript,content +33,117316,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1534,0,"",shellscript,selection_keyboard +34,117539,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1534,0,"n",shellscript,content +35,117540,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1535,0,"",shellscript,selection_keyboard +36,117984,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1534,1,"",shellscript,content +37,118115,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1533,1,"",shellscript,content +38,118536,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1530,3,"",shellscript,content +39,119086,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1530,0,"p",shellscript,content +40,119087,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1531,0,"",shellscript,selection_keyboard +41,119160,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1531,0,"a",shellscript,content +42,119161,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1532,0,"",shellscript,selection_keyboard +43,119365,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1532,0,"t",shellscript,content +44,119368,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1533,0,"",shellscript,selection_keyboard +45,119748,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1533,0,"c",shellscript,content +46,119749,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1534,0,"",shellscript,selection_keyboard +47,119824,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1534,0,"h",shellscript,content +48,119825,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1535,0,"",shellscript,selection_keyboard +49,121747,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1535,0,"_",shellscript,content +50,121749,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1536,0,"",shellscript,selection_keyboard +51,122042,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1536,0,"s",shellscript,content +52,122043,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1537,0,"",shellscript,selection_keyboard +53,122176,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1537,0,"i",shellscript,content +54,122177,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1538,0,"",shellscript,selection_keyboard +55,122296,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1538,0,"z",shellscript,content +56,122297,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1539,0,"",shellscript,selection_keyboard +57,122408,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1539,0,"e",shellscript,content +58,122409,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1540,0,"",shellscript,selection_keyboard +59,123840,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1540,0,"=",shellscript,content +60,123841,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1541,0,"",shellscript,selection_keyboard +61,123980,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1541,0,"!",shellscript,content +62,123981,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1542,0,"",shellscript,selection_keyboard +63,124342,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1542,0,"6",shellscript,content +64,124343,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1543,0,"",shellscript,selection_keyboard +65,124609,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1543,0," ",shellscript,content +66,124610,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1544,0,"",shellscript,selection_keyboard +67,124776,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1544,0,"\",shellscript,content +68,124777,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1545,0,"",shellscript,selection_keyboard +69,125596,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1544,0,"",shellscript,selection_command +70,125777,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1543,0,"",shellscript,selection_command +71,125976,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1542,0,"",shellscript,selection_command +72,126185,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1541,1,"",shellscript,content +73,126384,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1541,0,"1",shellscript,content +74,126385,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1542,0,"",shellscript,selection_keyboard +75,133104,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_tokenizer_repoduction.sbatch",0,0,"",shellscript,tab +76,135795,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_tokenizer_repoduction.sbatch",1163,0,"",shellscript,selection_mouse +77,135799,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_tokenizer_repoduction.sbatch",1162,0,"",shellscript,selection_command +78,136361,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_tokenizer_repoduction.sbatch",1147,0,"",shellscript,selection_mouse +79,138476,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",0,0,"",shellscript,tab +80,144570,"train_dynamics.py",0,0,"",python,tab +81,154004,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",0,0,"",shellscript,tab +82,155206,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1742,0,"",shellscript,selection_mouse +83,155898,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1805,0,"",shellscript,selection_mouse +84,227764,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",,terminal_command +85,227779,"TERMINAL",0,0,"]633;CSubmitted batch job 3472612\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar",,terminal_output +86,247842,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1806,0,"",shellscript,selection_mouse +87,255551,"TERMINAL",0,0,"idling",,terminal_command +88,255635,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1991.localdomain: Sun Sep 7 11:43:43 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 190 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated: 78 nodes idle\rPartition dev_accelerated-h100 :\t 1 nodes idle\rPartition accelerated-h100:\t 2 nodes idle\rPartition large:\t 3 nodes idle\rPartition accelerated-h200:\t 0 nodes idle",,terminal_output +89,256684,"TERMINAL",0,0,"4\t",,terminal_output +90,257677,"TERMINAL",0,0,"5\t",,terminal_output +91,258724,"TERMINAL",0,0,"6\t",,terminal_output +92,259791,"TERMINAL",0,0,"7\t",,terminal_output +93,260817,"TERMINAL",0,0,"8\t",,terminal_output +94,261819,"TERMINAL",0,0,"9\t",,terminal_output +95,262928,"TERMINAL",0,0,"50\t",,terminal_output +96,263893,"TERMINAL",0,0,"1\t",,terminal_output +97,264926,"TERMINAL",0,0,"2\t",,terminal_output +98,265954,"TERMINAL",0,0,"3\t",,terminal_output +99,267035,"TERMINAL",0,0,"4\t",,terminal_output +100,268064,"TERMINAL",0,0,"5\t",,terminal_output +101,268628,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",190,0,"",shellscript,selection_mouse +102,269150,"TERMINAL",0,0,"6\t",,terminal_output +103,270158,"TERMINAL",0,0,"8\t",,terminal_output +104,271195,"TERMINAL",0,0,"9\t",,terminal_output +105,271997,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction copy.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=2-00:00:00\n#SBATCH --cpus-per-task=8\n#SBATCH --gres=gpu:1\n#SBATCH --partition=accelerated\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics/%x_%j.log\n#SBATCH --job-name=train_dynamics_coinrun_og_reproduction\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\ntags=""coinrun_og dynanmics 10m_dataset repro_mihir""\n\nnpy_records_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m""\n\n# TODO mihir: update the tokenizer and lam checkpoints\ntokenizer_ckpt_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286""\nlam_ckpt_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/lam/train_lam_coinrun_og_reproduction/3466287""\n\nCHECKPOINT_DIR=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/dynamics/${job_name}/${slurm_job_id}""\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --tokenizer_checkpoint=""${tokenizer_ckpt_dir}"" \\n --patch_size=16 \\n --lam_checkpoint=""${lam_ckpt_dir}"" \\n --log_image_interval=1000 \\n --log \\n --name=""${job_name}_${slurm_job_id}"" \\n --tags ${tags} \\n --entity instant-uv \\n --project jafar \\n --data_dir $npy_records_dir \\n --wandb_id $slurm_job_id\n",shellscript,tab +106,272305,"TERMINAL",0,0,"4:00\t",,terminal_output +107,273223,"TERMINAL",0,0,"1\t",,terminal_output +108,274254,"TERMINAL",0,0,"2\t",,terminal_output +109,275290,"TERMINAL",0,0,"3\t",,terminal_output +110,276326,"TERMINAL",0,0,"4\t",,terminal_output +111,277364,"TERMINAL",0,0,"5\t",,terminal_output +112,278536,"TERMINAL",0,0,"6\t",,terminal_output +113,279441,"TERMINAL",0,0,"7\t",,terminal_output +114,280560,"TERMINAL",0,0,"8\t",,terminal_output +115,281516,"TERMINAL",0,0,"9\t",,terminal_output +116,283252,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=2-00:00:00\n#SBATCH --cpus-per-task=8\n#SBATCH --gres=gpu:1\n#SBATCH --partition=accelerated\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics/%x_%j.log\n#SBATCH --job-name=train_dynamics_coinrun_og_reproduction\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\ntags=""coinrun_og dynanmics 10m_dataset repro_mihir""\n\nnpy_records_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m""\n\n# TODO mihir: update the tokenizer and lam checkpoints\ntokenizer_ckpt_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286""\nlam_ckpt_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/lam/train_lam_coinrun_og_reproduction/3466287""\n\nCHECKPOINT_DIR=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/dynamics/${job_name}/${slurm_job_id}""\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --tokenizer_checkpoint=""${tokenizer_ckpt_dir}"" \\n --patch_size=16 \\n --lam_checkpoint=""${lam_ckpt_dir}"" \\n --log_image_interval=1000 \\n --log \\n --name=""${job_name}_${slurm_job_id}"" \\n --tags ${tags} \\n --entity instant-uv \\n --project jafar \\n --data_dir $npy_records_dir \\n --wandb_id $slurm_job_id\n",shellscript,tab +117,283319,"TERMINAL",0,0,"10\t",,terminal_output +118,283649,"TERMINAL",0,0,"1\t",,terminal_output +119,284657,"TERMINAL",0,0,"2\t",,terminal_output +120,284755,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",170,0,"",shellscript,selection_mouse +121,285365,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",171,0,"",shellscript,selection_command +122,285651,"TERMINAL",0,0,"3\t",,terminal_output +123,286694,"TERMINAL",0,0,"4\t",,terminal_output +124,287665,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",171,0,"-",shellscript,content +125,287666,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",172,0,"",shellscript,selection_keyboard +126,287742,"TERMINAL",0,0,"5\t",,terminal_output +127,288706,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",172,0,"h",shellscript,content +128,288707,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",173,0,"",shellscript,selection_keyboard +129,288800,"TERMINAL",0,0,"6\t",,terminal_output +130,289766,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",173,0,"1",shellscript,content +131,289767,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",174,0,"",shellscript,selection_keyboard +132,289832,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",174,0,"0",shellscript,content +133,289833,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",175,0,"",shellscript,selection_keyboard +134,289845,"TERMINAL",0,0,"7\t",,terminal_output +135,289954,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",175,0,"0",shellscript,content +136,289955,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",176,0,"",shellscript,selection_keyboard +137,290398,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",175,0,"",shellscript,selection_command +138,290841,"TERMINAL",0,0,"8\t",,terminal_output +139,291869,"TERMINAL",0,0,"9\t",,terminal_output +140,292912,"TERMINAL",0,0,"20\t",,terminal_output +141,293950,"TERMINAL",0,0,"1\t",,terminal_output +142,294752,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jafar",,terminal_output +143,303609,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",,terminal_command +144,303664,"TERMINAL",0,0,"]633;CSubmitted batch job 3472614\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar",,terminal_output +145,306239,"TERMINAL",0,0,"idling",,terminal_command +146,306343,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1991.localdomain: Sun Sep 7 11:44:34 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 190 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated: 78 nodes idle\rPartition dev_accelerated-h100 :\t 1 nodes idle\rPartition accelerated-h100:\t 2 nodes idle\rPartition large:\t 3 nodes idle\rPartition accelerated-h200:\t 0 nodes idle",,terminal_output +147,307348,"TERMINAL",0,0,"5\t",,terminal_output +148,308371,"TERMINAL",0,0,"6\t",,terminal_output +149,309401,"TERMINAL",0,0,"7\t",,terminal_output +150,309711,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jafar",,terminal_output +151,310753,"TERMINAL",0,0,"queue",,terminal_command +152,310850,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1991.localdomain: Sun Sep 7 11:44:38 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3472612 accelerat train_dy tum_cte0 R\t1:13\t 1 hkn08043472614 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)",,terminal_output +153,311885,"TERMINAL",0,0,"94",,terminal_output +154,313017,"TERMINAL",0,0,"405",,terminal_output +155,313949,"TERMINAL",0,0,"16",,terminal_output +156,315051,"TERMINAL",0,0,"27",,terminal_output +157,316176,"TERMINAL",0,0,"39",,terminal_output +158,317040,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jafar",,terminal_output +159,318671,"TERMINAL",0,0,"fqueue",,terminal_command +160,318734,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue -o ""%.10i %.16P %.30j %.8u %.8T %.10M %.9l %.6D %R""hkn1991.localdomain: Sun Sep 7 11:44:46 2025JOBIDPARTITIONNAME USER STATE\t TIME TIME_LIMI NODES NODELIST(REASON)3472612\taccelerated train_dynamics_coinrun_og_repr tum_cte0 RUNNING\t 1:21 2-00:00:00\t1 hkn08043472614 accelerated-h100 train_dynamics_coinrun_og_repr tum_cte0 PENDING\t 0:00 2-00:00:00\t1 (Priority)\t\t",,terminal_output +161,319864,"TERMINAL",0,0,"72",,terminal_output +162,320814,"TERMINAL",0,0,"83",,terminal_output +163,321804,"TERMINAL",0,0,"94",,terminal_output +164,322838,"TERMINAL",0,0,"505",,terminal_output +165,323966,"TERMINAL",0,0,"16",,terminal_output +166,324996,"TERMINAL",0,0,"27",,terminal_output +167,325708,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jafar",,terminal_output +168,404460,"TERMINAL",0,0,"logs",,terminal_command +169,406286,"TERMINAL",0,0,"cd jafar_og_reproduction/",,terminal_command +170,406324,"TERMINAL",0,0,"]633;C]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction",,terminal_output +171,408616,"TERMINAL",0,0,"cd dynamics/",,terminal_command +172,410467,"TERMINAL",0,0,"ls",,terminal_command +173,412256,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics/train_dynamics_coinrun_og_reproduction_3472612.log",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=2-00:00:00\n#SBATCH --cpus-per-task=8\n#SBATCH --gres=gpu:1\n#SBATCH --partition=accelerated\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics/%x_%j.log\n#SBATCH --job-name=train_dynamics_coinrun_og_reproduction\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\ntags=""coinrun_og dynanmics 10m_dataset repro_mihir""\n\nnpy_records_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_coinrun/coinrun_episodes_10m""\n\n# TODO mihir: update the tokenizer and lam checkpoints\ntokenizer_ckpt_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286""\nlam_ckpt_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/lam/train_lam_coinrun_og_reproduction/3466287""\n\nCHECKPOINT_DIR=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/dynamics/${job_name}/${slurm_job_id}""\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --tokenizer_checkpoint=""${tokenizer_ckpt_dir}"" \\n --patch_size=16 \\n --lam_checkpoint=""${lam_ckpt_dir}"" \\n --log_image_interval=1000 \\n --log \\n --name=""${job_name}_${slurm_job_id}"" \\n --tags ${tags} \\n --entity instant-uv \\n --project jafar \\n --data_dir $npy_records_dir \\n --wandb_id $slurm_job_id\nSLURM_JOB_USER=tum_cte0515\nSLURM_TASKS_PER_NODE=1\nSLURM_JOB_UID=999226\nSLURM_TASK_PID=3909444\nSLURM_JOB_GPUS=3\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\nSLURMD_NODENAME=hkn0804\nSLURM_JOB_START_TIME=1757238205\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1757411005\nSLURM_CPUS_ON_NODE=8\nSLURM_JOB_CPUS_PER_NODE=8\nSLURM_GPUS_ON_NODE=1\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=accelerated\nSLURM_TRES_PER_TASK=cpu=8\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=1\nSLURM_JOBID=3472612\nSLURM_JOB_QOS=normal\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=8\nSLURM_NTASKS=1\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e27.hkn0804\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_NODELIST=hkn0804\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=1\nSLURM_NNODES=1\nSLURM_SUBMIT_HOST=hkn1991.localdomain\nSLURM_JOB_ID=3472612\nSLURM_NODEID=0\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_dynamics_coinrun_og_reproduction\nSLURM_NTASKS_PER_NODE=1\nSLURM_JOB_GID=502226\nSLURM_JOB_NODELIST=hkn0804\nGpuFreq=control_disabled\nwandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\nwandb: creating run\nwandb: Tracking run with wandb version 0.21.3\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250907_114419-3472612\nwandb: Run `wandb offline` to turn off syncing.\nwandb: Syncing run train_dynamics_coinrun_og_reproduction_3472612\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/3472612\n2025-09-07 11:44:35.257771: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-09-07 11:44:48.740866: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\n2025-09-07 11:44:52.925493: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 164, in \n init_params = restore_genie_components(\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 202, in restore_genie_components\n PyTreeCheckpointer().restore(tokenizer)[""model""][""params""][""params""]\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 306, in restore\n restored = self._restore(directory, args=ckpt_args)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 328, in _restore\n return self._handler.restore(directory, args=args)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 842, in restore\n structure, use_zarr3_metadata = self._get_internal_metadata(directory)\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 1000, in _get_internal_metadata\n raise FileNotFoundError(\nFileNotFoundError: No structure could be identified for the checkpoint at /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286.\nwandb: \nwandb: 🚀 View run train_dynamics_coinrun_og_reproduction_3472612 at: https://wandb.ai/instant-uv/jafar/runs/3472612\nwandb: Find logs at: ../../../../../hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250907_114419-3472612/logs\nsrun: error: hkn0804: task 0: Exited with exit code 1\n\n============================= JOB FEEDBACK =============================\n\nJob ID: 3472612\nCluster: hk\nUser/Group: tum_cte0515/hk-project-p0023960\nAccount: hk-project-p0023960\nState: FAILED (exit code 1)\nPartition: accelerated\nNodes: 1\nCores per node: 8\nNodelist: hkn0804\nCPU Utilized: 00:01:25\nCPU Efficiency: 11.42% of 00:12:24 core-walltime\nJob Wall-clock time: 00:01:33\nStarttime: Sun Sep 7 11:43:25 2025\nEndtime: Sun Sep 7 11:44:58 2025\nMemory Utilized: 1.54 GB\nMemory Efficiency: 0.00% of 0.00 MB\nEnergy Consumed: 65527 Joule / 18.2019444444444 Watthours\nAverage node power draw: 704.591397849462 Watt\n",log,tab +174,421706,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",0,0,"",shellscript,tab +175,423484,"TERMINAL",0,0,"queue",,terminal_command +176,423546,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1991.localdomain: Sun Sep 7 11:46:31 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3472614 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)",,terminal_output +177,424584,"TERMINAL",0,0,"2\t",,terminal_output +178,425677,"TERMINAL",0,0,"3\t",,terminal_output +179,426424,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics",,terminal_output +180,428599,"TERMINAL",0,0,"scancel 3472614",,terminal_command +181,428621,"TERMINAL",0,0,"]633;C]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/jafar_og_reproduction/dynamics",,terminal_output +182,434653,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,0,"",shellscript,selection_mouse +183,434854,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,1,"/",shellscript,selection_mouse +184,434854,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,180,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286""\nlam_ckpt_dir=""/hkfs/work/",shellscript,selection_mouse +185,434855,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,188,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286""\nlam_ckpt_dir=""/hkfs/work/workspac",shellscript,selection_mouse +186,434855,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,197,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286""\nlam_ckpt_dir=""/hkfs/work/workspace/scratch",shellscript,selection_mouse +187,434855,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,30,"/hkfs/work/workspace/scratch/t",shellscript,selection_mouse +188,434886,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,35,"/hkfs/work/workspace/scratch/tum_in",shellscript,selection_mouse +189,434908,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,49,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_",shellscript,selection_mouse +190,434936,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,54,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_share",shellscript,selection_mouse +191,434964,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,59,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/che",shellscript,selection_mouse +192,434994,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,64,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoi",shellscript,selection_mouse +193,435024,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,67,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints",shellscript,selection_mouse +194,435052,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,71,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mih",shellscript,selection_mouse +195,435071,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,75,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/j",shellscript,selection_mouse +196,435101,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,78,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafa",shellscript,selection_mouse +197,435138,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,81,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_o",shellscript,selection_mouse +198,435159,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,85,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_re",shellscript,selection_mouse +199,435184,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,88,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_repro",shellscript,selection_mouse +200,435210,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,89,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reprod",shellscript,selection_mouse +201,435241,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,92,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduct",shellscript,selection_mouse +202,435273,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,269,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286""\nlam_ckpt_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/lam/",shellscript,selection_mouse +203,435288,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,271,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286""\nlam_ckpt_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/lam/tr",shellscript,selection_mouse +204,435326,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,272,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286""\nlam_ckpt_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/lam/tra",shellscript,selection_mouse +205,435351,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,311,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286""\nlam_ckpt_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/lam/train_lam_coinrun_og_reproduction/3466287""",shellscript,selection_mouse +206,435518,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,154,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286""",shellscript,selection_mouse +207,436032,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",899,153,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286",shellscript,selection_mouse +208,442975,"TERMINAL",0,0,"cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286",,terminal_command +209,443576,"TERMINAL",0,0,"ls",,terminal_command +210,443648,"TERMINAL",0,0,"]633;C",,terminal_output +211,443825,"TERMINAL",0,0,"tokenizer_1757013407_10000 tokenizer_1757013407_160000 tokenizer_1757013407_220000 tokenizer_1757013407_30000\r\ntokenizer_1757013407_100000 tokenizer_1757013407_170000 tokenizer_1757013407_230000 tokenizer_1757013407_40000\r\ntokenizer_1757013407_110000 tokenizer_1757013407_180000 tokenizer_1757013407_240000 tokenizer_1757013407_50000\r\ntokenizer_1757013407_120000 tokenizer_1757013407_190000 tokenizer_1757013407_250000 tokenizer_1757013407_60000\r\ntokenizer_1757013407_130000 tokenizer_1757013407_20000 tokenizer_1757013407_260000 tokenizer_1757013407_70000\r\ntokenizer_1757013407_140000 tokenizer_1757013407_200000 tokenizer_1757013407_270000 tokenizer_1757013407_80000\r\ntokenizer_1757013407_150000 tokenizer_1757013407_210000 tokenizer_1757013407_280000 tokenizer_1757013407_90000\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/tokenizer/train_tokenizer_coinrun_og_reproduction/3466286",,terminal_output +212,454251,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1052,0,"",shellscript,selection_mouse +213,459391,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1052,0,"/",shellscript,content +214,459392,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1053,0,"",shellscript,selection_keyboard +215,459680,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1053,0,"tokenizer_1757013407_280000",shellscript,content +216,462565,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1160,0,"",shellscript,selection_mouse +217,462700,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1152,11,"checkpoints",shellscript,selection_mouse +218,462908,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1082,157,"lam_ckpt_dir=""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/lam/train_lam_coinrun_og_reproduction/3466287""\n",shellscript,selection_mouse +219,465526,"TERMINAL",0,0,"cd ..",,terminal_command +220,467671,"TERMINAL",0,0,"cd ..",,terminal_command +221,469384,"TERMINAL",0,0,"cd ..",,terminal_command +222,471152,"TERMINAL",0,0,"cd lam/train_lam_coinrun_og_reproduction/",,terminal_command +223,471813,"TERMINAL",0,0,"ls",,terminal_command +224,480487,"TERMINAL",0,0,"cd 3466287/",,terminal_command +225,481104,"TERMINAL",0,0,"ls",,terminal_command +226,481137,"TERMINAL",0,0,"]633;C",,terminal_output +227,481182,"TERMINAL",0,0,"lam_1757013407_10000 lam_1757013407_130000 lam_1757013407_170000 lam_1757013407_200000 lam_1757013407_60000\r\nlam_1757013407_100000 lam_1757013407_140000 lam_1757013407_180000 lam_1757013407_30000 lam_1757013407_70000\r\nlam_1757013407_110000 lam_1757013407_150000 lam_1757013407_190000 lam_1757013407_40000 lam_1757013407_80000\r\nlam_1757013407_120000 lam_1757013407_160000 lam_1757013407_20000 lam_1757013407_50000 lam_1757013407_90000\r\n]0;tum_cte0515@hkn1991:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/mihir/jafar_og_reproduction/lam/train_lam_coinrun_og_reproduction/3466287",,terminal_output +228,485679,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1237,0,"",shellscript,selection_mouse +229,486642,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1237,0,"/",shellscript,content +230,486643,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1238,0,"",shellscript,selection_keyboard +231,486888,"slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",1238,0,"lam_1757013407_200000",shellscript,content +232,498367,"TERMINAL",0,0,"jafar",,terminal_command +233,509143,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction.sbatch",,terminal_command +234,509184,"TERMINAL",0,0,"]633;CSubmitted batch job 3472617\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar",,terminal_output +235,510587,"TERMINAL",0,0,"queue",,terminal_command +236,510717,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1991.localdomain: Sun Sep 7 11:47:58 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3472617 accelerat train_dy tum_cte0 PD\t0:00\t 1 (None)",,terminal_output +237,511781,"TERMINAL",0,0,"9\t",,terminal_output +238,512554,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jafar",,terminal_output +239,516121,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/jafar_og_reproduction/og_coinrun_dynamics_reproduction-h100.sbatch",,terminal_command +240,516153,"TERMINAL",0,0,"]633;CSubmitted batch job 3472618\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar",,terminal_output +241,517374,"TERMINAL",0,0,"queue",,terminal_command +242,517459,"TERMINAL",0,0,"]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1991.localdomain: Sun Sep 7 11:48:05 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3472617 accelerat train_dy tum_cte0 PD\t0:00\t 1 (None)3472618 accelerat train_dy tum_cte0 PD\t0:00\t 1 (Priority)",,terminal_output +243,518536,"TERMINAL",0,0,"6\t",,terminal_output +244,519563,"TERMINAL",0,0,"7\t",,terminal_output +245,520561,"TERMINAL",0,0,"8\t",,terminal_output +246,521718,"TERMINAL",0,0,"9 Rhkn0709",,terminal_output +247,522634,"TERMINAL",0,0,"101",,terminal_output +248,523670,"TERMINAL",0,0,"12",,terminal_output +249,524744,"TERMINAL",0,0,"23",,terminal_output +250,525816,"TERMINAL",0,0,"34",,terminal_output +251,526793,"TERMINAL",0,0,"45",,terminal_output +252,527836,"TERMINAL",0,0,"56",,terminal_output +253,528971,"TERMINAL",0,0,"67",,terminal_output +254,529919,"TERMINAL",0,0,"78",,terminal_output +255,531026,"TERMINAL",0,0,"89",,terminal_output +256,532043,"TERMINAL",0,0,"910",,terminal_output +257,533168,"TERMINAL",0,0,"202",,terminal_output +258,534197,"TERMINAL",0,0,"23",,terminal_output +259,535125,"TERMINAL",0,0,"34",,terminal_output +260,536167,"TERMINAL",0,0,"45",,terminal_output +261,537209,"TERMINAL",0,0,"56",,terminal_output +262,538278,"TERMINAL",0,0,"67",,terminal_output +263,539288,"TERMINAL",0,0,"78",,terminal_output +264,540339,"TERMINAL",0,0,"89",,terminal_output +265,541369,"TERMINAL",0,0,"920",,terminal_output +266,542417,"TERMINAL",0,0,"301",,terminal_output +267,543463,"TERMINAL",0,0,"12",,terminal_output +268,544535,"TERMINAL",0,0,"23",,terminal_output +269,545552,"TERMINAL",0,0,"34",,terminal_output +270,546695,"TERMINAL",0,0,"45",,terminal_output +271,547623,"TERMINAL",0,0,"56",,terminal_output +272,548657,"TERMINAL",0,0,"67",,terminal_output +273,549701,"TERMINAL",0,0,"78",,terminal_output +274,550777,"TERMINAL",0,0,"89",,terminal_output +275,551811,"TERMINAL",0,0,"930",,terminal_output +276,552827,"TERMINAL",0,0,"401",,terminal_output +277,553864,"TERMINAL",0,0,"12",,terminal_output +278,554924,"TERMINAL",0,0,"23",,terminal_output +279,555944,"TERMINAL",0,0,"34",,terminal_output +280,557036,"TERMINAL",0,0,"45",,terminal_output +281,558035,"TERMINAL",0,0,"56",,terminal_output +282,559189,"TERMINAL",0,0,"68",,terminal_output +283,560195,"TERMINAL",0,0,"89",,terminal_output +284,561162,"TERMINAL",0,0,"940",,terminal_output +285,562212,"TERMINAL",0,0,"501",,terminal_output +286,563250,"TERMINAL",0,0,"12",,terminal_output +287,564283,"TERMINAL",0,0,"23",,terminal_output +288,565321,"TERMINAL",0,0,"34",,terminal_output +289,566368,"TERMINAL",0,0,"45",,terminal_output +290,567423,"TERMINAL",0,0,"56",,terminal_output +291,568561,"TERMINAL",0,0,"67",,terminal_output +292,569492,"TERMINAL",0,0,"78",,terminal_output +293,570527,"TERMINAL",0,0,"89",,terminal_output +294,571683,"TERMINAL",0,0,"950",,terminal_output +295,572607,"TERMINAL",0,0,"9:001",,terminal_output +296,573645,"TERMINAL",0,0,"12",,terminal_output +297,574763,"TERMINAL",0,0,"23",,terminal_output +298,575763,"TERMINAL",0,0,"34",,terminal_output +299,576798,"TERMINAL",0,0,"45",,terminal_output +300,577802,"TERMINAL",0,0,"56",,terminal_output +301,578840,"TERMINAL",0,0,"67",,terminal_output +302,579882,"TERMINAL",0,0,"78",,terminal_output +303,580910,"TERMINAL",0,0,"89",,terminal_output +304,581957,"TERMINAL",0,0,"91:00",,terminal_output +305,583037,"TERMINAL",0,0,"101",,terminal_output +306,584018,"TERMINAL",0,0,"12",,terminal_output +307,585064,"TERMINAL",0,0,"24",,terminal_output +308,586098,"TERMINAL",0,0,"45",,terminal_output +309,587142,"TERMINAL",0,0,"56",,terminal_output +310,588159,"TERMINAL",0,0,"67",,terminal_output +311,589204,"TERMINAL",0,0,"78",,terminal_output +312,590282,"TERMINAL",0,0,"89",,terminal_output +313,591282,"TERMINAL",0,0,"910",,terminal_output +314,592330,"TERMINAL",0,0,"201",,terminal_output +315,593379,"TERMINAL",0,0,"12",,terminal_output +316,594408,"TERMINAL",0,0,"23",,terminal_output +317,595440,"TERMINAL",0,0,"34",,terminal_output +318,596553,"TERMINAL",0,0,"45",,terminal_output +319,597525,"TERMINAL",0,0,"56",,terminal_output +320,598565,"TERMINAL",0,0,"67",,terminal_output +321,599626,"TERMINAL",0,0,"78",,terminal_output +322,600649,"TERMINAL",0,0,"89",,terminal_output +323,601744,"TERMINAL",0,0,"920",,terminal_output +324,602732,"TERMINAL",0,0,"301",,terminal_output +325,603783,"TERMINAL",0,0,"12",,terminal_output +326,604851,"TERMINAL",0,0,"23",,terminal_output +327,605859,"TERMINAL",0,0,"34",,terminal_output +328,606900,"TERMINAL",0,0,"45",,terminal_output +329,608037,"TERMINAL",0,0,"56",,terminal_output +330,609053,"TERMINAL",0,0,"67",,terminal_output +331,610066,"TERMINAL",0,0,"78",,terminal_output +332,611210,"TERMINAL",0,0,"830",,terminal_output +333,612114,"TERMINAL",0,0,"401",,terminal_output +334,613245,"TERMINAL",0,0,"12",,terminal_output +335,614212,"TERMINAL",0,0,"23",,terminal_output +336,615247,"TERMINAL",0,0,"34",,terminal_output +337,616286,"TERMINAL",0,0,"45",,terminal_output +338,617347,"TERMINAL",0,0,"56",,terminal_output +339,618375,"TERMINAL",0,0,"67",,terminal_output +340,619421,"TERMINAL",0,0,"78",,terminal_output +341,620495,"TERMINAL",0,0,"89",,terminal_output +342,621518,"TERMINAL",0,0,"940",,terminal_output +343,622546,"TERMINAL",0,0,"501",,terminal_output +344,623612,"TERMINAL",0,0,"12",,terminal_output +345,624743,"TERMINAL",0,0,"23",,terminal_output +346,625733,"TERMINAL",0,0,"34",,terminal_output +347,626791,"TERMINAL",0,0,"45",,terminal_output +348,627783,"TERMINAL",0,0,"56",,terminal_output +349,628837,"TERMINAL",0,0,"67",,terminal_output +350,629880,"TERMINAL",0,0,"78",,terminal_output +351,630921,"TERMINAL",0,0,"89",,terminal_output +352,632083,"TERMINAL",0,0,"950",,terminal_output +353,633106,"TERMINAL",0,0,"50:001",,terminal_output +354,634036,"TERMINAL",0,0,"12",,terminal_output +355,635151,"TERMINAL",0,0,"24",,terminal_output +356,636183,"TERMINAL",0,0,"45",,terminal_output +357,637170,"TERMINAL",0,0,"56",,terminal_output +358,638200,"TERMINAL",0,0,"67",,terminal_output +359,639307,"TERMINAL",0,0,"78",,terminal_output +360,640298,"TERMINAL",0,0,"89",,terminal_output +361,641361,"TERMINAL",0,0,"92:00",,terminal_output +362,642384,"TERMINAL",0,0,"101",,terminal_output +363,643438,"TERMINAL",0,0,"12",,terminal_output +364,644469,"TERMINAL",0,0,"23",,terminal_output +365,645616,"TERMINAL",0,0,"34",,terminal_output +366,646625,"TERMINAL",0,0,"45",,terminal_output +367,647597,"TERMINAL",0,0,"56",,terminal_output +368,648634,"TERMINAL",0,0,"67",,terminal_output +369,649701,"TERMINAL",0,0,"78",,terminal_output +370,650830,"TERMINAL",0,0,"89",,terminal_output +371,651784,"TERMINAL",0,0,"910",,terminal_output +372,652796,"TERMINAL",0,0,"201",,terminal_output +373,653905,"TERMINAL",0,0,"12",,terminal_output +374,654890,"TERMINAL",0,0,"23",,terminal_output +375,655916,"TERMINAL",0,0,"34",,terminal_output +376,656973,"TERMINAL",0,0,"45",,terminal_output +377,657998,"TERMINAL",0,0,"56",,terminal_output +378,659127,"TERMINAL",0,0,"67",,terminal_output +379,660081,"TERMINAL",0,0,"79",,terminal_output +380,661164,"TERMINAL",0,0,"920",,terminal_output +381,662194,"TERMINAL",0,0,"301",,terminal_output +382,663221,"TERMINAL",0,0,"12",,terminal_output +383,664237,"TERMINAL",0,0,"23",,terminal_output +384,665322,"TERMINAL",0,0,"34",,terminal_output +385,666338,"TERMINAL",0,0,"45",,terminal_output +386,667359,"TERMINAL",0,0,"56",,terminal_output +387,668410,"TERMINAL",0,0,"67",,terminal_output +388,669437,"TERMINAL",0,0,"78",,terminal_output +389,670480,"TERMINAL",0,0,"89",,terminal_output +390,671612,"TERMINAL",0,0,"930",,terminal_output +391,672620,"TERMINAL",0,0,"401",,terminal_output +392,673610,"TERMINAL",0,0,"12",,terminal_output +393,674692,"TERMINAL",0,0,"23",,terminal_output +394,675725,"TERMINAL",0,0,"34",,terminal_output +395,676707,"TERMINAL",0,0,"45",,terminal_output +396,677747,"TERMINAL",0,0,"56",,terminal_output +397,678797,"TERMINAL",0,0,"67",,terminal_output +398,679897,"TERMINAL",0,0,"78",,terminal_output +399,680858,"TERMINAL",0,0,"89",,terminal_output +400,681896,"TERMINAL",0,0,"940",,terminal_output +401,682994,"TERMINAL",0,0,"501",,terminal_output +402,683972,"TERMINAL",0,0,"12",,terminal_output +403,685008,"TERMINAL",0,0,"23",,terminal_output +404,686054,"TERMINAL",0,0,"35",,terminal_output +405,687081,"TERMINAL",0,0,"56",,terminal_output +406,688126,"TERMINAL",0,0,"67",,terminal_output +407,689157,"TERMINAL",0,0,"78",,terminal_output +408,690245,"TERMINAL",0,0,"89",,terminal_output +409,691244,"TERMINAL",0,0,"950",,terminal_output +410,692331,"TERMINAL",0,0,"1:001",,terminal_output +411,693312,"TERMINAL",0,0,"12",,terminal_output +412,694345,"TERMINAL",0,0,"23",,terminal_output +413,695380,"TERMINAL",0,0,"34",,terminal_output +414,696427,"TERMINAL",0,0,"45",,terminal_output +415,697464,"TERMINAL",0,0,"56",,terminal_output +416,698494,"TERMINAL",0,0,"67",,terminal_output +417,699546,"TERMINAL",0,0,"78",,terminal_output +418,700596,"TERMINAL",0,0,"89",,terminal_output +419,701722,"TERMINAL",0,0,"93:00",,terminal_output +420,702645,"TERMINAL",0,0,"101",,terminal_output +421,703693,"TERMINAL",0,0,"12",,terminal_output +422,704789,"TERMINAL",0,0,"23",,terminal_output +423,705761,"TERMINAL",0,0,"34",,terminal_output +424,706810,"TERMINAL",0,0,"45",,terminal_output +425,707846,"TERMINAL",0,0,"56",,terminal_output +426,709015,"TERMINAL",0,0,"67",,terminal_output +427,709912,"TERMINAL",0,0,"78",,terminal_output +428,710946,"TERMINAL",0,0,"89",,terminal_output +429,712072,"TERMINAL",0,0,"910",,terminal_output +430,713030,"TERMINAL",0,0,"201",,terminal_output +431,714110,"TERMINAL",0,0,"13",,terminal_output +432,715097,"TERMINAL",0,0,"34",,terminal_output +433,716157,"TERMINAL",0,0,"45",,terminal_output +434,717178,"TERMINAL",0,0,"56",,terminal_output +435,718198,"TERMINAL",0,0,"67",,terminal_output +436,719334,"TERMINAL",0,0,"78",,terminal_output +437,720290,"TERMINAL",0,0,"89",,terminal_output +438,721327,"TERMINAL",0,0,"920",,terminal_output +439,722376,"TERMINAL",0,0,"301",,terminal_output +440,723397,"TERMINAL",0,0,"12",,terminal_output +441,724441,"TERMINAL",0,0,"23",,terminal_output +442,725508,"TERMINAL",0,0,"34",,terminal_output +443,726530,"TERMINAL",0,0,"45",,terminal_output +444,727579,"TERMINAL",0,0,"56",,terminal_output +445,728604,"TERMINAL",0,0,"67",,terminal_output +446,729674,"TERMINAL",0,0,"78",,terminal_output +447,730810,"TERMINAL",0,0,"89",,terminal_output +448,731763,"TERMINAL",0,0,"930",,terminal_output +449,732762,"TERMINAL",0,0,"401",,terminal_output +450,733808,"TERMINAL",0,0,"12",,terminal_output +451,734905,"TERMINAL",0,0,"23",,terminal_output +452,735913,"TERMINAL",0,0,"34",,terminal_output +453,737005,"TERMINAL",0,0,"45",,terminal_output +454,737998,"TERMINAL",0,0,"56",,terminal_output +455,739095,"TERMINAL",0,0,"67",,terminal_output +456,740147,"TERMINAL",0,0,"79",,terminal_output +457,741124,"TERMINAL",0,0,"940",,terminal_output +458,742144,"TERMINAL",0,0,"501",,terminal_output +459,743178,"TERMINAL",0,0,"12",,terminal_output +460,744826,"TERMINAL",0,0,"23",,terminal_output +461,745875,"TERMINAL",0,0,"34",,terminal_output +462,746849,"TERMINAL",0,0,"45",,terminal_output +463,747825,"TERMINAL",0,0,"56",,terminal_output +464,748925,"TERMINAL",0,0,"67",,terminal_output +465,749945,"TERMINAL",0,0,"78",,terminal_output +466,750942,"TERMINAL",0,0,"89",,terminal_output +467,751989,"TERMINAL",0,0,"950",,terminal_output +468,753166,"TERMINAL",0,0,"2:001",,terminal_output +469,754152,"TERMINAL",0,0,"13",,terminal_output +470,755095,"TERMINAL",0,0,"34",,terminal_output +471,756131,"TERMINAL",0,0,"45",,terminal_output +472,757176,"TERMINAL",0,0,"56",,terminal_output +473,758209,"TERMINAL",0,0,"67",,terminal_output +474,759266,"TERMINAL",0,0,"78",,terminal_output +475,760353,"TERMINAL",0,0,"89",,terminal_output +476,761323,"TERMINAL",0,0,"94:00",,terminal_output +477,762380,"TERMINAL",0,0,"101",,terminal_output +478,763390,"TERMINAL",0,0,"12",,terminal_output +479,764433,"TERMINAL",0,0,"23",,terminal_output +480,765468,"TERMINAL",0,0,"34",,terminal_output +481,766547,"TERMINAL",0,0,"45",,terminal_output +482,767580,"TERMINAL",0,0,"56",,terminal_output +483,768581,"TERMINAL",0,0,"67",,terminal_output +484,769712,"TERMINAL",0,0,"78",,terminal_output +485,770735,"TERMINAL",0,0,"89",,terminal_output +486,771729,"TERMINAL",0,0,"910",,terminal_output +487,772743,"TERMINAL",0,0,"201",,terminal_output +488,773777,"TERMINAL",0,0,"12",,terminal_output +489,774857,"TERMINAL",0,0,"23",,terminal_output +490,775898,"TERMINAL",0,0,"34",,terminal_output +491,776897,"TERMINAL",0,0,"45",,terminal_output +492,777945,"TERMINAL",0,0,"56",,terminal_output +493,779042,"TERMINAL",0,0,"67",,terminal_output +494,780014,"TERMINAL",0,0,"78",,terminal_output +495,781055,"TERMINAL",0,0,"820",,terminal_output +496,782081,"TERMINAL",0,0,"301",,terminal_output +497,783164,"TERMINAL",0,0,"12",,terminal_output +498,784162,"TERMINAL",0,0,"23",,terminal_output +499,785205,"TERMINAL",0,0,"34",,terminal_output +500,786251,"TERMINAL",0,0,"45",,terminal_output +501,787264,"TERMINAL",0,0,"56",,terminal_output +502,788367,"TERMINAL",0,0,"67",,terminal_output +503,789338,"TERMINAL",0,0,"78",,terminal_output +504,790376,"TERMINAL",0,0,"89",,terminal_output +505,791447,"TERMINAL",0,0,"930",,terminal_output +506,792468,"TERMINAL",0,0,"401",,terminal_output +507,793536,"TERMINAL",0,0,"12",,terminal_output +508,794601,"TERMINAL",0,0,"23",,terminal_output +509,795591,"TERMINAL",0,0,"34",,terminal_output +510,796640,"TERMINAL",0,0,"45",,terminal_output +511,797651,"TERMINAL",0,0,"56",,terminal_output +512,798681,"TERMINAL",0,0,"67",,terminal_output +513,799748,"TERMINAL",0,0,"78",,terminal_output +514,800787,"TERMINAL",0,0,"89",,terminal_output +515,801824,"TERMINAL",0,0,"940",,terminal_output +516,802861,"TERMINAL",0,0,"501",,terminal_output +517,803931,"TERMINAL",0,0,"12",,terminal_output +518,804942,"TERMINAL",0,0,"23",,terminal_output +519,806020,"TERMINAL",0,0,"34",,terminal_output +520,807010,"TERMINAL",0,0,"45",,terminal_output +521,808065,"TERMINAL",0,0,"57",,terminal_output +522,809113,"TERMINAL",0,0,"78",,terminal_output +523,810156,"TERMINAL",0,0,"89",,terminal_output +524,811175,"TERMINAL",0,0,"950",,terminal_output +525,812352,"TERMINAL",0,0,"3:001",,terminal_output +526,813251,"TERMINAL",0,0,"12",,terminal_output +527,814294,"TERMINAL",0,0,"23",,terminal_output +528,815407,"TERMINAL",0,0,"34",,terminal_output +529,816403,"TERMINAL",0,0,"45",,terminal_output +530,817408,"TERMINAL",0,0,"56",,terminal_output +531,818444,"TERMINAL",0,0,"67",,terminal_output +532,819487,"TERMINAL",0,0,"78",,terminal_output +533,820609,"TERMINAL",0,0,"89",,terminal_output +534,821575,"TERMINAL",0,0,"95:00",,terminal_output +535,822614,"TERMINAL",0,0,"101",,terminal_output +536,823652,"TERMINAL",0,0,"12",,terminal_output +537,824715,"TERMINAL",0,0,"23",,terminal_output +538,825729,"TERMINAL",0,0,"34",,terminal_output +539,826771,"TERMINAL",0,0,"45",,terminal_output +540,827808,"TERMINAL",0,0,"56",,terminal_output +541,828906,"TERMINAL",0,0,"67",,terminal_output +542,829885,"TERMINAL",0,0,"78",,terminal_output +543,830941,"TERMINAL",0,0,"89",,terminal_output +544,832075,"TERMINAL",0,0,"910",,terminal_output +545,833007,"TERMINAL",0,0,"201",,terminal_output +546,834146,"TERMINAL",0,0,"13",,terminal_output +547,835085,"TERMINAL",0,0,"34",,terminal_output +548,836175,"TERMINAL",0,0,"45",,terminal_output +549,837185,"TERMINAL",0,0,"56",,terminal_output +550,838222,"TERMINAL",0,0,"67",,terminal_output +551,839250,"TERMINAL",0,0,"78",,terminal_output +552,840286,"TERMINAL",0,0,"89",,terminal_output +553,841373,"TERMINAL",0,0,"920",,terminal_output +554,842409,"TERMINAL",0,0,"301",,terminal_output +555,843411,"TERMINAL",0,0,"12",,terminal_output +556,844470,"TERMINAL",0,0,"23",,terminal_output +557,845545,"TERMINAL",0,0,"34",,terminal_output +558,846538,"TERMINAL",0,0,"45",,terminal_output +559,847597,"TERMINAL",0,0,"56",,terminal_output +560,848635,"TERMINAL",0,0,"67",,terminal_output +561,849770,"TERMINAL",0,0,"78",,terminal_output +562,850809,"TERMINAL",0,0,"89",,terminal_output +563,851766,"TERMINAL",0,0,"930",,terminal_output +564,852800,"TERMINAL",0,0,"401",,terminal_output +565,853846,"TERMINAL",0,0,"12",,terminal_output +566,856653,"TERMINAL",0,0,"25",,terminal_output +567,857650,"TERMINAL",0,0,"56",,terminal_output +568,858687,"TERMINAL",0,0,"67",,terminal_output +569,859834,"TERMINAL",0,0,"78",,terminal_output +570,860788,"TERMINAL",0,0,"89",,terminal_output +571,861828,"TERMINAL",0,0,"940",,terminal_output +572,863005,"TERMINAL",0,0,"501",,terminal_output +573,864023,"TERMINAL",0,0,"12",,terminal_output +574,865101,"TERMINAL",0,0,"23",,terminal_output +575,866011,"TERMINAL",0,0,"34",,terminal_output +576,867099,"TERMINAL",0,0,"46",,terminal_output +577,868122,"TERMINAL",0,0,"67",,terminal_output +578,869241,"TERMINAL",0,0,"78",,terminal_output +579,870192,"TERMINAL",0,0,"89",,terminal_output +580,871291,"TERMINAL",0,0,"950",,terminal_output +581,872316,"TERMINAL",0,0,"4:001",,terminal_output +582,873316,"TERMINAL",0,0,"12",,terminal_output +583,874370,"TERMINAL",0,0,"23",,terminal_output +584,875419,"TERMINAL",0,0,"34",,terminal_output +585,876440,"TERMINAL",0,0,"45",,terminal_output +586,877486,"TERMINAL",0,0,"56",,terminal_output +587,878527,"TERMINAL",0,0,"67",,terminal_output +588,879570,"TERMINAL",0,0,"78",,terminal_output +589,880618,"TERMINAL",0,0,"89",,terminal_output +590,881660,"TERMINAL",0,0,"96:00",,terminal_output +591,882708,"TERMINAL",0,0,"101",,terminal_output +592,883746,"TERMINAL",0,0,"12",,terminal_output +593,884783,"TERMINAL",0,0,"23",,terminal_output +594,885845,"TERMINAL",0,0,"34",,terminal_output +595,886862,"TERMINAL",0,0,"45",,terminal_output +596,887983,"TERMINAL",0,0,"56",,terminal_output +597,889020,"TERMINAL",0,0,"67",,terminal_output +598,889987,"TERMINAL",0,0,"78",,terminal_output +599,891058,"TERMINAL",0,0,"89",,terminal_output +600,892189,"TERMINAL",0,0,"911",,terminal_output +601,893205,"TERMINAL",0,0,"212",,terminal_output +602,894147,"TERMINAL",0,0,"23",,terminal_output +603,895177,"TERMINAL",0,0,"34",,terminal_output +604,896236,"TERMINAL",0,0,"45",,terminal_output +605,897274,"TERMINAL",0,0,"56",,terminal_output +606,898275,"TERMINAL",0,0,"67",,terminal_output +607,899312,"TERMINAL",0,0,"78",,terminal_output +608,900357,"TERMINAL",0,0,"89",,terminal_output +609,901395,"TERMINAL",0,0,"920",,terminal_output +610,902447,"TERMINAL",0,0,"301",,terminal_output +611,903473,"TERMINAL",0,0,"12",,terminal_output +612,904730,"TERMINAL",0,0,"23",,terminal_output +613,905617,"TERMINAL",0,0,"34",,terminal_output +614,906620,"TERMINAL",0,0,"45",,terminal_output +615,907617,"TERMINAL",0,0,"56",,terminal_output +616,908665,"TERMINAL",0,0,"67",,terminal_output +617,909797,"TERMINAL",0,0,"78",,terminal_output +618,910920,"TERMINAL",0,0,"89",,terminal_output +619,911784,"TERMINAL",0,0,"930",,terminal_output +620,912814,"TERMINAL",0,0,"401",,terminal_output +621,913844,"TERMINAL",0,0,"12",,terminal_output +622,914971,"TERMINAL",0,0,"23",,terminal_output +623,915991,"TERMINAL",0,0,"34",,terminal_output +624,917073,"TERMINAL",0,0,"45",,terminal_output +625,918088,"TERMINAL",0,0,"56",,terminal_output +626,919111,"TERMINAL",0,0,"67",,terminal_output +627,920134,"TERMINAL",0,0,"79",,terminal_output +628,921101,"TERMINAL",0,0,"940",,terminal_output +629,922150,"TERMINAL",0,0,"501",,terminal_output +630,923224,"TERMINAL",0,0,"12",,terminal_output +631,924263,"TERMINAL",0,0,"23",,terminal_output +632,925418,"TERMINAL",0,0,"34",,terminal_output +633,926413,"TERMINAL",0,0,"45",,terminal_output +634,927387,"TERMINAL",0,0,"56",,terminal_output +635,928404,"TERMINAL",0,0,"67",,terminal_output +636,929445,"TERMINAL",0,0,"78",,terminal_output +637,930468,"TERMINAL",0,0,"89",,terminal_output +638,931522,"TERMINAL",0,0,"950",,terminal_output +639,932548,"TERMINAL",0,0,"5:001",,terminal_output +640,933596,"TERMINAL",0,0,"12",,terminal_output +641,934652,"TERMINAL",0,0,"23",,terminal_output +642,935726,"TERMINAL",0,0,"34",,terminal_output +643,936718,"TERMINAL",0,0,"45",,terminal_output +644,937735,"TERMINAL",0,0,"56",,terminal_output +645,938826,"TERMINAL",0,0,"67",,terminal_output +646,939841,"TERMINAL",0,0,"78",,terminal_output +647,940854,"TERMINAL",0,0,"89",,terminal_output +648,941880,"TERMINAL",0,0,"97:00",,terminal_output +649,942975,"TERMINAL",0,0,"101",,terminal_output +650,944001,"TERMINAL",0,0,"12",,terminal_output +651,945131,"TERMINAL",0,0,"23",,terminal_output +652,946150,"TERMINAL",0,0,"34",,terminal_output +653,947148,"TERMINAL",0,0,"46",,terminal_output +654,948207,"TERMINAL",0,0,"67",,terminal_output +655,949234,"TERMINAL",0,0,"78",,terminal_output +656,950193,"TERMINAL",0,0,"89",,terminal_output +657,951220,"TERMINAL",0,0,"910",,terminal_output +658,952297,"TERMINAL",0,0,"201",,terminal_output +659,953322,"TERMINAL",0,0,"12",,terminal_output +660,954330,"TERMINAL",0,0,"23",,terminal_output +661,955384,"TERMINAL",0,0,"34",,terminal_output +662,956414,"TERMINAL",0,0,"45",,terminal_output +663,957454,"TERMINAL",0,0,"56",,terminal_output +664,958494,"TERMINAL",0,0,"67",,terminal_output +665,959534,"TERMINAL",0,0,"78",,terminal_output +666,960573,"TERMINAL",0,0,"89",,terminal_output +667,961712,"TERMINAL",0,0,"920",,terminal_output +668,962644,"TERMINAL",0,0,"301",,terminal_output +669,963684,"TERMINAL",0,0,"12",,terminal_output +670,964799,"TERMINAL",0,0,"23",,terminal_output +671,965752,"TERMINAL",0,0,"34",,terminal_output +672,966821,"TERMINAL",0,0,"45",,terminal_output +673,967957,"TERMINAL",0,0,"56",,terminal_output +674,968871,"TERMINAL",0,0,"67",,terminal_output +675,969917,"TERMINAL",0,0,"78",,terminal_output +676,970953,"TERMINAL",0,0,"89",,terminal_output +677,972064,"TERMINAL",0,0,"930",,terminal_output +678,973079,"TERMINAL",0,0,"402",,terminal_output +679,974073,"TERMINAL",0,0,"23",,terminal_output +680,975228,"TERMINAL",0,0,"34",,terminal_output +681,976266,"TERMINAL",0,0,"45",,terminal_output +682,977191,"TERMINAL",0,0,"56",,terminal_output +683,978236,"TERMINAL",0,0,"67",,terminal_output +684,979262,"TERMINAL",0,0,"78",,terminal_output +685,980464,"TERMINAL",0,0,"89",,terminal_output +686,981332,"TERMINAL",0,0,"940",,terminal_output +687,982375,"TERMINAL",0,0,"501",,terminal_output +688,983443,"TERMINAL",0,0,"12",,terminal_output +689,984445,"TERMINAL",0,0,"23",,terminal_output +690,985484,"TERMINAL",0,0,"34",,terminal_output +691,986530,"TERMINAL",0,0,"45",,terminal_output +692,987566,"TERMINAL",0,0,"56",,terminal_output +693,988592,"TERMINAL",0,0,"67",,terminal_output +694,989687,"TERMINAL",0,0,"78",,terminal_output +695,990751,"TERMINAL",0,0,"89",,terminal_output +696,991725,"TERMINAL",0,0,"950",,terminal_output +697,992763,"TERMINAL",0,0,"6:001",,terminal_output +698,993788,"TERMINAL",0,0,"12",,terminal_output +699,994897,"TERMINAL",0,0,"23",,terminal_output +700,995913,"TERMINAL",0,0,"34",,terminal_output +701,996923,"TERMINAL",0,0,"45",,terminal_output +702,998065,"TERMINAL",0,0,"56",,terminal_output +703,999097,"TERMINAL",0,0,"67",,terminal_output +704,1000110,"TERMINAL",0,0,"79",,terminal_output +705,1001092,"TERMINAL",0,0,"98:00",,terminal_output +706,1002168,"TERMINAL",0,0,"101",,terminal_output +707,1003177,"TERMINAL",0,0,"12",,terminal_output +708,1004223,"TERMINAL",0,0,"23",,terminal_output +709,1005263,"TERMINAL",0,0,"34",,terminal_output +710,1006310,"TERMINAL",0,0,"45",,terminal_output +711,1007432,"TERMINAL",0,0,"56",,terminal_output +712,1008403,"TERMINAL",0,0,"67",,terminal_output +713,1009465,"TERMINAL",0,0,"78",,terminal_output +714,1010506,"TERMINAL",0,0,"89",,terminal_output +715,1011528,"TERMINAL",0,0,"910",,terminal_output +716,1012583,"TERMINAL",0,0,"201",,terminal_output +717,1013618,"TERMINAL",0,0,"12",,terminal_output +718,1014789,"TERMINAL",0,0,"23",,terminal_output +719,1015778,"TERMINAL",0,0,"34",,terminal_output +720,1016782,"TERMINAL",0,0,"45",,terminal_output +721,1017803,"TERMINAL",0,0,"56",,terminal_output +722,1018960,"TERMINAL",0,0,"67",,terminal_output +723,1019891,"TERMINAL",0,0,"78",,terminal_output +724,1020933,"TERMINAL",0,0,"89",,terminal_output +725,1022038,"TERMINAL",0,0,"920",,terminal_output +726,1023052,"TERMINAL",0,0,"301",,terminal_output +727,1024067,"TERMINAL",0,0,"13",,terminal_output +728,1025110,"TERMINAL",0,0,"34",,terminal_output +729,1026151,"TERMINAL",0,0,"45",,terminal_output +730,1027196,"TERMINAL",0,0,"56",,terminal_output +731,1028281,"TERMINAL",0,0,"67",,terminal_output +732,1029460,"TERMINAL",0,0,"78",,terminal_output +733,1030330,"TERMINAL",0,0,"89",,terminal_output +734,1031440,"TERMINAL",0,0,"930",,terminal_output +735,1032412,"TERMINAL",0,0,"401",,terminal_output +736,1033475,"TERMINAL",0,0,"12",,terminal_output +737,1034507,"TERMINAL",0,0,"23",,terminal_output +738,1035545,"TERMINAL",0,0,"34",,terminal_output +739,1036596,"TERMINAL",0,0,"45",,terminal_output +740,1037641,"TERMINAL",0,0,"56",,terminal_output +741,1038680,"TERMINAL",0,0,"67",,terminal_output +742,1039840,"TERMINAL",0,0,"78",,terminal_output +743,1040808,"TERMINAL",0,0,"89",,terminal_output +744,1041811,"TERMINAL",0,0,"940",,terminal_output +745,1042864,"TERMINAL",0,0,"501",,terminal_output +746,1043937,"TERMINAL",0,0,"12",,terminal_output +747,1044968,"TERMINAL",0,0,"23",,terminal_output +748,1045993,"TERMINAL",0,0,"34",,terminal_output +749,1047119,"TERMINAL",0,0,"45",,terminal_output +750,1048078,"TERMINAL",0,0,"57",,terminal_output +751,1049145,"TERMINAL",0,0,"78",,terminal_output +752,1050292,"TERMINAL",0,0,"89",,terminal_output +753,1051317,"TERMINAL",0,0,"950",,terminal_output +754,1052343,"TERMINAL",0,0,"7:001",,terminal_output +755,1053376,"TERMINAL",0,0,"12",,terminal_output +756,1054339,"TERMINAL",0,0,"23",,terminal_output +757,1055379,"TERMINAL",0,0,"34",,terminal_output +758,1056458,"TERMINAL",0,0,"45",,terminal_output +759,1057469,"TERMINAL",0,0,"56",,terminal_output +760,1058515,"TERMINAL",0,0,"67",,terminal_output +761,1059541,"TERMINAL",0,0,"78",,terminal_output +762,1060604,"TERMINAL",0,0,"89",,terminal_output +763,1061703,"TERMINAL",0,0,"99:00",,terminal_output +764,1062664,"TERMINAL",0,0,"101",,terminal_output +765,1063715,"TERMINAL",0,0,"12",,terminal_output +766,1064800,"TERMINAL",0,0,"23",,terminal_output +767,1065796,"TERMINAL",0,0,"34",,terminal_output +768,1066839,"TERMINAL",0,0,"45",,terminal_output +769,1067907,"TERMINAL",0,0,"56",,terminal_output +770,1068987,"TERMINAL",0,0,"67",,terminal_output +771,1070052,"TERMINAL",0,0,"78",,terminal_output +772,1070995,"TERMINAL",0,0,"89",,terminal_output +773,1072041,"TERMINAL",0,0,"910",,terminal_output +774,1073131,"TERMINAL",0,0,"212",,terminal_output +775,1074153,"TERMINAL",0,0,"23",,terminal_output +776,1075157,"TERMINAL",0,0,"34",,terminal_output +777,1076197,"TERMINAL",0,0,"45",,terminal_output +778,1077238,"TERMINAL",0,0,"56",,terminal_output +779,1078278,"TERMINAL",0,0,"67",,terminal_output +780,1079317,"TERMINAL",0,0,"78",,terminal_output +781,1080359,"TERMINAL",0,0,"89",,terminal_output +782,1081436,"TERMINAL",0,0,"920",,terminal_output +783,1082492,"TERMINAL",0,0,"301",,terminal_output +784,1083462,"TERMINAL",0,0,"12",,terminal_output +785,1084502,"TERMINAL",0,0,"23",,terminal_output +786,1085537,"TERMINAL",0,0,"34",,terminal_output +787,1086572,"TERMINAL",0,0,"45",,terminal_output +788,1087611,"TERMINAL",0,0,"56",,terminal_output +789,1088652,"TERMINAL",0,0,"67",,terminal_output +790,1089827,"TERMINAL",0,0,"78",,terminal_output +791,1090786,"TERMINAL",0,0,"89",,terminal_output +792,1091800,"TERMINAL",0,0,"930",,terminal_output +793,1092912,"TERMINAL",0,0,"401",,terminal_output +794,1093885,"TERMINAL",0,0,"12",,terminal_output +795,1095039,"TERMINAL",0,0,"23",,terminal_output +796,1096087,"TERMINAL",0,0,"34",,terminal_output +797,1097088,"TERMINAL",0,0,"45",,terminal_output +798,1098044,"TERMINAL",0,0,"56",,terminal_output +799,1099070,"TERMINAL",0,0,"68",,terminal_output +800,1100156,"TERMINAL",0,0,"89",,terminal_output +801,1101152,"TERMINAL",0,0,"940",,terminal_output +802,1102312,"TERMINAL",0,0,"501",,terminal_output +803,1103332,"TERMINAL",0,0,"12",,terminal_output +804,1104261,"TERMINAL",0,0,"23",,terminal_output +805,1105379,"TERMINAL",0,0,"34",,terminal_output +806,1106455,"TERMINAL",0,0,"45",,terminal_output +807,1107468,"TERMINAL",0,0,"56",,terminal_output +808,1108638,"TERMINAL",0,0,"67",,terminal_output +809,1109608,"TERMINAL",0,0,"78",,terminal_output +810,1110581,"TERMINAL",0,0,"89",,terminal_output +811,1111549,"TERMINAL",0,0,"950",,terminal_output +812,1112606,"TERMINAL",0,0,"8:001",,terminal_output +813,1113630,"TERMINAL",0,0,"12",,terminal_output +814,1114696,"TERMINAL",0,0,"23",,terminal_output +815,1115719,"TERMINAL",0,0,"34",,terminal_output +816,1116779,"TERMINAL",0,0,"45",,terminal_output +817,1117777,"TERMINAL",0,0,"56",,terminal_output +818,1118820,"TERMINAL",0,0,"67",,terminal_output +819,1119918,"TERMINAL",0,0,"78",,terminal_output +820,1120893,"TERMINAL",0,0,"89",,terminal_output +821,1122074,"TERMINAL",0,0,"910:00",,terminal_output +822,1123066,"TERMINAL",0,0,"101",,terminal_output +823,1124118,"TERMINAL",0,0,"13",,terminal_output +824,1125131,"TERMINAL",0,0,"34",,terminal_output +825,1126127,"TERMINAL",0,0,"45",,terminal_output +826,1127167,"TERMINAL",0,0,"56",,terminal_output +827,1128358,"TERMINAL",0,0,"67",,terminal_output +828,1129340,"TERMINAL",0,0,"78",,terminal_output +829,1130370,"TERMINAL",0,0,"89",,terminal_output +830,1131398,"TERMINAL",0,0,"910",,terminal_output +831,1132355,"TERMINAL",0,0,"201",,terminal_output +832,1133435,"TERMINAL",0,0,"12",,terminal_output +833,1134469,"TERMINAL",0,0,"23",,terminal_output +834,1135464,"TERMINAL",0,0,"34",,terminal_output +835,1136502,"TERMINAL",0,0,"45",,terminal_output +836,1137538,"TERMINAL",0,0,"56",,terminal_output +837,1138578,"TERMINAL",0,0,"67",,terminal_output +838,1139617,"TERMINAL",0,0,"78",,terminal_output +839,1140808,"TERMINAL",0,0,"89",,terminal_output +840,1141704,"TERMINAL",0,0,"920",,terminal_output +841,1142740,"TERMINAL",0,0,"301",,terminal_output +842,1143779,"TERMINAL",0,0,"12",,terminal_output +843,1144904,"TERMINAL",0,0,"23",,terminal_output +844,1145868,"TERMINAL",0,0,"34",,terminal_output +845,1146890,"TERMINAL",0,0,"45",,terminal_output +846,1147974,"TERMINAL",0,0,"56",,terminal_output +847,1149204,"TERMINAL",0,0,"67",,terminal_output +848,1150023,"TERMINAL",0,0,"78",,terminal_output +849,1151066,"TERMINAL",0,0,"830",,terminal_output +850,1152155,"TERMINAL",0,0,"401",,terminal_output +851,1153167,"TERMINAL",0,0,"12",,terminal_output +852,1154198,"TERMINAL",0,0,"23",,terminal_output +853,1155353,"TERMINAL",0,0,"34",,terminal_output +854,1156377,"TERMINAL",0,0,"45",,terminal_output +855,1157703,"TERMINAL",0,0,"56",,terminal_output +856,1158740,"TERMINAL",0,0,"67",,terminal_output +857,1159855,"TERMINAL",0,0,"78",,terminal_output +858,1160846,"TERMINAL",0,0,"89",,terminal_output +859,1161857,"TERMINAL",0,0,"940",,terminal_output +860,1162895,"TERMINAL",0,0,"501",,terminal_output +861,1164062,"TERMINAL",0,0,"12",,terminal_output +862,1165081,"TERMINAL",0,0,"23",,terminal_output +863,1166105,"TERMINAL",0,0,"34",,terminal_output +864,1167123,"TERMINAL",0,0,"46",,terminal_output +865,1168154,"TERMINAL",0,0,"67",,terminal_output +866,1169177,"TERMINAL",0,0,"78",,terminal_output +867,1170314,"TERMINAL",0,0,"89",,terminal_output +868,1171322,"TERMINAL",0,0,"950",,terminal_output +869,1172353,"TERMINAL",0,0,"9:001",,terminal_output +870,1173399,"TERMINAL",0,0,"12",,terminal_output +871,1174399,"TERMINAL",0,0,"23",,terminal_output +872,1175465,"TERMINAL",0,0,"34",,terminal_output +873,1176461,"TERMINAL",0,0,"45",,terminal_output +874,1177514,"TERMINAL",0,0,"56",,terminal_output +875,1178511,"TERMINAL",0,0,"67",,terminal_output +876,1179528,"TERMINAL",0,0,"78",,terminal_output +877,1180569,"TERMINAL",0,0,"89",,terminal_output +878,1181620,"TERMINAL",0,0,"91:00",,terminal_output +879,1182648,"TERMINAL",0,0,"101",,terminal_output +880,1183702,"TERMINAL",0,0,"12",,terminal_output +881,1184737,"TERMINAL",0,0,"23",,terminal_output +882,1185760,"TERMINAL",0,0,"34",,terminal_output +883,1186824,"TERMINAL",0,0,"45",,terminal_output +884,1187924,"TERMINAL",0,0,"56",,terminal_output +885,1188947,"TERMINAL",0,0,"67",,terminal_output +886,1189918,"TERMINAL",0,0,"78",,terminal_output +887,1190962,"TERMINAL",0,0,"89",,terminal_output +888,1191997,"TERMINAL",0,0,"910",,terminal_output +889,1193134,"TERMINAL",0,0,"201",,terminal_output +890,1194185,"TERMINAL",0,0,"23",,terminal_output +891,1195185,"TERMINAL",0,0,"34",,terminal_output +892,1196207,"TERMINAL",0,0,"45",,terminal_output +893,1197186,"TERMINAL",0,0,"56",,terminal_output +894,1198254,"TERMINAL",0,0,"67",,terminal_output +895,1199290,"TERMINAL",0,0,"78",,terminal_output +896,1200392,"TERMINAL",0,0,"89",,terminal_output +897,1201431,"TERMINAL",0,0,"920",,terminal_output +898,1202455,"TERMINAL",0,0,"301",,terminal_output +899,1203641,"TERMINAL",0,0,"12",,terminal_output +900,1204475,"TERMINAL",0,0,"23",,terminal_output +901,1205532,"TERMINAL",0,0,"34",,terminal_output +902,1206569,"TERMINAL",0,0,"45",,terminal_output +903,1207585,"TERMINAL",0,0,"56",,terminal_output +904,1208626,"TERMINAL",0,0,"67",,terminal_output +905,1209654,"TERMINAL",0,0,"78",,terminal_output +906,1210711,"TERMINAL",0,0,"89",,terminal_output +907,1211775,"TERMINAL",0,0,"930",,terminal_output +908,1212785,"TERMINAL",0,0,"401",,terminal_output +909,1213835,"TERMINAL",0,0,"12",,terminal_output +910,1214858,"TERMINAL",0,0,"23",,terminal_output +911,1215914,"TERMINAL",0,0,"34",,terminal_output +912,1216941,"TERMINAL",0,0,"45",,terminal_output +913,1218056,"TERMINAL",0,0,"56",,terminal_output +914,1219021,"TERMINAL",0,0,"67",,terminal_output +915,1220078,"TERMINAL",0,0,"79",,terminal_output +916,1221118,"TERMINAL",0,0,"940",,terminal_output +917,1222141,"TERMINAL",0,0,"501",,terminal_output +918,1223186,"TERMINAL",0,0,"12",,terminal_output +919,1224245,"TERMINAL",0,0,"23",,terminal_output +920,1225252,"TERMINAL",0,0,"34",,terminal_output +921,1226418,"TERMINAL",0,0,"45",,terminal_output +922,1227341,"TERMINAL",0,0,"56",,terminal_output +923,1228524,"TERMINAL",0,0,"67",,terminal_output +924,1229418,"TERMINAL",0,0,"78",,terminal_output +925,1230514,"TERMINAL",0,0,"89",,terminal_output +926,1231508,"TERMINAL",0,0,"950",,terminal_output +927,1232557,"TERMINAL",0,0,"2:00:001",,terminal_output +928,1233598,"TERMINAL",0,0,"12",,terminal_output +929,1234616,"TERMINAL",0,0,"23",,terminal_output +930,1235680,"TERMINAL",0,0,"34",,terminal_output +931,1236735,"TERMINAL",0,0,"45",,terminal_output +932,1237726,"TERMINAL",0,0,"56",,terminal_output +933,1238792,"TERMINAL",0,0,"67",,terminal_output +934,1239805,"TERMINAL",0,0,"78",,terminal_output +935,1240853,"TERMINAL",0,0,"89",,terminal_output +936,1241880,"TERMINAL",0,0,"92:00",,terminal_output +937,1242941,"TERMINAL",0,0,"101",,terminal_output +938,1244029,"TERMINAL",0,0,"12",,terminal_output +939,1244999,"TERMINAL",0,0,"23",,terminal_output +940,1246043,"TERMINAL",0,0,"34",,terminal_output +941,1247102,"TERMINAL",0,0,"46",,terminal_output +942,1248134,"TERMINAL",0,0,"67",,terminal_output +943,1249204,"TERMINAL",0,0,"78",,terminal_output +944,1250274,"TERMINAL",0,0,"89",,terminal_output +945,1251301,"TERMINAL",0,0,"910",,terminal_output +946,1252266,"TERMINAL",0,0,"201",,terminal_output +947,1253361,"TERMINAL",0,0,"12",,terminal_output +948,1254407,"TERMINAL",0,0,"23",,terminal_output +949,1255593,"TERMINAL",0,0,"34",,terminal_output +950,1256520,"TERMINAL",0,0,"45",,terminal_output +951,1257506,"TERMINAL",0,0,"56",,terminal_output +952,1258567,"TERMINAL",0,0,"67",,terminal_output +953,1259539,"TERMINAL",0,0,"78",,terminal_output +954,1260585,"TERMINAL",0,0,"89",,terminal_output +955,1261623,"TERMINAL",0,0,"920",,terminal_output +956,1262655,"TERMINAL",0,0,"301",,terminal_output +957,1263703,"TERMINAL",0,0,"12",,terminal_output +958,1264824,"TERMINAL",0,0,"23",,terminal_output +959,1265845,"TERMINAL",0,0,"34",,terminal_output +960,1266846,"TERMINAL",0,0,"45",,terminal_output +961,1267867,"TERMINAL",0,0,"56",,terminal_output +962,1269018,"TERMINAL",0,0,"67",,terminal_output +963,1269981,"TERMINAL",0,0,"78",,terminal_output +964,1270985,"TERMINAL",0,0,"89",,terminal_output +965,1272089,"TERMINAL",0,0,"930",,terminal_output +966,1273110,"TERMINAL",0,0,"402",,terminal_output +967,1274089,"TERMINAL",0,0,"23",,terminal_output +968,1275118,"TERMINAL",0,0,"34",,terminal_output +969,1276171,"TERMINAL",0,0,"45",,terminal_output +970,1277198,"TERMINAL",0,0,"56",,terminal_output +971,1278339,"TERMINAL",0,0,"67",,terminal_output +972,1279360,"TERMINAL",0,0,"78",,terminal_output +973,1280386,"TERMINAL",0,0,"89",,terminal_output +974,1281403,"TERMINAL",0,0,"940",,terminal_output +975,1282398,"TERMINAL",0,0,"501",,terminal_output +976,1283445,"TERMINAL",0,0,"12",,terminal_output +977,1284502,"TERMINAL",0,0,"23",,terminal_output +978,1285509,"TERMINAL",0,0,"34",,terminal_output +979,1286545,"TERMINAL",0,0,"45",,terminal_output +980,1287600,"TERMINAL",0,0,"56",,terminal_output +981,1288627,"TERMINAL",0,0,"67",,terminal_output +982,1289659,"TERMINAL",0,0,"78",,terminal_output +983,1290732,"TERMINAL",0,0,"89",,terminal_output +984,1291732,"TERMINAL",0,0,"950",,terminal_output +985,1292775,"TERMINAL",0,0,"1:001",,terminal_output +986,1293806,"TERMINAL",0,0,"12",,terminal_output +987,1294921,"TERMINAL",0,0,"23",,terminal_output +988,1295892,"TERMINAL",0,0,"34",,terminal_output +989,1296932,"TERMINAL",0,0,"45",,terminal_output +990,1298000,"TERMINAL",0,0,"56",,terminal_output +991,1299018,"TERMINAL",0,0,"67",,terminal_output +992,1300055,"TERMINAL",0,0,"79",,terminal_output +993,1301087,"TERMINAL",0,0,"93:00",,terminal_output +994,1302170,"TERMINAL",0,0,"101",,terminal_output +995,1303216,"TERMINAL",0,0,"12",,terminal_output +996,1304242,"TERMINAL",0,0,"23",,terminal_output +997,1305239,"TERMINAL",0,0,"34",,terminal_output +998,1306384,"TERMINAL",0,0,"45",,terminal_output +999,1307423,"TERMINAL",0,0,"56",,terminal_output +1000,1308353,"TERMINAL",0,0,"67",,terminal_output +1001,1309503,"TERMINAL",0,0,"78",,terminal_output +1002,1310490,"TERMINAL",0,0,"89",,terminal_output +1003,1311490,"TERMINAL",0,0,"910",,terminal_output +1004,1312553,"TERMINAL",0,0,"201",,terminal_output +1005,1313559,"TERMINAL",0,0,"12",,terminal_output +1006,1314593,"TERMINAL",0,0,"23",,terminal_output +1007,1315649,"TERMINAL",0,0,"34",,terminal_output +1008,1316676,"TERMINAL",0,0,"45",,terminal_output +1009,1317722,"TERMINAL",0,0,"56",,terminal_output +1010,1318754,"TERMINAL",0,0,"67",,terminal_output +1011,1319796,"TERMINAL",0,0,"78",,terminal_output +1012,1320854,"TERMINAL",0,0,"89",,terminal_output +1013,1321867,"TERMINAL",0,0,"920",,terminal_output +1014,1322986,"TERMINAL",0,0,"301",,terminal_output +1015,1324004,"TERMINAL",0,0,"12",,terminal_output +1016,1325036,"TERMINAL",0,0,"23",,terminal_output +1017,1326115,"TERMINAL",0,0,"34",,terminal_output +1018,1327122,"TERMINAL",0,0,"46",,terminal_output +1019,1328202,"TERMINAL",0,0,"67",,terminal_output +1020,1329225,"TERMINAL",0,0,"78",,terminal_output +1021,1330253,"TERMINAL",0,0,"89",,terminal_output +1022,1331299,"TERMINAL",0,0,"930",,terminal_output +1023,1332270,"TERMINAL",0,0,"401",,terminal_output +1024,1333310,"TERMINAL",0,0,"12",,terminal_output +1025,1334406,"TERMINAL",0,0,"23",,terminal_output +1026,1335393,"TERMINAL",0,0,"34",,terminal_output +1027,1336436,"TERMINAL",0,0,"45",,terminal_output +1028,1337464,"TERMINAL",0,0,"56",,terminal_output +1029,1338518,"TERMINAL",0,0,"67",,terminal_output +1030,1339551,"TERMINAL",0,0,"78",,terminal_output +1031,1340590,"TERMINAL",0,0,"89",,terminal_output +1032,1341624,"TERMINAL",0,0,"940",,terminal_output +1033,1342666,"TERMINAL",0,0,"501",,terminal_output +1034,1343699,"TERMINAL",0,0,"12",,terminal_output +1035,1344798,"TERMINAL",0,0,"23",,terminal_output +1036,1345787,"TERMINAL",0,0,"34",,terminal_output +1037,1346840,"TERMINAL",0,0,"45",,terminal_output +1038,1347867,"TERMINAL",0,0,"56",,terminal_output +1039,1348988,"TERMINAL",0,0,"67",,terminal_output +1040,1350011,"TERMINAL",0,0,"78",,terminal_output +1041,1351079,"TERMINAL",0,0,"89",,terminal_output +1042,1352062,"TERMINAL",0,0,"951",,terminal_output +1043,1353189,"TERMINAL",0,0,"2:012",,terminal_output +1044,1354215,"TERMINAL",0,0,"23",,terminal_output +1045,1355238,"TERMINAL",0,0,"34",,terminal_output +1046,1356194,"TERMINAL",0,0,"45",,terminal_output +1047,1357248,"TERMINAL",0,0,"56",,terminal_output +1048,1358310,"TERMINAL",0,0,"67",,terminal_output +1049,1359320,"TERMINAL",0,0,"78",,terminal_output +1050,1360459,"TERMINAL",0,0,"89",,terminal_output +1051,1361537,"TERMINAL",0,0,"94:00",,terminal_output +1052,1362508,"TERMINAL",0,0,"101",,terminal_output +1053,1363524,"TERMINAL",0,0,"12",,terminal_output +1054,1364560,"TERMINAL",0,0,"23",,terminal_output +1055,1365573,"TERMINAL",0,0,"34",,terminal_output +1056,1366627,"TERMINAL",0,0,"45",,terminal_output +1057,1367640,"TERMINAL",0,0,"56",,terminal_output +1058,1368675,"TERMINAL",0,0,"67",,terminal_output +1059,1369776,"TERMINAL",0,0,"78",,terminal_output +1060,1370801,"TERMINAL",0,0,"89",,terminal_output +1061,1371790,"TERMINAL",0,0,"910",,terminal_output +1062,1372827,"TERMINAL",0,0,"201",,terminal_output +1063,1373978,"TERMINAL",0,0,"12",,terminal_output +1064,1374908,"TERMINAL",0,0,"23",,terminal_output +1065,1375939,"TERMINAL",0,0,"34",,terminal_output +1066,1377047,"TERMINAL",0,0,"45",,terminal_output +1067,1378139,"TERMINAL",0,0,"56",,terminal_output +1068,1379104,"TERMINAL",0,0,"68",,terminal_output +1069,1380114,"TERMINAL",0,0,"89",,terminal_output +1070,1381251,"TERMINAL",0,0,"920",,terminal_output +1071,1382178,"TERMINAL",0,0,"301",,terminal_output +1072,1383203,"TERMINAL",0,0,"12",,terminal_output +1073,1384324,"TERMINAL",0,0,"23",,terminal_output +1074,1385343,"TERMINAL",0,0,"34",,terminal_output +1075,1386370,"TERMINAL",0,0,"45",,terminal_output +1076,1387389,"TERMINAL",0,0,"56",,terminal_output +1077,1388413,"TERMINAL",0,0,"67",,terminal_output +1078,1389539,"TERMINAL",0,0,"78",,terminal_output +1079,1390475,"TERMINAL",0,0,"89",,terminal_output +1080,1391538,"TERMINAL",0,0,"930",,terminal_output +1081,1392578,"TERMINAL",0,0,"401",,terminal_output +1082,1393652,"TERMINAL",0,0,"12",,terminal_output +1083,1394683,"TERMINAL",0,0,"23",,terminal_output +1084,1395721,"TERMINAL",0,0,"34",,terminal_output +1085,1396791,"TERMINAL",0,0,"45",,terminal_output +1086,1397807,"TERMINAL",0,0,"56",,terminal_output +1087,1398846,"TERMINAL",0,0,"67",,terminal_output +1088,1399986,"TERMINAL",0,0,"78",,terminal_output +1089,1400933,"TERMINAL",0,0,"89",,terminal_output +1090,1402049,"TERMINAL",0,0,"940",,terminal_output +1091,1403064,"TERMINAL",0,0,"501",,terminal_output +1092,1404081,"TERMINAL",0,0,"13",,terminal_output +1093,1405208,"TERMINAL",0,0,"34",,terminal_output +1094,1406239,"TERMINAL",0,0,"45",,terminal_output +1095,1407191,"TERMINAL",0,0,"56",,terminal_output +1096,1408278,"TERMINAL",0,0,"67",,terminal_output +1097,1409405,"TERMINAL",0,0,"78",,terminal_output +1098,1410332,"TERMINAL",0,0,"89",,terminal_output +1099,1411380,"TERMINAL",0,0,"950",,terminal_output +1100,1412420,"TERMINAL",0,0,"3:001",,terminal_output +1101,1413463,"TERMINAL",0,0,"12",,terminal_output +1102,1414566,"TERMINAL",0,0,"23",,terminal_output +1103,1415550,"TERMINAL",0,0,"34",,terminal_output +1104,1416604,"TERMINAL",0,0,"45",,terminal_output +1105,1417637,"TERMINAL",0,0,"56",,terminal_output +1106,1418674,"TERMINAL",0,0,"67",,terminal_output +1107,1419748,"TERMINAL",0,0,"78",,terminal_output +1108,1420773,"TERMINAL",0,0,"89",,terminal_output +1109,1421798,"TERMINAL",0,0,"95:00",,terminal_output +1110,1422922,"TERMINAL",0,0,"101",,terminal_output +1111,1423946,"TERMINAL",0,0,"12",,terminal_output +1112,1424976,"TERMINAL",0,0,"23",,terminal_output +1113,1425979,"TERMINAL",0,0,"34",,terminal_output +1114,1427122,"TERMINAL",0,0,"45",,terminal_output +1115,1428144,"TERMINAL",0,0,"57",,terminal_output +1116,1429180,"TERMINAL",0,0,"78",,terminal_output +1117,1430142,"TERMINAL",0,0,"89",,terminal_output +1118,1431264,"TERMINAL",0,0,"910",,terminal_output +1119,1432233,"TERMINAL",0,0,"201",,terminal_output +1120,1433384,"TERMINAL",0,0,"12",,terminal_output +1121,1434313,"TERMINAL",0,0,"23",,terminal_output +1122,1435359,"TERMINAL",0,0,"34",,terminal_output +1123,1436441,"TERMINAL",0,0,"45",,terminal_output +1124,1437468,"TERMINAL",0,0,"56",,terminal_output +1125,1438500,"TERMINAL",0,0,"67",,terminal_output +1126,1439537,"TERMINAL",0,0,"78",,terminal_output +1127,1440581,"TERMINAL",0,0,"89",,terminal_output +1128,1441637,"TERMINAL",0,0,"920",,terminal_output +1129,1442752,"TERMINAL",0,0,"301",,terminal_output +1130,1443761,"TERMINAL",0,0,"12",,terminal_output +1131,1444751,"TERMINAL",0,0,"23",,terminal_output +1132,1445852,"TERMINAL",0,0,"34",,terminal_output +1133,1446932,"TERMINAL",0,0,"45",,terminal_output +1134,1447962,"TERMINAL",0,0,"56",,terminal_output +1135,1449027,"TERMINAL",0,0,"67",,terminal_output +1136,1450103,"TERMINAL",0,0,"78",,terminal_output +1137,1451035,"TERMINAL",0,0,"89",,terminal_output +1138,1452186,"TERMINAL",0,0,"931",,terminal_output +1139,1453228,"TERMINAL",0,0,"412",,terminal_output +1140,1454233,"TERMINAL",0,0,"23",,terminal_output +1141,1455333,"TERMINAL",0,0,"34",,terminal_output +1142,1456393,"TERMINAL",0,0,"45",,terminal_output +1143,1457278,"TERMINAL",0,0,"56",,terminal_output +1144,1458440,"TERMINAL",0,0,"67",,terminal_output +1145,1459460,"TERMINAL",0,0,"78",,terminal_output +1146,1460545,"TERMINAL",0,0,"89",,terminal_output +1147,1461611,"TERMINAL",0,0,"940",,terminal_output +1148,1462511,"TERMINAL",0,0,"501",,terminal_output +1149,1463608,"TERMINAL",0,0,"12",,terminal_output +1150,1464652,"TERMINAL",0,0,"23",,terminal_output +1151,1465619,"TERMINAL",0,0,"34",,terminal_output +1152,1466662,"TERMINAL",0,0,"45",,terminal_output +1153,1467713,"TERMINAL",0,0,"56",,terminal_output +1154,1468749,"TERMINAL",0,0,"67",,terminal_output +1155,1469788,"TERMINAL",0,0,"78",,terminal_output +1156,1470911,"TERMINAL",0,0,"89",,terminal_output +1157,1471890,"TERMINAL",0,0,"950",,terminal_output +1158,1473019,"TERMINAL",0,0,"4:001",,terminal_output +1159,1474040,"TERMINAL",0,0,"12",,terminal_output +1160,1475147,"TERMINAL",0,0,"23",,terminal_output +1161,1476068,"TERMINAL",0,0,"35",,terminal_output +1162,1477123,"TERMINAL",0,0,"56",,terminal_output +1163,1478226,"TERMINAL",0,0,"67",,terminal_output +1164,1479183,"TERMINAL",0,0,"78",,terminal_output +1165,1480215,"TERMINAL",0,0,"89",,terminal_output +1166,1481294,"TERMINAL",0,0,"96:00",,terminal_output +1167,1482310,"TERMINAL",0,0,"101",,terminal_output +1168,1483953,"TERMINAL",0,0,"12",,terminal_output +1169,1484917,"TERMINAL",0,0,"23",,terminal_output +1170,1485933,"TERMINAL",0,0,"34",,terminal_output +1171,1486976,"TERMINAL",0,0,"45",,terminal_output +1172,1488015,"TERMINAL",0,0,"56",,terminal_output +1173,1489077,"TERMINAL",0,0,"68",,terminal_output +1174,1490108,"TERMINAL",0,0,"89",,terminal_output +1175,1491225,"TERMINAL",0,0,"910",,terminal_output +1176,1492253,"TERMINAL",0,0,"201",,terminal_output +1177,1493291,"TERMINAL",0,0,"12",,terminal_output +1178,1494276,"TERMINAL",0,0,"23",,terminal_output +1179,1495333,"TERMINAL",0,0,"34",,terminal_output +1180,1496403,"TERMINAL",0,0,"45",,terminal_output +1181,1497404,"TERMINAL",0,0,"56",,terminal_output +1182,1498454,"TERMINAL",0,0,"67",,terminal_output +1183,1499587,"TERMINAL",0,0,"78",,terminal_output +1184,1500567,"TERMINAL",0,0,"89",,terminal_output +1185,1501582,"TERMINAL",0,0,"920",,terminal_output +1186,1502617,"TERMINAL",0,0,"301",,terminal_output +1187,1503666,"TERMINAL",0,0,"12",,terminal_output +1188,1504732,"TERMINAL",0,0,"23",,terminal_output +1189,1505810,"TERMINAL",0,0,"34",,terminal_output +1190,1506801,"TERMINAL",0,0,"45",,terminal_output +1191,1507920,"TERMINAL",0,0,"56",,terminal_output +1192,1508940,"TERMINAL",0,0,"67",,terminal_output +1193,1509976,"TERMINAL",0,0,"78",,terminal_output +1194,1510970,"TERMINAL",0,0,"89",,terminal_output +1195,1512117,"TERMINAL",0,0,"930",,terminal_output +1196,1513055,"TERMINAL",0,0,"402",,terminal_output +1197,1514169,"TERMINAL",0,0,"23",,terminal_output +1198,1515191,"TERMINAL",0,0,"34",,terminal_output +1199,1516209,"TERMINAL",0,0,"45",,terminal_output +1200,1517217,"TERMINAL",0,0,"56",,terminal_output +1201,1518256,"TERMINAL",0,0,"67",,terminal_output +1202,1519387,"TERMINAL",0,0,"78",,terminal_output +1203,1520408,"TERMINAL",0,0,"89",,terminal_output +1204,1521388,"TERMINAL",0,0,"940",,terminal_output +1205,1522466,"TERMINAL",0,0,"501",,terminal_output +1206,1523500,"TERMINAL",0,0,"12",,terminal_output +1207,1524607,"TERMINAL",0,0,"23",,terminal_output +1208,1525580,"TERMINAL",0,0,"34",,terminal_output +1209,1526615,"TERMINAL",0,0,"45",,terminal_output +1210,1527669,"TERMINAL",0,0,"56",,terminal_output +1211,1528722,"TERMINAL",0,0,"67",,terminal_output +1212,1529755,"TERMINAL",0,0,"78",,terminal_output +1213,1530938,"TERMINAL",0,0,"89",,terminal_output +1214,1531829,"TERMINAL",0,0,"950",,terminal_output +1215,1532868,"TERMINAL",0,0,"5:001",,terminal_output +1216,1534034,"TERMINAL",0,0,"12",,terminal_output +1217,1535053,"TERMINAL",0,0,"23",,terminal_output +1218,1535995,"TERMINAL",0,0,"34",,terminal_output +1219,1537100,"TERMINAL",0,0,"45",,terminal_output +1220,1538079,"TERMINAL",0,0,"57",,terminal_output +1221,1539115,"TERMINAL",0,0,"78",,terminal_output +1222,1540167,"TERMINAL",0,0,"89",,terminal_output +1223,1541198,"TERMINAL",0,0,"97:00",,terminal_output +1224,1542322,"TERMINAL",0,0,"101",,terminal_output +1225,1543347,"TERMINAL",0,0,"12",,terminal_output +1226,1544319,"TERMINAL",0,0,"23",,terminal_output +1227,1545393,"TERMINAL",0,0,"34",,terminal_output +1228,1546519,"TERMINAL",0,0,"45",,terminal_output +1229,1547611,"TERMINAL",0,0,"56",,terminal_output +1230,1548519,"TERMINAL",0,0,"67",,terminal_output +1231,1549529,"TERMINAL",0,0,"78",,terminal_output +1232,1550582,"TERMINAL",0,0,"89",,terminal_output +1233,1551613,"TERMINAL",0,0,"910",,terminal_output +1234,1552649,"TERMINAL",0,0,"201",,terminal_output +1235,1553697,"TERMINAL",0,0,"12",,terminal_output +1236,1554735,"TERMINAL",0,0,"23",,terminal_output +1237,1555848,"TERMINAL",0,0,"34",,terminal_output +1238,1556821,"TERMINAL",0,0,"45",,terminal_output +1239,1557863,"TERMINAL",0,0,"56",,terminal_output +1240,1558907,"TERMINAL",0,0,"67",,terminal_output +1241,1559941,"TERMINAL",0,0,"78",,terminal_output +1242,1560981,"TERMINAL",0,0,"89",,terminal_output +1243,1562196,"TERMINAL",0,0,"920",,terminal_output +1244,1563063,"TERMINAL",0,0,"302",,terminal_output +1245,1564139,"TERMINAL",0,0,"23",,terminal_output +1246,1565258,"TERMINAL",0,0,"34",,terminal_output +1247,1566290,"TERMINAL",0,0,"45",,terminal_output +1248,1567234,"TERMINAL",0,0,"56",,terminal_output +1249,1568334,"TERMINAL",0,0,"67",,terminal_output +1250,1569324,"TERMINAL",0,0,"78",,terminal_output +1251,1570388,"TERMINAL",0,0,"89",,terminal_output +1252,1571548,"TERMINAL",0,0,"930",,terminal_output +1253,1572605,"TERMINAL",0,0,"401",,terminal_output +1254,1573523,"TERMINAL",0,0,"12",,terminal_output +1255,1574515,"TERMINAL",0,0,"23",,terminal_output +1256,1575586,"TERMINAL",0,0,"34",,terminal_output +1257,1576632,"TERMINAL",0,0,"45",,terminal_output +1258,1577684,"TERMINAL",0,0,"56",,terminal_output +1259,1578680,"TERMINAL",0,0,"67",,terminal_output +1260,1579800,"TERMINAL",0,0,"78",,terminal_output +1261,1580775,"TERMINAL",0,0,"89",,terminal_output +1262,1581809,"TERMINAL",0,0,"940",,terminal_output +1263,1582978,"TERMINAL",0,0,"501",,terminal_output +1264,1584000,"TERMINAL",0,0,"12",,terminal_output +1265,1584967,"TERMINAL",0,0,"23",,terminal_output +1266,1585982,"TERMINAL",0,0,"34",,terminal_output +1267,1587136,"TERMINAL",0,0,"45",,terminal_output +1268,1588100,"TERMINAL",0,0,"57",,terminal_output +1269,1589145,"TERMINAL",0,0,"78",,terminal_output +1270,1590348,"TERMINAL",0,0,"89",,terminal_output +1271,1591379,"TERMINAL",0,0,"950",,terminal_output +1272,1592307,"TERMINAL",0,0,"6:001",,terminal_output +1273,1593328,"TERMINAL",0,0,"12",,terminal_output +1274,1594347,"TERMINAL",0,0,"23",,terminal_output +1275,1595365,"TERMINAL",0,0,"34",,terminal_output +1276,1596460,"TERMINAL",0,0,"45",,terminal_output +1277,1597522,"TERMINAL",0,0,"56",,terminal_output +1278,1598519,"TERMINAL",0,0,"67",,terminal_output +1279,1599573,"TERMINAL",0,0,"78",,terminal_output +1280,1600616,"TERMINAL",0,0,"89",,terminal_output +1281,1601614,"TERMINAL",0,0,"98:00",,terminal_output +1282,1602659,"TERMINAL",0,0,"101",,terminal_output +1283,1603702,"TERMINAL",0,0,"12",,terminal_output +1284,1604739,"TERMINAL",0,0,"23",,terminal_output +1285,1605821,"TERMINAL",0,0,"34",,terminal_output +1286,1606818,"TERMINAL",0,0,"45",,terminal_output +1287,1607963,"TERMINAL",0,0,"56",,terminal_output +1288,1608986,"TERMINAL",0,0,"67",,terminal_output +1289,1610113,"TERMINAL",0,0,"78",,terminal_output +1290,1610989,"TERMINAL",0,0,"89",,terminal_output +1291,1612057,"TERMINAL",0,0,"910",,terminal_output +1292,1613120,"TERMINAL",0,0,"202",,terminal_output +1293,1614147,"TERMINAL",0,0,"23",,terminal_output +1294,1615167,"TERMINAL",0,0,"34",,terminal_output +1295,1616257,"TERMINAL",0,0,"45",,terminal_output +1296,1617252,"TERMINAL",0,0,"56",,terminal_output +1297,1618299,"TERMINAL",0,0,"67",,terminal_output +1298,1619340,"TERMINAL",0,0,"78",,terminal_output +1299,1620461,"TERMINAL",0,0,"89",,terminal_output +1300,1621481,"TERMINAL",0,0,"920",,terminal_output +1301,1622523,"TERMINAL",0,0,"301",,terminal_output +1302,1623519,"TERMINAL",0,0,"12",,terminal_output +1303,1624596,"TERMINAL",0,0,"23",,terminal_output +1304,1625601,"TERMINAL",0,0,"34",,terminal_output +1305,1626757,"TERMINAL",0,0,"45",,terminal_output +1306,1627681,"TERMINAL",0,0,"56",,terminal_output +1307,1628727,"TERMINAL",0,0,"67",,terminal_output +1308,1629874,"TERMINAL",0,0,"78",,terminal_output +1309,1630837,"TERMINAL",0,0,"89",,terminal_output +1310,1631860,"TERMINAL",0,0,"930",,terminal_output +1311,1632933,"TERMINAL",0,0,"401",,terminal_output +1312,1633939,"TERMINAL",0,0,"12",,terminal_output +1313,1635097,"TERMINAL",0,0,"23",,terminal_output +1314,1636057,"TERMINAL",0,0,"35",,terminal_output +1315,1637128,"TERMINAL",0,0,"56",,terminal_output +1316,1638138,"TERMINAL",0,0,"67",,terminal_output +1317,1639182,"TERMINAL",0,0,"78",,terminal_output +1318,1640220,"TERMINAL",0,0,"89",,terminal_output +1319,1641261,"TERMINAL",0,0,"940",,terminal_output +1320,1642313,"TERMINAL",0,0,"501",,terminal_output +1321,1643347,"TERMINAL",0,0,"12",,terminal_output +1322,1644425,"TERMINAL",0,0,"23",,terminal_output +1323,1645541,"TERMINAL",0,0,"34",,terminal_output +1324,1646606,"TERMINAL",0,0,"45",,terminal_output +1325,1647567,"TERMINAL",0,0,"56",,terminal_output +1326,1648553,"TERMINAL",0,0,"67",,terminal_output +1327,1649638,"TERMINAL",0,0,"78",,terminal_output +1328,1650666,"TERMINAL",0,0,"89",,terminal_output +1329,1651714,"TERMINAL",0,0,"950",,terminal_output +1330,1652805,"TERMINAL",0,0,"7:001",,terminal_output +1331,1653760,"TERMINAL",0,0,"12",,terminal_output +1332,1654798,"TERMINAL",0,0,"23",,terminal_output +1333,1655911,"TERMINAL",0,0,"34",,terminal_output +1334,1656881,"TERMINAL",0,0,"45",,terminal_output +1335,1658039,"TERMINAL",0,0,"56",,terminal_output +1336,1659057,"TERMINAL",0,0,"67",,terminal_output +1337,1660089,"TERMINAL",0,0,"78",,terminal_output +1338,1661045,"TERMINAL",0,0,"89",,terminal_output +1339,1662130,"TERMINAL",0,0,"109:01",,terminal_output +1340,1663120,"TERMINAL",0,0,"12",,terminal_output +1341,1664177,"TERMINAL",0,0,"23",,terminal_output +1342,1665201,"TERMINAL",0,0,"34",,terminal_output +1343,1666331,"TERMINAL",0,0,"45",,terminal_output +1344,1667358,"TERMINAL",0,0,"56",,terminal_output +1345,1668379,"TERMINAL",0,0,"67",,terminal_output +1346,1669401,"TERMINAL",0,0,"78",,terminal_output +1347,1670412,"TERMINAL",0,0,"89",,terminal_output +1348,1671551,"TERMINAL",0,0,"910",,terminal_output +1349,1672561,"TERMINAL",0,0,"201",,terminal_output +1350,1673519,"TERMINAL",0,0,"12",,terminal_output +1351,1674632,"TERMINAL",0,0,"23",,terminal_output +1352,1675642,"TERMINAL",0,0,"34",,terminal_output +1353,1676645,"TERMINAL",0,0,"45",,terminal_output +1354,1677673,"TERMINAL",0,0,"56",,terminal_output +1355,1678714,"TERMINAL",0,0,"67",,terminal_output +1356,1679753,"TERMINAL",0,0,"78",,terminal_output +1357,1680793,"TERMINAL",0,0,"89",,terminal_output +1358,1681844,"TERMINAL",0,0,"920",,terminal_output +1359,1781594,"TERMINAL",0,0,"301",,terminal_output +1360,1782039,"TERMINAL",0,0,"1223344667788993040112233445566778",,terminal_output +1361,1782368,"TERMINAL",0,0,"8994050112233445566778899508:00223344556677889920:0010112233445566778899102011223344556677882030112233445566778899304011223344556677889940501122334455778899509:001122334455667788991:00",,terminal_output +1362,1782582,"TERMINAL",0,0,"101",,terminal_output +1363,1783639,"TERMINAL",0,0,"12",,terminal_output +1364,1784757,"TERMINAL",0,0,"23",,terminal_output +1365,1785766,"TERMINAL",0,0,"34",,terminal_output +1366,1786755,"TERMINAL",0,0,"45",,terminal_output +1367,1787789,"TERMINAL",0,0,"56",,terminal_output +1368,1788814,"TERMINAL",0,0,"67",,terminal_output +1369,1789886,"TERMINAL",0,0,"78",,terminal_output +1370,1790912,"TERMINAL",0,0,"89",,terminal_output +1371,1791930,"TERMINAL",0,0,"910",,terminal_output +1372,1792972,"TERMINAL",0,0,"201",,terminal_output +1373,1794006,"TERMINAL",0,0,"12",,terminal_output +1374,1795043,"TERMINAL",0,0,"23",,terminal_output +1375,1796078,"TERMINAL",0,0,"45",,terminal_output +1376,1797156,"TERMINAL",0,0,"56",,terminal_output +1377,1798148,"TERMINAL",0,0,"67",,terminal_output +1378,1799202,"TERMINAL",0,0,"78",,terminal_output +1379,1800236,"TERMINAL",0,0,"89",,terminal_output +1380,1801262,"TERMINAL",0,0,"920",,terminal_output +1381,1802311,"TERMINAL",0,0,"301",,terminal_output +1382,1803350,"TERMINAL",0,0,"12",,terminal_output +1383,1804478,"TERMINAL",0,0,"23",,terminal_output +1384,1805414,"TERMINAL",0,0,"34",,terminal_output +1385,1806451,"TERMINAL",0,0,"45",,terminal_output +1386,1807489,"TERMINAL",0,0,"56",,terminal_output +1387,1808537,"TERMINAL",0,0,"67",,terminal_output +1388,1809563,"TERMINAL",0,0,"78",,terminal_output +1389,1810602,"TERMINAL",0,0,"89",,terminal_output +1390,1811662,"TERMINAL",0,0,"930",,terminal_output +1391,1812674,"TERMINAL",0,0,"401",,terminal_output +1392,1813770,"TERMINAL",0,0,"12",,terminal_output +1393,1814757,"TERMINAL",0,0,"23",,terminal_output +1394,1815808,"TERMINAL",0,0,"34",,terminal_output +1395,1816858,"TERMINAL",0,0,"45",,terminal_output +1396,1817896,"TERMINAL",0,0,"56",,terminal_output +1397,1818976,"TERMINAL",0,0,"67",,terminal_output +1398,1819970,"TERMINAL",0,0,"78",,terminal_output +1399,1821018,"TERMINAL",0,0,"89",,terminal_output +1400,1822110,"TERMINAL",0,0,"941",,terminal_output +1401,1823083,"TERMINAL",0,0,"512",,terminal_output +1402,1824136,"TERMINAL",0,0,"23",,terminal_output +1403,1825161,"TERMINAL",0,0,"34",,terminal_output +1404,1826217,"TERMINAL",0,0,"45",,terminal_output +1405,1827322,"TERMINAL",0,0,"56",,terminal_output +1406,1828335,"TERMINAL",0,0,"67",,terminal_output +1407,1829333,"TERMINAL",0,0,"78",,terminal_output +1408,1830372,"TERMINAL",0,0,"89",,terminal_output +1409,1831404,"TERMINAL",0,0,"950",,terminal_output +1410,1832442,"TERMINAL",0,0,"10:001",,terminal_output +1411,1833554,"TERMINAL",0,0,"12",,terminal_output +1412,1834512,"TERMINAL",0,0,"23",,terminal_output +1413,1835546,"TERMINAL",0,0,"34",,terminal_output +1414,1836598,"TERMINAL",0,0,"45",,terminal_output +1415,1837646,"TERMINAL",0,0,"56",,terminal_output +1416,1838684,"TERMINAL",0,0,"67",,terminal_output +1417,1839694,"TERMINAL",0,0,"\r73472733 accelerat train_la tum_cte0 PD\t0:00\t 1 (None)8",,terminal_output +1418,1840774,"TERMINAL",0,0,"89",,terminal_output +1419,1841810,"TERMINAL",0,0,"92:00",,terminal_output +1420,1842819,"TERMINAL",0,0,"101",,terminal_output +1421,1843854,"TERMINAL",0,0,"12",,terminal_output +1422,1844889,"TERMINAL",0,0,"2 Rhkn08043",,terminal_output +1423,1845932,"TERMINAL",0,0,"314",,terminal_output +1424,1846978,"TERMINAL",0,0,"425",,terminal_output +1425,1848103,"TERMINAL",0,0,"536",,terminal_output +1426,1849050,"TERMINAL",0,0,"658",,terminal_output +1427,1850082,"TERMINAL",0,0,"869",,terminal_output +1428,1851178,"TERMINAL",0,0,"9710",,terminal_output +1429,1852161,"TERMINAL",0,0,"2081",,terminal_output +1430,1853201,"TERMINAL",0,0,"192",,terminal_output +1431,1854231,"TERMINAL",0,0,"2103",,terminal_output +1432,1855377,"TERMINAL",0,0,"314",,terminal_output +1433,1856334,"TERMINAL",0,0,"425",,terminal_output +1434,1857428,"TERMINAL",0,0,"536",,terminal_output +1435,1858448,"TERMINAL",0,0,"647",,terminal_output +1436,1859416,"TERMINAL",0,0,"758",,terminal_output +1437,1860494,"TERMINAL",0,0,"869",,terminal_output +1438,1861488,"TERMINAL",0,0,"9720",,terminal_output +1439,1862685,"TERMINAL",0,0,"3081",,terminal_output +1440,1863558,"TERMINAL",0,0,"192",,terminal_output +1441,1864594,"TERMINAL",0,0,"2203",,terminal_output +1442,1865692,"TERMINAL",0,0,"314",,terminal_output +1443,1866732,"TERMINAL",0,0,"425",,terminal_output +1444,1867741,"TERMINAL",0,0,"536",,terminal_output +1445,1868750,"TERMINAL",0,0,"647",,terminal_output +1446,1869779,"TERMINAL",0,0,"758",,terminal_output +1447,1870825,"TERMINAL",0,0,"869",,terminal_output +1448,1871969,"TERMINAL",0,0,"9730",,terminal_output +1449,1872890,"TERMINAL",0,0,"4081",,terminal_output +1450,1873985,"TERMINAL",0,0,"192",,terminal_output +1451,1874966,"TERMINAL",0,0,"2303",,terminal_output +1452,1876011,"TERMINAL",0,0,"314",,terminal_output +1453,1877043,"TERMINAL",0,0,"436",,terminal_output +1454,1878105,"TERMINAL",0,0,"647",,terminal_output +1455,1879145,"TERMINAL",0,0,"758",,terminal_output +1456,1880170,"TERMINAL",0,0,"869",,terminal_output +1457,1881270,"TERMINAL",0,0,"9740",,terminal_output +1458,1882232,"TERMINAL",0,0,"5081",,terminal_output +1459,1883260,"TERMINAL",0,0,"192",,terminal_output +1460,1884369,"TERMINAL",0,0,"2403",,terminal_output +1461,1885468,"TERMINAL",0,0,"314",,terminal_output +1462,1886383,"TERMINAL",0,0,"425",,terminal_output +1463,1887418,"TERMINAL",0,0,"536",,terminal_output +1464,1889777,"TERMINAL",0,0,"658",,terminal_output +1465,1890815,"TERMINAL",0,0,"869",,terminal_output +1466,1891861,"TERMINAL",0,0,"9750",,terminal_output +1467,1892886,"TERMINAL",0,0,"1:0081",,terminal_output +1468,1893979,"TERMINAL",0,0,"192",,terminal_output +1469,1894970,"TERMINAL",0,0,"2503",,terminal_output +1470,1896000,"TERMINAL",0,0,"314",,terminal_output +1471,1897039,"TERMINAL",0,0,"425",,terminal_output +1472,1898122,"TERMINAL",0,0,"547",,terminal_output +1473,1899106,"TERMINAL",0,0,"758",,terminal_output +1474,1900157,"TERMINAL",0,0,"869",,terminal_output +1475,1901245,"TERMINAL",0,0,"973:00",,terminal_output +1476,1902270,"TERMINAL",0,0,"1081",,terminal_output +1477,1903276,"TERMINAL",0,0,"192",,terminal_output +1478,1904434,"TERMINAL",0,0,"21:003",,terminal_output +1479,1905343,"TERMINAL",0,0,"314",,terminal_output +1480,1906474,"TERMINAL",0,0,"425",,terminal_output +1481,1907512,"TERMINAL",0,0,"536",,terminal_output +1482,1908516,"TERMINAL",0,0,"647",,terminal_output +1483,1909489,"TERMINAL",0,0,"758",,terminal_output +1484,1910590,"TERMINAL",0,0,"869",,terminal_output +1485,1911716,"TERMINAL",0,0,"9710",,terminal_output +1486,1912613,"TERMINAL",0,0,"2081",,terminal_output +1487,1913744,"TERMINAL",0,0,"192",,terminal_output +1488,1914709,"TERMINAL",0,0,"2103",,terminal_output +1489,1915736,"TERMINAL",0,0,"314",,terminal_output +1490,1916811,"TERMINAL",0,0,"425",,terminal_output +1491,1917812,"TERMINAL",0,0,"536",,terminal_output +1492,1918853,"TERMINAL",0,0,"647",,terminal_output +1493,1919944,"TERMINAL",0,0,"758",,terminal_output +1494,1920934,"TERMINAL",0,0,"869",,terminal_output +1495,1921980,"TERMINAL",0,0,"9720",,terminal_output +1496,1923058,"TERMINAL",0,0,"3081",,terminal_output +1497,1924068,"TERMINAL",0,0,"1203",,terminal_output +1498,1925102,"TERMINAL",0,0,"314",,terminal_output +1499,1926154,"TERMINAL",0,0,"425",,terminal_output +1500,1927254,"TERMINAL",0,0,"536",,terminal_output +1501,1928300,"TERMINAL",0,0,"647",,terminal_output +1502,1929254,"TERMINAL",0,0,"758",,terminal_output +1503,1930408,"TERMINAL",0,0,"869",,terminal_output +1504,1931477,"TERMINAL",0,0,"9730",,terminal_output +1505,1932375,"TERMINAL",0,0,"4081",,terminal_output +1506,1933499,"TERMINAL",0,0,"192",,terminal_output +1507,1934455,"TERMINAL",0,0,"2303",,terminal_output +1508,1935542,"TERMINAL",0,0,"314",,terminal_output +1509,1936564,"TERMINAL",0,0,"425",,terminal_output +1510,1937576,"TERMINAL",0,0,"536",,terminal_output +1511,1938718,"TERMINAL",0,0,"647",,terminal_output +1512,1939742,"TERMINAL",0,0,"758",,terminal_output +1513,1940766,"TERMINAL",0,0,"869",,terminal_output +1514,1941749,"TERMINAL",0,0,"9740",,terminal_output +1515,1942878,"TERMINAL",0,0,"5081",,terminal_output +1516,1943827,"TERMINAL",0,0,"192",,terminal_output +1517,1944872,"TERMINAL",0,0,"2403",,terminal_output +1518,1945894,"TERMINAL",0,0,"314",,terminal_output +1519,1946942,"TERMINAL",0,0,"425",,terminal_output +1520,1948045,"TERMINAL",0,0,"536",,terminal_output +1521,1949023,"TERMINAL",0,0,"647",,terminal_output +1522,1950068,"TERMINAL",0,0,"769",,terminal_output +1523,1951137,"TERMINAL",0,0,"9750",,terminal_output +1524,1952155,"TERMINAL",0,0,"2:0081",,terminal_output +1525,1953176,"TERMINAL",0,0,"192",,terminal_output +1526,1954212,"TERMINAL",0,0,"2503",,terminal_output +1527,1955304,"TERMINAL",0,0,"314",,terminal_output +1528,1956342,"TERMINAL",0,0,"425",,terminal_output +1529,1957363,"TERMINAL",0,0,"536",,terminal_output +1530,1958408,"TERMINAL",0,0,"647",,terminal_output +1531,1959421,"TERMINAL",0,0,"758",,terminal_output +1532,1960539,"TERMINAL",0,0,"869",,terminal_output +1533,1961662,"TERMINAL",0,0,"974:00",,terminal_output +1534,1962577,"TERMINAL",0,0,"1081",,terminal_output +1535,1963566,"TERMINAL",0,0,"192",,terminal_output +1536,1964758,"TERMINAL",0,0,"22:003",,terminal_output +1537,1965800,"TERMINAL",0,0,"314",,terminal_output +1538,1966837,"TERMINAL",0,0,"425",,terminal_output +1539,1967878,"TERMINAL",0,0,"536",,terminal_output +1540,1968956,"TERMINAL",0,0,"647",,terminal_output +1541,1969950,"TERMINAL",0,0,"758",,terminal_output +1542,1970990,"TERMINAL",0,0,"869",,terminal_output +1543,1972027,"TERMINAL",0,0,"9710",,terminal_output +1544,1973119,"TERMINAL",0,0,"2092",,terminal_output +1545,1974141,"TERMINAL",0,0,"2103",,terminal_output +1546,1975182,"TERMINAL",0,0,"314",,terminal_output +1547,1976204,"TERMINAL",0,0,"425",,terminal_output +1548,1977332,"TERMINAL",0,0,"536",,terminal_output +1549,1978252,"TERMINAL",0,0,"647",,terminal_output +1550,1979291,"TERMINAL",0,0,"758",,terminal_output +1551,1980414,"TERMINAL",0,0,"869",,terminal_output +1552,1981379,"TERMINAL",0,0,"9720",,terminal_output +1553,1982454,"TERMINAL",0,0,"3081",,terminal_output +1554,1983446,"TERMINAL",0,0,"192",,terminal_output +1555,1984520,"TERMINAL",0,0,"2203",,terminal_output +1556,1985622,"TERMINAL",0,0,"314",,terminal_output +1557,1986654,"TERMINAL",0,0,"425",,terminal_output +1558,1987666,"TERMINAL",0,0,"536",,terminal_output +1559,1988735,"TERMINAL",0,0,"647",,terminal_output +1560,1989699,"TERMINAL",0,0,"758",,terminal_output +1561,1990710,"TERMINAL",0,0,"869",,terminal_output +1562,1991771,"TERMINAL",0,0,"9730",,terminal_output +1563,1992796,"TERMINAL",0,0,"4081",,terminal_output +1564,1993812,"TERMINAL",0,0,"192",,terminal_output +1565,1994851,"TERMINAL",0,0,"2303",,terminal_output +1566,1995901,"TERMINAL",0,0,"314",,terminal_output +1567,1996935,"TERMINAL",0,0,"425",,terminal_output +1568,1997967,"TERMINAL",0,0,"536",,terminal_output +1569,1999042,"TERMINAL",0,0,"647",,terminal_output +1570,2000043,"TERMINAL",0,0,"769",,terminal_output +1571,2001087,"TERMINAL",0,0,"9740",,terminal_output +1572,2002127,"TERMINAL",0,0,"5081",,terminal_output +1573,2003160,"TERMINAL",0,0,"192",,terminal_output +1574,2004202,"TERMINAL",0,0,"2403",,terminal_output +1575,2005296,"TERMINAL",0,0,"314",,terminal_output +1576,2006294,"TERMINAL",0,0,"425",,terminal_output +1577,2007321,"TERMINAL",0,0,"536",,terminal_output +1578,2008455,"TERMINAL",0,0,"647",,terminal_output +1579,2009401,"TERMINAL",0,0,"758",,terminal_output +1580,2010451,"TERMINAL",0,0,"869",,terminal_output +1581,2011533,"TERMINAL",0,0,"9750",,terminal_output +1582,2012552,"TERMINAL",0,0,"3:0081",,terminal_output +1583,2013597,"TERMINAL",0,0,"192",,terminal_output +1584,2014772,"TERMINAL",0,0,"2503",,terminal_output +1585,2015626,"TERMINAL",0,0,"314",,terminal_output +1586,2016776,"TERMINAL",0,0,"425",,terminal_output +1587,2017780,"TERMINAL",0,0,"536",,terminal_output +1588,2018746,"TERMINAL",0,0,"647",,terminal_output +1589,2019799,"TERMINAL",0,0,"758",,terminal_output +1590,2020877,"TERMINAL",0,0,"869",,terminal_output +1591,2021910,"TERMINAL",0,0,"975:00",,terminal_output +1592,2022971,"TERMINAL",0,0,"1081",,terminal_output +1593,2024002,"TERMINAL",0,0,"192",,terminal_output +1594,2025036,"TERMINAL",0,0,"23:003",,terminal_output +1595,2026089,"TERMINAL",0,0,"325",,terminal_output +1596,2027128,"TERMINAL",0,0,"536",,terminal_output +1597,2028179,"TERMINAL",0,0,"647",,terminal_output +1598,2029195,"TERMINAL",0,0,"758",,terminal_output +1599,2030236,"TERMINAL",0,0,"869",,terminal_output +1600,2031271,"TERMINAL",0,0,"9710",,terminal_output +1601,2032312,"TERMINAL",0,0,"2081",,terminal_output +1602,2033358,"TERMINAL",0,0,"192",,terminal_output +1603,2034394,"TERMINAL",0,0,"2103",,terminal_output +1604,2035498,"TERMINAL",0,0,"314",,terminal_output +1605,2036436,"TERMINAL",0,0,"425",,terminal_output +1606,2037501,"TERMINAL",0,0,"536",,terminal_output +1607,2038545,"TERMINAL",0,0,"647",,terminal_output +1608,2039584,"TERMINAL",0,0,"758",,terminal_output +1609,2040617,"TERMINAL",0,0,"869",,terminal_output +1610,2041634,"TERMINAL",0,0,"9720",,terminal_output +1611,2042781,"TERMINAL",0,0,"3081",,terminal_output +1612,2043798,"TERMINAL",0,0,"192",,terminal_output +1613,2044818,"TERMINAL",0,0,"2203",,terminal_output +1614,2045799,"TERMINAL",0,0,"314",,terminal_output +1615,2046837,"TERMINAL",0,0,"425",,terminal_output +1616,2047878,"TERMINAL",0,0,"536",,terminal_output +1617,2048918,"TERMINAL",0,0,"647",,terminal_output +1618,2049975,"TERMINAL",0,0,"758",,terminal_output +1619,2050985,"TERMINAL",0,0,"869",,terminal_output +1620,2052029,"TERMINAL",0,0,"9730",,terminal_output +1621,2053089,"TERMINAL",0,0,"4092",,terminal_output +1622,2054110,"TERMINAL",0,0,"2303",,terminal_output +1623,2055154,"TERMINAL",0,0,"314",,terminal_output +1624,2056219,"TERMINAL",0,0,"425",,terminal_output +1625,2057312,"TERMINAL",0,0,"536",,terminal_output +1626,2058263,"TERMINAL",0,0,"647",,terminal_output +1627,2059348,"TERMINAL",0,0,"758",,terminal_output +1628,2060369,"TERMINAL",0,0,"869",,terminal_output +1629,2061511,"TERMINAL",0,0,"9740",,terminal_output +1630,2062532,"TERMINAL",0,0,"5081",,terminal_output +1631,2063549,"TERMINAL",0,0,"192",,terminal_output +1632,2064495,"TERMINAL",0,0,"2403",,terminal_output +1633,2065568,"TERMINAL",0,0,"314",,terminal_output +1634,2066587,"TERMINAL",0,0,"425",,terminal_output +1635,2067625,"TERMINAL",0,0,"536",,terminal_output +1636,2068869,"TERMINAL",0,0,"647",,terminal_output +1637,2069702,"TERMINAL",0,0,"758",,terminal_output +1638,2070764,"TERMINAL",0,0,"869",,terminal_output +1639,2071835,"TERMINAL",0,0,"9750",,terminal_output +1640,2072822,"TERMINAL",0,0,"4:0081",,terminal_output +1641,2073885,"TERMINAL",0,0,"192",,terminal_output +1642,2074920,"TERMINAL",0,0,"2503",,terminal_output +1643,2075936,"TERMINAL",0,0,"314",,terminal_output +1644,2076980,"TERMINAL",0,0,"425",,terminal_output +1645,2078046,"TERMINAL",0,0,"536",,terminal_output +1646,2079084,"TERMINAL",0,0,"658",,terminal_output +1647,2080119,"TERMINAL",0,0,"869",,terminal_output +1648,2081181,"TERMINAL",0,0,"976:00",,terminal_output +1649,2082195,"TERMINAL",0,0,"1081",,terminal_output +1650,2083214,"TERMINAL",0,0,"192",,terminal_output +1651,2084252,"TERMINAL",0,0,"24:003",,terminal_output +1652,2085353,"TERMINAL",0,0,"314",,terminal_output +1653,2086447,"TERMINAL",0,0,"425",,terminal_output +1654,2087448,"TERMINAL",0,0,"536",,terminal_output +1655,2088416,"TERMINAL",0,0,"647",,terminal_output +1656,2089562,"TERMINAL",0,0,"758",,terminal_output +1657,2090494,"TERMINAL",0,0,"869",,terminal_output +1658,2091602,"TERMINAL",0,0,"9710",,terminal_output +1659,2092603,"TERMINAL",0,0,"2081",,terminal_output +1660,2093662,"TERMINAL",0,0,"192",,terminal_output +1661,2094726,"TERMINAL",0,0,"2103",,terminal_output +1662,2095709,"TERMINAL",0,0,"314",,terminal_output +1663,2096848,"TERMINAL",0,0,"425",,terminal_output +1664,2097796,"TERMINAL",0,0,"536",,terminal_output +1665,2098860,"TERMINAL",0,0,"647",,terminal_output +1666,2099889,"TERMINAL",0,0,"758",,terminal_output +1667,2100931,"TERMINAL",0,0,"869",,terminal_output +1668,2101971,"TERMINAL",0,0,"9720",,terminal_output +1669,2103008,"TERMINAL",0,0,"3081",,terminal_output +1670,2104054,"TERMINAL",0,0,"192",,terminal_output +1671,2105107,"TERMINAL",0,0,"2214",,terminal_output +1672,2106092,"TERMINAL",0,0,"425",,terminal_output +1673,2107175,"TERMINAL",0,0,"536",,terminal_output +1674,2108180,"TERMINAL",0,0,"647",,terminal_output +1675,2109248,"TERMINAL",0,0,"758",,terminal_output +1676,2110348,"TERMINAL",0,0,"869",,terminal_output +1677,2111282,"TERMINAL",0,0,"9730",,terminal_output +1678,2112320,"TERMINAL",0,0,"4081",,terminal_output +1679,2113368,"TERMINAL",0,0,"192",,terminal_output +1680,2114428,"TERMINAL",0,0,"2303",,terminal_output +1681,2115439,"TERMINAL",0,0,"314",,terminal_output +1682,2116563,"TERMINAL",0,0,"425",,terminal_output +1683,2117618,"TERMINAL",0,0,"536",,terminal_output +1684,2118569,"TERMINAL",0,0,"647",,terminal_output +1685,2119667,"TERMINAL",0,0,"758",,terminal_output +1686,2120683,"TERMINAL",0,0,"869",,terminal_output +1687,2121831,"TERMINAL",0,0,"9740",,terminal_output +1688,2122786,"TERMINAL",0,0,"5081",,terminal_output +1689,2123788,"TERMINAL",0,0,"192",,terminal_output +1690,2124820,"TERMINAL",0,0,"2403",,terminal_output +1691,2125885,"TERMINAL",0,0,"314",,terminal_output +1692,2126902,"TERMINAL",0,0,"425",,terminal_output +1693,2127959,"TERMINAL",0,0,"536",,terminal_output +1694,2129088,"TERMINAL",0,0,"647",,terminal_output +1695,2130072,"TERMINAL",0,0,"758",,terminal_output +1696,2131112,"TERMINAL",0,0,"8750",,terminal_output +1697,2132151,"TERMINAL",0,0,"5:0081",,terminal_output +1698,2133195,"TERMINAL",0,0,"192",,terminal_output +1699,2134207,"TERMINAL",0,0,"2503",,terminal_output +1700,2135330,"TERMINAL",0,0,"314",,terminal_output +1701,2136355,"TERMINAL",0,0,"425",,terminal_output +1702,2137386,"TERMINAL",0,0,"536",,terminal_output +1703,2138406,"TERMINAL",0,0,"647",,terminal_output +1704,2139424,"TERMINAL",0,0,"758",,terminal_output +1705,2140560,"TERMINAL",0,0,"869",,terminal_output +1706,2141568,"TERMINAL",0,0,"977:00",,terminal_output +1707,2142605,"TERMINAL",0,0,"1081",,terminal_output +1708,2143786,"TERMINAL",0,0,"192",,terminal_output +1709,2144855,"TERMINAL",0,0,"25:003",,terminal_output +1710,2145832,"TERMINAL",0,0,"314",,terminal_output +1711,2146800,"TERMINAL",0,0,"425",,terminal_output +1712,2147821,"TERMINAL",0,0,"536",,terminal_output +1713,2148841,"TERMINAL",0,0,"647",,terminal_output +1714,2149884,"TERMINAL",0,0,"758",,terminal_output +1715,2150927,"TERMINAL",0,0,"869",,terminal_output +1716,2151967,"TERMINAL",0,0,"9710",,terminal_output +1717,2153082,"TERMINAL",0,0,"2081",,terminal_output +1718,2154141,"TERMINAL",0,0,"1103",,terminal_output +1719,2155144,"TERMINAL",0,0,"314",,terminal_output +1720,2156222,"TERMINAL",0,0,"425",,terminal_output +1721,2157249,"TERMINAL",0,0,"536",,terminal_output +1722,2158199,"TERMINAL",0,0,"647",,terminal_output +1723,2159252,"TERMINAL",0,0,"758",,terminal_output +1724,2160333,"TERMINAL",0,0,"869",,terminal_output +1725,2161366,"TERMINAL",0,0,"9720",,terminal_output +1726,2162467,"TERMINAL",0,0,"3081",,terminal_output +1727,2163704,"TERMINAL",0,0,"192",,terminal_output +1728,2164726,"TERMINAL",0,0,"2203",,terminal_output +1729,2165791,"TERMINAL",0,0,"314",,terminal_output +1730,2166720,"TERMINAL",0,0,"425",,terminal_output +1731,2167752,"TERMINAL",0,0,"536",,terminal_output +1732,2168832,"TERMINAL",0,0,"647",,terminal_output +1733,2169893,"TERMINAL",0,0,"758",,terminal_output +1734,2170884,"TERMINAL",0,0,"869",,terminal_output +1735,2171896,"TERMINAL",0,0,"9730",,terminal_output +1736,2172938,"TERMINAL",0,0,"4081",,terminal_output +1737,2173976,"TERMINAL",0,0,"192",,terminal_output +1738,2175017,"TERMINAL",0,0,"2303",,terminal_output +1739,2176151,"TERMINAL",0,0,"325",,terminal_output +1740,2177131,"TERMINAL",0,0,"536",,terminal_output +1741,2178156,"TERMINAL",0,0,"647",,terminal_output +1742,2179205,"TERMINAL",0,0,"758",,terminal_output +1743,2180244,"TERMINAL",0,0,"869",,terminal_output +1744,2181308,"TERMINAL",0,0,"9740",,terminal_output +1745,2182319,"TERMINAL",0,0,"5081",,terminal_output +1746,2183397,"TERMINAL",0,0,"192",,terminal_output +1747,2184378,"TERMINAL",0,0,"2403",,terminal_output +1748,2185414,"TERMINAL",0,0,"314",,terminal_output +1749,2186469,"TERMINAL",0,0,"425",,terminal_output +1750,2187533,"TERMINAL",0,0,"536",,terminal_output +1751,2188583,"TERMINAL",0,0,"647",,terminal_output +1752,2189546,"TERMINAL",0,0,"758",,terminal_output +1753,2190589,"TERMINAL",0,0,"869",,terminal_output +1754,2191678,"TERMINAL",0,0,"9750",,terminal_output +1755,2192794,"TERMINAL",0,0,"6:0081",,terminal_output +1756,2193701,"TERMINAL",0,0,"192",,terminal_output +1757,2194834,"TERMINAL",0,0,"2503",,terminal_output +1758,2195853,"TERMINAL",0,0,"314",,terminal_output +1759,2196840,"TERMINAL",0,0,"425",,terminal_output +1760,2198285,"TERMINAL",0,0,"547",,terminal_output +1761,2199333,"TERMINAL",0,0,"758",,terminal_output +1762,2200464,"TERMINAL",0,0,"869",,terminal_output +1763,2201502,"TERMINAL",0,0,"978:00",,terminal_output +1764,2202509,"TERMINAL",0,0,"1081",,terminal_output +1765,2203532,"TERMINAL",0,0,"192",,terminal_output +1766,2204565,"TERMINAL",0,0,"26:003",,terminal_output +1767,2205676,"TERMINAL",0,0,"314",,terminal_output +1768,2206706,"TERMINAL",0,0,"425",,terminal_output +1769,2207737,"TERMINAL",0,0,"536",,terminal_output +1770,2208853,"TERMINAL",0,0,"647",,terminal_output +1771,2209787,"TERMINAL",0,0,"758",,terminal_output +1772,2210831,"TERMINAL",0,0,"869",,terminal_output +1773,2211832,"TERMINAL",0,0,"9710",,terminal_output +1774,2212837,"TERMINAL",0,0,"2081",,terminal_output +1775,2213876,"TERMINAL",0,0,"192",,terminal_output +1776,2214931,"TERMINAL",0,0,"2103",,terminal_output +1777,2215965,"TERMINAL",0,0,"314",,terminal_output +1778,2217056,"TERMINAL",0,0,"425",,terminal_output +1779,2218090,"TERMINAL",0,0,"547",,terminal_output +1780,2219122,"TERMINAL",0,0,"758",,terminal_output +1781,2220170,"TERMINAL",0,0,"869",,terminal_output +1782,2221274,"TERMINAL",0,0,"9720",,terminal_output +1783,2222255,"TERMINAL",0,0,"3081",,terminal_output +1784,2223294,"TERMINAL",0,0,"192",,terminal_output +1785,2224550,"TERMINAL",0,0,"2203",,terminal_output +1786,2225362,"TERMINAL",0,0,"314",,terminal_output +1787,2226464,"TERMINAL",0,0,"425",,terminal_output +1788,2227494,"TERMINAL",0,0,"536",,terminal_output +1789,2228511,"TERMINAL",0,0,"647",,terminal_output +1790,2229496,"TERMINAL",0,0,"758",,terminal_output +1791,2230575,"TERMINAL",0,0,"869",,terminal_output +1792,2231593,"TERMINAL",0,0,"9730",,terminal_output +1793,2232714,"TERMINAL",0,0,"4081",,terminal_output +1794,2233737,"TERMINAL",0,0,"192",,terminal_output +1795,2234805,"TERMINAL",0,0,"2303",,terminal_output +1796,2235832,"TERMINAL",0,0,"314",,terminal_output +1797,2236812,"TERMINAL",0,0,"425",,terminal_output +1798,2237833,"TERMINAL",0,0,"536",,terminal_output +1799,2238835,"TERMINAL",0,0,"647",,terminal_output +1800,2239897,"TERMINAL",0,0,"758",,terminal_output +1801,2240915,"TERMINAL",0,0,"869",,terminal_output +1802,2241950,"TERMINAL",0,0,"9740",,terminal_output +1803,2242998,"TERMINAL",0,0,"5081",,terminal_output +1804,2244029,"TERMINAL",0,0,"192",,terminal_output +1805,2245080,"TERMINAL",0,0,"2414",,terminal_output +1806,2246164,"TERMINAL",0,0,"425",,terminal_output +1807,2247163,"TERMINAL",0,0,"536",,terminal_output +1808,2248268,"TERMINAL",0,0,"647",,terminal_output +1809,2249287,"TERMINAL",0,0,"758",,terminal_output +1810,2250329,"TERMINAL",0,0,"869",,terminal_output +1811,2251364,"TERMINAL",0,0,"9750",,terminal_output +1812,2252490,"TERMINAL",0,0,"7:0081",,terminal_output +1813,2253405,"TERMINAL",0,0,"192",,terminal_output +1814,2254520,"TERMINAL",0,0,"2503",,terminal_output +1815,2255545,"TERMINAL",0,0,"314",,terminal_output +1816,2256570,"TERMINAL",0,0,"425",,terminal_output +1817,2257558,"TERMINAL",0,0,"536",,terminal_output +1818,2258611,"TERMINAL",0,0,"647",,terminal_output +1819,2259651,"TERMINAL",0,0,"758",,terminal_output +1820,2260786,"TERMINAL",0,0,"869",,terminal_output +1821,2261838,"TERMINAL",0,0,"979:00",,terminal_output +1822,2262814,"TERMINAL",0,0,"1081",,terminal_output +1823,2263845,"TERMINAL",0,0,"192",,terminal_output +1824,2264864,"TERMINAL",0,0,"27:003",,terminal_output +1825,2265920,"TERMINAL",0,0,"314",,terminal_output +1826,2266942,"TERMINAL",0,0,"425",,terminal_output +1827,2267995,"TERMINAL",0,0,"536",,terminal_output +1828,2269029,"TERMINAL",0,0,"647",,terminal_output +1829,2270081,"TERMINAL",0,0,"769",,terminal_output +1830,2271277,"TERMINAL",0,0,"9710",,terminal_output +1831,2272218,"TERMINAL",0,0,"2081",,terminal_output +1832,2273282,"TERMINAL",0,0,"192",,terminal_output +1833,2274232,"TERMINAL",0,0,"2103",,terminal_output +1834,2275308,"TERMINAL",0,0,"314",,terminal_output +1835,2276335,"TERMINAL",0,0,"425",,terminal_output +1836,2277366,"TERMINAL",0,0,"536",,terminal_output +1837,2278383,"TERMINAL",0,0,"647",,terminal_output +1838,2279513,"TERMINAL",0,0,"758",,terminal_output +1839,2280540,"TERMINAL",0,0,"869",,terminal_output +1840,2281552,"TERMINAL",0,0,"9720",,terminal_output +1841,2282540,"TERMINAL",0,0,"3081",,terminal_output +1842,2283550,"TERMINAL",0,0,"192",,terminal_output +1843,2284598,"TERMINAL",0,0,"2203",,terminal_output +1844,2285642,"TERMINAL",0,0,"314",,terminal_output +1845,2286691,"TERMINAL",0,0,"425",,terminal_output +1846,2287704,"TERMINAL",0,0,"536",,terminal_output +1847,2288832,"TERMINAL",0,0,"647",,terminal_output +1848,2289957,"TERMINAL",0,0,"758",,terminal_output +1849,2290848,"TERMINAL",0,0,"869",,terminal_output +1850,2291899,"TERMINAL",0,0,"9730",,terminal_output +1851,2292977,"TERMINAL",0,0,"4081",,terminal_output +1852,2293937,"TERMINAL",0,0,"192",,terminal_output +1853,2294970,"TERMINAL",0,0,"2303",,terminal_output +1854,2296026,"TERMINAL",0,0,"314",,terminal_output +1855,2297076,"TERMINAL",0,0,"436",,terminal_output +1856,2298104,"TERMINAL",0,0,"647",,terminal_output +1857,2299140,"TERMINAL",0,0,"758",,terminal_output +1858,2300189,"TERMINAL",0,0,"869",,terminal_output +1859,2301332,"TERMINAL",0,0,"9740",,terminal_output +1860,2302243,"TERMINAL",0,0,"5081",,terminal_output +1861,2303392,"TERMINAL",0,0,"192",,terminal_output +1862,2304400,"TERMINAL",0,0,"2403",,terminal_output +1863,2305416,"TERMINAL",0,0,"314",,terminal_output +1864,2306410,"TERMINAL",0,0,"425",,terminal_output +1865,2307566,"TERMINAL",0,0,"536",,terminal_output +1866,2308591,"TERMINAL",0,0,"647",,terminal_output +1867,2309529,"TERMINAL",0,0,"758",,terminal_output +1868,2310563,"TERMINAL",0,0,"869",,terminal_output +1869,2311602,"TERMINAL",0,0,"9750",,terminal_output +1870,2312692,"TERMINAL",0,0,"8:0081",,terminal_output +1871,2313684,"TERMINAL",0,0,"192",,terminal_output +1872,2314719,"TERMINAL",0,0,"2503",,terminal_output +1873,2315764,"TERMINAL",0,0,"314",,terminal_output +1874,2316868,"TERMINAL",0,0,"425",,terminal_output +1875,2317860,"TERMINAL",0,0,"536",,terminal_output +1876,2318864,"TERMINAL",0,0,"647",,terminal_output +1877,2319907,"TERMINAL",0,0,"758",,terminal_output +1878,2320944,"TERMINAL",0,0,"869",,terminal_output +1879,2321986,"TERMINAL",0,0,"9730:00",,terminal_output +1880,2323078,"TERMINAL",0,0,"1081",,terminal_output +1881,2324062,"TERMINAL",0,0,"18:003",,terminal_output +1882,2325106,"TERMINAL",0,0,"314",,terminal_output +1883,2326154,"TERMINAL",0,0,"425",,terminal_output +1884,2327229,"TERMINAL",0,0,"536",,terminal_output +1885,2328221,"TERMINAL",0,0,"647",,terminal_output +1886,2329322,"TERMINAL",0,0,"758",,terminal_output +1887,2330306,"TERMINAL",0,0,"869",,terminal_output +1888,2331427,"TERMINAL",0,0,"9710",,terminal_output +1889,2332395,"TERMINAL",0,0,"2081",,terminal_output +1890,2333430,"TERMINAL",0,0,"192",,terminal_output +1891,2334541,"TERMINAL",0,0,"2103",,terminal_output +1892,2335613,"TERMINAL",0,0,"314",,terminal_output +1893,2336545,"TERMINAL",0,0,"425",,terminal_output +1894,2337670,"TERMINAL",0,0,"536",,terminal_output +1895,2338693,"TERMINAL",0,0,"647",,terminal_output +1896,2339716,"TERMINAL",0,0,"758",,terminal_output +1897,2340851,"TERMINAL",0,0,"869",,terminal_output +1898,2341870,"TERMINAL",0,0,"9720",,terminal_output +1899,2342894,"TERMINAL",0,0,"3081",,terminal_output +1900,2343885,"TERMINAL",0,0,"192",,terminal_output +1901,2344895,"TERMINAL",0,0,"2203",,terminal_output +1902,2345931,"TERMINAL",0,0,"314",,terminal_output +1903,2347087,"TERMINAL",0,0,"425",,terminal_output +1904,2348029,"TERMINAL",0,0,"536",,terminal_output +1905,2349087,"TERMINAL",0,0,"658",,terminal_output +1906,2350336,"TERMINAL",0,0,"869",,terminal_output +1907,2351170,"TERMINAL",0,0,"9730",,terminal_output +1908,2352196,"TERMINAL",0,0,"4081",,terminal_output +1909,2353231,"TERMINAL",0,0,"192",,terminal_output +1910,2354351,"TERMINAL",0,0,"2303",,terminal_output +1911,2355384,"TERMINAL",0,0,"314",,terminal_output +1912,2356411,"TERMINAL",0,0,"425",,terminal_output +1913,2357444,"TERMINAL",0,0,"536",,terminal_output +1914,2358457,"TERMINAL",0,0,"647",,terminal_output +1915,2359499,"TERMINAL",0,0,"758",,terminal_output +1916,2360609,"TERMINAL",0,0,"869",,terminal_output +1917,2361574,"TERMINAL",0,0,"9740",,terminal_output +1918,2362587,"TERMINAL",0,0,"5081",,terminal_output +1919,2363686,"TERMINAL",0,0,"192",,terminal_output +1920,2364684,"TERMINAL",0,0,"2403",,terminal_output +1921,2365721,"TERMINAL",0,0,"314",,terminal_output +1922,2366763,"TERMINAL",0,0,"425",,terminal_output +1923,2367885,"TERMINAL",0,0,"536",,terminal_output +1924,2368830,"TERMINAL",0,0,"647",,terminal_output +1925,2369858,"TERMINAL",0,0,"758",,terminal_output +1926,2370911,"TERMINAL",0,0,"869",,terminal_output +1927,2371931,"TERMINAL",0,0,"9750",,terminal_output +1928,2372975,"TERMINAL",0,0,"9:0081",,terminal_output +1929,2374049,"TERMINAL",0,0,"192",,terminal_output +1930,2375066,"TERMINAL",0,0,"2514",,terminal_output +1931,2376197,"TERMINAL",0,0,"425",,terminal_output +1932,2377126,"TERMINAL",0,0,"536",,terminal_output +1933,2378167,"TERMINAL",0,0,"647",,terminal_output +1934,2379251,"TERMINAL",0,0,"758",,terminal_output +1935,2380377,"TERMINAL",0,0,"869",,terminal_output +1936,2382024,"TERMINAL",0,0,"971:00",,terminal_output +1937,2382374,"TERMINAL",0,0,"1081",,terminal_output +1938,2383454,"TERMINAL",0,0,"192",,terminal_output +1939,2384405,"TERMINAL",0,0,"29:003",,terminal_output +1940,2385443,"TERMINAL",0,0,"314",,terminal_output +1941,2386517,"TERMINAL",0,0,"425",,terminal_output +1942,2387511,"TERMINAL",0,0,"536",,terminal_output +1943,2388555,"TERMINAL",0,0,"647",,terminal_output +1944,2389696,"TERMINAL",0,0,"758",,terminal_output +1945,2390685,"TERMINAL",0,0,"869",,terminal_output +1946,2391878,"TERMINAL",0,0,"9710",,terminal_output +1947,2392707,"TERMINAL",0,0,"2081",,terminal_output +1948,2393791,"TERMINAL",0,0,"192",,terminal_output +1949,2394855,"TERMINAL",0,0,"2103",,terminal_output +1950,2395821,"TERMINAL",0,0,"314",,terminal_output +1951,2396863,"TERMINAL",0,0,"425",,terminal_output +1952,2398006,"TERMINAL",0,0,"536",,terminal_output +1953,2398938,"TERMINAL",0,0,"647",,terminal_output +1954,2399971,"TERMINAL",0,0,"758",,terminal_output +1955,2401013,"TERMINAL",0,0,"869",,terminal_output +1956,2402086,"TERMINAL",0,0,"9821",,terminal_output +1957,2403102,"TERMINAL",0,0,"3192",,terminal_output +1958,2404156,"TERMINAL",0,0,"2203",,terminal_output +1959,2405214,"TERMINAL",0,0,"314",,terminal_output +1960,2406239,"TERMINAL",0,0,"425",,terminal_output +1961,2407302,"TERMINAL",0,0,"536",,terminal_output +1962,2408343,"TERMINAL",0,0,"647",,terminal_output +1963,2409453,"TERMINAL",0,0,"758",,terminal_output +1964,2410387,"TERMINAL",0,0,"869",,terminal_output +1965,2411449,"TERMINAL",0,0,"9730",,terminal_output +1966,2412559,"TERMINAL",0,0,"4081",,terminal_output +1967,2413553,"TERMINAL",0,0,"192",,terminal_output +1968,2414695,"TERMINAL",0,0,"2303",,terminal_output +1969,2415621,"TERMINAL",0,0,"314",,terminal_output +1970,2416834,"TERMINAL",0,0,"425",,terminal_output +1971,2417719,"TERMINAL",0,0,"536",,terminal_output +1972,2418732,"TERMINAL",0,0,"647",,terminal_output +1973,2419812,"TERMINAL",0,0,"758",,terminal_output +1974,2420836,"TERMINAL",0,0,"869",,terminal_output +1975,2421842,"TERMINAL",0,0,"9740",,terminal_output +1976,2422850,"TERMINAL",0,0,"5081",,terminal_output +1977,2423888,"TERMINAL",0,0,"192",,terminal_output +1978,2424928,"TERMINAL",0,0,"2403",,terminal_output +1979,2426014,"TERMINAL",0,0,"314",,terminal_output +1980,2427043,"TERMINAL",0,0,"425",,terminal_output +1981,2428041,"TERMINAL",0,0,"547",,terminal_output +1982,2429076,"TERMINAL",0,0,"758",,terminal_output +1983,2430125,"TERMINAL",0,0,"869",,terminal_output +1984,2431173,"TERMINAL",0,0,"9750",,terminal_output +1985,2432195,"TERMINAL",0,0,"20:0081",,terminal_output +1986,2433247,"TERMINAL",0,0,"192",,terminal_output +1987,2434344,"TERMINAL",0,0,"2503",,terminal_output +1988,2435364,"TERMINAL",0,0,"314",,terminal_output +1989,2436382,"TERMINAL",0,0,"425",,terminal_output +1990,2437395,"TERMINAL",0,0,"536",,terminal_output +1991,2438545,"TERMINAL",0,0,"647",,terminal_output +1992,2439563,"TERMINAL",0,0,"758",,terminal_output +1993,2440584,"TERMINAL",0,0,"869",,terminal_output +1994,2441549,"TERMINAL",0,0,"972:00",,terminal_output +1995,2442641,"TERMINAL",0,0,"1081",,terminal_output +1996,2443627,"TERMINAL",0,0,"192",,terminal_output +1997,2444674,"TERMINAL",0,0,"210:003",,terminal_output +1998,2445811,"TERMINAL",0,0,"314",,terminal_output +1999,2446747,"TERMINAL",0,0,"425",,terminal_output +2000,2447772,"TERMINAL",0,0,"536",,terminal_output +2001,2448814,"TERMINAL",0,0,"647",,terminal_output +2002,2449904,"TERMINAL",0,0,"758",,terminal_output +2003,2450908,"TERMINAL",0,0,"869",,terminal_output +2004,2451926,"TERMINAL",0,0,"9710",,terminal_output +2005,2452986,"TERMINAL",0,0,"2081",,terminal_output +2006,2454021,"TERMINAL",0,0,"192",,terminal_output +2007,2455056,"TERMINAL",0,0,"2114",,terminal_output +2008,2456148,"TERMINAL",0,0,"425",,terminal_output +2009,2457213,"TERMINAL",0,0,"536",,terminal_output +2010,2458170,"TERMINAL",0,0,"647",,terminal_output +2011,2459210,"TERMINAL",0,0,"758",,terminal_output +2012,2460243,"TERMINAL",0,0,"869",,terminal_output +2013,2461296,"TERMINAL",0,0,"9720",,terminal_output +2014,2462321,"TERMINAL",0,0,"3081",,terminal_output +2015,2463391,"TERMINAL",0,0,"192",,terminal_output +2016,2464468,"TERMINAL",0,0,"2203",,terminal_output +2017,2465470,"TERMINAL",0,0,"314",,terminal_output +2018,2466503,"TERMINAL",0,0,"425",,terminal_output +2019,2467615,"TERMINAL",0,0,"536",,terminal_output +2020,2468644,"TERMINAL",0,0,"647",,terminal_output +2021,2469687,"TERMINAL",0,0,"758",,terminal_output +2022,2470660,"TERMINAL",0,0,"869",,terminal_output +2023,2471698,"TERMINAL",0,0,"9730",,terminal_output +2024,2472722,"TERMINAL",0,0,"4081",,terminal_output +2025,2473768,"TERMINAL",0,0,"192",,terminal_output +2026,2474798,"TERMINAL",0,0,"2303",,terminal_output +2027,2475950,"TERMINAL",0,0,"314",,terminal_output +2028,2476958,"TERMINAL",0,0,"425",,terminal_output +2029,2477918,"TERMINAL",0,0,"536",,terminal_output +2030,2478955,"TERMINAL",0,0,"647",,terminal_output +2031,2480036,"TERMINAL",0,0,"758",,terminal_output +2032,2481041,"TERMINAL",0,0,"869",,terminal_output +2033,2482065,"TERMINAL",0,0,"9841",,terminal_output +2034,2483113,"TERMINAL",0,0,"5192",,terminal_output +2035,2484166,"TERMINAL",0,0,"2403",,terminal_output +2036,2485201,"TERMINAL",0,0,"314",,terminal_output +2037,2486256,"TERMINAL",0,0,"425",,terminal_output +2038,2487268,"TERMINAL",0,0,"536",,terminal_output +2039,2488402,"TERMINAL",0,0,"647",,terminal_output +2040,2489348,"TERMINAL",0,0,"758",,terminal_output +2041,2490398,"TERMINAL",0,0,"869",,terminal_output +2042,2491485,"TERMINAL",0,0,"9750",,terminal_output +2043,2492507,"TERMINAL",0,0,"1:0081",,terminal_output +2044,2493525,"TERMINAL",0,0,"192",,terminal_output +2045,2494651,"TERMINAL",0,0,"2503",,terminal_output +2046,2495674,"TERMINAL",0,0,"314",,terminal_output +2047,2496703,"TERMINAL",0,0,"425",,terminal_output +2048,2497724,"TERMINAL",0,0,"536",,terminal_output +2049,2498748,"TERMINAL",0,0,"647",,terminal_output +2050,2499727,"TERMINAL",0,0,"758",,terminal_output +2051,2500775,"TERMINAL",0,0,"869",,terminal_output +2052,2501854,"TERMINAL",0,0,"973:00",,terminal_output +2053,2502881,"TERMINAL",0,0,"1081",,terminal_output +2054,2503902,"TERMINAL",0,0,"192",,terminal_output +2055,2504918,"TERMINAL",0,0,"21:003",,terminal_output +2056,2505959,"TERMINAL",0,0,"314",,terminal_output +2057,2506994,"TERMINAL",0,0,"425",,terminal_output +2058,2508037,"TERMINAL",0,0,"536",,terminal_output +2059,2509072,"TERMINAL",0,0,"758",,terminal_output +2060,2510112,"TERMINAL",0,0,"869",,terminal_output +2061,2511242,"TERMINAL",0,0,"9710",,terminal_output +2062,2512190,"TERMINAL",0,0,"2081",,terminal_output +2063,2513230,"TERMINAL",0,0,"192",,terminal_output +2064,2514267,"TERMINAL",0,0,"2103",,terminal_output +2065,2515333,"TERMINAL",0,0,"314",,terminal_output +2066,2516385,"TERMINAL",0,0,"425",,terminal_output +2067,2517475,"TERMINAL",0,0,"536",,terminal_output +2068,2518472,"TERMINAL",0,0,"647",,terminal_output +2069,2519469,"TERMINAL",0,0,"758",,terminal_output +2070,2520580,"TERMINAL",0,0,"869",,terminal_output +2071,2521596,"TERMINAL",0,0,"9720",,terminal_output +2072,2522662,"TERMINAL",0,0,"3081",,terminal_output +2073,2523731,"TERMINAL",0,0,"192",,terminal_output +2074,2524782,"TERMINAL",0,0,"2203",,terminal_output +2075,2525789,"TERMINAL",0,0,"314",,terminal_output +2076,2526817,"TERMINAL",0,0,"425",,terminal_output +2077,2527929,"TERMINAL",0,0,"536",,terminal_output +2078,2528902,"TERMINAL",0,0,"647",,terminal_output +2079,2529965,"TERMINAL",0,0,"758",,terminal_output +2080,2530997,"TERMINAL",0,0,"869",,terminal_output +2081,2532045,"TERMINAL",0,0,"9730",,terminal_output +2082,2533072,"TERMINAL",0,0,"4092",,terminal_output +2083,2534126,"TERMINAL",0,0,"2303",,terminal_output +2084,2535155,"TERMINAL",0,0,"314",,terminal_output +2085,2536196,"TERMINAL",0,0,"425",,terminal_output +2086,2537233,"TERMINAL",0,0,"536",,terminal_output +2087,2538276,"TERMINAL",0,0,"647",,terminal_output +2088,2539400,"TERMINAL",0,0,"758",,terminal_output +2089,2540339,"TERMINAL",0,0,"869",,terminal_output +2090,2541450,"TERMINAL",0,0,"9740",,terminal_output +2091,2542471,"TERMINAL",0,0,"5081",,terminal_output +2092,2543505,"TERMINAL",0,0,"192",,terminal_output +2093,2544521,"TERMINAL",0,0,"2403",,terminal_output +2094,2545602,"TERMINAL",0,0,"314",,terminal_output +2095,2546608,"TERMINAL",0,0,"425",,terminal_output +2096,2547699,"TERMINAL",0,0,"536",,terminal_output +2097,2548718,"TERMINAL",0,0,"647",,terminal_output +2098,2549749,"TERMINAL",0,0,"758",,terminal_output +2099,2550775,"TERMINAL",0,0,"869",,terminal_output +2100,2551794,"TERMINAL",0,0,"9750",,terminal_output +2101,2552921,"TERMINAL",0,0,"2:0081",,terminal_output +2102,2553896,"TERMINAL",0,0,"192",,terminal_output +2103,2554900,"TERMINAL",0,0,"2503",,terminal_output +2104,2555960,"TERMINAL",0,0,"314",,terminal_output +2105,2557351,"TERMINAL",0,0,"436",,terminal_output +2106,2558454,"TERMINAL",0,0,"647",,terminal_output +2107,2559398,"TERMINAL",0,0,"758",,terminal_output +2108,2560499,"TERMINAL",0,0,"869",,terminal_output +2109,2561524,"TERMINAL",0,0,"974:00",,terminal_output +2110,2562583,"TERMINAL",0,0,"1081",,terminal_output +2111,2563672,"TERMINAL",0,0,"192",,terminal_output +2112,2564709,"TERMINAL",0,0,"22:003",,terminal_output +2113,2565923,"TERMINAL",0,0,"314",,terminal_output +2114,2566967,"TERMINAL",0,0,"425",,terminal_output +2115,2568025,"TERMINAL",0,0,"536",,terminal_output +2116,2569030,"TERMINAL",0,0,"647",,terminal_output +2117,2570077,"TERMINAL",0,0,"758",,terminal_output +2118,2571143,"TERMINAL",0,0,"8710",,terminal_output +2119,2572140,"TERMINAL",0,0,"2081",,terminal_output +2120,2573169,"TERMINAL",0,0,"192",,terminal_output +2121,2574226,"TERMINAL",0,0,"2103",,terminal_output +2122,2575262,"TERMINAL",0,0,"314",,terminal_output +2123,2576311,"TERMINAL",0,0,"425",,terminal_output +2124,2577355,"TERMINAL",0,0,"536",,terminal_output +2125,2578434,"TERMINAL",0,0,"647",,terminal_output +2126,2579428,"TERMINAL",0,0,"758",,terminal_output +2127,2580462,"TERMINAL",0,0,"869",,terminal_output +2128,2581594,"TERMINAL",0,0,"9720",,terminal_output +2129,2582562,"TERMINAL",0,0,"3081",,terminal_output +2130,2583637,"TERMINAL",0,0,"192",,terminal_output +2131,2584664,"TERMINAL",0,0,"2203",,terminal_output +2132,2585686,"TERMINAL",0,0,"314",,terminal_output +2133,2586821,"TERMINAL",0,0,"425",,terminal_output +2134,2587837,"TERMINAL",0,0,"536",,terminal_output +2135,2588806,"TERMINAL",0,0,"647",,terminal_output +2136,2589898,"TERMINAL",0,0,"758",,terminal_output +2137,2590928,"TERMINAL",0,0,"869",,terminal_output +2138,2591924,"TERMINAL",0,0,"9730",,terminal_output +2139,2592953,"TERMINAL",0,0,"4081",,terminal_output +2140,2594008,"TERMINAL",0,0,"192",,terminal_output +2141,2595081,"TERMINAL",0,0,"2314",,terminal_output +2142,2596128,"TERMINAL",0,0,"425",,terminal_output +2143,2597165,"TERMINAL",0,0,"536",,terminal_output +2144,2598189,"TERMINAL",0,0,"647",,terminal_output +2145,2599254,"TERMINAL",0,0,"758",,terminal_output +2146,2600266,"TERMINAL",0,0,"869",,terminal_output +2147,2601365,"TERMINAL",0,0,"9740",,terminal_output +2148,2602376,"TERMINAL",0,0,"5081",,terminal_output +2149,2603506,"TERMINAL",0,0,"192",,terminal_output +2150,2604527,"TERMINAL",0,0,"2403",,terminal_output +2151,2605553,"TERMINAL",0,0,"314",,terminal_output +2152,2606510,"TERMINAL",0,0,"425",,terminal_output +2153,2607598,"TERMINAL",0,0,"536",,terminal_output +2154,2608604,"TERMINAL",0,0,"647",,terminal_output +2155,2609649,"TERMINAL",0,0,"758",,terminal_output +2156,2610756,"TERMINAL",0,0,"869",,terminal_output +2157,2611799,"TERMINAL",0,0,"9750",,terminal_output +2158,2612760,"TERMINAL",0,0,"3:0081",,terminal_output +2159,2613848,"TERMINAL",0,0,"192",,terminal_output +2160,2614888,"TERMINAL",0,0,"2503",,terminal_output +2161,2615895,"TERMINAL",0,0,"314",,terminal_output +2162,2616924,"TERMINAL",0,0,"425",,terminal_output +2163,2617963,"TERMINAL",0,0,"536",,terminal_output +2164,2619230,"TERMINAL",0,0,"647",,terminal_output +2165,2620038,"TERMINAL",0,0,"758",,terminal_output +2166,2621150,"TERMINAL",0,0,"875:00",,terminal_output +2167,2622118,"TERMINAL",0,0,"1081",,terminal_output +2168,2623203,"TERMINAL",0,0,"192",,terminal_output +2169,2624188,"TERMINAL",0,0,"23:003",,terminal_output +2170,2625224,"TERMINAL",0,0,"314",,terminal_output +2171,2626270,"TERMINAL",0,0,"425",,terminal_output +2172,2627317,"TERMINAL",0,0,"536",,terminal_output +2173,2628344,"TERMINAL",0,0,"647",,terminal_output +2174,2629418,"TERMINAL",0,0,"758",,terminal_output +2175,2630445,"TERMINAL",0,0,"869",,terminal_output +2176,2631469,"TERMINAL",0,0,"9710",,terminal_output +2177,2632500,"TERMINAL",0,0,"2081",,terminal_output +2178,2633538,"TERMINAL",0,0,"192",,terminal_output +2179,2634680,"TERMINAL",0,0,"2103",,terminal_output +2180,2635611,"TERMINAL",0,0,"314",,terminal_output +2181,2636652,"TERMINAL",0,0,"425",,terminal_output +2182,2637734,"TERMINAL",0,0,"536",,terminal_output +2183,2638845,"TERMINAL",0,0,"647",,terminal_output +2184,2639882,"TERMINAL",0,0,"758",,terminal_output +2185,2640898,"TERMINAL",0,0,"869",,terminal_output +2186,2641907,"TERMINAL",0,0,"9720",,terminal_output +2187,2642928,"TERMINAL",0,0,"3081",,terminal_output +2188,2643925,"TERMINAL",0,0,"192",,terminal_output +2189,2644975,"TERMINAL",0,0,"2203",,terminal_output +2190,2646008,"TERMINAL",0,0,"314",,terminal_output +2191,2647065,"TERMINAL",0,0,"436",,terminal_output +2192,2648087,"TERMINAL",0,0,"647",,terminal_output +2193,2649131,"TERMINAL",0,0,"758",,terminal_output +2194,2650160,"TERMINAL",0,0,"869",,terminal_output +2195,2651222,"TERMINAL",0,0,"9730",,terminal_output +2196,2652245,"TERMINAL",0,0,"4081",,terminal_output +2197,2653275,"TERMINAL",0,0,"192",,terminal_output +2198,2654400,"TERMINAL",0,0,"2303",,terminal_output +2199,2655424,"TERMINAL",0,0,"314",,terminal_output +2200,2656492,"TERMINAL",0,0,"425",,terminal_output +2201,2657470,"TERMINAL",0,0,"536",,terminal_output +2202,2658594,"TERMINAL",0,0,"647",,terminal_output +2203,2659617,"TERMINAL",0,0,"758",,terminal_output +2204,2660642,"TERMINAL",0,0,"869",,terminal_output +2205,2661609,"TERMINAL",0,0,"9740",,terminal_output +2206,2662694,"TERMINAL",0,0,"5081",,terminal_output +2207,2663722,"TERMINAL",0,0,"192",,terminal_output +2208,2664812,"TERMINAL",0,0,"2403",,terminal_output +2209,2665760,"TERMINAL",0,0,"314",,terminal_output +2210,2666793,"TERMINAL",0,0,"425",,terminal_output +2211,2667827,"TERMINAL",0,0,"536",,terminal_output +2212,2668865,"TERMINAL",0,0,"647",,terminal_output +2213,2669958,"TERMINAL",0,0,"758",,terminal_output +2214,2670957,"TERMINAL",0,0,"869",,terminal_output +2215,2671992,"TERMINAL",0,0,"9750",,terminal_output +2216,2673040,"TERMINAL",0,0,"4:0081",,terminal_output +2217,2674071,"TERMINAL",0,0,"2503",,terminal_output +2218,2675133,"TERMINAL",0,0,"314",,terminal_output +2219,2676225,"TERMINAL",0,0,"425",,terminal_output +2220,2677440,"TERMINAL",0,0,"536",,terminal_output +2221,2678221,"TERMINAL",0,0,"647",,terminal_output +2222,2679409,"TERMINAL",0,0,"758",,terminal_output +2223,2680294,"TERMINAL",0,0,"869",,terminal_output +2224,2681656,"TERMINAL",0,0,"976:00",,terminal_output +2225,2682469,"TERMINAL",0,0,"1081",,terminal_output +2226,2683413,"TERMINAL",0,0,"192",,terminal_output +2227,2684447,"TERMINAL",0,0,"24:003",,terminal_output +2228,2685486,"TERMINAL",0,0,"314",,terminal_output +2229,2686549,"TERMINAL",0,0,"425",,terminal_output +2230,2687580,"TERMINAL",0,0,"536",,terminal_output +2231,2688705,"TERMINAL",0,0,"647",,terminal_output +2232,2689647,"TERMINAL",0,0,"758",,terminal_output +2233,2690754,"TERMINAL",0,0,"869",,terminal_output +2234,2691709,"TERMINAL",0,0,"9710",,terminal_output +2235,2692750,"TERMINAL",0,0,"2081",,terminal_output +2236,2693857,"TERMINAL",0,0,"192",,terminal_output +2237,2694893,"TERMINAL",0,0,"2103",,terminal_output +2238,2695934,"TERMINAL",0,0,"314",,terminal_output +2239,2696934,"TERMINAL",0,0,"425",,terminal_output +2240,2698026,"TERMINAL",0,0,"536",,terminal_output +2241,2699003,"TERMINAL",0,0,"647",,terminal_output +2242,2700069,"TERMINAL",0,0,"769",,terminal_output +2243,2701082,"TERMINAL",0,0,"9720",,terminal_output +2244,2702119,"TERMINAL",0,0,"3081",,terminal_output +2245,2703177,"TERMINAL",0,0,"192",,terminal_output +2246,2704267,"TERMINAL",0,0,"2203",,terminal_output +2247,2705277,"TERMINAL",0,0,"314",,terminal_output +2248,2706306,"TERMINAL",0,0,"425",,terminal_output +2249,2707336,"TERMINAL",0,0,"536",,terminal_output +2250,2708395,"TERMINAL",0,0,"647",,terminal_output +2251,2709492,"TERMINAL",0,0,"758",,terminal_output +2252,2710466,"TERMINAL",0,0,"869",,terminal_output +2253,2711553,"TERMINAL",0,0,"9730",,terminal_output +2254,2712521,"TERMINAL",0,0,"4081",,terminal_output +2255,2713585,"TERMINAL",0,0,"192",,terminal_output +2256,2714652,"TERMINAL",0,0,"2303",,terminal_output +2257,2715739,"TERMINAL",0,0,"314",,terminal_output +2258,2716706,"TERMINAL",0,0,"425",,terminal_output +2259,2717713,"TERMINAL",0,0,"536",,terminal_output +2260,2718814,"TERMINAL",0,0,"647",,terminal_output +2261,2719849,"TERMINAL",0,0,"758",,terminal_output +2262,2720844,"TERMINAL",0,0,"869",,terminal_output +2263,2721925,"TERMINAL",0,0,"9740",,terminal_output +2264,2723004,"TERMINAL",0,0,"5081",,terminal_output +2265,2723966,"TERMINAL",0,0,"192",,terminal_output +2266,2725019,"TERMINAL",0,0,"2403",,terminal_output +2267,2726053,"TERMINAL",0,0,"325",,terminal_output +2268,2727206,"TERMINAL",0,0,"536",,terminal_output +2269,2728209,"TERMINAL",0,0,"647",,terminal_output +2270,2729184,"TERMINAL",0,0,"758",,terminal_output +2271,2730228,"TERMINAL",0,0,"869",,terminal_output +2272,2731311,"TERMINAL",0,0,"9750",,terminal_output +2273,2732322,"TERMINAL",0,0,"5:0081",,terminal_output +2274,2733356,"TERMINAL",0,0,"192",,terminal_output +2275,2734473,"TERMINAL",0,0,"2503",,terminal_output +2276,2735495,"TERMINAL",0,0,"314",,terminal_output +2277,2736523,"TERMINAL",0,0,"425",,terminal_output +2278,2737528,"TERMINAL",0,0,"536",,terminal_output +2279,2738585,"TERMINAL",0,0,"647",,terminal_output +2280,2739698,"TERMINAL",0,0,"758",,terminal_output +2281,2740725,"TERMINAL",0,0,"869",,terminal_output +2282,2741754,"TERMINAL",0,0,"977:00",,terminal_output +2283,2742779,"TERMINAL",0,0,"1081",,terminal_output +2284,2743788,"TERMINAL",0,0,"192",,terminal_output +2285,2744878,"TERMINAL",0,0,"25:003",,terminal_output +2286,2745942,"TERMINAL",0,0,"314",,terminal_output +2287,2746943,"TERMINAL",0,0,"425",,terminal_output +2288,2747992,"TERMINAL",0,0,"536",,terminal_output +2289,2749003,"TERMINAL",0,0,"647",,terminal_output +2290,2750047,"TERMINAL",0,0,"769",,terminal_output +2291,2751076,"TERMINAL",0,0,"9710",,terminal_output +2292,2752153,"TERMINAL",0,0,"2081",,terminal_output +2293,2753183,"TERMINAL",0,0,"192",,terminal_output +2294,2754310,"TERMINAL",0,0,"2103",,terminal_output +2295,2755267,"TERMINAL",0,0,"314",,terminal_output +2296,2756334,"TERMINAL",0,0,"425",,terminal_output +2297,2757400,"TERMINAL",0,0,"536",,terminal_output +2298,2758445,"TERMINAL",0,0,"647",,terminal_output +2299,2759478,"TERMINAL",0,0,"758",,terminal_output +2300,2760541,"TERMINAL",0,0,"869",,terminal_output +2301,2761608,"TERMINAL",0,0,"9720",,terminal_output +2302,2762562,"TERMINAL",0,0,"3081",,terminal_output +2303,2763603,"TERMINAL",0,0,"192",,terminal_output +2304,2764689,"TERMINAL",0,0,"2203",,terminal_output +2305,2765755,"TERMINAL",0,0,"314",,terminal_output +2306,2766838,"TERMINAL",0,0,"425",,terminal_output +2307,2767857,"TERMINAL",0,0,"536",,terminal_output +2308,2768787,"TERMINAL",0,0,"647",,terminal_output +2309,2769938,"TERMINAL",0,0,"758",,terminal_output +2310,2770874,"TERMINAL",0,0,"869",,terminal_output +2311,2771955,"TERMINAL",0,0,"9730",,terminal_output +2312,2772943,"TERMINAL",0,0,"4081",,terminal_output +2313,2774034,"TERMINAL",0,0,"192",,terminal_output +2314,2775020,"TERMINAL",0,0,"2303",,terminal_output +2315,2776180,"TERMINAL",0,0,"325",,terminal_output +2316,2777135,"TERMINAL",0,0,"536",,terminal_output +2317,2778178,"TERMINAL",0,0,"647",,terminal_output +2318,2779251,"TERMINAL",0,0,"758",,terminal_output +2319,2780275,"TERMINAL",0,0,"869",,terminal_output +2320,2781261,"TERMINAL",0,0,"9740",,terminal_output +2321,2782306,"TERMINAL",0,0,"5081",,terminal_output +2322,2783422,"TERMINAL",0,0,"192",,terminal_output +2323,2784446,"TERMINAL",0,0,"2403",,terminal_output +2324,2785473,"TERMINAL",0,0,"314",,terminal_output +2325,2786596,"TERMINAL",0,0,"425",,terminal_output +2326,2787517,"TERMINAL",0,0,"536",,terminal_output +2327,2788645,"TERMINAL",0,0,"647",,terminal_output +2328,2789588,"TERMINAL",0,0,"758",,terminal_output +2329,2790638,"TERMINAL",0,0,"869",,terminal_output +2330,2791750,"TERMINAL",0,0,"9750",,terminal_output +2331,2792739,"TERMINAL",0,0,"6:0081",,terminal_output +2332,2793748,"TERMINAL",0,0,"192",,terminal_output +2333,2794792,"TERMINAL",0,0,"2503",,terminal_output +2334,2795919,"TERMINAL",0,0,"314",,terminal_output +2335,2796948,"TERMINAL",0,0,"425",,terminal_output +2336,2797965,"TERMINAL",0,0,"536",,terminal_output +2337,2798999,"TERMINAL",0,0,"647",,terminal_output +2338,2799999,"TERMINAL",0,0,"758",,terminal_output +2339,2801058,"TERMINAL",0,0,"878:00",,terminal_output +2340,2802136,"TERMINAL",0,0,"1081",,terminal_output +2341,2803131,"TERMINAL",0,0,"192",,terminal_output +2342,2804218,"TERMINAL",0,0,"26:003",,terminal_output +2343,2805214,"TERMINAL",0,0,"314",,terminal_output +2344,2806261,"TERMINAL",0,0,"425",,terminal_output +2345,2807390,"TERMINAL",0,0,"536",,terminal_output +2346,2808412,"TERMINAL",0,0,"647",,terminal_output +2347,2809480,"TERMINAL",0,0,"758",,terminal_output +2348,2810463,"TERMINAL",0,0,"869",,terminal_output +2349,2811484,"TERMINAL",0,0,"9710",,terminal_output +2350,2812567,"TERMINAL",0,0,"2081",,terminal_output +2351,2813642,"TERMINAL",0,0,"192",,terminal_output diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-b85ae4be-e003-4b5f-b92d-85f73081a0e71757580020189-2025_09_11-10.41.19.400/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-b85ae4be-e003-4b5f-b92d-85f73081a0e71757580020189-2025_09_11-10.41.19.400/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..0fb2161671403fdf9b4d1b5508ff8cd309d121e2 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-b85ae4be-e003-4b5f-b92d-85f73081a0e71757580020189-2025_09_11-10.41.19.400/source.csv @@ -0,0 +1,4 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +1,15,"train_dynamics.py",0,0,"from dataclasses import dataclass, field\nimport os\nimport time\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nimport optax\nimport orbax\nimport numpy as np\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\n\nts = int(time.time())\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_resolution: int = 64\n data_dir: str = ""data/coinrun_episodes""\n # Optimization\n batch_size: int = 36\n min_lr: float = 3e-6\n max_lr: float = 3e-5\n warmup_steps: int = 5000\n # Tokenizer\n tokenizer_dim: int = 512\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 8\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 8\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_dim: int = 512\n dyna_num_blocks: int = 12\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_gradients: bool = False\n name: str = """"\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n wandb_id: str = """"\n\nargs = tyro.cli(Args)\n\n\ndef dynamics_loss_fn(params, state, inputs):\n """"""Compute masked dynamics loss""""""\n outputs = state.apply_fn(\n params, inputs, training=True, rngs={""dropout"": inputs[""dropout_rng""]}\n )\n mask = outputs[""mask""]\n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n outputs[""token_logits""], outputs[""video_tokens""]\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = outputs[""token_logits""].argmax(-1) == outputs[""video_tokens""]\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(outputs[""token_logits""])\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=outputs[""token_logits""].max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n """"""Update state and compute metrics""""""\n grad_fn = jax.value_and_grad(dynamics_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n rng = jax.random.PRNGKey(args.seed)\n if args.log:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n # --- Initialize model ---\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_resolution, args.image_resolution, args.image_channels)\n dummy_inputs = dict(\n videos=jnp.zeros(\n (args.batch_size, args.seq_len, *image_shape), dtype=jnp.float32\n ),\n mask_rng=_rng,\n )\n rng, _rng = jax.random.split(rng)\n init_params = genie.init(_rng, dummy_inputs)\n init_params = restore_genie_components(\n init_params, args.tokenizer_checkpoint, args.lam_checkpoint\n )\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=genie.apply, params=init_params, tx=tx)\n\n # --- TRAIN LOOP ---\n dataloader = get_dataloader(args.data_dir, args.seq_len, args.batch_size)\n step = 0\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _mask_rng = jax.random.split(rng, 3)\n inputs = dict(\n videos=videos,\n action=jnp.zeros((args.batch_size, args.seq_len), dtype=jnp.float32),\n dropout_rng=_rng,\n mask_rng=_mask_rng,\n )\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0:\n wandb.log({""loss"": loss, ""step"": step, **metrics})\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[15])),\n recon=wandb.Image(np.asarray(recon_seq[15])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""genie_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break\n",python,tab +2,910,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"10:41:19 AM [info] Activating crowd-code\n10:41:19 AM [info] Recording started\n10:41:19 AM [info] Initializing git provider using file system watchers...\n",Log,tab +3,1419,"extension-output-pdoom-org.crowd-code-#1-crowd-code",153,0,"10:41:19 AM [info] Git repository found\n10:41:19 AM [info] Git provider initialized successfully\n10:41:19 AM [info] Initial git state: [object Object]\n",Log,content diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c0873732-2064-4df8-be5c-5c4fe03049aa1751367864693-2025_07_01-13.05.12.330/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c0873732-2064-4df8-be5c-5c4fe03049aa1751367864693-2025_07_01-13.05.12.330/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..87d6c9f3402b5409a1e3fbabe52632c561f040e8 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c0873732-2064-4df8-be5c-5c4fe03049aa1751367864693-2025_07_01-13.05.12.330/source.csv @@ -0,0 +1,204 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +2,663,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"1:05:12 PM [info] Activating crowd-code\n1:05:12 PM [info] Recording started\n1:05:12 PM [info] Initializing git provider using file system watchers...\n1:05:12 PM [info] Git repository found\n1:05:12 PM [info] Git provider initialized successfully\n1:05:12 PM [info] Initial git state: [object Object]\n",Log,tab +3,3981,"TERMINAL",0,0,"/bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt",,terminal_command +4,4032,"TERMINAL",0,0,"]633;E;2025-07-01 13:05:16 /bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt;9ca2de01-018b-4814-8d7c-cf21dc51d48c]633;C",,terminal_output +5,4078,"TERMINAL",0,0,"]0;tum_cte0515@hkn1991:/hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash]633;D;0",,terminal_output +6,140668,"TERMINAL",0,0,"ls $ws_dir",,terminal_command +7,140713,"TERMINAL",0,0,"]633;E;2025-07-01 13:07:32 ls $ws_dir;4d11dbdc-690b-4257-b927-bbd493ebfa56]633;C",,terminal_output +8,140977,"TERMINAL",0,0,"checkpoints knoms_arrayrecords_500_shards knoms_tfrecords knoms_tfrecords_500_shards_overfit_1 open_ai_minecraft_first_try_npy overfit_dir\r\ncoinrun knoms_mp4 knoms_tfrecords_200_shards knoms_tfrecords_500_shards_overfit_10 open_ai_minecraft_first_try_tfrecord procgen_env_16_episodes_20000\r\ndata_knoms knoms_mp4_clips knoms_tfrecords_2_shards_overfit open_ai_minecraft open_ai_minecraft_npy\r\ndummy knoms_npy knoms_tfrecords_500_shards open_ai_minecraft_first_try open_ai_minecraft_tfrecord\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output +9,145213,"TERMINAL",0,0,"ls $ws_dir/../",,terminal_command +10,145252,"TERMINAL",0,0,"]633;E;2025-07-01 13:07:37 ls $ws_dir/../;4d11dbdc-690b-4257-b927-bbd493ebfa56]633;Ccheckpoints count_items.sh data logs scripts\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output +11,147530,"TERMINAL",0,0,"ls $ws_dir/../logs",,terminal_command +12,147592,"TERMINAL",0,0,"]633;E;2025-07-01 13:07:39 ls $ws_dir/../logs;4d11dbdc-690b-4257-b927-bbd493ebfa56]633;C3306965 logs_alfred\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output +13,153045,"TERMINAL",0,0,"ls $ws_dir/../logs/logs_alfred",,terminal_command +14,153102,"TERMINAL",0,0,"]633;E;2025-07-01 13:07:45 ls $ws_dir/../logs/logs_alfred;4d11dbdc-690b-4257-b927-bbd493ebfa56]633;Clogs_training_dyn logs_training_tokenizer\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output +15,162155,"TERMINAL",0,0,"ls $ws_dir/../logs/logs_alfred/logs_training_tokenizer",,terminal_command +16,162198,"TERMINAL",0,0,"]633;E;2025-07-01 13:07:54 ls $ws_dir/../logs/logs_alfred/logs_training_tokenizer;4d11dbdc-690b-4257-b927-bbd493ebfa56]633;C",,terminal_output +17,162274,"TERMINAL",0,0,"train_tokenizer_batch_size_scaling_16_node_3307475.log train_tokenizer_batch_size_scaling_2_node_3307422.log train_tokenizer_overfit_sample_size_0_5_21mio_3297706.log\r\ntrain_tokenizer_batch_size_scaling_1_node_3307372.log train_tokenizer_batch_size_scaling_2_node_3307472.log train_tokenizer_overfit_sample_size_0_5_3297582.log\r\ntrain_tokenizer_batch_size_scaling_1_node_3307421.log train_tokenizer_batch_size_scaling_4_node_3307473.log train_tokenizer_overfit_sample_size_0_5_3297586.log\r\ntrain_tokenizer_batch_size_scaling_1_node_3307470.log train_tokenizer_batch_size_scaling_4_node_3307688.log train_tokenizer_overfit_sample_size_0_5_9mio_3297693.log\r\ntrain_tokenizer_batch_size_scaling_2_node_3307416.log train_tokenizer_batch_size_scaling_8_node_3307474.log\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output +18,174522,"TERMINAL",0,0,"ls $ws_dir/../logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",,terminal_command +19,174550,"TERMINAL",0,0,"]633;E;2025-07-01 13:08:06 ls $ws_dir/../logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log;4d11dbdc-690b-4257-b927-bbd493ebfa56]633;C/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/../logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output +20,177362,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=10:00:00\n#SBATCH --account=hk-project-p0023960\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=8\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/%x_%j.log\n#SBATCH --mail-user=avocadoaling@gmail.com\n#SBATCH --mail-type=ALL\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv_jafar/bin/activate\n\nJOB_NAME=$SLURM_JOB_NAME\nSLURM_JOB_ID=$SLURM_JOB_ID\nTAGS=(tokenizer batch-size-scaling 8-node time-step adjusted-lr)\n\nWS_DIR='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\nTF_RECORDS_DIR=$WS_DIR/data/open_ai_minecraft_tfrecord\nCHECKPOINT_DIR=$WS_DIR/checkpoints/tokenizer/${JOB_NAME}_${SLURM_JOB_ID}\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=8.49e-4 \\n --max_lr=8.49e-4 \\n --log_image_interval=250 \\n --log \\n --name=${JOB_NAME}-${SLURM_JOB_ID} \\n --tags $TAGS \\n --entity instant-uv \\n --project jafar \\n --log_checkpoint_interval=1000 \\n --data_dir $TF_RECORDS_DIR\nSLURM_JOB_USER=tum_ind3695\nSLURM_TASKS_PER_NODE=4(x8)\nSLURM_JOB_UID=991285\nSLURM_TASK_PID=634526\nSLURM_JOB_GPUS=0,1,2,3\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling\nSLURMD_NODENAME=hkn0420\nSLURM_JOB_START_TIME=1751318063\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1751354063\nSLURM_CPUS_ON_NODE=32\nSLURM_JOB_CPUS_PER_NODE=32(x8)\nSLURM_GPUS_ON_NODE=4\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=accelerated\nSLURM_TRES_PER_TASK=cpu=8\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=8\nSLURM_JOBID=3307474\nSLURM_JOB_QOS=normal\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=8\nSLURM_NTASKS=32\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e12.hkn0420\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_NODELIST=hkn[0420,0516,0520,0523,0704,0717,0810,0820]\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=32\nSLURM_NNODES=8\nSLURM_SUBMIT_HOST=hkn1991.localdomain\nSLURM_JOB_ID=3307474\nSLURM_NODEID=0\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_tokenizer_batch_size_scaling_8_node\nSLURM_NTASKS_PER_NODE=4\nSLURM_JOB_GID=502289\nSLURM_JOB_NODELIST=hkn[0420,0516,0520,0523,0704,0717,0810,0820]\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\nGpuFreq=control_disabled\n2025-06-30 23:14:53.456775: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:14:53.456867: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:14:53.457066: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:14:53.457602: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318093.469657 2856352 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318093.469881 2856354 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318093.470016 2856351 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318093.470869 2856353 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nE0000 00:00:1751318093.474025 2856352 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318093.474537 2856354 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318093.474636 2856351 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318093.475228 2856353 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nW0000 00:00:1751318093.487901 2856353 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.487916 2856353 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.487918 2856353 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.487919 2856353 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.488292 2856352 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.488308 2856352 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.488310 2856352 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.488312 2856352 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.488485 2856351 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.488502 2856351 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.488503 2856351 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.488505 2856351 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.488598 2856354 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.488615 2856354 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.488617 2856354 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318093.488619 2856354 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\n2025-06-30 23:14:54.547413: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:14:54.547731: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:14:54.547735: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:14:54.547955: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318094.560520 2160604 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318094.560830 2160602 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318094.560803 2160603 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318094.561295 2160605 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nE0000 00:00:1751318094.565025 2160604 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318094.565476 2160602 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318094.565428 2160603 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318094.566045 2160605 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2025-06-30 23:14:54.566659: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:14:54.566657: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:14:54.566878: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:14:54.567170: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\nW0000 00:00:1751318094.578832 2160604 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.578847 2160604 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.578849 2160604 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.578850 2160604 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.579318 2160602 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.579333 2160602 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.579335 2160602 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.579337 2160602 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.579319 2160603 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.579333 2160603 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.579336 2160603 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.579337 2160603 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.579563 2160605 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.579580 2160605 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.579582 2160605 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.579583 2160605 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318094.579697 2427863 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318094.579746 2427864 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318094.579766 2427865 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318094.580755 2427862 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nE0000 00:00:1751318094.584398 2427863 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318094.584385 2427864 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318094.584400 2427865 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318094.585575 2427862 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nW0000 00:00:1751318094.598710 2427863 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.598724 2427863 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.598725 2427863 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.598727 2427863 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.598695 2427864 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.598711 2427864 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.598713 2427864 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.598715 2427864 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.598698 2427865 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.598713 2427865 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.598716 2427865 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.598717 2427865 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.599187 2427862 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.599205 2427862 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.599207 2427862 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.599208 2427862 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\n2025-06-30 23:14:54.661486: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:14:54.661507: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:14:54.661685: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:14:54.661718: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318094.674450 634594 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318094.674449 634595 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318094.674517 634596 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318094.674538 634593 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nE0000 00:00:1751318094.678694 634596 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318094.678813 634594 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318094.679115 634593 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318094.679028 634595 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nW0000 00:00:1751318094.692678 634593 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692692 634593 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692694 634593 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692696 634593 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692652 634594 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692665 634594 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692667 634594 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692668 634594 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692666 634595 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692682 634595 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692683 634595 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692685 634595 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692627 634596 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692643 634596 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692645 634596 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318094.692646 634596 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318099.464209 2856354 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318099.464573 2856353 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318099.464844 2856351 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318099.467389 2856352 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318099.527601 2160602 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318099.527693 2160603 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318099.527574 2160604 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318099.527854 2160605 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318099.612797 2427862 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318099.612805 2427863 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318099.612785 2427864 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318099.612795 2427865 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318100.065098 634594 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318100.065192 634595 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318100.065238 634593 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318100.065260 634596 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\n2025-06-30 23:15:01.213957: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:15:01.213951: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:15:01.213959: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:15:01.213955: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318101.384363 1578014 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318101.384257 1578015 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318101.384268 1578017 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318101.384491 1578016 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nE0000 00:00:1751318101.391716 1578014 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318101.391711 1578015 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318101.391718 1578016 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318101.391711 1578017 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nW0000 00:00:1751318101.879162 1578014 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879190 1578014 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879193 1578014 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879194 1578014 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879159 1578015 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879189 1578015 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879190 1578015 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879192 1578015 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879165 1578016 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879191 1578016 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879193 1578016 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879195 1578016 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879160 1578017 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879185 1578017 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879187 1578017 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318101.879188 1578017 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\n2025-06-30 23:15:02.184361: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:15:02.184363: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:15:02.184357: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:15:02.184357: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318102.214725 1012654 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318102.214631 1012655 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318102.214813 1012656 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318102.214638 1012657 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nE0000 00:00:1751318102.234122 1012655 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318102.234118 1012657 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318102.234789 1012654 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318102.234772 1012656 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nW0000 00:00:1751318102.361690 1012654 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361715 1012654 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361717 1012654 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361719 1012654 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361694 1012655 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361718 1012655 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361720 1012655 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361721 1012655 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361687 1012656 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361712 1012656 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361715 1012656 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361716 1012656 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361696 1012657 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361714 1012657 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361717 1012657 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.361718 1012657 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\n2025-06-30 23:15:02.368657: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:15:02.368656: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:15:02.368655: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:15:02.368654: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318102.398233 3005151 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318102.398299 3005153 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318102.398548 3005152 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318102.398657 3005154 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nE0000 00:00:1751318102.404511 3005151 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318102.404506 3005152 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318102.404509 3005153 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318102.404508 3005154 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2025-06-30 23:15:02.421302: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:15:02.421417: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:15:02.421303: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2025-06-30 23:15:02.421425: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:467] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318102.455392 3894030 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318102.455330 3894031 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318102.455513 3894032 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nWARNING: All log messages before absl::InitializeLog() is called are written to STDERR\nE0000 00:00:1751318102.455482 3894033 cuda_dnn.cc:8579] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\nE0000 00:00:1751318102.461184 3894031 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318102.461184 3894032 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318102.461181 3894033 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nE0000 00:00:1751318102.462009 3894030 cuda_blas.cc:1407] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\nW0000 00:00:1751318102.504381 3005151 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504408 3005151 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504410 3005151 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504412 3005151 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504377 3005152 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504402 3005152 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504405 3005152 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504407 3005152 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504382 3005153 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504409 3005153 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504412 3005153 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504414 3005153 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504380 3005154 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504405 3005154 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504407 3005154 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.504409 3005154 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534858 3894030 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534885 3894030 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534887 3894030 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534889 3894030 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534860 3894031 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534886 3894031 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534888 3894031 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534890 3894031 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534857 3894032 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534883 3894032 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534885 3894032 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534887 3894032 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534855 3894033 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534878 3894033 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534881 3894033 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318102.534883 3894033 computation_placer.cc:177] computation placer already registered. Please check linkage and avoid linking the same target more than once.\nW0000 00:00:1751318118.615248 1578017 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318118.615450 1578015 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318118.615611 1578014 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318118.615637 1578016 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318120.944360 3894031 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318120.944433 3894032 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318120.944359 3894033 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318120.944517 3894030 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318122.638730 3005151 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318122.638731 3005153 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318122.638863 3005154 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318122.638879 3005152 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318123.581820 1012654 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318123.581743 1012655 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318123.581753 1012657 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nW0000 00:00:1751318123.581744 1012656 gpu_device.cc:2341] Cannot dlopen some GPU libraries. Please make sure the missing libraries mentioned above are installed properly if you would like to use GPU. Follow the guide at https://www.tensorflow.org/install/gpu for how to download and setup the required libraries for your platform.\nSkipping registering GPU devices...\nwandb: Currently logged in as: avocadoali (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\nwandb: Tracking run with wandb version 0.19.11\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/wandb/run-20250630_231525-hzw9mfvm\nwandb: Run `wandb offline` to turn off syncing.\nwandb: Syncing run train_tokenizer_batch_size_scaling_8_node-3307474\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/hzw9mfvm\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 115, loss: 0.17339496314525604\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 115, loss: 0.17339496314525604\nStep 115, loss: 0.17339496314525604\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 115, loss: 0.17339496314525604\nStep 115, loss: 0.17339496314525604\nStep 115, loss: 0.17339496314525604\nStep 115, loss: 0.17339496314525604\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 115, loss: 0.17339496314525604\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 115, loss: 0.17339496314525604\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 115, loss: 0.17339496314525604\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 115, loss: 0.17339496314525604\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 115, loss: 0.17339496314525604\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 115, loss: 0.17339496314525604\nStep 115, loss: 0.17339496314525604\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 115, loss: 0.17339496314525604\nStep 115, loss: 0.17339496314525604\nStep 115, loss: 0.17339496314525604\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 115, loss: 0.17339496314525604\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 115, loss: 0.17339496314525604\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 115, loss: 0.17339496314525604\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 115, loss: 0.17339496314525604\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 115, loss: 0.17339496314525604\nStep 115, loss: 0.17339496314525604\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 115, loss: 0.17339496314525604\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 115, loss: 0.17339496314525604\nStep 115, loss: 0.17339496314525604\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 115, loss: 0.17339496314525604\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 115, loss: 0.17339496314525604\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nRunning on 32 devices.\nStarting training from step 0...\nStep 0, loss: 0.3471579849720001\nStep 1, loss: 0.26770803332328796\nStep 2, loss: 0.14104889333248138\nStep 3, loss: 0.13309144973754883\nStep 4, loss: 0.13091082870960236\nStep 5, loss: 0.13509593904018402\nStep 6, loss: 0.15321968495845795\nStep 7, loss: 0.14519496262073517\nStep 8, loss: 0.14099560678005219\nStep 9, loss: 0.1463150680065155\nStep 10, loss: 0.14669667184352875\nStep 11, loss: 0.1430281549692154\nStep 12, loss: 0.13978496193885803\nStep 13, loss: 0.13720174133777618\nStep 14, loss: 0.14147496223449707\nStep 15, loss: 0.12998127937316895\nStep 16, loss: 0.12920445203781128\nStep 17, loss: 0.13227719068527222\nStep 18, loss: 0.125120609998703\nStep 19, loss: 0.13470731675624847\nStep 20, loss: 0.12474726885557175\nStep 21, loss: 0.1333005726337433\nStep 22, loss: 0.1423025131225586\nStep 23, loss: 0.1377020627260208\nStep 24, loss: 0.1437397450208664\nStep 25, loss: 0.13958904147148132\nStep 26, loss: 0.13821439445018768\nStep 27, loss: 0.14022579789161682\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 115, loss: 0.17339496314525604\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 115, loss: 0.17339496314525604\nStep 115, loss: 0.17339496314525604\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 28, loss: 0.13478153944015503\nStep 29, loss: 0.1434846669435501\nStep 30, loss: 0.13330970704555511\nStep 31, loss: 0.14295107126235962\nStep 32, loss: 0.14157991111278534\nStep 33, loss: 0.13778111338615417\nStep 34, loss: 0.1367509663105011\nStep 35, loss: 0.14291462302207947\nStep 36, loss: 0.13713470101356506\nStep 37, loss: 0.1495264619588852\nStep 38, loss: 0.15450292825698853\nStep 39, loss: 0.1477455347776413\nStep 40, loss: 0.14942966401576996\nStep 41, loss: 0.15012970566749573\nStep 42, loss: 0.14514152705669403\nStep 43, loss: 0.1656753122806549\nStep 44, loss: 0.15873289108276367\nStep 45, loss: 0.17154277861118317\nStep 46, loss: 0.17314167320728302\nStep 47, loss: 0.17424646019935608\nStep 48, loss: 0.17465342581272125\nStep 49, loss: 0.17834438383579254\nStep 50, loss: 0.1750008463859558\nStep 51, loss: 0.17383235692977905\nStep 52, loss: 0.1738968789577484\nStep 53, loss: 0.1788439154624939\nStep 54, loss: 0.17455320060253143\nStep 55, loss: 0.17899788916110992\nStep 56, loss: 0.17514869570732117\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 57, loss: 0.17526665329933167\nStep 58, loss: 0.17442293465137482\nStep 59, loss: 0.1755731999874115\nStep 60, loss: 0.17686791718006134\nStep 61, loss: 0.1774514615535736\nStep 62, loss: 0.17734190821647644\nStep 63, loss: 0.17445750534534454\nStep 64, loss: 0.1791164129972458\nStep 65, loss: 0.17443639039993286\nStep 66, loss: 0.1721915602684021\nStep 67, loss: 0.17724011838436127\nStep 68, loss: 0.1726222038269043\nStep 69, loss: 0.17514818906784058\nStep 70, loss: 0.17387129366397858\nStep 71, loss: 0.17663298547267914\nStep 72, loss: 0.17917156219482422\nStep 73, loss: 0.17480577528476715\nStep 74, loss: 0.17429769039154053\nStep 75, loss: 0.17326530814170837\nStep 76, loss: 0.17588405311107635\nStep 77, loss: 0.16884107887744904\nStep 78, loss: 0.17379479110240936\nStep 79, loss: 0.17513428628444672\nStep 80, loss: 0.17434917390346527\nStep 81, loss: 0.17318695783615112\nStep 82, loss: 0.17380864918231964\nStep 83, loss: 0.17578117549419403\nStep 84, loss: 0.1710706204175949\nStep 85, loss: 0.17886455357074738\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 86, loss: 0.17388099431991577\nStep 87, loss: 0.17131978273391724\nStep 88, loss: 0.17590537667274475\nStep 89, loss: 0.1696556806564331\nStep 90, loss: 0.1746663749217987\nStep 91, loss: 0.1718718260526657\nStep 92, loss: 0.17620952427387238\nStep 93, loss: 0.17497602105140686\nStep 94, loss: 0.17112667858600616\nStep 95, loss: 0.17390812933444977\nStep 96, loss: 0.17191694676876068\nStep 97, loss: 0.17629854381084442\nStep 98, loss: 0.17539462447166443\nStep 99, loss: 0.173387348651886\nStep 100, loss: 0.17558738589286804\nStep 101, loss: 0.17344261705875397\nStep 102, loss: 0.17063868045806885\nStep 103, loss: 0.17445780336856842\nStep 104, loss: 0.17477689683437347\nStep 105, loss: 0.1743089109659195\nStep 106, loss: 0.17661947011947632\nStep 107, loss: 0.17260797321796417\nStep 108, loss: 0.17181360721588135\nStep 109, loss: 0.17290376126766205\nStep 110, loss: 0.17456933856010437\nStep 111, loss: 0.17321565747261047\nStep 112, loss: 0.17227257788181305\nStep 113, loss: 0.1743168830871582\nStep 114, loss: 0.16588141024112701\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 115, loss: 0.17339496314525604\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 116, loss: 0.17650632560253143\nStep 117, loss: 0.1721123307943344\nStep 118, loss: 0.17484426498413086\nStep 119, loss: 0.17425554990768433\nStep 120, loss: 0.17355427145957947\nStep 121, loss: 0.17475546896457672\nStep 122, loss: 0.1722601056098938\nStep 123, loss: 0.17617866396903992\nStep 124, loss: 0.1784822940826416\nStep 125, loss: 0.17228849232196808\nStep 126, loss: 0.1726805716753006\nStep 127, loss: 0.17028960585594177\nStep 128, loss: 0.1759348213672638\nStep 129, loss: 0.1740720421075821\nStep 130, loss: 0.17981602251529694\nStep 131, loss: 0.1703324317932129\nStep 132, loss: 0.17444254457950592\nStep 133, loss: 0.17495402693748474\nStep 134, loss: 0.1774214208126068\nStep 135, loss: 0.16589730978012085\nStep 136, loss: 0.17302198708057404\nStep 137, loss: 0.17555880546569824\nStep 138, loss: 0.1742158681154251\nStep 139, loss: 0.172081857919693\nStep 140, loss: 0.17138473689556122\nStep 141, loss: 0.17552287876605988\nStep 142, loss: 0.17444556951522827\nStep 143, loss: 0.1737712323665619\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 144, loss: 0.17160776257514954\nStep 145, loss: 0.17339970171451569\nStep 146, loss: 0.17081758379936218\nStep 147, loss: 0.16918820142745972\nStep 148, loss: 0.1746884435415268\nStep 149, loss: 0.17227376997470856\nStep 150, loss: 0.17421938478946686\nStep 151, loss: 0.1723541021347046\nStep 152, loss: 0.17495852708816528\nStep 153, loss: 0.1736605316400528\nStep 154, loss: 0.1731863170862198\nStep 155, loss: 0.173353374004364\nStep 156, loss: 0.17461292445659637\nStep 157, loss: 0.17833106219768524\nStep 158, loss: 0.1696862280368805\nStep 159, loss: 0.17354346811771393\nStep 160, loss: 0.17869757115840912\nStep 161, loss: 0.16943944990634918\nStep 162, loss: 0.17502373456954956\nStep 163, loss: 0.17490755021572113\nStep 164, loss: 0.1724388152360916\nStep 165, loss: 0.1717279553413391\nStep 166, loss: 0.1696639358997345\nStep 167, loss: 0.1763082593679428\nStep 168, loss: 0.17012852430343628\nStep 169, loss: 0.1756332814693451\nStep 170, loss: 0.17349863052368164\nStep 171, loss: 0.17180608212947845\nStep 172, loss: 0.1735956072807312\nStep 173, loss: 0.16950492560863495\nStep 174, loss: 0.1710001528263092\nStep 175, loss: 0.1720818281173706\nStep 176, loss: 0.17734099924564362\nStep 177, loss: 0.17544645071029663\nStep 178, loss: 0.17164328694343567\nStep 179, loss: 0.17365553975105286\nStep 180, loss: 0.17731669545173645\nStep 181, loss: 0.16690389811992645\nStep 182, loss: 0.1742502599954605\nStep 183, loss: 0.1716139167547226\nStep 184, loss: 0.1698140799999237\nStep 185, loss: 0.17759019136428833\nStep 186, loss: 0.17442889511585236\nStep 187, loss: 0.17123228311538696\nStep 188, loss: 0.17149075865745544\nStep 189, loss: 0.17389832437038422\nStep 190, loss: 0.1708046942949295\nStep 191, loss: 0.171993225812912\nStep 192, loss: 0.17431600391864777\nStep 193, loss: 0.17453940212726593\nStep 194, loss: 0.1709374040365219\nStep 195, loss: 0.17524591088294983\nStep 196, loss: 0.17337879538536072\nStep 197, loss: 0.1743336170911789\nStep 198, loss: 0.1775231659412384\nStep 199, loss: 0.16916300356388092\nStep 200, loss: 0.17302356660366058\nStep 201, loss: 0.17107906937599182\nStep 202, loss: 0.17300015687942505\nStep 203, loss: 0.1730821281671524\nStep 204, loss: 0.17554880678653717\nStep 205, loss: 0.1745760142803192\nStep 206, loss: 0.17443722486495972\nStep 207, loss: 0.17716984450817108\nStep 208, loss: 0.178369402885437\nStep 209, loss: 0.17524774372577667\nStep 210, loss: 0.17759115993976593\nStep 211, loss: 0.17509549856185913\nStep 212, loss: 0.1736840456724167\nStep 213, loss: 0.1769413948059082\nStep 214, loss: 0.17503388226032257\nStep 215, loss: 0.17699705064296722\nStep 216, loss: 0.17280948162078857\nStep 217, loss: 0.17379453778266907\nStep 218, loss: 0.17551490664482117\nStep 219, loss: 0.1724730134010315\nStep 220, loss: 0.17256437242031097\nStep 221, loss: 0.17449802160263062\nStep 222, loss: 0.1751682460308075\nStep 223, loss: 0.17546094954013824\nStep 224, loss: 0.17379115521907806\nStep 225, loss: 0.1778661012649536\nStep 226, loss: 0.17416249215602875\nStep 227, loss: 0.17225857079029083\nStep 228, loss: 0.17188940942287445\nStep 229, loss: 0.17189325392246246\nStep 230, loss: 0.17526964843273163\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 459, loss: 0.17431780695915222\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 459, loss: 0.17431780695915222\nStep 459, loss: 0.17431780695915222\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 459, loss: 0.17431780695915222\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 459, loss: 0.17431780695915222\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 459, loss: 0.17431780695915222\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 459, loss: 0.17431780695915222\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 459, loss: 0.17431780695915222\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 459, loss: 0.17431780695915222\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 459, loss: 0.17431780695915222\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 459, loss: 0.17431780695915222\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 459, loss: 0.17431780695915222\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 459, loss: 0.17431780695915222\nStep 459, loss: 0.17431780695915222\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 459, loss: 0.17431780695915222\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 459, loss: 0.17431780695915222\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 459, loss: 0.17431780695915222\nStep 459, loss: 0.17431780695915222\nStep 459, loss: 0.17431780695915222\nStep 459, loss: 0.17431780695915222\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 459, loss: 0.17431780695915222\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 231, loss: 0.17420436441898346\nStep 232, loss: 0.1751464307308197\nStep 233, loss: 0.17173068225383759\nStep 234, loss: 0.17404291033744812\nStep 235, loss: 0.174527108669281\nStep 236, loss: 0.17618782818317413\nStep 237, loss: 0.17089204490184784\nStep 238, loss: 0.17432117462158203\nStep 239, loss: 0.17289616167545319\nStep 240, loss: 0.17036113142967224\nStep 241, loss: 0.17384111881256104\nStep 242, loss: 0.17676493525505066\nStep 243, loss: 0.17606355249881744\nStep 244, loss: 0.17296987771987915\nStep 245, loss: 0.17211146652698517\nStep 246, loss: 0.16899186372756958\nStep 247, loss: 0.1751350313425064\nStep 248, loss: 0.17139440774917603\nStep 249, loss: 0.175702303647995\nStep 250, loss: 0.17180339992046356\nStep 251, loss: 0.1714021861553192\nStep 252, loss: 0.17811693251132965\nStep 253, loss: 0.1751716434955597\nStep 254, loss: 0.17288176715373993\nStep 255, loss: 0.1767214834690094\nStep 256, loss: 0.17531660199165344\nStep 257, loss: 0.1743871569633484\nStep 258, loss: 0.17458227276802063\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 259, loss: 0.1700548678636551\nStep 260, loss: 0.17513228952884674\nStep 261, loss: 0.16957631707191467\nStep 262, loss: 0.17602184414863586\nStep 263, loss: 0.16820436716079712\nStep 264, loss: 0.17445197701454163\nStep 265, loss: 0.17325876653194427\nStep 266, loss: 0.17151372134685516\nStep 267, loss: 0.17281821370124817\nStep 268, loss: 0.17472431063652039\nStep 269, loss: 0.17343536019325256\nStep 270, loss: 0.174852654337883\nStep 271, loss: 0.17498376965522766\nStep 272, loss: 0.17547272145748138\nStep 273, loss: 0.17591245472431183\nStep 274, loss: 0.1728111356496811\nStep 275, loss: 0.17593519389629364\nStep 276, loss: 0.17476886510849\nStep 277, loss: 0.17651814222335815\nStep 278, loss: 0.17160353064537048\nStep 279, loss: 0.16974833607673645\nStep 280, loss: 0.1757923811674118\nStep 281, loss: 0.1767502725124359\nStep 282, loss: 0.16855189204216003\nStep 283, loss: 0.1713992804288864\nStep 284, loss: 0.17589080333709717\nStep 285, loss: 0.17468930780887604\nStep 286, loss: 0.17121709883213043\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 287, loss: 0.17423152923583984\nStep 288, loss: 0.17463481426239014\nStep 289, loss: 0.17352049052715302\nStep 290, loss: 0.17575746774673462\nStep 291, loss: 0.17079386115074158\nStep 292, loss: 0.1753157377243042\nStep 293, loss: 0.17148235440254211\nStep 294, loss: 0.17589227855205536\nStep 295, loss: 0.17230594158172607\nStep 296, loss: 0.17561212182044983\nStep 297, loss: 0.17357952892780304\nStep 298, loss: 0.17742855846881866\nStep 299, loss: 0.17149633169174194\nStep 300, loss: 0.17210690677165985\nStep 301, loss: 0.17151214182376862\nStep 302, loss: 0.17388419806957245\nStep 303, loss: 0.1740708202123642\nStep 304, loss: 0.17560219764709473\nStep 305, loss: 0.17067687213420868\nStep 306, loss: 0.16773159801959991\nStep 307, loss: 0.171116903424263\nStep 308, loss: 0.1745900809764862\nStep 309, loss: 0.17308343946933746\nStep 310, loss: 0.17476658523082733\nStep 311, loss: 0.17111080884933472\nStep 312, loss: 0.17190101742744446\nStep 313, loss: 0.16911008954048157\nStep 314, loss: 0.1785719096660614\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 315, loss: 0.17609170079231262\nStep 316, loss: 0.1770344078540802\nStep 317, loss: 0.17321272194385529\nStep 318, loss: 0.1736145317554474\nStep 319, loss: 0.17849916219711304\nStep 320, loss: 0.16995446383953094\nStep 321, loss: 0.17075978219509125\nStep 322, loss: 0.1684204787015915\nStep 323, loss: 0.1707623153924942\nStep 324, loss: 0.16800695657730103\nStep 325, loss: 0.1711212694644928\nStep 326, loss: 0.17421497404575348\nStep 327, loss: 0.17173270881175995\nStep 328, loss: 0.1733817309141159\nStep 329, loss: 0.1723349690437317\nStep 330, loss: 0.17115896940231323\nStep 331, loss: 0.17340612411499023\nStep 332, loss: 0.17287622392177582\nStep 333, loss: 0.17327342927455902\nStep 334, loss: 0.17215271294116974\nStep 335, loss: 0.17460770905017853\nStep 336, loss: 0.17455795407295227\nStep 337, loss: 0.17333874106407166\nStep 338, loss: 0.1695828139781952\nStep 339, loss: 0.17012254893779755\nStep 340, loss: 0.17383894324302673\nStep 341, loss: 0.1735590249300003\nStep 342, loss: 0.17276807129383087\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 343, loss: 0.16952309012413025\nStep 344, loss: 0.17104601860046387\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 345, loss: 0.17573249340057373\nStep 346, loss: 0.1728946715593338\nStep 347, loss: 0.17545942962169647\nStep 348, loss: 0.17130595445632935\nStep 349, loss: 0.17322826385498047\nStep 350, loss: 0.1729235053062439\nStep 351, loss: 0.17295922338962555\nStep 352, loss: 0.1735742837190628\nStep 353, loss: 0.1697748452425003\nStep 354, loss: 0.17621870338916779\nStep 355, loss: 0.1742292046546936\nStep 356, loss: 0.17279407382011414\nStep 357, loss: 0.17330560088157654\nStep 358, loss: 0.17479054629802704\nStep 359, loss: 0.17317044734954834\nStep 360, loss: 0.17544831335544586\nStep 361, loss: 0.17814604938030243\nStep 362, loss: 0.17601491510868073\nStep 363, loss: 0.17304247617721558\nStep 364, loss: 0.17508576810359955\nStep 365, loss: 0.1717367172241211\nStep 366, loss: 0.1739933043718338\nStep 367, loss: 0.17015735805034637\nStep 368, loss: 0.16894015669822693\nStep 369, loss: 0.17233464121818542\nStep 370, loss: 0.17584733664989471\nStep 371, loss: 0.17348554730415344\nStep 372, loss: 0.17454135417938232\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 373, loss: 0.1766464114189148\nStep 374, loss: 0.17427922785282135\nStep 375, loss: 0.16924604773521423\nStep 376, loss: 0.1737813949584961\nStep 377, loss: 0.17719173431396484\nStep 378, loss: 0.17514315247535706\nStep 379, loss: 0.173935204744339\nStep 380, loss: 0.17518827319145203\nStep 381, loss: 0.17630116641521454\nStep 382, loss: 0.17496182024478912\nStep 383, loss: 0.17837847769260406\nStep 384, loss: 0.1724771112203598\nStep 385, loss: 0.17744266986846924\nStep 386, loss: 0.17381663620471954\nStep 387, loss: 0.17394840717315674\nStep 388, loss: 0.17453908920288086\nStep 389, loss: 0.1730494201183319\nStep 390, loss: 0.17195409536361694\nStep 391, loss: 0.17394614219665527\nStep 392, loss: 0.17362229526042938\nStep 393, loss: 0.1768149733543396\nStep 394, loss: 0.16851279139518738\nStep 395, loss: 0.17069624364376068\nStep 396, loss: 0.17005404829978943\nStep 397, loss: 0.1732345074415207\nStep 398, loss: 0.17623025178909302\nStep 399, loss: 0.17276795208454132\nStep 400, loss: 0.1697579026222229\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 459, loss: 0.17431780695915222\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 459, loss: 0.17431780695915222\nStep 401, loss: 0.1745413988828659\nStep 402, loss: 0.17456986010074615\nStep 403, loss: 0.17597635090351105\nStep 404, loss: 0.17094317078590393\nStep 405, loss: 0.17680500447750092\nStep 406, loss: 0.17142018675804138\nStep 407, loss: 0.1751982569694519\nStep 408, loss: 0.17486493289470673\nStep 409, loss: 0.17512650787830353\nStep 410, loss: 0.17006459832191467\nStep 411, loss: 0.16990330815315247\nStep 412, loss: 0.17894293367862701\nStep 413, loss: 0.1735304743051529\nStep 414, loss: 0.17329593002796173\nStep 415, loss: 0.17388761043548584\nStep 416, loss: 0.17462217807769775\nStep 417, loss: 0.17329254746437073\nStep 418, loss: 0.17006121575832367\nStep 419, loss: 0.17535844445228577\nStep 420, loss: 0.17794457077980042\nStep 421, loss: 0.17409290373325348\nStep 422, loss: 0.1734360158443451\nStep 423, loss: 0.17678727209568024\nStep 424, loss: 0.17418088018894196\nStep 425, loss: 0.17600953578948975\nStep 426, loss: 0.17062707245349884\nStep 427, loss: 0.17593447864055634\nStep 428, loss: 0.177713543176651\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 459, loss: 0.17431780695915222\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 459, loss: 0.17431780695915222\nStep 429, loss: 0.17125695943832397\nStep 430, loss: 0.17908503115177155\nStep 431, loss: 0.17312182486057281\nStep 432, loss: 0.17226175963878632\nStep 433, loss: 0.17499332129955292\nStep 434, loss: 0.1692870408296585\nStep 435, loss: 0.17121432721614838\nStep 436, loss: 0.17658698558807373\nStep 437, loss: 0.17520497739315033\nStep 438, loss: 0.1722518801689148\nStep 439, loss: 0.16925294697284698\nStep 440, loss: 0.17579831182956696\nStep 441, loss: 0.17457902431488037\nStep 442, loss: 0.17101430892944336\nStep 443, loss: 0.17209774255752563\nStep 444, loss: 0.17333845794200897\nStep 445, loss: 0.17015598714351654\nStep 446, loss: 0.17298994958400726\nStep 447, loss: 0.17452651262283325\nStep 448, loss: 0.17778415977954865\nStep 449, loss: 0.17184945940971375\nStep 450, loss: 0.1754949539899826\nStep 451, loss: 0.1755322515964508\nStep 452, loss: 0.17120981216430664\nStep 453, loss: 0.16953344643115997\nStep 454, loss: 0.17265845835208893\nStep 455, loss: 0.1755368858575821\nStep 456, loss: 0.1723787635564804\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 459, loss: 0.17431780695915222\nStep 459, loss: 0.17431780695915222\nStep 459, loss: 0.17431780695915222\nStep 457, loss: 0.17290666699409485\nStep 458, loss: 0.1749171018600464\nStep 459, loss: 0.17431780695915222\nStep 459, loss: 0.17431780695915222\nStep 459, loss: 0.17431780695915222\nStep 459, loss: 0.17431780695915222\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 688, loss: 0.168025940656662\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 688, loss: 0.168025940656662\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 688, loss: 0.168025940656662\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 688, loss: 0.168025940656662\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 688, loss: 0.168025940656662\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 688, loss: 0.168025940656662\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 688, loss: 0.168025940656662\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 688, loss: 0.168025940656662\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 460, loss: 0.1741330772638321\nStep 461, loss: 0.17441166937351227\nStep 462, loss: 0.1756756603717804\nStep 463, loss: 0.17418205738067627\nStep 464, loss: 0.17287975549697876\nStep 465, loss: 0.17231808602809906\nStep 466, loss: 0.17125974595546722\nStep 467, loss: 0.17578507959842682\nStep 468, loss: 0.1705021858215332\nStep 469, loss: 0.17248119413852692\nStep 470, loss: 0.17409169673919678\nStep 471, loss: 0.17181046307086945\nStep 472, loss: 0.17418991029262543\nStep 473, loss: 0.1714387834072113\nStep 474, loss: 0.17240971326828003\nStep 475, loss: 0.17331163585186005\nStep 476, loss: 0.17074371874332428\nStep 477, loss: 0.17336028814315796\nStep 478, loss: 0.17391328513622284\nStep 479, loss: 0.17445111274719238\nStep 480, loss: 0.17107941210269928\nStep 481, loss: 0.17514654994010925\nStep 482, loss: 0.17364846169948578\nStep 483, loss: 0.16913525760173798\nStep 484, loss: 0.17099237442016602\nStep 485, loss: 0.17201489210128784\nStep 486, loss: 0.17490939795970917\nStep 487, loss: 0.17496316134929657\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 688, loss: 0.168025940656662\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 488, loss: 0.17247094213962555\nStep 489, loss: 0.17306217551231384\nStep 490, loss: 0.17290695011615753\nStep 491, loss: 0.174490824341774\nStep 492, loss: 0.17366597056388855\nStep 493, loss: 0.17133276164531708\nStep 494, loss: 0.17247363924980164\nStep 495, loss: 0.17463356256484985\nStep 496, loss: 0.17164246737957\nStep 497, loss: 0.1705310344696045\nStep 498, loss: 0.17419204115867615\nStep 499, loss: 0.1729225218296051\nStep 500, loss: 0.17634651064872742\nStep 501, loss: 0.17864353954792023\nStep 502, loss: 0.17430740594863892\nStep 503, loss: 0.17458714544773102\nStep 504, loss: 0.17079102993011475\nStep 505, loss: 0.1753505915403366\nStep 506, loss: 0.17569056153297424\nStep 507, loss: 0.17542502284049988\nStep 508, loss: 0.17260128259658813\nStep 509, loss: 0.1728348284959793\nStep 510, loss: 0.16836656630039215\nStep 511, loss: 0.1756671667098999\nStep 512, loss: 0.17208218574523926\nStep 513, loss: 0.17600437998771667\nStep 514, loss: 0.16866090893745422\nStep 515, loss: 0.17926707863807678\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 688, loss: 0.168025940656662\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 516, loss: 0.17697672545909882\nStep 517, loss: 0.17678968608379364\nStep 518, loss: 0.17956872284412384\nStep 519, loss: 0.17370600998401642\nStep 520, loss: 0.17517651617527008\nStep 521, loss: 0.18297064304351807\nStep 522, loss: 0.17328175902366638\nStep 523, loss: 0.1723300963640213\nStep 524, loss: 0.17295172810554504\nStep 525, loss: 0.17543092370033264\nStep 526, loss: 0.17585870623588562\nStep 527, loss: 0.17599253356456757\nStep 528, loss: 0.17656224966049194\nStep 529, loss: 0.17330434918403625\nStep 530, loss: 0.17364926636219025\nStep 531, loss: 0.17312079668045044\nStep 532, loss: 0.17132186889648438\nStep 533, loss: 0.175296813249588\nStep 534, loss: 0.17150579392910004\nStep 535, loss: 0.17446093261241913\nStep 536, loss: 0.1724652647972107\nStep 537, loss: 0.1731881946325302\nStep 538, loss: 0.1703382283449173\nStep 539, loss: 0.17619144916534424\nStep 540, loss: 0.17391523718833923\nStep 541, loss: 0.1682422012090683\nStep 542, loss: 0.1671500951051712\nStep 543, loss: 0.17543551325798035\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 544, loss: 0.17303699254989624\nStep 545, loss: 0.1722414791584015\nStep 546, loss: 0.17534324526786804\nStep 547, loss: 0.1776745766401291\nStep 548, loss: 0.17317381501197815\nStep 549, loss: 0.1715957671403885\nStep 550, loss: 0.17733988165855408\nStep 551, loss: 0.17435169219970703\nStep 552, loss: 0.17444191873073578\nStep 553, loss: 0.17420919239521027\nStep 554, loss: 0.17075781524181366\nStep 555, loss: 0.17201998829841614\nStep 556, loss: 0.17111505568027496\nStep 557, loss: 0.17426453530788422\nStep 558, loss: 0.17481622099876404\nStep 559, loss: 0.17275071144104004\nStep 560, loss: 0.17486710846424103\nStep 561, loss: 0.17817048728466034\nStep 562, loss: 0.1738932728767395\nStep 563, loss: 0.17490266263484955\nStep 564, loss: 0.1721874177455902\nStep 565, loss: 0.1751432716846466\nStep 566, loss: 0.17336255311965942\nStep 567, loss: 0.17610953748226166\nStep 568, loss: 0.17619186639785767\nStep 569, loss: 0.17549742758274078\nStep 570, loss: 0.177082821726799\nStep 571, loss: 0.17325030267238617\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 572, loss: 0.17269103229045868\nStep 573, loss: 0.16923318803310394\nStep 688, loss: 0.168025940656662\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 574, loss: 0.17252901196479797\nStep 575, loss: 0.1739131659269333\nStep 576, loss: 0.17262126505374908\nStep 577, loss: 0.17376677691936493\nStep 578, loss: 0.17164166271686554\nStep 579, loss: 0.17398510873317719\nStep 580, loss: 0.17701764404773712\nStep 581, loss: 0.17249749600887299\nStep 582, loss: 0.17545673251152039\nStep 583, loss: 0.1745530217885971\nStep 584, loss: 0.1704486757516861\nStep 585, loss: 0.17150364816188812\nStep 586, loss: 0.1709315925836563\nStep 587, loss: 0.1745835691690445\nStep 588, loss: 0.17887862026691437\nStep 589, loss: 0.17313744127750397\nStep 590, loss: 0.17206576466560364\nStep 591, loss: 0.17699959874153137\nStep 592, loss: 0.17157968878746033\nStep 593, loss: 0.16830675303936005\nStep 594, loss: 0.17418381571769714\nStep 595, loss: 0.1751183122396469\nStep 596, loss: 0.1756243109703064\nStep 597, loss: 0.17006143927574158\nStep 598, loss: 0.1699618548154831\nStep 599, loss: 0.17193539440631866\nStep 600, loss: 0.17631924152374268\nStep 601, loss: 0.17033645510673523\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 602, loss: 0.1693136841058731\nStep 603, loss: 0.17088831961154938\nStep 604, loss: 0.17239458858966827\nStep 605, loss: 0.17504796385765076\nStep 606, loss: 0.17697462439537048\nStep 607, loss: 0.17530332505702972\nStep 608, loss: 0.17112182080745697\nStep 609, loss: 0.1731833964586258\nStep 610, loss: 0.17709961533546448\nStep 611, loss: 0.17229171097278595\nStep 612, loss: 0.17470520734786987\nStep 613, loss: 0.1714342087507248\nStep 614, loss: 0.17176499962806702\nStep 615, loss: 0.17025373876094818\nStep 616, loss: 0.174561008810997\nStep 617, loss: 0.17275869846343994\nStep 618, loss: 0.17185458540916443\nStep 619, loss: 0.17359009385108948\nStep 620, loss: 0.17360955476760864\nStep 621, loss: 0.17254018783569336\nStep 622, loss: 0.17021723091602325\nStep 623, loss: 0.17180855572223663\nStep 624, loss: 0.17233887314796448\nStep 625, loss: 0.1753343939781189\nStep 626, loss: 0.17674151062965393\nStep 627, loss: 0.17399686574935913\nStep 628, loss: 0.17107972502708435\nStep 629, loss: 0.17392827570438385\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 630, loss: 0.1768183708190918\nStep 631, loss: 0.1721939742565155\nStep 632, loss: 0.16967353224754333\nStep 633, loss: 0.17319253087043762\nStep 634, loss: 0.17231595516204834\nStep 635, loss: 0.17189306020736694\nStep 636, loss: 0.1779624968767166\nStep 637, loss: 0.1734601855278015\nStep 638, loss: 0.17567768692970276\nStep 639, loss: 0.17271453142166138\nStep 640, loss: 0.1713656783103943\nStep 641, loss: 0.17186495661735535\nStep 642, loss: 0.1760798543691635\nStep 643, loss: 0.17462579905986786\nStep 644, loss: 0.17327368259429932\nStep 645, loss: 0.1764570027589798\nStep 646, loss: 0.17871925234794617\nStep 647, loss: 0.17270927131175995\nStep 648, loss: 0.175270676612854\nStep 649, loss: 0.17492403090000153\nStep 650, loss: 0.17410653829574585\nStep 651, loss: 0.17691680788993835\nStep 652, loss: 0.17220641672611237\nStep 653, loss: 0.17202720046043396\nStep 654, loss: 0.1755136102437973\nStep 655, loss: 0.17424365878105164\nStep 656, loss: 0.1737462878227234\nStep 657, loss: 0.16914884746074677\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 658, loss: 0.17008645832538605\nStep 659, loss: 0.1742316037416458\nStep 660, loss: 0.17355559766292572\nStep 661, loss: 0.16900521516799927\nStep 662, loss: 0.17213037610054016\nStep 663, loss: 0.1717776209115982\nStep 664, loss: 0.1679377555847168\nStep 665, loss: 0.1741892397403717\nStep 666, loss: 0.1769883781671524\nStep 667, loss: 0.16878068447113037\nStep 668, loss: 0.1752023845911026\nStep 669, loss: 0.17089705169200897\nStep 670, loss: 0.1746937334537506\nStep 671, loss: 0.17643389105796814\nStep 672, loss: 0.1757017821073532\nStep 673, loss: 0.1753089427947998\nStep 674, loss: 0.175148144364357\nStep 675, loss: 0.17286920547485352\nStep 676, loss: 0.18160521984100342\nStep 677, loss: 0.1749887466430664\nStep 678, loss: 0.17353855073451996\nStep 679, loss: 0.17348110675811768\nStep 680, loss: 0.17679621279239655\nStep 681, loss: 0.1688816398382187\nStep 682, loss: 0.17490769922733307\nStep 683, loss: 0.17182523012161255\nStep 684, loss: 0.17375357449054718\nStep 685, loss: 0.17375998198986053\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 686, loss: 0.1722177267074585\nStep 687, loss: 0.17532184720039368\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 688, loss: 0.168025940656662\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 917, loss: 0.1942889243364334\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 917, loss: 0.1942889243364334\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 917, loss: 0.1942889243364334\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 917, loss: 0.1942889243364334\nStep 917, loss: 0.1942889243364334\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 917, loss: 0.1942889243364334\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 917, loss: 0.1942889243364334\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 917, loss: 0.1942889243364334\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 917, loss: 0.1942889243364334\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 917, loss: 0.1942889243364334\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 917, loss: 0.1942889243364334\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 917, loss: 0.1942889243364334\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 917, loss: 0.1942889243364334\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 917, loss: 0.1942889243364334\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 917, loss: 0.1942889243364334\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 917, loss: 0.1942889243364334\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 917, loss: 0.1942889243364334\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 917, loss: 0.1942889243364334\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 689, loss: 0.17899134755134583\nStep 690, loss: 0.17709007859230042\nStep 691, loss: 0.17748837172985077\nStep 692, loss: 0.17005573213100433\nStep 693, loss: 0.17669154703617096\nStep 694, loss: 0.1746414601802826\nStep 695, loss: 0.16808484494686127\nStep 696, loss: 0.17436909675598145\nStep 697, loss: 0.17107637226581573\nStep 698, loss: 0.1729474663734436\nStep 699, loss: 0.17476464807987213\nStep 700, loss: 0.17165909707546234\nStep 701, loss: 0.17324410378932953\nStep 702, loss: 0.17189548909664154\nStep 703, loss: 0.1744406521320343\nStep 704, loss: 0.17370614409446716\nStep 705, loss: 0.17620445787906647\nStep 706, loss: 0.174727201461792\nStep 707, loss: 0.17428751289844513\nStep 708, loss: 0.17151211202144623\nStep 709, loss: 0.1715441644191742\nStep 710, loss: 0.17064987123012543\nStep 711, loss: 0.17504172027111053\nStep 712, loss: 0.17415207624435425\nStep 713, loss: 0.17689689993858337\nStep 714, loss: 0.1732635200023651\nStep 715, loss: 0.17517362534999847\nStep 716, loss: 0.1733883023262024\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 717, loss: 0.17322330176830292\nStep 718, loss: 0.17422790825366974\nStep 719, loss: 0.17621846497058868\nStep 720, loss: 0.1731739491224289\nStep 721, loss: 0.1730724275112152\nStep 722, loss: 0.17115963995456696\nStep 723, loss: 0.17356044054031372\nStep 724, loss: 0.17796820402145386\nStep 725, loss: 0.17152009904384613\nStep 726, loss: 0.17454403638839722\nStep 727, loss: 0.1752878725528717\nStep 728, loss: 0.17705191671848297\nStep 729, loss: 0.17480850219726562\nStep 730, loss: 0.17321400344371796\nStep 731, loss: 0.1733068823814392\nStep 732, loss: 0.17345958948135376\nStep 733, loss: 0.1765996515750885\nStep 734, loss: 0.1709475815296173\nStep 735, loss: 0.17511475086212158\nStep 736, loss: 0.17417703568935394\nStep 737, loss: 0.17453403770923615\nStep 738, loss: 0.176243856549263\nStep 739, loss: 0.17172643542289734\nStep 740, loss: 0.17579515278339386\nStep 741, loss: 0.17285293340682983\nStep 742, loss: 0.17573365569114685\nStep 743, loss: 0.17104139924049377\nStep 744, loss: 0.1719278246164322\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 917, loss: 0.1942889243364334\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 745, loss: 0.16906842589378357\nStep 746, loss: 0.17420704662799835\nStep 747, loss: 0.17533841729164124\nStep 748, loss: 0.17181938886642456\nStep 749, loss: 0.17634889483451843\nStep 750, loss: 0.17465147376060486\nStep 751, loss: 0.1787661761045456\nStep 752, loss: 0.17374488711357117\nStep 753, loss: 0.17162586748600006\nStep 754, loss: 0.1724761724472046\nStep 755, loss: 0.17726847529411316\nStep 756, loss: 0.17167915403842926\nStep 757, loss: 0.17300309240818024\nStep 758, loss: 0.172225683927536\nStep 759, loss: 0.16727131605148315\nStep 760, loss: 0.17071224749088287\nStep 761, loss: 0.17563322186470032\nStep 762, loss: 0.17502345144748688\nStep 763, loss: 0.17246218025684357\nStep 764, loss: 0.1696949303150177\nStep 765, loss: 0.17512226104736328\nStep 766, loss: 0.17293953895568848\nStep 767, loss: 0.17304810881614685\nStep 768, loss: 0.17117135226726532\nStep 769, loss: 0.17588680982589722\nStep 770, loss: 0.17896541953086853\nStep 771, loss: 0.17421455681324005\nStep 772, loss: 0.17155921459197998\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 773, loss: 0.1709890216588974\nStep 774, loss: 0.1724115014076233\nStep 775, loss: 0.1722429394721985\nStep 776, loss: 0.1747157871723175\nStep 777, loss: 0.1790749728679657\nStep 778, loss: 0.1717798113822937\nStep 779, loss: 0.1735348254442215\nStep 780, loss: 0.17521053552627563\nStep 781, loss: 0.17568501830101013\nStep 782, loss: 0.17057661712169647\nStep 783, loss: 0.17483651638031006\nStep 784, loss: 0.17737214267253876\nStep 785, loss: 0.17214278876781464\nStep 786, loss: 0.1768963485956192\nStep 787, loss: 0.1741919070482254\nStep 788, loss: 0.17316724359989166\nStep 789, loss: 0.17093606293201447\nStep 790, loss: 0.175703763961792\nStep 791, loss: 0.17170217633247375\nStep 792, loss: 0.17211130261421204\nStep 793, loss: 0.17337283492088318\nStep 794, loss: 0.17829380929470062\nStep 795, loss: 0.17962464690208435\nStep 796, loss: 0.17689000070095062\nStep 797, loss: 0.170718714594841\nStep 798, loss: 0.1725749671459198\nStep 799, loss: 0.17247985303401947\nStep 800, loss: 0.17255917191505432\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 801, loss: 0.17473682761192322\nStep 802, loss: 0.17221669852733612\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 803, loss: 0.16969819366931915\nStep 804, loss: 0.17372442781925201\nStep 805, loss: 0.16981947422027588\nStep 806, loss: 0.17430536448955536\nStep 807, loss: 0.1727956086397171\nStep 808, loss: 0.1761023849248886\nStep 809, loss: 0.17369487881660461\nStep 810, loss: 0.17208799719810486\nStep 811, loss: 0.17292466759681702\nStep 812, loss: 0.17395928502082825\nStep 813, loss: 0.17069312930107117\nStep 814, loss: 0.1750108152627945\nStep 815, loss: 0.17069949209690094\nStep 816, loss: 0.18419882655143738\nStep 817, loss: 0.1798088252544403\nStep 818, loss: 0.18515458703041077\nStep 819, loss: 0.18215692043304443\nStep 820, loss: 0.19395676255226135\nStep 821, loss: 0.18590819835662842\nStep 822, loss: 0.18910640478134155\nStep 823, loss: 0.19816628098487854\nStep 824, loss: 0.1925753653049469\nStep 825, loss: 0.19668900966644287\nStep 826, loss: 0.1993279755115509\nStep 827, loss: 0.19620366394519806\nStep 828, loss: 0.19355981051921844\nStep 829, loss: 0.19763712584972382\nStep 830, loss: 0.19579480588436127\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 831, loss: 0.19501915574073792\nStep 832, loss: 0.19334354996681213\nStep 833, loss: 0.194542795419693\nStep 834, loss: 0.19733156263828278\nStep 835, loss: 0.19719944894313812\nStep 836, loss: 0.19262580573558807\nStep 837, loss: 0.19488126039505005\nStep 838, loss: 0.20572589337825775\nStep 839, loss: 0.19548571109771729\nStep 840, loss: 0.19498057663440704\nStep 841, loss: 0.19680288434028625\nStep 842, loss: 0.19567465782165527\nStep 843, loss: 0.19134315848350525\nStep 844, loss: 0.19077661633491516\nStep 845, loss: 0.19448570907115936\nStep 846, loss: 0.19932132959365845\nStep 847, loss: 0.1971946656703949\nStep 848, loss: 0.19468267261981964\nStep 849, loss: 0.1999991238117218\nStep 850, loss: 0.19588442146778107\nStep 851, loss: 0.19512861967086792\nStep 852, loss: 0.19713445007801056\nStep 853, loss: 0.19512681663036346\nStep 854, loss: 0.19677601754665375\nStep 855, loss: 0.19457361102104187\nStep 856, loss: 0.1932019740343094\nStep 857, loss: 0.19812166690826416\nStep 858, loss: 0.1941109299659729\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 859, loss: 0.1927156299352646\nStep 860, loss: 0.1943768411874771\nStep 861, loss: 0.19775605201721191\nStep 862, loss: 0.19254070520401\nStep 863, loss: 0.19562996923923492\nStep 864, loss: 0.1940685212612152\nStep 865, loss: 0.19106583297252655\nStep 866, loss: 0.19553005695343018\nStep 867, loss: 0.19323907792568207\nStep 868, loss: 0.1959063708782196\nStep 869, loss: 0.1980343759059906\nStep 870, loss: 0.19228437542915344\nStep 871, loss: 0.19487135112285614\nStep 872, loss: 0.19422484934329987\nStep 873, loss: 0.19584865868091583\nStep 874, loss: 0.19885213673114777\nStep 875, loss: 0.19261327385902405\nStep 876, loss: 0.19618886709213257\nStep 877, loss: 0.19166012108325958\nStep 878, loss: 0.19807542860507965\nStep 879, loss: 0.19533373415470123\nStep 880, loss: 0.19554272294044495\nStep 881, loss: 0.1983797401189804\nStep 882, loss: 0.19314245879650116\nStep 883, loss: 0.1953534185886383\nStep 884, loss: 0.19479385018348694\nStep 885, loss: 0.19552813470363617\nStep 886, loss: 0.19319650530815125\nStep 917, loss: 0.1942889243364334\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 917, loss: 0.1942889243364334\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 887, loss: 0.19092334806919098\nStep 888, loss: 0.19556617736816406\nStep 889, loss: 0.19262593984603882\nStep 890, loss: 0.19435875117778778\nStep 891, loss: 0.1943531036376953\nStep 892, loss: 0.19408383965492249\nStep 893, loss: 0.1942259818315506\nStep 894, loss: 0.19630619883537292\nStep 895, loss: 0.19281530380249023\nStep 896, loss: 0.19543924927711487\nStep 897, loss: 0.2001473754644394\nStep 898, loss: 0.19750091433525085\nStep 899, loss: 0.195859894156456\nStep 900, loss: 0.19264118373394012\nStep 901, loss: 0.19377319514751434\nStep 902, loss: 0.19348706305027008\nStep 903, loss: 0.19239738583564758\nStep 904, loss: 0.19326980412006378\nStep 905, loss: 0.1943923681974411\nStep 906, loss: 0.1941441297531128\nStep 907, loss: 0.19398580491542816\nStep 908, loss: 0.1948879212141037\nStep 909, loss: 0.1920291930437088\nStep 910, loss: 0.19403284788131714\nStep 911, loss: 0.1945066601037979\nStep 912, loss: 0.197319895029068\nStep 913, loss: 0.19126680493354797\nStep 914, loss: 0.19808724522590637\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 917, loss: 0.1942889243364334\nStep 917, loss: 0.1942889243364334\nStep 917, loss: 0.1942889243364334\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 917, loss: 0.1942889243364334\nStep 915, loss: 0.19713211059570312\nStep 916, loss: 0.19838380813598633\nStep 917, loss: 0.1942889243364334\nStep 917, loss: 0.1942889243364334\nStep 917, loss: 0.1942889243364334\nStep 917, loss: 0.1942889243364334\nStep 917, loss: 0.1942889243364334\nStep 917, loss: 0.1942889243364334\nStep 917, loss: 0.1942889243364334\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\nValueError: [process_index=11] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318099_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\nValueError: [process_index=9] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318099_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\nValueError: [process_index=8] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318099_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\nValueError: [process_index=10] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318099_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\nValueError: [process_index=25] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318099_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\nValueError: [process_index=5] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318099_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nValueError: [process_index=27] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318099_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nValueError: [process_index=7] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318099_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nValueError: [process_index=4] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318099_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\nValueError: [process_index=24] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318099_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nValueError: [process_index=6] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318099_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\nValueError: [process_index=26] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318099_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nStep 918, loss: 0.19523441791534424\nStep 919, loss: 0.19546496868133545\nStep 920, loss: 0.19409704208374023\nStep 921, loss: 0.19166922569274902\nStep 922, loss: 0.1975981593132019\nStep 923, loss: 0.19526618719100952\nStep 924, loss: 0.19312119483947754\nStep 925, loss: 0.19502730667591095\nStep 926, loss: 0.19699089229106903\nStep 927, loss: 0.19339591264724731\nStep 928, loss: 0.19219785928726196\nStep 929, loss: 0.19684770703315735\nStep 930, loss: 0.1927916556596756\nStep 931, loss: 0.1985294222831726\nStep 932, loss: 0.1954779475927353\nStep 933, loss: 0.19568152725696564\nStep 934, loss: 0.19383743405342102\nStep 935, loss: 0.19561338424682617\nStep 936, loss: 0.19460172951221466\nStep 937, loss: 0.19666717946529388\nStep 938, loss: 0.1971161663532257\nStep 939, loss: 0.19739972054958344\nStep 940, loss: 0.19577215611934662\nStep 941, loss: 0.19569700956344604\nStep 942, loss: 0.1939944624900818\nStep 943, loss: 0.1971275806427002\nStep 944, loss: 0.19501261413097382\nStep 945, loss: 0.19413445889949799\nStep 946, loss: 0.1940343976020813\nStep 947, loss: 0.19624318182468414\nStep 948, loss: 0.19629356265068054\nStep 949, loss: 0.19340308010578156\nStep 950, loss: 0.193202942609787\nStep 951, loss: 0.19428031146526337\nStep 952, loss: 0.19374477863311768\nStep 953, loss: 0.1919095665216446\nStep 954, loss: 0.19433891773223877\nStep 955, loss: 0.19187457859516144\nStep 956, loss: 0.19621175527572632\nStep 957, loss: 0.19424453377723694\nStep 958, loss: 0.19666950404644012\nStep 959, loss: 0.19401375949382782\nStep 960, loss: 0.19302953779697418\nStep 961, loss: 0.19898679852485657\nStep 962, loss: 0.19154222309589386\nStep 963, loss: 0.1965298056602478\nStep 964, loss: 0.19366717338562012\nStep 965, loss: 0.19513016939163208\nStep 966, loss: 0.19310520589351654\nStep 967, loss: 0.19546397030353546\nStep 968, loss: 0.19591833651065826\nStep 969, loss: 0.19176341593265533\nStep 970, loss: 0.19398139417171478\nStep 971, loss: 0.19647134840488434\nStep 972, loss: 0.1946771889925003\nStep 973, loss: 0.19320639967918396\nStep 974, loss: 0.19349446892738342\nStep 975, loss: 0.19277448952198029\nStep 976, loss: 0.19457297027111053\nStep 977, loss: 0.19542548060417175\nStep 978, loss: 0.19564610719680786\nStep 979, loss: 0.19077078998088837\nStep 980, loss: 0.19871804118156433\nStep 981, loss: 0.1972716599702835\nStep 982, loss: 0.19280266761779785\nStep 983, loss: 0.19526249170303345\nStep 984, loss: 0.19640924036502838\nStep 985, loss: 0.19558152556419373\nStep 986, loss: 0.19950821995735168\nStep 987, loss: 0.19390521943569183\nStep 988, loss: 0.19487619400024414\nStep 989, loss: 0.19438578188419342\nStep 990, loss: 0.19506306946277618\nStep 991, loss: 0.19303816556930542\nStep 992, loss: 0.19323404133319855\nStep 993, loss: 0.1923423707485199\nStep 994, loss: 0.1945382058620453\nStep 995, loss: 0.19345350563526154\nStep 996, loss: 0.1991584599018097\nStep 997, loss: 0.19994023442268372\nStep 998, loss: 0.1919410675764084\nStep 999, loss: 0.19577297568321228\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\nValueError: [process_index=23] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318122_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nValueError: [process_index=16] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318118_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nValueError: [process_index=18] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318118_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\nValueError: [process_index=31] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318120_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nValueError: [process_index=21] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318122_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nValueError: [process_index=14] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318123_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\nValueError: [process_index=12] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318123_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nValueError: [process_index=17] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318118_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\nValueError: [process_index=28] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318120_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nValueError: [process_index=20] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318122_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nValueError: [process_index=19] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318118_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\nValueError: [process_index=22] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318122_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nValueError: [process_index=13] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318123_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\nValueError: [process_index=15] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318123_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nValueError: [process_index=30] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318120_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\nValueError: [process_index=29] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318120_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n2025-07-01 00:12:31.357123: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1689] Shutdown barrier in coordination service has failed:\nDEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n2025-07-01 00:12:31.357635: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\nThis suggests that the workers are out of sync. Either at least one worker (a) crashed early due to program error or scheduler events (e.g. preemption, eviction), (b) was too fast in its execution, or (c) too slow / hanging. Check the logs (both the program and scheduler events) for an earlier error to identify the root cause.\n2025-07-01 00:12:31.357150: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1730] Use error polling to propagate the following error to all tasks: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n2025-07-01 00:12:31.357272: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357408: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n2025-07-01 00:12:31.357511: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.357537: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357838: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n2025-07-01 00:12:31.357879: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357487: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357789: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.358027: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.357624: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.357937: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.357864: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.357965: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.358050: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357759: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357856: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.358181: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.357705: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.358251: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.357880: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.357838: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357950: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n2025-07-01 00:12:31.358247: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n2025-07-01 00:12:31.358561: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.358682: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.357944: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.358160: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.358277: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.358191: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388132: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.389752: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.387984: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.389659: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388310: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.389838: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388302: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.389938: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388225: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390670: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\n2025-07-01 00:12:31.374109: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390981: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388191: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390626: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\n2025-07-01 00:12:31.374262: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.391021: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388298: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390927: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\n2025-07-01 00:12:31.374546: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.391311: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.387946: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390776: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388298: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390889: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.374395: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.391196: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388167: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390957: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388074: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390919: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388180: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.391065: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\nsrun: error: hkn0520: task 10: Aborted (core dumped)\nsrun: error: hkn0523: task 15: Aborted (core dumped)\nsrun: error: hkn0820: task 31: Aborted (core dumped)\nsrun: error: hkn0810: task 27: Aborted (core dumped)\nsrun: error: hkn0420: tasks 1-3: Aborted (core dumped)\nsrun: error: hkn0516: tasks 5,7: Aborted (core dumped)\nsrun: error: hkn0717: tasks 20,22: Aborted (core dumped)\nsrun: error: hkn0520: task 11: Aborted (core dumped)\nsrun: error: hkn0704: tasks 17,19: Aborted (core dumped)\nsrun: error: hkn0820: task 29: Aborted (core dumped)\nsrun: error: hkn0810: task 24: Aborted (core dumped)\nsrun: error: hkn0520: task 9: Aborted (core dumped)\nsrun: error: hkn0820: task 28: Aborted (core dumped)\nsrun: error: hkn0516: task 4: Aborted (core dumped)\nsrun: error: hkn0704: task 16: Aborted (core dumped)\nsrun: error: hkn0523: tasks 13-14: Aborted (core dumped)\nsrun: error: hkn0717: task 21: Aborted (core dumped)\nsrun: error: hkn0717: task 23: Aborted (core dumped)\nsrun: error: hkn0820: task 30: Aborted (core dumped)\nsrun: error: hkn0810: task 26: Aborted (core dumped)\nsrun: error: hkn0520: task 8: Aborted (core dumped)\nsrun: error: hkn0516: task 6: Aborted (core dumped)\nsrun: error: hkn0810: task 25: Aborted (core dumped)\nsrun: error: hkn0523: task 12: Aborted (core dumped)\nsrun: error: hkn0704: task 18: Aborted (core dumped)\nsrun: error: hkn0420: task 0: Aborted (core dumped)\n\n============================= JOB FEEDBACK =============================\n\nJob ID: 3307474\nCluster: hk\nUser/Group: tum_ind3695/hk-project-pai00039\nAccount: hk-project-p0023960\nState: FAILED (exit code 134)\nPartition: accelerated\nNodes: 8\nCores per node: 32\nNodelist: hkn[0420,0516,0520,0523,0704,0717,0810,0820]\nCPU Utilized: 22:56:17\nCPU Efficiency: 9.24% of 10-08:10:40 core-walltime\nJob Wall-clock time: 00:58:10\nStarttime: Mon Jun 30 23:14:23 2025\nEndtime: Tue Jul 1 00:12:33 2025\nMemory Utilized: 423.14 GB (estimated maximum)\nMemory Efficiency: 0.00% of 0.00 MB (0.00 MB/node)\nEnergy Consumed: 35390279 Joule / 9830.63305555556 Watthours\nAverage node power draw: 10140.4810888252 Watt\n",log,tab +21,177754,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",645,0,"",log,selection_mouse +22,177789,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",644,0,"",log,selection_command +23,178578,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",1441133,0,"",log,selection_command +24,184664,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",1438043,0,"",log,selection_mouse +25,184801,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",1438030,17,"DEADLINE_EXCEEDED",log,selection_mouse +26,194680,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",1384536,0,"",log,selection_mouse +27,194841,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",1384530,15,"array_metadatas",log,selection_mouse +28,216890,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",1273273,0,"",log,selection_mouse +29,216914,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",1273272,0,"",log,selection_command +30,217576,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",1273308,0,"",log,selection_mouse +31,217590,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",1273307,0,"",log,selection_command +32,225590,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",1273274,34,"Traceback (most recent call last):",log,selection_command +33,227875,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",1273274,167859,"Traceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\nValueError: [process_index=23] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318122_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nValueError: [process_index=16] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318118_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nValueError: [process_index=18] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318118_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\nValueError: [process_index=31] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318120_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nValueError: [process_index=21] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318122_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nValueError: [process_index=14] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318123_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\nValueError: [process_index=12] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318123_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nValueError: [process_index=17] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318118_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\nValueError: [process_index=28] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318120_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nValueError: [process_index=20] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318122_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 159, in wait_for_base_dir_creation\n await asyncio.sleep(0.25)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 605, in sleep\n return await future\nasyncio.exceptions.CancelledError\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 456, in wait_for\n return fut.result()\nasyncio.exceptions.CancelledError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 162, in _maybe_create_base_dir\n await asyncio.wait_for(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/tasks.py"", line 458, in wait_for\n raise exceptions.TimeoutError() from exc\nasyncio.exceptions.TimeoutError\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/train_tokenizer.py"", line 262, in \n orbax_checkpointer.save(\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/checkpointers/checkpointer.py"", line 255, in save\n self._handler.save(tmpdir.get(), args=ckpt_args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\nValueError: [process_index=19] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318118_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/pytree_checkpoint_handler.py"", line 594, in save\n self._handler_impl.save(directory, args=args)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 554, in save\n asyncio_utils.run_sync(async_save(directory, *args, **kwargs))\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/handlers/base_pytree_checkpoint_handler.py"", line 552, in async_save\n f.result() # Block on result.\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 174, in result\n f.result(timeout=time_remaining)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 397, in result\n return self._f.result(timeout=timeout)\nValueError: [process_index=22] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318122_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nValueError: [process_index=13] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318123_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 348, in result\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\nValueError: [process_index=15] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318123_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\nValueError: [process_index=30] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318120_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n return self._t.result(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 297, in result\n self.join(timeout=timeout)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 294, in join\n raise self._exception\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 281, in run\n super().run()\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/threading.py"", line 953, in run\n self._target(*self._args, **self._kwargs)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 230, in _target_setting_result\n self._result = target()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/futures/future.py"", line 341, in \n target=lambda: asyncio_utils.run_sync(coro),\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/asyncio_utils.py"", line 50, in run_sync\n return asyncio.run(coro)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/runners.py"", line 44, in run\n return loop.run_until_complete(main)\n File ""/home/hk-project-pai00039/tum_ind3695/.local/share/uv/python/cpython-3.10.17-linux-x86_64-gnu/lib/python3.10/asyncio/base_events.py"", line 649, in run_until_complete\n return future.result()\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/serialization/type_handlers.py"", line 1105, in _background_serialize\n await asyncio.gather(*write_coros)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 194, in write\n await self._maybe_create_base_dir(file_path.parent)\n File ""/hkfs/home/project/hk-project-pai00039/tum_ind3695/projects/jafar_batch_size_scaling/.venv_jafar/lib/python3.10/site-packages/orbax/checkpoint/_src/metadata/array_metadata_store.py"", line 169, in _maybe_create_base_dir\n raise ValueError(\nValueError: [process_index=29] Timed out waiting for array_metadatas base directory creation: /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474/tokenizer_1751318120_1000.orbax-checkpoint-tmp-0/array_metadatas. timeout=600 seconds. primary_process=0\n2025-07-01 00:12:31.357123: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1689] Shutdown barrier in coordination service has failed:\nDEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n2025-07-01 00:12:31.357635: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\nThis suggests that the workers are out of sync. Either at least one worker (a) crashed early due to program error or scheduler events (e.g. preemption, eviction), (b) was too fast in its execution, or (c) too slow / hanging. Check the logs (both the program and scheduler events) for an earlier error to identify the root cause.\n2025-07-01 00:12:31.357150: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service.cc:1730] Use error polling to propagate the following error to all tasks: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n2025-07-01 00:12:31.357272: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357408: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n2025-07-01 00:12:31.357511: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380'] [type.googleapis.com/tensorflow.CoordinationServiceError='']\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.357537: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357838: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n2025-07-01 00:12:31.357879: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357487: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357789: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.358027: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.357624: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.357937: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.357864: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.357965: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.358050: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357759: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357856: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.358181: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.357705: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.358251: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.357880: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.357838: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.357950: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n2025-07-01 00:12:31.358247: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n2025-07-01 00:12:31.358561: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.358682: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.357944: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.358160: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\n2025-07-01 00:12:31.358277: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.358191: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/PollForError [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388132: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.389752: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.387984: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.389659: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388310: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.389838: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388302: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.389938: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388225: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390670: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\n2025-07-01 00:12:31.374109: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390981: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388191: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390626: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\n2025-07-01 00:12:31.374262: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.391021: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388298: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390927: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\n2025-07-01 00:12:31.374546: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.391311: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.387946: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390776: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388298: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390889: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.374395: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.391196: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388167: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390957: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388074: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.390919: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\n2025-07-01 00:12:31.388180: E external/xla/xla/tsl/distributed_runtime/coordination/coordination_service_agent.cc:427] Polled an error from coordination service (this can be an error from this or another task).\n2025-07-01 00:12:31.391065: F external/xla/xla/pjrt/distributed/client.h:80] Terminating process because the JAX distributed service detected fatal errors. This most likely indicates that another task died; see the other task logs for more details. Disable Python buffering, i.e. `python -u`, to be sure to see all the previous output. absl::Status: INTERNAL: Failed to disconnect from coordination service with status: INTERNAL: Shutdown barrier has failed.\nBarrier result: 'DEADLINE_EXCEEDED: Barrier timed out. Id: Shutdown::16634615291091153380::0. This usually happens because a task triggered the barrier too early or too slowly. Please look at the task logs (both timed out and first task) to debug further.\n# of tasks that reached the barrier: 28/32.\nThe first task at the barrier: /job:jax_worker/replica:0/task:5. Some timed out task names:\n/job:jax_worker/replica:0/task:2\n/job:jax_worker/replica:0/task:0\n [type.googleapis.com/tensorflow.CoordinationServiceError=''] [type.googleapis.com/tensorflow.BarrierError='\n\x1eShutdown::16634615291091153380']\n\nRPC: /tensorflow.CoordinationService/ShutdownTask [type.googleapis.com/tensorflow.CoordinationServiceError='']Proceeding with agent shutdown anyway. This is usually caused by an earlier error during execution. Check the logs of (a) this task, (b) the leader (usually slice 0 task 0) and (c) the scheduler (e.g. preemption, eviction) for an earlier error to debug further. [type.googleapis.com/tensorflow.CoordinationServiceError='']\nsrun: error: hkn0520: task 10: Aborted (core dumped)\nsrun: error: hkn0523: task 15: Aborted (core dumped)\nsrun: error: hkn0820: task 31: Aborted (core dumped)\nsrun: error: hkn0810: task 27: Aborted (core dumped)\nsrun: error: hkn0420: tasks 1-3: Aborted (core dumped)\nsrun: error: hkn0516: tasks 5,7: Aborted (core dumped)\nsrun: error: hkn0717: tasks 20,22: Aborted (core dumped)\nsrun: error: hkn0520: task 11: Aborted (core dumped)\nsrun: error: hkn0704: tasks 17,19: Aborted (core dumped)\nsrun: error: hkn0820: task 29: Aborted (core dumped)\nsrun: error: hkn0810: task 24: Aborted (core dumped)\nsrun: error: hkn0520: task 9: Aborted (core dumped)\nsrun: error: hkn0820: task 28: Aborted (core dumped)\nsrun: error: hkn0516: task 4: Aborted (core dumped)\nsrun: error: hkn0704: task 16: Aborted (core dumped)\nsrun: error: hkn0523: tasks 13-14: Aborted (core dumped)\nsrun: error: hkn0717: task 21: Aborted (core dumped)\nsrun: error: hkn0717: task 23: Aborted (core dumped)\nsrun: error: hkn0820: task 30: Aborted (core dumped)\nsrun: error: hkn0810: task 26: Aborted (core dumped)\nsrun: error: hkn0520: task 8: Aborted (core dumped)\nsrun: error: hkn0516: task 6: Aborted (core dumped)\nsrun: error: hkn0810: task 25: Aborted (core dumped)\nsrun: error: hkn0523: task 12: Aborted (core dumped)\nsrun: error: hkn0704: task 18: Aborted (core dumped)\nsrun: error: hkn0420: task 0: Aborted (core dumped)\n\n============================= JOB FEEDBACK =============================\n\nJob ID: 3307474\nCluster: hk\nUser/Group: tum_ind3695/hk-project-pai00039\nAccount: hk-project-p0023960\nState: FAILED (exit code 134)\nPartition: accelerated\nNodes: 8\nCores per node: 32\nNodelist: hkn[0420,0516,0520,0523,0704,0717,0810,0820]\nCPU Utilized: 22:56:17\nCPU Efficiency: 9.24% of 10-08:10:40 core-walltime\nJob Wall-clock time: 00:58:10\nStarttime: Mon Jun 30 23:14:23 2025\nEndtime: Tue Jul 1 00:12:33 2025\nMemory Utilized: 423.14 GB (estimated maximum)\nMemory Efficiency: 0.00% of 0.00 MB (0.00 MB/node)\nEnergy Consumed: 35390279 Joule / 9830.63305555556 Watthours\nAverage node power draw: 10140.4810888252 Watt\n",log,selection_command +34,299049,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_alfred/logs_training_tokenizer/train_tokenizer_batch_size_scaling_8_node_3307474.log",1441133,0,"",log,selection_mouse +35,326035,"scripts_horeka/train_tokenizer.sh",0,0,"#!/usr/bin/env bash\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=""debug""\nslurm_job_id=""0000""\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=16 \\n --min_lr=4.24e-4 \\n --max_lr=4.24e-4 \\n --log_image_interval=100 \\n --log \\n --name=test-wandb-tags-$slurm_job_id \\n --tags test tokenizer debug \\n --entity instant-uv \\n --project jafar \\n --data_dir $tf_records_dir",shellscript,tab +36,337973,"train_tokenizer.py",0,0,"from dataclasses import dataclass, field\nimport os\nimport time\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax\nfrom orbax.checkpoint import PyTreeCheckpointer\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\n\nts = int(time.time())\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data_tfrecords/coinrun""\n checkpoint: str = """"\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n min_lr: float = 3e-4\n max_lr: float = 3e-4\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 8\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_gradients: bool = False\n\n\nargs = tyro.cli(Args)\n\n\ndef tokenizer_loss_fn(params, state, inputs):\n # --- Compute loss ---\n outputs = state.apply_fn(\n params,\n inputs,\n training=True,\n rngs={""params"": inputs[""rng""], ""dropout"": inputs[""dropout_rng""]},\n )\n mse = jnp.square(inputs[""videos""] - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = inputs[""videos""].clip(0, 1).reshape(-1, *inputs[""videos""].shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = pix.psnr(gt, recon).mean()\n ssim = pix.ssim(gt, recon).mean()\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n grad_fn = jax.value_and_grad(tokenizer_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n if args.log and jax.process_index() == 0:\n wandb.init(\n entity=args.entity,\n project=args.project,\n name=args.name,\n tags=args.tags,\n group=""debug"",\n config=args\n )\n\n # --- Initialize model ---\n tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_height, args.image_width, args.image_channels)\n inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=jnp.float32,\n ),\n )\n init_params = tokenizer.init(_rng, inputs)\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=tokenizer.apply, params=init_params, tx=tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Load checkpoint ---\n step = 0\n if args.checkpoint:\n restore_target = {""model"": train_state}\n restore_args = orbax_utils.restore_args_from_target(restore_target)\n train_state.params[""params""].update(\n PyTreeCheckpointer()\n .restore(args.checkpoint, item=restore_target, restore_args=restore_args)[\n ""model""\n ]\n .params[""params""]\n )\n # Assume checkpoint is of the form tokenizer__\n step += int(args.checkpoint.split(""_"")[-1])\n\n # --- TRAIN LOOP ---\n tfrecord_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".tfrecord"")\n ]\n dataloader = get_dataloader(\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n tfrecord_files,\n args.seq_len,\n args.batch_size,\n *image_shape,\n )\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n # for videos in dataloader:\n npy_path = ""overfit_dir/single_sample_corner.npy""\n # npy_path = ""overfit_dir/single_batch_12_elems.npy""\n videos = np.load(npy_path)\n print(""batch shape: "", videos.shape)\n while(True):\n # --- Train step ---\n rng, _rng, _rng_dropout = jax.random.split(rng, 3)\n\n videos_sharding = NamedSharding(\n mesh, PartitionSpec(""data"", None, None, None, None)\n )\n videos = jax.make_array_from_process_local_data(videos_sharding, videos)\n\n inputs = dict(videos=videos, rng=_rng, dropout_rng=_rng_dropout)\n start_time = time.time()\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n # jax.block_until_ready(loss)\n elapsed_time = (time.time() - start_time) * 1000\n print(f""Step {step}, loss: {loss}, step time: {elapsed_time}ms"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n ""step_time_ms"": elapsed_time,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break\n",python,tab +37,340875,"TERMINAL",0,0,"bash",,terminal_focus +38,342066,"TERMINAL",0,0,"idling",,terminal_command +39,342153,"TERMINAL",0,0,"]633;E;2025-07-01 13:10:54 idling;d7d28c43-3fca-451b-bc76-771109467dda]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1991.localdomain: Tue Jul 1 13:10:54 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 32 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated: 10 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 7 nodes idle",,terminal_output +40,343235,"TERMINAL",0,0,"5",,terminal_output +41,343633,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output +42,355197,"train_tokenizer.py",2394,0,"",python,selection_mouse +43,355208,"train_tokenizer.py",2393,0,"",python,selection_command +44,355807,"train_tokenizer.py",1743,0,"",python,selection_mouse +45,360688,"train_tokenizer.py",9298,0,"",python,selection_mouse +46,361292,"train_tokenizer.py",9290,0,"",python,selection_mouse +47,361455,"train_tokenizer.py",9289,4,"ckpt",python,selection_mouse +48,361735,"train_tokenizer.py",9289,5,"ckpt ",python,selection_mouse +49,361736,"train_tokenizer.py",9289,8,"ckpt = {",python,selection_mouse +50,361736,"train_tokenizer.py",9289,14,"ckpt = {""model",python,selection_mouse +51,361799,"train_tokenizer.py",9289,17,"ckpt = {""model"": ",python,selection_mouse +52,361800,"train_tokenizer.py",9289,28,"ckpt = {""model"": train_state",python,selection_mouse +53,361858,"train_tokenizer.py",9289,67,"ckpt = {""model"": train_state}\n orbax_checkpointer = ",python,selection_mouse +54,361865,"train_tokenizer.py",9289,72,"ckpt = {""model"": train_state}\n orbax_checkpointer = orbax",python,selection_mouse +55,361914,"train_tokenizer.py",9289,73,"ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.",python,selection_mouse +56,361941,"train_tokenizer.py",9289,83,"ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint",python,selection_mouse +57,362482,"train_tokenizer.py",9369,0,"",python,selection_mouse +58,363001,"train_tokenizer.py",9502,0,"",python,selection_mouse +59,363016,"train_tokenizer.py",9501,0,"",python,selection_command +60,363180,"train_tokenizer.py",9502,0,"",python,selection_mouse +61,363188,"train_tokenizer.py",9501,0,"",python,selection_command +62,363390,"train_tokenizer.py",9501,1,"(",python,selection_mouse +63,363394,"train_tokenizer.py",9502,0,"",python,selection_command +64,363555,"train_tokenizer.py",9501,1,"(",python,selection_mouse +65,363556,"train_tokenizer.py",9497,5,"save(",python,selection_mouse +66,363585,"train_tokenizer.py",9478,24,"orbax_checkpointer.save(",python,selection_mouse +67,363789,"train_tokenizer.py",9410,92,"save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +68,363919,"train_tokenizer.py",9335,167,"orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +69,363954,"train_tokenizer.py",9334,168," orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +70,363986,"train_tokenizer.py",9333,169," orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +71,364042,"train_tokenizer.py",9286,216," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +72,364377,"train_tokenizer.py",9286,0,"",python,selection_mouse +73,364378,"train_tokenizer.py",9273,16," ",python,selection_mouse +74,364670,"train_tokenizer.py",9273,60," ckpt = {""model"": train_state}\n ",python,selection_mouse +75,364671,"train_tokenizer.py",9273,62," ckpt = {""model"": train_state}\n ",python,selection_mouse +76,364671,"train_tokenizer.py",9273,80," ckpt = {""model"": train_state}\n orbax_checkpointer",python,selection_mouse +77,364672,"train_tokenizer.py",9273,146," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args",python,selection_mouse +78,364672,"train_tokenizer.py",9273,148," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args =",python,selection_mouse +79,364673,"train_tokenizer.py",9273,160," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils",python,selection_mouse +80,364726,"train_tokenizer.py",9273,228," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save",python,selection_mouse +81,364751,"train_tokenizer.py",9273,229," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +82,365126,"train_tokenizer.py",9502,0,"",python,selection_mouse +83,365134,"train_tokenizer.py",9501,0,"",python,selection_command +84,365285,"train_tokenizer.py",9502,0,"",python,selection_mouse +85,365305,"train_tokenizer.py",9501,0,"",python,selection_command +86,365449,"train_tokenizer.py",9501,1,"(",python,selection_mouse +87,365545,"train_tokenizer.py",9502,0,"",python,selection_command +88,365546,"train_tokenizer.py",9420,82,"= orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +89,365547,"train_tokenizer.py",9410,92,"save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +90,365631,"train_tokenizer.py",9335,167,"orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +91,365632,"train_tokenizer.py",9332,170," orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +92,365664,"train_tokenizer.py",9331,171," orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +93,365695,"train_tokenizer.py",9330,172," orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +94,365863,"train_tokenizer.py",9284,218," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +95,365929,"train_tokenizer.py",9285,217," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +96,365931,"train_tokenizer.py",9228,274,"if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(",python,selection_mouse +97,366275,"train_tokenizer.py",9228,0,"",python,selection_mouse +98,372384,"train_tokenizer.py",9190,0,"",python,selection_mouse +99,372969,"train_tokenizer.py",9240,0,"",python,selection_mouse +100,374612,"train_tokenizer.py",9318,0,"",python,selection_mouse +101,374625,"train_tokenizer.py",9317,0,"",python,selection_command +102,375186,"train_tokenizer.py",9249,0,"",python,selection_mouse +103,376066,"train_tokenizer.py",9349,0,"",python,selection_mouse +104,376770,"train_tokenizer.py",9367,0,"",python,selection_mouse +105,377467,"train_tokenizer.py",9340,0,"",python,selection_mouse +106,377630,"train_tokenizer.py",9335,18,"orbax_checkpointer",python,selection_mouse +107,378608,"train_tokenizer.py",9415,0,"",python,selection_mouse +108,378785,"train_tokenizer.py",9410,9,"save_args",python,selection_mouse +109,379646,"train_tokenizer.py",9339,0,"",python,selection_mouse +110,379806,"train_tokenizer.py",9335,18,"orbax_checkpointer",python,selection_mouse +111,380740,"train_tokenizer.py",9523,0,"",python,selection_mouse +112,381379,"train_tokenizer.py",9423,0,"",python,selection_mouse +113,381578,"train_tokenizer.py",9422,11,"orbax_utils",python,selection_mouse +114,382177,"train_tokenizer.py",9532,0,"",python,selection_mouse +115,382352,"train_tokenizer.py",9531,4,"join",python,selection_mouse +116,382942,"train_tokenizer.py",9493,0,"",python,selection_mouse +117,383096,"train_tokenizer.py",9478,18,"orbax_checkpointer",python,selection_mouse +118,384150,"train_tokenizer.py",9436,0,"",python,selection_mouse +119,417320,"train_tokenizer.py",9736,0,"",python,selection_mouse +120,417321,"train_tokenizer.py",9735,0,"",python,selection_command +121,417512,"train_tokenizer.py",9735,1,"k",python,selection_mouse +122,417550,"train_tokenizer.py",9736,0,"",python,selection_command +123,417551,"train_tokenizer.py",9706,30,"m_steps:\n break",python,selection_mouse +124,417613,"train_tokenizer.py",9640,96,"e_args=save_args,\n )\n if step >= args.num_steps:\n break",python,selection_mouse +125,417614,"train_tokenizer.py",9478,258,"orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break",python,selection_mouse +126,417670,"train_tokenizer.py",9330,406," orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break",python,selection_mouse +127,417750,"train_tokenizer.py",9282,454," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break",python,selection_mouse +128,417789,"train_tokenizer.py",9279,457," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break",python,selection_mouse +129,417834,"train_tokenizer.py",9275,461," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break",python,selection_mouse +130,417892,"train_tokenizer.py",9216,520," if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break",python,selection_mouse +131,428788,"train_tokenizer.py",9737,0,"",python,selection_mouse +132,434939,"train_tokenizer.py",9652,0,"",python,selection_mouse +133,435109,"train_tokenizer.py",9647,9,"save_args",python,selection_mouse +134,439895,"train_tokenizer.py",9311,0,"",python,selection_mouse +135,446256,"train_tokenizer.py",9310,0,"",python,selection_mouse +136,448728,"train_tokenizer.py",0,0,"",python,tab +137,448729,"train_tokenizer.py",7477,0,"",python,selection_mouse +138,450504,"train_tokenizer.py",7480,0,"",python,selection_mouse +139,450782,"train_tokenizer.py",0,0,"",python,tab +140,452942,"train_tokenizer.py",3466,0,"",python,selection_mouse +141,453142,"train_tokenizer.py",3465,5,"state",python,selection_mouse +142,454655,"train_tokenizer.py",3025,0,"",python,selection_mouse +143,454802,"train_tokenizer.py",3024,5,"state",python,selection_mouse +144,455085,"train_tokenizer.py",3024,21,"state.apply_gradients",python,selection_mouse +145,455500,"train_tokenizer.py",3037,0,"",python,selection_mouse +146,455500,"train_tokenizer.py",3030,15,"apply_gradients",python,selection_mouse +147,456145,"train_tokenizer.py",3025,0,"",python,selection_mouse +148,456304,"train_tokenizer.py",3024,5,"state",python,selection_mouse +149,456907,"train_tokenizer.py",3037,0,"",python,selection_mouse +150,457062,"train_tokenizer.py",3030,15,"apply_gradients",python,selection_mouse +151,457725,"train_tokenizer.py",3028,0,"",python,selection_mouse +152,457890,"train_tokenizer.py",3024,5,"state",python,selection_mouse +153,463644,"train_tokenizer.py",2836,0,"",python,selection_mouse +154,465456,"train_tokenizer.py",7488,0,"",python,selection_mouse +155,484792,"train_tokenizer.py",9616,0,"",python,selection_mouse +156,484794,"train_tokenizer.py",9615,0,"",python,selection_command +157,485406,"train_tokenizer.py",9675,0,"",python,selection_mouse +158,485422,"train_tokenizer.py",9674,0,"",python,selection_command +159,485579,"train_tokenizer.py",9674,1,")",python,selection_mouse +160,485689,"train_tokenizer.py",9675,0,"",python,selection_command +161,485690,"train_tokenizer.py",9657,18,"\n )",python,selection_mouse +162,485690,"train_tokenizer.py",9345,330,"kpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )",python,selection_mouse +163,485690,"train_tokenizer.py",9343,332,"eckpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,\n )",python,selection_mouse +164,486461,"train_tokenizer.py",9283,0,"",python,selection_mouse +165,486596,"train_tokenizer.py",9273,16," ",python,selection_mouse +166,486770,"train_tokenizer.py",9273,80," ckpt = {""model"": train_state}\n orbax_checkpointer",python,selection_mouse +167,486770,"train_tokenizer.py",9273,146," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args",python,selection_mouse +168,486813,"train_tokenizer.py",9273,228," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save",python,selection_mouse +169,486813,"train_tokenizer.py",9273,343," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,",python,selection_mouse +170,486908,"train_tokenizer.py",9273,384," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,",python,selection_mouse +171,487089,"train_tokenizer.py",9657,0,"",python,selection_mouse +172,487121,"train_tokenizer.py",9656,0,"",python,selection_command +173,487362,"train_tokenizer.py",9657,0,"",python,selection_mouse +174,487389,"train_tokenizer.py",9656,0,"",python,selection_command +175,487516,"train_tokenizer.py",9656,1,",",python,selection_mouse +176,487518,"train_tokenizer.py",9657,0,"",python,selection_command +177,487659,"train_tokenizer.py",9616,41,"\n save_args=save_args,",python,selection_mouse +178,487659,"train_tokenizer.py",9520,137," os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,",python,selection_mouse +179,487660,"train_tokenizer.py",9470,187," orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,",python,selection_mouse +180,487660,"train_tokenizer.py",9399,258," save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,",python,selection_mouse +181,487661,"train_tokenizer.py",9396,261," save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,",python,selection_mouse +182,487661,"train_tokenizer.py",9395,262," save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,",python,selection_mouse +183,487683,"train_tokenizer.py",9394,263," save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,",python,selection_mouse +184,487763,"train_tokenizer.py",9319,338," orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,",python,selection_mouse +185,487840,"train_tokenizer.py",9273,384," ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,",python,selection_mouse +186,487931,"train_tokenizer.py",9216,441," if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{ts}_{step}""),\n ckpt,\n save_args=save_args,",python,selection_mouse +187,593439,"train_tokenizer.py",9714,0,"",python,selection_mouse +188,593504,"train_tokenizer.py",9713,0,"",python,selection_command +189,801184,"train_tokenizer.py",9215,0,"",python,selection_mouse +190,801199,"train_tokenizer.py",9214,0,"",python,selection_command +191,802197,"train_tokenizer.py",9452,0,"",python,selection_mouse +192,803383,"train_tokenizer.py",9272,0,"",python,selection_mouse +193,803400,"train_tokenizer.py",9271,0,"",python,selection_command +194,803959,"train_tokenizer.py",9318,0,"",python,selection_mouse +195,803963,"train_tokenizer.py",9317,0,"",python,selection_command +196,895557,"train_tokenizer.py",9639,0,"",python,selection_mouse +197,895748,"train_tokenizer.py",9637,9,"save_args",python,selection_mouse +198,896339,"train_tokenizer.py",9675,0,"",python,selection_mouse +199,896376,"train_tokenizer.py",9674,0,"",python,selection_command +200,898019,"train_tokenizer.py",9483,0,"",python,selection_mouse +201,898177,"train_tokenizer.py",9478,18,"orbax_checkpointer",python,selection_mouse +202,899127,"train_tokenizer.py",9675,0,"",python,selection_mouse +203,899178,"train_tokenizer.py",9674,0,"",python,selection_command +204,1530973,"TERMINAL",0,0,"# save_checkpoint_multiprocess",,terminal_command diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c190e976-4e73-4096-9dfa-66c3a0cce2061752157080469-2025_07_10-16.18.14.262/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c190e976-4e73-4096-9dfa-66c3a0cce2061752157080469-2025_07_10-16.18.14.262/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..a3866264332de5a60d97bf341f366e92f2212ee8 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c190e976-4e73-4096-9dfa-66c3a0cce2061752157080469-2025_07_10-16.18.14.262/source.csv @@ -0,0 +1,4206 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +2,695,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"4:18:14 PM [info] Activating crowd-code\n4:18:14 PM [info] Recording started\n4:18:14 PM [info] Initializing git provider using file system watchers...\n4:18:14 PM [info] Git repository found\n4:18:14 PM [info] Git provider initialized successfully\n",Log,tab +3,697,"extension-output-pdoom-org.crowd-code-#1-crowd-code",245,0,"4:18:14 PM [info] Initial git state: [object Object]\n",Log,content +4,3747,"TERMINAL",0,0,"/bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt",,terminal_command +5,3773,"TERMINAL",0,0,"]633;E;2025-07-10 16:18:17 /bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt;44d013a9-d4b7-491b-9109-05938bc2908f]633;C]0;tum_cte0515@hkn1993:/hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash]633;D;0",,terminal_output +6,198600,"TERMINAL",0,0,"bash",,terminal_focus +7,386141,"TERMINAL",0,0,"watch",,terminal_focus +8,459809,"TERMINAL",0,0,"bash",,terminal_focus +9,737837,"TERMINAL",0,0,"salloc",,terminal_focus +10,739105,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=""debug""\nslurm_job_id=""debug-mihir""\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/train_lam_action_space_scaling_20/3318547/lam_1751657975_200000/\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=82 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n # --lam_checkpoint=$lam_ckpt_dir\n ",shellscript,tab +11,739533,"TERMINAL",0,0,"bash",,terminal_focus +12,749343,"TERMINAL",0,0,"watch",,terminal_focus +13,751007,"TERMINAL",0,0,"bash",,terminal_focus +14,754607,"TERMINAL",0,0,"watch",,terminal_focus +15,757748,"TERMINAL",0,0,"idling",,terminal_command +16,757821,"TERMINAL",0,0,"]633;E;2025-07-10 16:30:51 idling;d0af5434-9a5b-4b8f-9412-c82c37ee36e4]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Jul 10 16:30:51 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 36 nodes idle\rPartition dev_accelerated:\t 0 nodes idle\rPartition accelerated:\t 0 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 6 nodes idle",,terminal_output +17,758468,"TERMINAL",0,0,"bash",,terminal_focus +18,758847,"TERMINAL",0,0,"2",,terminal_output +19,759883,"TERMINAL",0,0,"42",,terminal_output +20,760989,"TERMINAL",0,0,"5",,terminal_output +21,762011,"TERMINAL",0,0,"6",,terminal_output +22,763008,"TERMINAL",0,0,"7",,terminal_output +23,764060,"TERMINAL",0,0,"8",,terminal_output +24,765083,"TERMINAL",0,0,"9",,terminal_output +25,765189,"TERMINAL",0,0,"salloc",,terminal_focus +26,766212,"TERMINAL",0,0,"1:00",,terminal_output +27,767141,"TERMINAL",0,0,"1",,terminal_output +28,768261,"TERMINAL",0,0,"2",,terminal_output +29,769218,"TERMINAL",0,0,"3",,terminal_output +30,770309,"TERMINAL",0,0,"4",,terminal_output +31,771304,"TERMINAL",0,0,"5",,terminal_output +32,772356,"TERMINAL",0,0,"6",,terminal_output +33,773376,"TERMINAL",0,0,"7",,terminal_output +34,774508,"TERMINAL",0,0,"8",,terminal_output +35,775442,"TERMINAL",0,0,"9",,terminal_output +36,776479,"TERMINAL",0,0,"10",,terminal_output +37,777515,"TERMINAL",0,0,"1",,terminal_output +38,778550,"TERMINAL",0,0,"2",,terminal_output +39,779583,"TERMINAL",0,0,"3",,terminal_output +40,780648,"TERMINAL",0,0,"4",,terminal_output +41,781675,"TERMINAL",0,0,"5",,terminal_output +42,782697,"TERMINAL",0,0,"6",,terminal_output +43,783826,"TERMINAL",0,0,"7",,terminal_output +44,784772,"TERMINAL",0,0,"8",,terminal_output +45,785873,"TERMINAL",0,0,"9",,terminal_output +46,786900,"TERMINAL",0,0,"20",,terminal_output +47,787920,"TERMINAL",0,0,"2",,terminal_output +48,788944,"TERMINAL",0,0,"3",,terminal_output +49,790071,"TERMINAL",0,0,"4",,terminal_output +50,791095,"TERMINAL",0,0,"5",,terminal_output +51,792038,"TERMINAL",0,0,"6",,terminal_output +52,793158,"TERMINAL",0,0,"7",,terminal_output +53,794128,"TERMINAL",0,0,"8",,terminal_output +54,795193,"TERMINAL",0,0,"9",,terminal_output +55,796215,"TERMINAL",0,0,"30",,terminal_output +56,797373,"TERMINAL",0,0,"1",,terminal_output +57,798365,"TERMINAL",0,0,"2",,terminal_output +58,799389,"TERMINAL",0,0,"3",,terminal_output +59,800333,"TERMINAL",0,0,"4",,terminal_output +60,801374,"TERMINAL",0,0,"5",,terminal_output +61,802436,"TERMINAL",0,0,"6",,terminal_output +62,803444,"TERMINAL",0,0,"7",,terminal_output +63,804477,"TERMINAL",0,0,"8",,terminal_output +64,805519,"TERMINAL",0,0,"9",,terminal_output +65,806591,"TERMINAL",0,0,"40",,terminal_output +66,807632,"TERMINAL",0,0,"1",,terminal_output +67,808715,"TERMINAL",0,0,"2",,terminal_output +68,809733,"TERMINAL",0,0,"3",,terminal_output +69,810754,"TERMINAL",0,0,"4",,terminal_output +70,811785,"TERMINAL",0,0,"5",,terminal_output +71,812907,"TERMINAL",0,0,"6",,terminal_output +72,813984,"TERMINAL",0,0,"8",,terminal_output +73,814953,"TERMINAL",0,0,"9",,terminal_output +74,815978,"TERMINAL",0,0,"50",,terminal_output +75,817003,"TERMINAL",0,0,"1",,terminal_output +76,818005,"TERMINAL",0,0,"2",,terminal_output +77,819049,"TERMINAL",0,0,"3",,terminal_output +78,820078,"TERMINAL",0,0,"4",,terminal_output +79,821215,"TERMINAL",0,0,"5",,terminal_output +80,822169,"TERMINAL",0,0,"6",,terminal_output +81,823252,"TERMINAL",0,0,"72",,terminal_output +82,824275,"TERMINAL",0,0,"8",,terminal_output +83,825296,"TERMINAL",0,0,"9",,terminal_output +84,826298,"TERMINAL",0,0,"2:00",,terminal_output +85,827339,"TERMINAL",0,0,"1",,terminal_output +86,828382,"TERMINAL",0,0,"2",,terminal_output +87,829495,"TERMINAL",0,0,"3",,terminal_output +88,830523,"TERMINAL",0,0,"4",,terminal_output +89,831492,"TERMINAL",0,0,"5",,terminal_output +90,832529,"TERMINAL",0,0,"6",,terminal_output +91,833564,"TERMINAL",0,0,"7",,terminal_output +92,834600,"TERMINAL",0,0,"8",,terminal_output +93,835637,"TERMINAL",0,0,"9",,terminal_output +94,836703,"TERMINAL",0,0,"10",,terminal_output +95,837712,"TERMINAL",0,0,"1",,terminal_output +96,838765,"TERMINAL",0,0,"2",,terminal_output +97,839816,"TERMINAL",0,0,"3",,terminal_output +98,840825,"TERMINAL",0,0,"4",,terminal_output +99,841860,"TERMINAL",0,0,"6",,terminal_output +100,842914,"TERMINAL",0,0,"7",,terminal_output +101,842977,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=dev_accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5",,terminal_command +102,843027,"TERMINAL",0,0,"]633;E;2025-07-10 16:32:17 salloc --time=01:00:00 --partition=dev_accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5;cd331af2-3b46-4f78-aca4-9dda848374ca]633;Csalloc: Pending job allocation 3335323\r\nsalloc: job 3335323 queued and waiting for resources\r\n",,terminal_output +103,843929,"TERMINAL",0,0,"8",,terminal_output +104,845060,"TERMINAL",0,0,"9",,terminal_output +105,846003,"TERMINAL",0,0,"20",,terminal_output +106,847112,"TERMINAL",0,0,"1",,terminal_output +107,848078,"TERMINAL",0,0,"2",,terminal_output +108,849156,"TERMINAL",0,0,"3",,terminal_output +109,850179,"TERMINAL",0,0,"4",,terminal_output +110,851205,"TERMINAL",0,0,"5",,terminal_output +111,852333,"TERMINAL",0,0,"6",,terminal_output +112,853268,"TERMINAL",0,0,"7",,terminal_output +113,854378,"TERMINAL",0,0,"8",,terminal_output +114,855408,"TERMINAL",0,0,"9",,terminal_output +115,856382,"TERMINAL",0,0,"30",,terminal_output +116,857464,"TERMINAL",0,0,"1",,terminal_output +117,857714,"TERMINAL",0,0,"^C",,terminal_output +118,857901,"TERMINAL",0,0,"^[[A",,terminal_output +119,858567,"TERMINAL",0,0,"^C",,terminal_output +120,858764,"TERMINAL",0,0,"^C",,terminal_output +121,858931,"TERMINAL",0,0,"^C",,terminal_output +122,859051,"TERMINAL",0,0,"salloc: Job allocation 3335323 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;1",,terminal_output +123,859166,"TERMINAL",0,0,"20",,terminal_output +124,860115,"TERMINAL",0,0,"4",,terminal_output +125,861173,"TERMINAL",0,0,"5",,terminal_output +126,862191,"TERMINAL",0,0,"6",,terminal_output +127,863246,"TERMINAL",0,0,"7",,terminal_output +128,864274,"TERMINAL",0,0,"8",,terminal_output +129,865315,"TERMINAL",0,0,"9",,terminal_output +130,865331,"TERMINAL",0,0,"salloc --time=05:00:00 --partition=accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5",,terminal_command +131,865386,"TERMINAL",0,0,"]633;E;2025-07-10 16:32:39 salloc --time=05:00:00 --partition=accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5;cd331af2-3b46-4f78-aca4-9dda848374ca]633;Csalloc: Pending job allocation 3335324\r\nsalloc: job 3335324 queued and waiting for resources\r\n",,terminal_output +132,866345,"TERMINAL",0,0,"40",,terminal_output +133,867379,"TERMINAL",0,0,"1",,terminal_output +134,868413,"TERMINAL",0,0,"2",,terminal_output +135,869534,"TERMINAL",0,0,"3",,terminal_output +136,869908,"TERMINAL",0,0,"^Csalloc: Job allocation 3335324 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;1]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +137,870512,"TERMINAL",0,0,"4",,terminal_output +138,871536,"TERMINAL",0,0,"5",,terminal_output +139,872568,"TERMINAL",0,0,"6",,terminal_output +140,873637,"TERMINAL",0,0,"7",,terminal_output +141,874686,"TERMINAL",0,0,"8",,terminal_output +142,875296,"TERMINAL",0,0,"salloc --time=01:00:00 --partition=dev_accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5",,terminal_command +143,875342,"TERMINAL",0,0,"]633;E;2025-07-10 16:32:49 salloc --time=01:00:00 --partition=dev_accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5;cd331af2-3b46-4f78-aca4-9dda848374ca]633;Csalloc: Pending job allocation 3335325\r\nsalloc: job 3335325 queued and waiting for resources\r\n",,terminal_output +144,875886,"TERMINAL",0,0,"9",,terminal_output +145,876817,"TERMINAL",0,0,"50",,terminal_output +146,877765,"TERMINAL",0,0,"1",,terminal_output +147,878802,"TERMINAL",0,0,"2",,terminal_output +148,879851,"TERMINAL",0,0,"3",,terminal_output +149,880879,"TERMINAL",0,0,"5",,terminal_output +150,881928,"TERMINAL",0,0,"6",,terminal_output +151,883052,"TERMINAL",0,0,"7",,terminal_output +152,884078,"TERMINAL",0,0,"8",,terminal_output +153,885101,"TERMINAL",0,0,"9",,terminal_output +154,886125,"TERMINAL",0,0,"3:00",,terminal_output +155,887148,"TERMINAL",0,0,"1",,terminal_output +156,888140,"TERMINAL",0,0,"2",,terminal_output +157,889194,"TERMINAL",0,0,"3",,terminal_output +158,890361,"TERMINAL",0,0,"4",,terminal_output +159,891328,"TERMINAL",0,0,"5",,terminal_output +160,892338,"TERMINAL",0,0,"6",,terminal_output +161,893325,"TERMINAL",0,0,"7",,terminal_output +162,894363,"TERMINAL",0,0,"8",,terminal_output +163,895460,"TERMINAL",0,0,"9",,terminal_output +164,896479,"TERMINAL",0,0,"10",,terminal_output +165,897488,"TERMINAL",0,0,"1",,terminal_output +166,898508,"TERMINAL",0,0,"2",,terminal_output +167,899549,"TERMINAL",0,0,"3",,terminal_output +168,900582,"TERMINAL",0,0,"4",,terminal_output +169,901615,"TERMINAL",0,0,"5",,terminal_output +170,902694,"TERMINAL",0,0,"6",,terminal_output +171,903735,"TERMINAL",0,0,"7",,terminal_output +172,904734,"TERMINAL",0,0,"8",,terminal_output +173,905783,"TERMINAL",0,0,"9",,terminal_output +174,906912,"TERMINAL",0,0,"20",,terminal_output +175,907937,"TERMINAL",0,0,"1",,terminal_output +176,908882,"TERMINAL",0,0,"3",,terminal_output +177,909984,"TERMINAL",0,0,"4",,terminal_output +178,910957,"TERMINAL",0,0,"5",,terminal_output +179,911998,"TERMINAL",0,0,"6",,terminal_output +180,913056,"TERMINAL",0,0,"7",,terminal_output +181,914082,"TERMINAL",0,0,"8",,terminal_output +182,915207,"TERMINAL",0,0,"9",,terminal_output +183,916228,"TERMINAL",0,0,"30",,terminal_output +184,917196,"TERMINAL",0,0,"1",,terminal_output +185,918254,"train_tokenizer.py",0,0,"from dataclasses import dataclass, field\nimport os\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n min_lr: float = 0.0\n max_lr: float = 3e-4\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 8\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_gradients: bool = False\n\n\nargs = tyro.cli(Args)\n\n\ndef tokenizer_loss_fn(params, state, inputs):\n # --- Compute loss ---\n outputs = state.apply_fn(\n params,\n inputs,\n training=True,\n rngs={""params"": inputs[""rng""], ""dropout"": inputs[""dropout_rng""]},\n )\n mse = jnp.square(inputs[""videos""] - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = inputs[""videos""].clip(0, 1).reshape(-1, *inputs[""videos""].shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = pix.psnr(gt, recon).mean() # type: ignore\n ssim = pix.ssim(gt, recon).mean() # type: ignore\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n grad_fn = jax.value_and_grad(tokenizer_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n\n # --- Initialize model ---\n tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_height, args.image_width, args.image_channels)\n inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=jnp.float32,\n ),\n )\n init_params = tokenizer.init(_rng, inputs)\n\n param_counts = count_parameters_by_component(init_params)\n\n if args.log and jax.process_index() == 0:\n wandb.init(\n entity=args.entity,\n project=args.project,\n name=args.name,\n tags=args.tags,\n group=""debug"",\n config=args,\n )\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=tokenizer.apply, params=init_params, tx=tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(\n mesh, PartitionSpec(""data"", None, None, None, None)\n )\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add('model_state', ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler)\n handler_registry.add('model_state', ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler)\n handler_registry.add('dataloader_state', grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add('dataloader_state', grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler) # type: ignore\n \n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=10,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n \n checkpoint_manager = ocp.CheckpointManager(\n os.path.abspath(args.ckpt_dir),\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n \n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_train_state = jax.tree_util.tree_map(ocp.utils.to_shape_dtype_struct, train_state)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_train_state),\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator),\n )\n )\n train_state = restored[""model_state""]\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader = (jax.make_array_from_process_local_data(videos_sharding, elem) for elem in grain_iterator) # type: ignore\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _rng_dropout = jax.random.split(rng, 3)\n\n inputs = dict(videos=videos, rng=_rng, dropout_rng=_rng_dropout)\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.StandardSave(train_state),\n dataloader_state=grain.checkpoint.CheckpointSave(grain_iterator),\n )\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()",python,tab +186,918378,"TERMINAL",0,0,"2",,terminal_output +187,919300,"TERMINAL",0,0,"3",,terminal_output +188,920294,"TERMINAL",0,0,"4",,terminal_output +189,921344,"TERMINAL",0,0,"5",,terminal_output +190,922443,"TERMINAL",0,0,"60",,terminal_output +191,922595,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n# tf_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/open_ai_minecraft_tfrecord\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/train_lam_action_space_scaling_20/3318547/lam_1751657975_200000/\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --lam_checkpoint=$lam_ckpt_dir\n",shellscript,tab +192,923522,"TERMINAL",0,0,"7",,terminal_output +193,924528,"TERMINAL",0,0,"8",,terminal_output +194,925547,"TERMINAL",0,0,"9",,terminal_output +195,926424,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1787,0,"",shellscript,selection_mouse +196,926426,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1786,0,"",shellscript,selection_command +197,926532,"TERMINAL",0,0,"40",,terminal_output +198,927607,"TERMINAL",0,0,"1",,terminal_output +199,927687,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1750,0,"",shellscript,selection_command +200,928093,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1752,0,"",shellscript,selection_command +201,928193,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1751,1,"",shellscript,content +202,928339,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1750,1,"",shellscript,content +203,928613,"TERMINAL",0,0,"2",,terminal_output +204,928627,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1749,0,"",shellscript,selection_command +205,928800,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1783,0,"",shellscript,selection_command +206,929059,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1755,0,"",shellscript,selection_command +207,929650,"TERMINAL",0,0,"3",,terminal_output +208,929779,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1755,0,"#",shellscript,content +209,929781,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1756,0,"",shellscript,selection_keyboard +210,929864,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1756,0," ",shellscript,content +211,929865,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1757,0,"",shellscript,selection_keyboard +212,930586,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1756,0,"",shellscript,selection_command +213,930700,"TERMINAL",0,0,"4",,terminal_output +214,931724,"TERMINAL",0,0,"5",,terminal_output +215,932768,"TERMINAL",0,0,"6",,terminal_output +216,933803,"TERMINAL",0,0,"7",,terminal_output +217,933822,"TERMINAL",0,0,"watch",,terminal_focus +218,934840,"TERMINAL",0,0,"8",,terminal_output +219,935548,"TERMINAL",0,0,"salloc",,terminal_focus +220,935878,"TERMINAL",0,0,"50",,terminal_output +221,936968,"TERMINAL",0,0,"1",,terminal_output +222,937958,"TERMINAL",0,0,"2",,terminal_output +223,938992,"TERMINAL",0,0,"3",,terminal_output +224,940031,"TERMINAL",0,0,"4",,terminal_output +225,941072,"TERMINAL",0,0,"5",,terminal_output +226,942133,"TERMINAL",0,0,"6",,terminal_output +227,943164,"TERMINAL",0,0,"7",,terminal_output +228,944289,"TERMINAL",0,0,"8",,terminal_output +229,945227,"TERMINAL",0,0,"9",,terminal_output +230,946295,"TERMINAL",0,0,"4:00",,terminal_output +231,947319,"TERMINAL",0,0,"1",,terminal_output +232,948387,"TERMINAL",0,0,"2",,terminal_output +233,949407,"TERMINAL",0,0,"3",,terminal_output +234,950433,"TERMINAL",0,0,"4",,terminal_output +235,951450,"TERMINAL",0,0,"5",,terminal_output +236,952515,"TERMINAL",0,0,"6",,terminal_output +237,953520,"TERMINAL",0,0,"7",,terminal_output +238,954558,"TERMINAL",0,0,"8",,terminal_output +239,955603,"TERMINAL",0,0,"9",,terminal_output +240,956680,"TERMINAL",0,0,"10",,terminal_output +241,957702,"TERMINAL",0,0,"1",,terminal_output +242,958697,"TERMINAL",0,0,"2",,terminal_output +243,959735,"TERMINAL",0,0,"3",,terminal_output +244,960776,"TERMINAL",0,0,"4",,terminal_output +245,961806,"TERMINAL",0,0,"5",,terminal_output +246,962842,"TERMINAL",0,0,"6",,terminal_output +247,963878,"TERMINAL",0,0,"8",,terminal_output +248,964916,"TERMINAL",0,0,"9",,terminal_output +249,965956,"TERMINAL",0,0,"20",,terminal_output +250,966996,"TERMINAL",0,0,"1",,terminal_output +251,968150,"TERMINAL",0,0,"2",,terminal_output +252,969170,"TERMINAL",0,0,"3",,terminal_output +253,970110,"TERMINAL",0,0,"4",,terminal_output +254,971225,"TERMINAL",0,0,"5",,terminal_output +255,972186,"TERMINAL",0,0,"6",,terminal_output +256,973220,"TERMINAL",0,0,"7",,terminal_output +257,974292,"TERMINAL",0,0,"8",,terminal_output +258,975314,"TERMINAL",0,0,"9",,terminal_output +259,976328,"TERMINAL",0,0,"30",,terminal_output +260,977468,"TERMINAL",0,0,"1",,terminal_output +261,978409,"TERMINAL",0,0,"2",,terminal_output +262,979702,"TERMINAL",0,0,"3",,terminal_output +263,980742,"TERMINAL",0,0,"4",,terminal_output +264,981868,"TERMINAL",0,0,"5",,terminal_output +265,982892,"TERMINAL",0,0,"6",,terminal_output +266,983916,"TERMINAL",0,0,"8",,terminal_output +267,984890,"TERMINAL",0,0,"9",,terminal_output +268,985965,"TERMINAL",0,0,"40",,terminal_output +269,986995,"TERMINAL",0,0,"1",,terminal_output +270,988013,"TERMINAL",0,0,"2",,terminal_output +271,989064,"TERMINAL",0,0,"3",,terminal_output +272,990171,"TERMINAL",0,0,"4",,terminal_output +273,991101,"TERMINAL",0,0,"5",,terminal_output +274,992215,"TERMINAL",0,0,"6",,terminal_output +275,993177,"TERMINAL",0,0,"7",,terminal_output +276,994213,"TERMINAL",0,0,"8",,terminal_output +277,995281,"TERMINAL",0,0,"9",,terminal_output +278,996295,"TERMINAL",0,0,"50",,terminal_output +279,997332,"TERMINAL",0,0,"1",,terminal_output +280,998381,"TERMINAL",0,0,"2",,terminal_output +281,999401,"TERMINAL",0,0,"3",,terminal_output +282,1000431,"TERMINAL",0,0,"4",,terminal_output +283,1001486,"TERMINAL",0,0,"5",,terminal_output +284,1002269,"TERMINAL",0,0,"bash",,terminal_focus +285,1002506,"TERMINAL",0,0,"6",,terminal_output +286,1003577,"TERMINAL",0,0,"7",,terminal_output +287,1003940,"TERMINAL",0,0,"queue",,terminal_command +288,1004012,"TERMINAL",0,0,"]633;E;2025-07-10 16:34:58 queue;16c08be6-3885-420f-ac53-f5272aed6e54]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Jul 10 16:34:58 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3333584 accelerat interact tum_cte0 R 18:14:52\t 1 hkn05233331283 accelerat train_la tum_cte0 R 1-15:56:45\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 1-15:56:45\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 1-15:56:45\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 1-15:56:45\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 1-15:56:45\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 1-15:56:45\t 2 hkn[0809,0811]3335325 dev_accel interact tum_cte0 PD\t0:00\t 2 (Priority)",,terminal_output +289,1004597,"TERMINAL",0,0,"8",,terminal_output +290,1005055,"TERMINAL",0,0,"93666666",,terminal_output +291,1005616,"TERMINAL",0,0,"9",,terminal_output +292,1006136,"TERMINAL",0,0,"5:004777777",,terminal_output +293,1006749,"TERMINAL",0,0,"5:00",,terminal_output +294,1007145,"TERMINAL",0,0,"15888888",,terminal_output +295,1007695,"TERMINAL",0,0,"1",,terminal_output +296,1008184,"TERMINAL",0,0,"26999999",,terminal_output +297,1008732,"TERMINAL",0,0,"2",,terminal_output +298,1008957,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +299,1009775,"TERMINAL",0,0,"3",,terminal_output +300,1010836,"TERMINAL",0,0,"4",,terminal_output +301,1011874,"TERMINAL",0,0,"6",,terminal_output +302,1012895,"TERMINAL",0,0,"7",,terminal_output +303,1013953,"TERMINAL",0,0,"8",,terminal_output +304,1014980,"TERMINAL",0,0,"9",,terminal_output +305,1015277,"TERMINAL",0,0,"srun --overlap --jobid=3333584 --pty /bin/bash",,terminal_command +306,1015328,"TERMINAL",0,0,"]633;E;2025-07-10 16:35:09 srun --overlap --jobid=3333584 --pty /bin/bash;16c08be6-3885-420f-ac53-f5272aed6e54]633;C",,terminal_output +307,1015472,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +308,1016021,"TERMINAL",0,0,"10",,terminal_output +309,1016305,"TERMINAL",0,0,"]0;tum_cte0515@hkn0523:~/Projects/jafar[?2004h]633;A(jafar) ]633;Ajafar[tum_cte0515@hkn0523 jafar]$ ]633;B]633;B",,terminal_output +310,1017107,"TERMINAL",0,0,"1",,terminal_output +311,1017471,"TERMINAL",0,0,"q",,terminal_output +312,1018106,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +313,1018106,"TERMINAL",0,0,"2",,terminal_output +314,1018177,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +315,1018310,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +316,1018378,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +317,1018549,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: squeue --mehkn0523.localdomain: Thu Jul 10 16:35:12 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3333584 accelerat interact tum_cte0 R 18:15:06\t 1 hkn05233331283 accelerat train_la tum_cte0 R 1-15:56:59\t 2 hkn[0533,0536]3331284 accelerat train_la tum_cte0 R 1-15:56:59\t 2 hkn[0601-0602]3331285 accelerat train_la tum_cte0 R 1-15:56:59\t 2 hkn[0701,0710]3331286 accelerat train_la tum_cte0 R 1-15:56:59\t 2 hkn[0425,0428]3331287 accelerat train_la tum_cte0 R 1-15:56:59\t 2 hkn[0430-0431]3331288 accelerat train_la tum_cte0 R 1-15:56:59\t 2 hkn[0809,0811]3335325 dev_accel interact tum_cte0 PD\t0:00\t 2 (Priority)",,terminal_output +318,1019141,"TERMINAL",0,0,"3",,terminal_output +319,1019531,"TERMINAL",0,0,"377:007:007:007:007:007:00",,terminal_output +320,1020155,"TERMINAL",0,0,"4",,terminal_output +321,1020425,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0523:~/Projects/jafar[?2004h]633;A(jafar) ]633;Ajafar[tum_cte0515@hkn0523 jafar]$ ]633;B]633;B",,terminal_output +322,1020651,"TERMINAL",0,0,"s",,terminal_output +323,1020718,"TERMINAL",0,0,"[?25lm[?25h",,terminal_output +324,1020886,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +325,1021072,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +326,1021193,"TERMINAL",0,0,"5",,terminal_output +327,1021464,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: nvidia-smihkn0523.localdomain: Thu Jul 10 16:35:15 2025Thu Jul 10 16:35:15 2025\r+-----------------------------------------------------------------------------------------+\r| NVIDIA-SMI 570.133.20Driver Version: 570.133.20 CUDA Version: 12.8 |\r|-----------------------------------------+------------------------+----------------------+\r| GPU NamePersistence-M | Bus-IdDisp.A | Volatile Uncorr. ECC |\r| Fan Temp PerfPwr:Usage/Cap |Memory-Usage | GPU-Util Compute M. |\r|||MIG M. |\r|=========================================+========================+======================|\r| 0 NVIDIA A100-SXM4-40GBOn | 00000000:31:00.0 Off |0 |\r| N/A 46C P057W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r+-----------------------------------------------------------------------------------------+\r| Processes:|\r| GPU GI CIPID Type Process nameGPU Memory |\r|ID IDUsage\t |\r|=========================================================================================|\r| 0 N/A N/A2617G /usr/libexec/Xorg17MiB |\r+-----------------------------------------------------------------------------------------+",,terminal_output +328,1022318,"TERMINAL",0,0,"6",,terminal_output +329,1022738,"TERMINAL",0,0,"6663",,terminal_output +330,1022849,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0523:~/Projects/jafar[?2004h]633;A(jafar) ]633;Ajafar[tum_cte0515@hkn0523 jafar]$ ]633;B]633;B",,terminal_output +331,1023340,"TERMINAL",0,0,"7",,terminal_output +332,1024332,"TERMINAL",0,0,"8",,terminal_output +333,1025391,"TERMINAL",0,0,"9",,terminal_output +334,1026396,"TERMINAL",0,0,"20",,terminal_output +335,1027415,"TERMINAL",0,0,"1",,terminal_output +336,1028461,"TERMINAL",0,0,"2",,terminal_output +337,1029524,"TERMINAL",0,0,"3",,terminal_output +338,1030537,"TERMINAL",0,0,"4",,terminal_output +339,1031573,"TERMINAL",0,0,"5",,terminal_output +340,1032618,"TERMINAL",0,0,"6",,terminal_output +341,1033659,"TERMINAL",0,0,"7",,terminal_output +342,1034694,"TERMINAL",0,0,"8",,terminal_output +343,1035734,"TERMINAL",0,0,"9",,terminal_output +344,1036776,"TERMINAL",0,0,"30",,terminal_output +345,1037882,"TERMINAL",0,0,"1",,terminal_output +346,1038915,"TERMINAL",0,0,"2",,terminal_output +347,1039885,"TERMINAL",0,0,"4",,terminal_output +348,1040954,"TERMINAL",0,0,"5",,terminal_output +349,1041956,"TERMINAL",0,0,"6",,terminal_output +350,1043009,"TERMINAL",0,0,"7",,terminal_output +351,1044129,"TERMINAL",0,0,"8",,terminal_output +352,1045081,"TERMINAL",0,0,"9",,terminal_output +353,1046152,"TERMINAL",0,0,"40",,terminal_output +354,1047198,"TERMINAL",0,0,"1",,terminal_output +355,1048221,"TERMINAL",0,0,"2",,terminal_output +356,1049247,"TERMINAL",0,0,"3",,terminal_output +357,1050260,"TERMINAL",0,0,"4",,terminal_output +358,1051294,"TERMINAL",0,0,"5",,terminal_output +359,1052423,"TERMINAL",0,0,"6",,terminal_output +360,1053448,"TERMINAL",0,0,"7",,terminal_output +361,1054470,"TERMINAL",0,0,"8",,terminal_output +362,1055496,"TERMINAL",0,0,"9",,terminal_output +363,1056479,"TERMINAL",0,0,"50",,terminal_output +364,1057558,"TERMINAL",0,0,"1",,terminal_output +365,1058553,"TERMINAL",0,0,"2",,terminal_output +366,1059599,"TERMINAL",0,0,"3",,terminal_output +367,1060630,"TERMINAL",0,0,"4",,terminal_output +368,1061667,"TERMINAL",0,0,"5",,terminal_output +369,1062703,"TERMINAL",0,0,"6",,terminal_output +370,1063744,"TERMINAL",0,0,"7",,terminal_output +371,1064783,"TERMINAL",0,0,"8",,terminal_output +372,1065838,"TERMINAL",0,0,"9",,terminal_output +373,1066862,"TERMINAL",0,0,"6:012",,terminal_output +374,1067898,"TERMINAL",0,0,"2",,terminal_output +375,1069007,"TERMINAL",0,0,"3",,terminal_output +376,1070033,"TERMINAL",0,0,"4",,terminal_output +377,1071061,"TERMINAL",0,0,"5",,terminal_output +378,1072085,"TERMINAL",0,0,"6",,terminal_output +379,1073080,"TERMINAL",0,0,"7",,terminal_output +380,1074132,"TERMINAL",0,0,"8",,terminal_output +381,1075285,"TERMINAL",0,0,"9",,terminal_output +382,1076200,"TERMINAL",0,0,"10",,terminal_output +383,1077306,"TERMINAL",0,0,"1",,terminal_output +384,1078279,"TERMINAL",0,0,"2",,terminal_output +385,1079355,"TERMINAL",0,0,"3",,terminal_output +386,1080382,"TERMINAL",0,0,"4",,terminal_output +387,1081395,"TERMINAL",0,0,"5",,terminal_output +388,1082441,"TERMINAL",0,0,"6",,terminal_output +389,1083488,"TERMINAL",0,0,"7",,terminal_output +390,1084567,"TERMINAL",0,0,"8",,terminal_output +391,1085574,"TERMINAL",0,0,"9",,terminal_output +392,1086600,"TERMINAL",0,0,"20",,terminal_output +393,1087635,"TERMINAL",0,0,"1",,terminal_output +394,1088674,"TERMINAL",0,0,"2",,terminal_output +395,1089738,"TERMINAL",0,0,"3",,terminal_output +396,1090748,"TERMINAL",0,0,"4",,terminal_output +397,1091103,"train_tokenizer.py",0,0,"",python,tab +398,1091796,"TERMINAL",0,0,"5",,terminal_output +399,1092669,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"",shellscript,tab +400,1092832,"TERMINAL",0,0,"6",,terminal_output +401,1093377,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1645,0,"",shellscript,selection_mouse +402,1093383,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1644,0,"",shellscript,selection_command +403,1093868,"TERMINAL",0,0,"8",,terminal_output +404,1094556,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1337,0,"",shellscript,selection_mouse +405,1094921,"TERMINAL",0,0,"9",,terminal_output +406,1095090,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1370,0,"",shellscript,selection_mouse +407,1095950,"TERMINAL",0,0,"30",,terminal_output +408,1095999,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1369,1,"",shellscript,content +409,1096119,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1368,1,"",shellscript,content +410,1096209,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1368,0,"1",shellscript,content +411,1096210,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1369,0,"",shellscript,selection_keyboard +412,1096302,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1369,0,"2",shellscript,content +413,1096303,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1370,0,"",shellscript,selection_keyboard +414,1096894,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1369,0,"",shellscript,selection_command +415,1097023,"TERMINAL",0,0,"1",,terminal_output +416,1098092,"TERMINAL",0,0,"2",,terminal_output +417,1099067,"TERMINAL",0,0,"3",,terminal_output +418,1100581,"TERMINAL",0,0,"40",,terminal_output +419,1101684,"TERMINAL",0,0,"5",,terminal_output +420,1102641,"TERMINAL",0,0,"6",,terminal_output +421,1103023,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1368,2,"",shellscript,content +422,1103376,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1368,0,"96",shellscript,content +423,1103692,"TERMINAL",0,0,"7",,terminal_output +424,1104743,"TERMINAL",0,0,"8",,terminal_output +425,1105760,"TERMINAL",0,0,"9",,terminal_output +426,1105914,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +427,1106801,"TERMINAL",0,0,"40",,terminal_output +428,1107556,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"",shellscript,tab +429,1107851,"TERMINAL",0,0,"1",,terminal_output +430,1108947,"TERMINAL",0,0,"3",,terminal_output +431,1109971,"TERMINAL",0,0,"4",,terminal_output +432,1110846,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1788,0,"",shellscript,selection_mouse +433,1110959,"TERMINAL",0,0,"5",,terminal_output +434,1111740,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1750,0,"",shellscript,selection_mouse +435,1111995,"TERMINAL",0,0,"6",,terminal_output +436,1112010,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1749,1,"r",shellscript,selection_mouse +437,1112026,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1748,2,"ir",shellscript,selection_mouse +438,1112041,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1746,4,"_dir",shellscript,selection_mouse +439,1112056,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1745,5,"s_dir",shellscript,selection_mouse +440,1112070,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1744,6,"ds_dir",shellscript,selection_mouse +441,1112093,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1743,7,"rds_dir",shellscript,selection_mouse +442,1112115,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1741,9,"cords_dir",shellscript,selection_mouse +443,1112130,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1739,11,"records_dir",shellscript,selection_mouse +444,1112146,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1687,63,"ckpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +445,1112162,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1685,65,"heckpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +446,1112184,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1661,89,"afar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +447,1112199,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1590,160,"ynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +448,1112215,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1533,217,"=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +449,1112230,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1464,286,"m_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +450,1112245,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1451,299,"og \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +451,1112262,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1450,300,"log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +452,1112279,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1417,333,"-log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +453,1112334,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1395,355,"-max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +454,1112336,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1394,356,"--max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +455,1112338,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1393,357," --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +456,1112353,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1376,374," --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +457,1112380,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1375,375," --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +458,1112403,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1352,398," --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +459,1112418,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1319,431," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +460,1112433,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,432," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +461,1112491,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1286,464,"srun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +462,1112904,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,432," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +463,1113037,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1351,399," --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +464,1113053,"TERMINAL",0,0,"7",,terminal_output +465,1114138,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1321,0,"",shellscript,selection_mouse +466,1114167,"TERMINAL",0,0,"8",,terminal_output +467,1114288,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,4," ",shellscript,selection_mouse +468,1114491,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,37," --ckpt_dir $CHECKPOINT_DIR \\n ",shellscript,selection_mouse +469,1114513,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,84," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr",shellscript,selection_mouse +470,1114528,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,137," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \",shellscript,selection_mouse +471,1114543,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,162," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions",shellscript,selection_mouse +472,1114563,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,197," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval",shellscript,selection_mouse +473,1114578,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,232," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun",shellscript,selection_mouse +474,1114643,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,288," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run",shellscript,selection_mouse +475,1114644,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,289," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run ",shellscript,selection_mouse +476,1114656,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,327," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \",shellscript,selection_mouse +477,1114714,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,349," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \",shellscript,selection_mouse +478,1114733,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,396," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir",shellscript,selection_mouse +479,1114791,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1318,432," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +480,1115109,"TERMINAL",0,0,"9",,terminal_output +481,1116215,"TERMINAL",0,0,"50",,terminal_output +482,1117189,"TERMINAL",0,0,"1",,terminal_output +483,1118231,"TERMINAL",0,0,"2",,terminal_output +484,1119263,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +485,1119331,"TERMINAL",0,0,"3",,terminal_output +486,1120302,"TERMINAL",0,0,"4",,terminal_output +487,1120415,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1265,0,"",shellscript,selection_mouse +488,1120433,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1264,0,"",shellscript,selection_command +489,1120608,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1253,12,"lam_ckpt_dir",shellscript,selection_mouse +490,1120610,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1254,11,"am_ckpt_dir",shellscript,selection_command +491,1121164,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1228,0,"",shellscript,selection_mouse +492,1121165,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1227,0,"",shellscript,selection_command +493,1121298,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1227,1,"r",shellscript,selection_mouse +494,1121301,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1228,0,"",shellscript,selection_command +495,1121328,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1177,51,"enizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +496,1121345,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1169,59,"int=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +497,1121363,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1109,119,"y instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +498,1121397,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1061,167,"-tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +499,1121398,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1002,226," --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +500,1122792,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",934,294," --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +501,1122793,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",922,306," --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +502,1122794,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",890,338," --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +503,1122794,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",868,360," --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +504,1122794,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",851,377," --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +505,1122794,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",829,399," --batch_size=82 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +506,1122795,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",796,432," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=82 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,selection_mouse +507,1122796,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",796,432,"",shellscript,content +508,1122878,"TERMINAL",0,0,"56",,terminal_output +509,1122981,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",796,0," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,content +510,1123415,"TERMINAL",0,0,"7",,terminal_output +511,1123695,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",986,0,"",shellscript,selection_mouse +512,1124452,"TERMINAL",0,0,"8",,terminal_output +513,1125539,"TERMINAL",0,0,"9",,terminal_output +514,1126530,"TERMINAL",0,0,"7:00",,terminal_output +515,1127608,"TERMINAL",0,0,"1",,terminal_output +516,1128609,"TERMINAL",0,0,"2",,terminal_output +517,1129645,"TERMINAL",0,0,"3",,terminal_output +518,1130694,"TERMINAL",0,0,"4",,terminal_output +519,1131721,"TERMINAL",0,0,"5",,terminal_output +520,1132769,"TERMINAL",0,0,"6",,terminal_output +521,1133830,"TERMINAL",0,0,"7",,terminal_output +522,1134956,"TERMINAL",0,0,"8",,terminal_output +523,1135979,"TERMINAL",0,0,"10",,terminal_output +524,1137003,"TERMINAL",0,0,"1",,terminal_output +525,1137966,"TERMINAL",0,0,"2",,terminal_output +526,1139055,"TERMINAL",0,0,"3",,terminal_output +527,1140046,"TERMINAL",0,0,"4",,terminal_output +528,1141100,"TERMINAL",0,0,"5",,terminal_output +529,1142130,"TERMINAL",0,0,"6",,terminal_output +530,1143253,"TERMINAL",0,0,"7",,terminal_output +531,1144207,"TERMINAL",0,0,"8",,terminal_output +532,1145302,"TERMINAL",0,0,"9",,terminal_output +533,1146282,"TERMINAL",0,0,"20",,terminal_output +534,1147348,"TERMINAL",0,0,"1",,terminal_output +535,1148475,"TERMINAL",0,0,"2",,terminal_output +536,1149402,"TERMINAL",0,0,"3",,terminal_output +537,1150446,"TERMINAL",0,0,"4",,terminal_output +538,1151483,"TERMINAL",0,0,"5",,terminal_output +539,1152523,"TERMINAL",0,0,"6",,terminal_output +540,1153604,"TERMINAL",0,0,"7",,terminal_output +541,1154601,"TERMINAL",0,0,"8",,terminal_output +542,1154635,"TERMINAL",0,0,"watch",,terminal_focus +543,1155135,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +544,1157048,"TERMINAL",0,0,"git branch",,terminal_command +545,1157100,"TERMINAL",0,0,"]633;E;2025-07-10 16:37:31 git branch;d0af5434-9a5b-4b8f-9412-c82c37ee36e4]633;C[?1h=\r add-wandb-name-and-tags\r\n convert-to-jax-array-in-iter\r\n correct-batched-sampling\r\n dev\r\n dont-let-tf-see-gpu\r\n feat/explicit-image-dims\r\n fix-sampling\r\n* grain-dataloader\r\n logging-variants\r\n main\r\n preprocess_video\r\n revised-dataloader\r\n runner\r\n runner-grain\r\n:",,terminal_output +546,1159192,"TERMINAL",0,0,"srun",,terminal_focus +547,1160967,"train_tokenizer.py",0,0,"",python,tab +548,1165746,"train_tokenizer.py",807,0,"",python,selection_mouse +549,1165881,"train_tokenizer.py",802,9,"save_ckpt",python,selection_mouse +550,1197273,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +551,1198507,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1267,0,"",shellscript,selection_mouse +552,1199016,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1228,0,"",shellscript,selection_mouse +553,1202286,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +554,1207932,"TERMINAL",0,0,"git",,terminal_focus +555,1208630,"TERMINAL",0,0," add-wandb-name-and-tags\r\n convert-to-jax-array-in-iter\r\n correct-batched-sampling\r\n dev\r\n dont-let-tf-see-gpu\r\n feat/explicit-image-dims\r\n fix-sampling\r\n* grain-dataloader\r\n logging-variants\r\n main\r\n preprocess_video\r\n revised-dataloader\r\n runner\r\n runner-grain\r\n:",,terminal_output +556,1209044,"TERMINAL",0,0,"\r\r:",,terminal_output +557,1209957,"TERMINAL",0,0,"\r[?1l>\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;130",,terminal_output +558,1211256,"TERMINAL",0,0,"git pull",,terminal_command +559,1211285,"TERMINAL",0,0,"]633;E;2025-07-10 16:38:25 git pull;d0af5434-9a5b-4b8f-9412-c82c37ee36e4]633;C",,terminal_output +560,1212799,"TERMINAL",0,0,"remote: Enumerating objects: 17, done.\r\nremote: Counting objects: 5% (1/17)\rremote: Counting objects: 11% (2/17)\rremote: Counting objects: 17% (3/17)\rremote: Counting objects: 23% (4/17)\rremote: Counting objects: 29% (5/17)\rremote: Counting objects: 35% (6/17)\rremote: Counting objects: 41% (7/17)\rremote: Counting objects: 47% (8/17)\rremote: Counting objects: 52% (9/17)\rremote: Counting objects: 58% (10/17)\rremote: Counting objects: 64% (11/17)\rremote: Counting objects: 70% (12/17)\rremote: Counting objects: 76% (13/17)\rremote: Counting objects: 82% (14/17)\rremote: Counting objects: 88% (15/17)\rremote: Counting objects: 94% (16/17)\rremote: Counting objects: 100% (17/17)\rremote: Counting objects: 100% (17/17), done.\r\nremote: Compressing objects: 10% (1/10)\rremote: Compressing objects: 20% (2/10)\rremote: Compressing objects: 30% (3/10)\rremote: Compressing objects: 40% (4/10)\rremote: Compressing objects: 50% (5/10)\rremote: Compressing objects: 60% (6/10)\rremote: Compressing objects: 70% (7/10)\rremote: Compressing objects: 80% (8/10)\rremote: Compressing objects: 90% (9/10)\rremote: Compressing objects: 100% (10/10)\rremote: Compressing objects: 100% (10/10), done.\r\n",,terminal_output +561,1212964,"TERMINAL",0,0,"remote: Total 17 (delta 7), reused 16 (delta 7), pack-reused 0 (from 0)\r\nUnpacking objects: 5% (1/17)\rUnpacking objects: 11% (2/17)\rUnpacking objects: 17% (3/17)\rUnpacking objects: 23% (4/17)\rUnpacking objects: 29% (5/17)\rUnpacking objects: 35% (6/17)\rUnpacking objects: 41% (7/17)\rUnpacking objects: 47% (8/17)\rUnpacking objects: 52% (9/17)\rUnpacking objects: 58% (10/17)\r",,terminal_output +562,1213032,"TERMINAL",0,0,"Unpacking objects: 64% (11/17)\rUnpacking objects: 70% (12/17)\rUnpacking objects: 76% (13/17)\rUnpacking objects: 82% (14/17)\rUnpacking objects: 88% (15/17)\rUnpacking objects: 94% (16/17)\rUnpacking objects: 100% (17/17)\rUnpacking objects: 100% (17/17), 18.20 KiB | 170.00 KiB/s, done.\r\n",,terminal_output +563,1213140,"TERMINAL",0,0,"From github.com:p-doom/jafar\r\n 75647c1..dae8cf6 grain-dataloader -> origin/grain-dataloader\r\n",,terminal_output +564,1213207,"TERMINAL",0,0," b4614ee..eb5163c main -> origin/main\r\n",,terminal_output +565,1213237,"TERMINAL",0,0,"Updating 75647c1..dae8cf6\r\nerror: Your local changes to the following files would be overwritten by merge:\r\n\ttrain_tokenizer.py\r\nPlease commit your changes or stash them before you merge.\r\nAborting\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;1",,terminal_output +566,1218092,"TERMINAL",0,0,"git stash",,terminal_command +567,1218126,"TERMINAL",0,0,"]633;E;2025-07-10 16:38:32 git stash;d0af5434-9a5b-4b8f-9412-c82c37ee36e4]633;C",,terminal_output +568,1218211,"TERMINAL",0,0,"Saved working directory and index state WIP on grain-dataloader: 75647c1 Merge branch 'main' into grain-dataloader\r\n",,terminal_output +569,1218231,"TERMINAL",0,0,"]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +570,1219310,"TERMINAL",0,0,"git pull",,terminal_command +571,1219361,"TERMINAL",0,0,"]633;E;2025-07-10 16:38:33 git pull;d0af5434-9a5b-4b8f-9412-c82c37ee36e4]633;C",,terminal_output +572,1220759,"TERMINAL",0,0,"Updating 75647c1..dae8cf6\r\n",,terminal_output +573,1220822,"TERMINAL",0,0,"Fast-forward\r\n genie.py | 68 +++++++++++++++++++++++++++++++++++++++++---------------------------\r\n train_dynamics.py | 88 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++----------------------\r\n train_lam.py | 4 +++-\r\n train_tokenizer.py | 4 +++-\r\n 4 files changed, 113 insertions(+), 51 deletions(-)\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +574,1445021,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",984,0,"",shellscript,selection_mouse +575,1445769,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1249,0,"",shellscript,selection_mouse +576,1575712,"TERMINAL",0,0,"srun",,terminal_focus +577,1577163,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +578,1974174,"TERMINAL",0,0,"i",,terminal_output +579,1974322,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +580,1974367,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +581,1974591,"TERMINAL",0,0,"[?25li[?25h[?25ln[?25h",,terminal_output +582,1975334,"TERMINAL",0,0,"[?25lg[?25h",,terminal_output +583,1975401,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +584,1975465,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn0523.localdomain: Thu Jul 10 16:51:09 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 38 nodes idle\rPartition dev_accelerated:\t 0 nodes idle\rPartition accelerated:\t 5 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 6 nodes idle",,terminal_output +585,1976480,"TERMINAL",0,0,"10",,terminal_output +586,1977461,"TERMINAL",0,0,"1",,terminal_output +587,1978481,"TERMINAL",0,0,"2",,terminal_output +588,1978693,"TERMINAL",0,0,"salloc",,terminal_focus +589,1979506,"TERMINAL",0,0,"3",,terminal_output +590,1980538,"TERMINAL",0,0,"4",,terminal_output +591,1981553,"TERMINAL",0,0,"5",,terminal_output +592,1981680,"TERMINAL",0,0,"^Csalloc: Job allocation 3335325 has been revoked.\r\nsalloc: Job aborted due to signal\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;1]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +593,1982633,"TERMINAL",0,0,"6",,terminal_output +594,1983603,"TERMINAL",0,0,"7",,terminal_output +595,1984605,"TERMINAL",0,0,"8",,terminal_output +596,1985653,"TERMINAL",0,0,"9",,terminal_output +597,1986672,"TERMINAL",0,0,"20",,terminal_output +598,1987668,"TERMINAL",0,0,"1",,terminal_output +599,1988274,"TERMINAL",0,0,"salloc --time=10:00:00 --partition=accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5",,terminal_command +600,1988327,"TERMINAL",0,0,"]633;E;2025-07-10 16:51:22 salloc --time=10:00:00 --partition=accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5;cd331af2-3b46-4f78-aca4-9dda848374ca]633;Csalloc: Granted job allocation 3335335\r\n",,terminal_output +601,1988446,"TERMINAL",0,0,"salloc: Waiting for resource configuration\r\n",,terminal_output +602,1988688,"TERMINAL",0,0,"23",,terminal_output +603,1989799,"TERMINAL",0,0,"3",,terminal_output +604,1990841,"TERMINAL",0,0,"4",,terminal_output +605,1991850,"TERMINAL",0,0,"5",,terminal_output +606,1992919,"TERMINAL",0,0,"6",,terminal_output +607,1993897,"TERMINAL",0,0,"7",,terminal_output +608,1994821,"TERMINAL",0,0,"8",,terminal_output +609,1995842,"TERMINAL",0,0,"9",,terminal_output +610,1996866,"TERMINAL",0,0,"31",,terminal_output +611,1997873,"TERMINAL",0,0,"2",,terminal_output +612,1998913,"TERMINAL",0,0,"3",,terminal_output +613,1999911,"TERMINAL",0,0,"4",,terminal_output +614,2000980,"TERMINAL",0,0,"5",,terminal_output +615,2001990,"TERMINAL",0,0,"6",,terminal_output +616,2003011,"TERMINAL",0,0,"7",,terminal_output +617,2004037,"TERMINAL",0,0,"8",,terminal_output +618,2005062,"TERMINAL",0,0,"9",,terminal_output +619,2006084,"TERMINAL",0,0,"40",,terminal_output +620,2007106,"TERMINAL",0,0,"1",,terminal_output +621,2008132,"TERMINAL",0,0,"2",,terminal_output +622,2009172,"TERMINAL",0,0,"3",,terminal_output +623,2010165,"TERMINAL",0,0,"4",,terminal_output +624,2011207,"TERMINAL",0,0,"5",,terminal_output +625,2012225,"TERMINAL",0,0,"6",,terminal_output +626,2013226,"TERMINAL",0,0,"7",,terminal_output +627,2014268,"TERMINAL",0,0,"8",,terminal_output +628,2015299,"TERMINAL",0,0,"9",,terminal_output +629,2015516,"TERMINAL",0,0,"salloc: Nodes hkn[0425,0428] are ready for job\r\n",,terminal_output +630,2016279,"TERMINAL",0,0,"50",,terminal_output +631,2016342,"TERMINAL",0,0,"]0;tum_cte0515@hkn0425:~/Projects/jafar[?2004h[tum_cte0515@hkn0425 jafar]$ ",,terminal_output +632,2017757,"TERMINAL",0,0,"1",,terminal_output +633,2018300,"TERMINAL",0,0,"2",,terminal_output +634,2019320,"TERMINAL",0,0,"3",,terminal_output +635,2020533,"TERMINAL",0,0,"4",,terminal_output +636,2021545,"TERMINAL",0,0,"5",,terminal_output +637,2022770,"TERMINAL",0,0,"6",,terminal_output +638,2023386,"TERMINAL",0,0,"7",,terminal_output +639,2024417,"TERMINAL",0,0,"8",,terminal_output +640,2025436,"TERMINAL",0,0,"9",,terminal_output +641,2026460,"TERMINAL",0,0,"2:00",,terminal_output +642,2027489,"TERMINAL",0,0,"1",,terminal_output +643,2028490,"TERMINAL",0,0,"2",,terminal_output +644,2029516,"TERMINAL",0,0,"3",,terminal_output +645,2030563,"TERMINAL",0,0,"4",,terminal_output +646,2031716,"TERMINAL",0,0,"5",,terminal_output +647,2032605,"TERMINAL",0,0,"6",,terminal_output +648,2033628,"TERMINAL",0,0,"7",,terminal_output +649,2034655,"TERMINAL",0,0,"8",,terminal_output +650,2035888,"TERMINAL",0,0,"9",,terminal_output +651,2036905,"TERMINAL",0,0,"10",,terminal_output +652,2037728,"TERMINAL",0,0,"1",,terminal_output +653,2038696,"TERMINAL",0,0,"2",,terminal_output +654,2039720,"TERMINAL",0,0,"3",,terminal_output +655,2039747,"TERMINAL",0,0,"[?25lqu[?25h[?25lu[?25h",,terminal_output +656,2039854,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +657,2039932,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +658,2040015,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +659,2040118,"TERMINAL",0,0,"[?25l[?2004l\r[?25h[?1049h(B[?7hEvery 1.0s: squeue --mehkn0425.localdomain: Thu Jul 10 16:52:14 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3335335 accelerat interact tum_cte0 R\t0:52\t 2 hkn[0425,0428]3333584 accelerat interact tum_cte0 R 18:32:08\t 1 hkn0523",,terminal_output +660,2040796,"TERMINAL",0,0,"4",,terminal_output +661,2041106,"TERMINAL",0,0,"539",,terminal_output +662,2041984,"TERMINAL",0,0,"5",,terminal_output +663,2042137,"TERMINAL",0,0,"6410",,terminal_output +664,2042845,"TERMINAL",0,0,"6",,terminal_output +665,2043223,"TERMINAL",0,0,"751[?1049l\r[?1l>]0;tum_cte0515@hkn0425:~/Projects/jafar[?2004h[tum_cte0515@hkn0425 jafar]$ ",,terminal_output +666,2043787,"TERMINAL",0,0,"7",,terminal_output +667,2044852,"TERMINAL",0,0,"8",,terminal_output +668,2046227,"TERMINAL",0,0,"9",,terminal_output +669,2046942,"TERMINAL",0,0,"21",,terminal_output +670,2047904,"TERMINAL",0,0,"2",,terminal_output +671,2048909,"TERMINAL",0,0,"3",,terminal_output +672,2049984,"TERMINAL",0,0,"4",,terminal_output +673,2050947,"TERMINAL",0,0,"5",,terminal_output +674,2051961,"TERMINAL",0,0,"6",,terminal_output +675,2052985,"TERMINAL",0,0,"7",,terminal_output +676,2054018,"TERMINAL",0,0,"8",,terminal_output +677,2055032,"TERMINAL",0,0,"9",,terminal_output +678,2056054,"TERMINAL",0,0,"30",,terminal_output +679,2057078,"TERMINAL",0,0,"1",,terminal_output +680,2058091,"TERMINAL",0,0,"2",,terminal_output +681,2059124,"TERMINAL",0,0,"3",,terminal_output +682,2060103,"TERMINAL",0,0,"4",,terminal_output +683,2061144,"TERMINAL",0,0,"5",,terminal_output +684,2062199,"TERMINAL",0,0,"62",,terminal_output +685,2063207,"TERMINAL",0,0,"7",,terminal_output +686,2064248,"TERMINAL",0,0,"8",,terminal_output +687,2065276,"TERMINAL",0,0,"9",,terminal_output +688,2066296,"TERMINAL",0,0,"40",,terminal_output +689,2067319,"TERMINAL",0,0,"1",,terminal_output +690,2068334,"TERMINAL",0,0,"2",,terminal_output +691,2069354,"TERMINAL",0,0,"3",,terminal_output +692,2070374,"TERMINAL",0,0,"4",,terminal_output +693,2071393,"TERMINAL",0,0,"5",,terminal_output +694,2072441,"TERMINAL",0,0,"6",,terminal_output +695,2073434,"TERMINAL",0,0,"7",,terminal_output +696,2074457,"TERMINAL",0,0,"8",,terminal_output +697,2075477,"TERMINAL",0,0,"9",,terminal_output +698,2076535,"TERMINAL",0,0,"50",,terminal_output +699,2077518,"TERMINAL",0,0,"1",,terminal_output +700,2078582,"TERMINAL",0,0,"2",,terminal_output +701,2079617,"TERMINAL",0,0,"3",,terminal_output +702,2080577,"TERMINAL",0,0,"4",,terminal_output +703,2081594,"TERMINAL",0,0,"5",,terminal_output +704,2082679,"TERMINAL",0,0,"6",,terminal_output +705,2083703,"TERMINAL",0,0,"7",,terminal_output +706,2084653,"TERMINAL",0,0,"8",,terminal_output +707,2085673,"TERMINAL",0,0,"9",,terminal_output +708,2086700,"TERMINAL",0,0,"3:00",,terminal_output +709,2087712,"TERMINAL",0,0,"1",,terminal_output +710,2088823,"TERMINAL",0,0,"2",,terminal_output +711,2089849,"TERMINAL",0,0,"3",,terminal_output +712,2090871,"TERMINAL",0,0,"4",,terminal_output +713,2091896,"TERMINAL",0,0,"5",,terminal_output +714,2092817,"TERMINAL",0,0,"6",,terminal_output +715,2093844,"TERMINAL",0,0,"7",,terminal_output +716,2094860,"TERMINAL",0,0,"9",,terminal_output +717,2095889,"TERMINAL",0,0,"103",,terminal_output +718,2096897,"TERMINAL",0,0,"1",,terminal_output +719,2097917,"TERMINAL",0,0,"2",,terminal_output +720,2098936,"TERMINAL",0,0,"3",,terminal_output +721,2099989,"TERMINAL",0,0,"4",,terminal_output +722,2100983,"TERMINAL",0,0,"5",,terminal_output +723,2102033,"TERMINAL",0,0,"6",,terminal_output +724,2102360,"TERMINAL",0,0,"[?25lso[?25h",,terminal_output +725,2102435,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +726,2102502,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +727,2102569,"TERMINAL",0,0,"[?25lr[?25h",,terminal_output +728,2102727,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +729,2102899,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +730,2103027,"TERMINAL",0,0,"7",,terminal_output +731,2103080,"TERMINAL",0,0,"[?25l [?25h[?25l.[?25h[?25lv[?25h",,terminal_output +732,2103323,"TERMINAL",0,0,"env/",,terminal_output +733,2103486,"TERMINAL",0,0,"[?25lb[?25h",,terminal_output +734,2103556,"TERMINAL",0,0,"in/",,terminal_output +735,2103801,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +736,2103870,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +737,2104027,"TERMINAL",0,0,"tivate",,terminal_output +738,2104041,"TERMINAL",0,0,"8",,terminal_output +739,2104208,"TERMINAL",0,0,"[?25l[?2004l\r[?25h]0;tum_cte0515@hkn0425:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +740,2105107,"TERMINAL",0,0,"9",,terminal_output +741,2106127,"TERMINAL",0,0,"20",,terminal_output +742,2106924,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +743,2107147,"TERMINAL",0,0,"1",,terminal_output +744,2107764,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1228,0,"",shellscript,selection_mouse +745,2108121,"TERMINAL",0,0,"2",,terminal_output +746,2108360,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",889,0,"",shellscript,selection_mouse +747,2108863,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",862,0,"",shellscript,selection_mouse +748,2109215,"TERMINAL",0,0,"3",,terminal_output +749,2109374,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",884,0,"",shellscript,selection_mouse +750,2110165,"TERMINAL",0,0,"4",,terminal_output +751,2110404,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1267,0,"",shellscript,selection_mouse +752,2110559,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1250,17,"t=$lam_ckpt_dir\n ",shellscript,selection_mouse +753,2110579,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1253,14,"lam_ckpt_dir\n ",shellscript,selection_mouse +754,2110594,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",1267,0,"",shellscript,selection_mouse +755,2111250,"TERMINAL",0,0,"5",,terminal_output +756,2112273,"TERMINAL",0,0,"6",,terminal_output +757,2113225,"TERMINAL",0,0,"7",,terminal_output +758,2114324,"TERMINAL",0,0,"8",,terminal_output +759,2115272,"TERMINAL",0,0,"9",,terminal_output +760,2116286,"TERMINAL",0,0,"30",,terminal_output +761,2117303,"TERMINAL",0,0,"1",,terminal_output +762,2118324,"TERMINAL",0,0,"2",,terminal_output +763,2119396,"TERMINAL",0,0,"srun",,terminal_focus +764,2119409,"TERMINAL",0,0,"3",,terminal_output +765,2120364,"TERMINAL",0,0,"4",,terminal_output +766,2120809,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0523:~/Projects/jafar[?2004h]633;A(jafar) ]633;Ajafar[tum_cte0515@hkn0523 jafar]$ ]633;B]633;B",,terminal_output +767,2121506,"TERMINAL",0,0,"[?25lgi[?25h",,terminal_output +768,2121552,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +769,2121652,"TERMINAL",0,0,"[?25lt[?25h[?25l [?25h",,terminal_output +770,2121850,"TERMINAL",0,0,"[?25lb[?25h[?25lr[?25h",,terminal_output +771,2122068,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +772,2122136,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +773,2122290,"TERMINAL",0,0,"[?25lc[?25h[?25lh[?25h",,terminal_output +774,2122472,"TERMINAL",0,0,"\r\n[?2004l\r[?1h=\r",,terminal_output +775,2122555,"TERMINAL",0,0," add-wandb-name-and-tags\r\n convert-to-jax-array-in-iter\r\n correct-batched-sampling\r\n dev\r\n dont-let-tf-see-gpu\r\n feat/explicit-image-dims\r\n fix-sampling\r\n* grain-dataloader\r\n logging-variants\r\n main\r\n preprocess_video\r\n revised-dataloader\r\n runner\r\n runner-grain\r\n speedup-tfrecord-preprocessing\r\n tmp\r\n\r[?1l>]0;tum_cte0515@hkn0523:~/Projects/jafar[?2004h]633;A(jafar) ]633;Ajafar[tum_cte0515@hkn0523 jafar]$ ]633;B]633;B",,terminal_output +776,2124123,"TERMINAL",0,0,"g",,terminal_output +777,2124192,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +778,2124260,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +779,2124313,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +780,2124481,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +781,2124823,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +782,2124942,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +783,2125015,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +784,2125081,"TERMINAL",0,0,"[?25lj[?25h",,terminal_output +785,2125234,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +786,2125359,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +787,2126415,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +788,2126605,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +789,2126671,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +790,2126817,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +791,2126884,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +792,2128987,"TERMINAL",0,0,"[?25lr[?25h",,terminal_output +793,2129099,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +794,2129366,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +795,2129559,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +796,2129613,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +797,2129709,"TERMINAL",0,0,"[?25lr[?25h",,terminal_output +798,2129773,"TERMINAL",0,0,"[?25l-[?25h",,terminal_output +799,2130089,"TERMINAL",0,0,"[?25lg[?25h",,terminal_output +800,2130143,"TERMINAL",0,0,"[?25lr[?25h",,terminal_output +801,2130333,"TERMINAL",0,0,"[?25la[?25h[?25li[?25h",,terminal_output +802,2130402,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +803,2130578,"TERMINAL",0,0,"\r\n[?2004l\rgit: 'chechout' is not a git command. See 'git --help'.\r\n\r\nThe most similar command is\r\n\tcheckout\r\n]0;tum_cte0515@hkn0523:~/Projects/jafar[?2004h]633;A(jafar) ]633;Ajafar[tum_cte0515@hkn0523 jafar]$ ]633;B]633;B",,terminal_output +804,2132360,"TERMINAL",0,0,"git chechout runner-grain",,terminal_output +805,2132472,"TERMINAL",0,0,"branch",,terminal_output +806,2132964,"TERMINAL",0,0,"chechout runner-grain",,terminal_output +807,2135341,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +808,2135870,"TERMINAL",0,0,"[1@k",,terminal_output +809,2135969,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +810,2136529,"TERMINAL",0,0,"Switched to branch 'runner-grain'\r\n]0;tum_cte0515@hkn0523:~/Projects/jafar[?2004h]633;A(jafar) ]633;Ajafar[tum_cte0515@hkn0523 jafar]$ ]633;B]633;B",,terminal_output +811,2139969,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +812,2140363,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"Switched from branch 'grain-dataloader' to 'runner-grain'",shellscript,git_branch_checkout +813,2145277,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +814,2146536,"TERMINAL",0,0,"[?25lslurm/jobs/mihir/horeka/yolo-runs/tester.sh[?25h",,terminal_output +815,2148990,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +816,2149055,"TERMINAL",0,0,"[?25lshslurm/jobs/mihir/horeka/yolo-runs/tester.sh[?25h slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +817,2149361,"TERMINAL",0,0,"\r\n[?2004l\r\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\n\r\njob_name=""debug""\r\nslurm_job_id=""debug-mihir""\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/train_lam_action_space_scaling_20/3318547/lam_1751657975_200000/\r\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\r\n\r\nenv | grep SLURM\r\n\r\nsrun python train_dynamics.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --min_lr=0 \\r\n --max_lr=1.4e-4 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --num_latent_actions=20 \\r\n --log_checkpoint_interval=1000 \\r\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\r\n --tags dynamics yolo-run tf_records \\r\n --entity instant-uv \\r\n --project jafar \\r\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\r\n --data_dir $array_records_dir\r\n # --lam_checkpoint=$lam_ckpt_dir\r\n ",,terminal_output +818,2149551,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=1\r\nSLURM_JOB_UID=999226\r\nSLURM_STEP_GPUS=0\r\nSLURM_CPU_BIND=quiet,mask_cpu:0x00000003FC00000007F00000003FC00000007F\r\nSLURM_TASK_PID=1458022\r\nSLURM_LOCALID=0\r\nSLURM_CPU_BIND_VERBOSE=quiet\r\nSLURMD_NODENAME=hkn0523\r\nSLURM_JOB_START_TIME=1752092406\r\nSLURM_STEP_NODELIST=hkn0523\r\nSLURM_JOB_END_TIME=1752178806\r\nSLURM_CPUS_ON_NODE=30\r\nSLURM_UMASK=0022\r\nSLURM_JOB_CPUS_PER_NODE=30\r\nSLURM_GPUS_ON_NODE=1\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_STEPID=0\r\nSLURM_CPU_BIND_LIST=0x00000003FC00000007F00000003FC00000007F\r\nSLURM_JOBID=3333584\r\nSLURM_PTY_PORT=43975\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=48\r\nSLURM_CPU_BIND_TYPE=mask_cpu:\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_NTASKS=1\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e11.hkn0523\r\nSLURM_DISTRIBUTION=cyclic\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=228\r\nSLURM_NODELIST=hkn0523\r\nSLURM_SRUN_COMM_PORT=42223\r\nSLURM_STEP_ID=0\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=1\r\nSLURM_NNODES=1\r\nSLURM_JOB_ID=3333584\r\nSLURM_NODEID=0\r\nSLURMD_TRES_FREQ=gpu:high,memory=high\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=bash\r\nSLURM_STEP_LAUNCHER_PORT=42223\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn0523\r\n",,terminal_output +819,2149696,"TERMINAL",0,0,"GpuFreq=control_disabled\r\n",,terminal_output +820,2152260,"TERMINAL",0,0,"srun",,terminal_focus +821,2154328,"TERMINAL",0,0,"srun",,terminal_focus +822,2155300,"TERMINAL",0,0,"srun",,terminal_focus +823,2156270,"TERMINAL",0,0,"[?25lsm[?25h[?25lm[?25h[?25li[?25h",,terminal_output +824,2156480,"TERMINAL",0,0,"[?25l[?2004l\r[?25h",,terminal_output +825,2158057,"TERMINAL",0,0,"[?1049h(B[?7hEvery 1.0s: nvidia-smihkn0425.localdomain: Thu Jul 10 16:54:10 2025Thu Jul 10 16:54:10 2025\r+-----------------------------------------------------------------------------------------+\r| NVIDIA-SMI 570.133.20Driver Version: 570.133.20 CUDA Version: 12.8 |\r|-----------------------------------------+------------------------+----------------------+\r| GPU NamePersistence-M | Bus-IdDisp.A | Volatile Uncorr. ECC |\r| Fan Temp PerfPwr:Usage/Cap |Memory-Usage | GPU-Util Compute M. |\r|||MIG M. |\r|=========================================+========================+======================|\r| 0 NVIDIA A100-SXM4-40GBOn | 00000000:31:00.0 Off |0 |\r| N/A 43C P052W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r| 1 NVIDIA A100-SXM4-40GBOn | 00000000:4B:00.0 Off |0 |\r| N/A 44C P060W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r| 2 NVIDIA A100-SXM4-40GBOn | 00000000:CA:00.0 Off |0 |\r| N/A 44C P059W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r| 3 NVIDIA A100-SXM4-40GBOn | 00000000:E3:00.0 Off |0 |\r| N/A 43C P054W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r+-----------------------------------------------------------------------------------------+\r| Processes:|\r| GPU GI CIPID Type Process nameGPU Memory |\r|ID IDUsage\t |\r|=========================================================================================|\r| 0 N/A N/A2945G /usr/libexec/Xorg17MiB |\r| 1 N/A N/A2945G /usr/libexec/Xorg17MiB |\r| 2 N/A N/A2945G /usr/libexec/Xorg17MiB |\r| 3 N/A N/A2945G /usr/libexec/Xorg17MiB |\r+-----------------------------------------------------------------------------------------+",,terminal_output +826,2159127,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0425:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +827,2160011,"TERMINAL",0,0,"srun",,terminal_focus +828,2172672,"TERMINAL",0,0,"srun",,terminal_focus +829,2173601,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +830,2173905,"TERMINAL",0,0,"2025-07-10 16:54:28.040417: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:654] The byte size of input/output arguments (47390392320) exceeds the base limit (33925523046). This indicates an error in the calculation!\r\n2025-07-10 16:54:28.040460: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:654] The byte size of input/output arguments (47390392320) exceeds the base limit (33925523046). This indicates an error in the calculation!\r\n2025-07-10 16:54:28.040560: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:654] The byte size of input/output arguments (47390392320) exceeds the base limit (33925523046). This indicates an error in the calculation!\r\n2025-07-10 16:54:28.040577: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:654] The byte size of input/output arguments (47390392320) exceeds the base limit (33925523046). This indicates an error in the calculation!\r\n2025-07-10 16:54:28.040670: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3022] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 38.75GiB (41602252800 bytes), down from 38.75GiB (41602252800 bytes) originally\r\n2025-07-10 16:54:28.040680: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3022] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 38.75GiB (41602252800 bytes), down from 38.75GiB (41602252800 bytes) originally\r\n2025-07-10 16:54:28.040805: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3022] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 38.75GiB (41602252800 bytes), down from 38.75GiB (41602252800 bytes) originally\r\n2025-07-10 16:54:28.040817: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3022] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 38.75GiB (41602252800 bytes), down from 38.75GiB (41602252800 bytes) originally\r\n2025-07-10 16:54:28.040832: E external/xla/xla/service/gpu/gpu_hlo_schedule.cc:654] The byte size of input/output arguments (47390392320) exceeds the base limit (33925523046). This indicates an error in the calculation!\r\n2025-07-10 16:54:28.041113: W external/xla/xla/hlo/transforms/simplifiers/hlo_rematerialization.cc:3022] Can't reduce memory use below 0B (0 bytes) by rematerialization; only reduced to 38.75GiB (41606447120 bytes), down from 38.75GiB (41606447120 bytes) originally\r\n",,terminal_output +831,2173979,"TERMINAL",0,0,"[?25l\rsh slurm/jobs/mihir/horeka/yolo-runs/tester.sh\r\n[?2004l\r[?25h\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\r\n\r\njob_name=""debug""\r\nslurm_job_id=""debug-mihir""\r\n\r\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/train_lam_action_space_scaling_20/3318547/lam_1751657975_200000/\r\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\r\n\r\nenv | grep SLURM\r\n\r\nsrun python train_dynamics.py \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --min_lr=0 \\r\n --max_lr=1.4e-4 \\r\n --log_image_interval=1000 \\r\n --log \\r\n --num_latent_actions=20 \\r\n --log_checkpoint_interval=1000 \\r\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\r\n --tags dynamics yolo-run tf_records \\r\n --entity instant-uv \\r\n --project jafar \\r\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\r\n --data_dir $array_records_dir\r\n # --lam_checkpoint=$lam_ckpt_dir\r\n ",,terminal_output +832,2174130,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1196654\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\r\nSLURMD_NODENAME=hkn0425\r\nSLURM_JOB_START_TIME=1752159082\r\nSLURM_STEP_NODELIST=hkn0425\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752195082\r\nSLURM_PMI2_SRUN_PORT=33587\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3335335\r\nSLURM_PTY_PORT=45071\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=48\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e13.hkn0425\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=228\r\nSLURM_NODELIST=hkn[0425,0428]\r\nSLURM_SRUN_COMM_PORT=37803\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3335335\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0425\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_STEP_LAUNCHER_PORT=37803\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0425,0428]\r\n",,terminal_output +833,2174283,"TERMINAL",0,0,"GpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +834,2175231,"TERMINAL",0,0,"srun",,terminal_focus +835,2178881,"TERMINAL",0,0,"srun",,terminal_focus +836,2184280,"TERMINAL",0,0,"2025-07-10 16:54:38.342206: W external/xla/xla/tsl/framework/bfc_allocator.cc:501] Allocator (GPU_0_bfc) ran out of memory trying to allocate 38.75GiB (rounded to 41602252800)requested by op \r\n2025-07-10 16:54:38.342328: W external/xla/xla/tsl/framework/bfc_allocator.cc:512] ******************************________*****************************_________________________________\r\nRunning on 1 devices.\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 172, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 76, in __call__\r\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/tokenizer.py"", line 57, in vq_encode\r\n x = self.encoder(x) # (B, T, N, E)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 87, in __call__\r\n x = STBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 41, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 674, in __call__\r\n x = self.attention_fn(*attn_args, **attn_kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 266, in dot_product_attention\r\n attn_weights = dot_product_attention_weights(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 132, in dot_product_attention_weights\r\n attn_weights = einsum('...qhd,...khd->...hqk', query, key)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/einsum.py"", line 315, in einsum\r\n return jit_einsum(operand_arrays, contractions, precision,\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/contextlib.py"", line 79, in inner\r\n",,terminal_output +837,2184334,"TERMINAL",0,0," return func(*args, **kwds)\r\njax._src.source_info_util.JaxStackTraceBeforeTransformation: jaxlib._jax.XlaRuntimeError: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes.\r\n\r\nThe preceding stack trace is the source of the JAX operation that, once transformed by JAX, triggered the following exception.\r\n\r\n--------------------\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_dynamics.py"", line 172, in \r\n init_params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 76, in __call__\r\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/tokenizer.py"", line 57, in vq_encode\r\n x = self.encoder(x) # (B, T, N, E)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 87, in __call__\r\n x = STBlock(\r\njaxlib._jax.XlaRuntimeError: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 41602252800 bytes.\r\n",,terminal_output +838,2185337,"TERMINAL",0,0,"srun: error: hkn0523: task 0: Exited with exit code 1\r\n]0;tum_cte0515@hkn0523:~/Projects/jafar[?2004h]633;A(jafar) ]633;Ajafar[tum_cte0515@hkn0523 jafar]$ ]633;B]633;B",,terminal_output +839,2239661,"TERMINAL",0,0,"2025-07-10 16:55:33.736977: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +840,2239891,"TERMINAL",0,0,"2025-07-10 16:55:34.077240: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +841,2240033,"TERMINAL",0,0,"2025-07-10 16:55:34.220355: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +842,2240582,"TERMINAL",0,0,"2025-07-10 16:55:34.707786: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +843,2242526,"TERMINAL",0,0,"2025-07-10 16:55:36.639745: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +844,2243060,"TERMINAL",0,0,"2025-07-10 16:55:37.128454: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-10 16:55:37.247320: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +845,2243553,"TERMINAL",0,0,"2025-07-10 16:55:37.714220: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +846,2244146,"TERMINAL",0,0,"2025-07-10 16:55:38.333436: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +847,2245400,"TERMINAL",0,0,"2025-07-10 16:55:39.469636: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-10 16:55:39.566128: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +848,2246300,"TERMINAL",0,0,"2025-07-10 16:55:40.486438: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +849,2247355,"TERMINAL",0,0,"2025-07-10 16:55:41.542271: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +850,2248470,"TERMINAL",0,0,"2025-07-10 16:55:42.657769: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +851,2248581,"TERMINAL",0,0,"2025-07-10 16:55:42.750127: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +852,2249522,"TERMINAL",0,0,"2025-07-10 16:55:43.708810: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +853,2257250,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +854,2257935,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.19.11\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250710_165551-dihg8uvq\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run dynamics-yolorun-tf-records-debug-mihir\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/dihg8uvq\r\n",,terminal_output +855,2353560,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +856,2353734,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +857,2353867,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['tokenizer', 'lam', 'dynamics']\r\nParameter counts:\r\n{'tokenizer': 37989616, 'lam': 19349920, 'dynamics': 29735424, 'total': 87074960}\r\n",,terminal_output +858,2353931,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['tokenizer', 'lam', 'dynamics']\r\nParameter counts:\r\n{'tokenizer': 37989616, 'lam': 19349920, 'dynamics': 29735424, 'total': 87074960}\r\n",,terminal_output +859,2354040,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['tokenizer', 'lam', 'dynamics']\r\nParameter counts:\r\n{'tokenizer': 37989616, 'lam': 19349920, 'dynamics': 29735424, 'total': 87074960}\r\n",,terminal_output +860,2354212,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['tokenizer', 'lam', 'dynamics']\r\nParameter counts:\r\n{'tokenizer': 37989616, 'lam': 19349920, 'dynamics': 29735424, 'total': 87074960}\r\n",,terminal_output +861,2359145,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +862,2359570,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['tokenizer', 'lam', 'dynamics']\r\nParameter counts:\r\n{'tokenizer': 37989616, 'lam': 19349920, 'dynamics': 29735424, 'total': 87074960}\r\nRunning on 8 devices.\r\nCounting all components: ['tokenizer', 'lam', 'dynamics']\r\nParameter counts:\r\n{'tokenizer': 37989616, 'lam': 19349920, 'dynamics': 29735424, 'total': 87074960}\r\n",,terminal_output +863,2361463,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +864,2361839,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['tokenizer', 'lam', 'dynamics']\r\nParameter counts:\r\n{'tokenizer': 37989616, 'lam': 19349920, 'dynamics': 29735424, 'total': 87074960}\r\n",,terminal_output +865,2363968,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +866,2364334,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['tokenizer', 'lam', 'dynamics']\r\nParameter counts:\r\n{'tokenizer': 37989616, 'lam': 19349920, 'dynamics': 29735424, 'total': 87074960}\r\n",,terminal_output +867,2377217,"TERMINAL",0,0,"srun",,terminal_focus +868,2378318,"TERMINAL",0,0,"[?25lid[?25h",,terminal_output +869,2378387,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +870,2378504,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +871,2378689,"TERMINAL",0,0,"[?25li[?25h[?25ln[?25h",,terminal_output +872,2378754,"TERMINAL",0,0,"[?25lg[?25h",,terminal_output +873,2378945,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn0523.localdomain: Thu Jul 10 16:57:53 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly:\t 2 nodes idle\rPartition dev_accelerated:\t 0 nodes idle\rPartition accelerated:\t 2 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 6 nodes idle",,terminal_output +874,2379949,"TERMINAL",0,0,"4",,terminal_output +875,2380966,"TERMINAL",0,0,"5",,terminal_output +876,2381034,"TERMINAL",0,0,"srun",,terminal_focus +877,2382000,"TERMINAL",0,0,"6",,terminal_output +878,2383009,"TERMINAL",0,0,"7",,terminal_output +879,2384031,"TERMINAL",0,0,"8",,terminal_output +880,2385070,"TERMINAL",0,0,"9",,terminal_output +881,2386100,"TERMINAL",0,0,"8:00",,terminal_output +882,2387089,"TERMINAL",0,0,"1",,terminal_output +883,2388110,"TERMINAL",0,0,"2",,terminal_output +884,2389167,"TERMINAL",0,0,"3",,terminal_output +885,2390190,"TERMINAL",0,0,"4",,terminal_output +886,2392680,"TERMINAL",0,0,"56",,terminal_output +887,2393214,"TERMINAL",0,0,"7",,terminal_output +888,2394242,"TERMINAL",0,0,"8",,terminal_output +889,2395312,"TERMINAL",0,0,"9",,terminal_output +890,2396289,"TERMINAL",0,0,"10",,terminal_output +891,2397357,"TERMINAL",0,0,"1",,terminal_output +892,2398323,"TERMINAL",0,0,"2",,terminal_output +893,2399406,"TERMINAL",0,0,"3",,terminal_output +894,2400443,"TERMINAL",0,0,"4",,terminal_output +895,2401509,"TERMINAL",0,0,"5",,terminal_output +896,2402588,"TERMINAL",0,0,"6",,terminal_output +897,2403505,"TERMINAL",0,0,"7",,terminal_output +898,2404442,"TERMINAL",0,0,"8",,terminal_output +899,2405462,"TERMINAL",0,0,"9",,terminal_output +900,2405758,"TERMINAL",0,0,"2025-07-10 16:58:19.887258: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +901,2406514,"TERMINAL",0,0,"20",,terminal_output +902,2407503,"TERMINAL",0,0,"1",,terminal_output +903,2408523,"TERMINAL",0,0,"2",,terminal_output +904,2409548,"TERMINAL",0,0,"3",,terminal_output +905,2410157,"TERMINAL",0,0,"2025-07-10 16:58:24.280770: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-10 16:58:24.281913: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +906,2410569,"TERMINAL",0,0,"4",,terminal_output +907,2411592,"TERMINAL",0,0,"5",,terminal_output +908,2412716,"TERMINAL",0,0,"6",,terminal_output +909,2413740,"TERMINAL",0,0,"7",,terminal_output +910,2414657,"TERMINAL",0,0,"8",,terminal_output +911,2415562,"TERMINAL",0,0,"2025-07-10 16:58:29.730335: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-10 16:58:29.730370: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +912,2415677,"TERMINAL",0,0,"9",,terminal_output +913,2416714,"TERMINAL",0,0,"30",,terminal_output +914,2417744,"TERMINAL",0,0,"1",,terminal_output +915,2418758,"TERMINAL",0,0,"2",,terminal_output +916,2419773,"TERMINAL",0,0,"3",,terminal_output +917,2420777,"TERMINAL",0,0,"4",,terminal_output +918,2421797,"TERMINAL",0,0,"5",,terminal_output +919,2422857,"TERMINAL",0,0,"6",,terminal_output +920,2423883,"TERMINAL",0,0,"8",,terminal_output +921,2424864,"TERMINAL",0,0,"9",,terminal_output +922,2425926,"TERMINAL",0,0,"40",,terminal_output +923,2426948,"TERMINAL",0,0,"1",,terminal_output +924,2427922,"TERMINAL",0,0,"2",,terminal_output +925,2428942,"TERMINAL",0,0,"3",,terminal_output +926,2429967,"TERMINAL",0,0,"4",,terminal_output +927,2430982,"TERMINAL",0,0,"5",,terminal_output +928,2432075,"TERMINAL",0,0,"6",,terminal_output +929,2433080,"TERMINAL",0,0,"7",,terminal_output +930,2434044,"TERMINAL",0,0,"8",,terminal_output +931,2435061,"TERMINAL",0,0,"9",,terminal_output +932,2436083,"TERMINAL",0,0,"50",,terminal_output +933,2437104,"TERMINAL",0,0,"1",,terminal_output +934,2438127,"TERMINAL",0,0,"2",,terminal_output +935,2439148,"TERMINAL",0,0,"3",,terminal_output +936,2440168,"TERMINAL",0,0,"4",,terminal_output +937,2441190,"TERMINAL",0,0,"5",,terminal_output +938,2442217,"TERMINAL",0,0,"6",,terminal_output +939,2443230,"TERMINAL",0,0,"7",,terminal_output +940,2444256,"TERMINAL",0,0,"8",,terminal_output +941,2445273,"TERMINAL",0,0,"9",,terminal_output +942,2446294,"TERMINAL",0,0,"9:00",,terminal_output +943,2447316,"TERMINAL",0,0,"1",,terminal_output +944,2448333,"TERMINAL",0,0,"2",,terminal_output +945,2449353,"TERMINAL",0,0,"3",,terminal_output +946,2450375,"TERMINAL",0,0,"4",,terminal_output +947,2451394,"TERMINAL",0,0,"5",,terminal_output +948,2452441,"TERMINAL",0,0,"6",,terminal_output +949,2453432,"TERMINAL",0,0,"7",,terminal_output +950,2454506,"TERMINAL",0,0,"8",,terminal_output +951,2455478,"TERMINAL",0,0,"9",,terminal_output +952,2456545,"TERMINAL",0,0,"10",,terminal_output +953,2457518,"TERMINAL",0,0,"1",,terminal_output +954,2458542,"TERMINAL",0,0,"2",,terminal_output +955,2459557,"TERMINAL",0,0,"3",,terminal_output +956,2460579,"TERMINAL",0,0,"4",,terminal_output +957,2461601,"TERMINAL",0,0,"5",,terminal_output +958,2462624,"TERMINAL",0,0,"6",,terminal_output +959,2463645,"TERMINAL",0,0,"7",,terminal_output +960,2464736,"TERMINAL",0,0,"8",,terminal_output +961,2465761,"TERMINAL",0,0,"9",,terminal_output +962,2466786,"TERMINAL",0,0,"20",,terminal_output +963,2467731,"TERMINAL",0,0,"1",,terminal_output +964,2468748,"TERMINAL",0,0,"2",,terminal_output +965,2469856,"TERMINAL",0,0,"3",,terminal_output +966,2470884,"TERMINAL",0,0,"4",,terminal_output +967,2471909,"TERMINAL",0,0,"5",,terminal_output +968,2472873,"TERMINAL",0,0,"6",,terminal_output +969,2473900,"TERMINAL",0,0,"8",,terminal_output +970,2474877,"TERMINAL",0,0,"9",,terminal_output +971,2475899,"TERMINAL",0,0,"30",,terminal_output +972,2476915,"TERMINAL",0,0,"1",,terminal_output +973,2477940,"TERMINAL",0,0,"2",,terminal_output +974,2478960,"TERMINAL",0,0,"3",,terminal_output +975,2479982,"TERMINAL",0,0,"4",,terminal_output +976,2481008,"TERMINAL",0,0,"5",,terminal_output +977,2482017,"TERMINAL",0,0,"6",,terminal_output +978,2483069,"TERMINAL",0,0,"7",,terminal_output +979,2484092,"TERMINAL",0,0,"8",,terminal_output +980,2485114,"TERMINAL",0,0,"9",,terminal_output +981,2486142,"TERMINAL",0,0,"40",,terminal_output +982,2487163,"TERMINAL",0,0,"1",,terminal_output +983,2488142,"TERMINAL",0,0,"2",,terminal_output +984,2489212,"TERMINAL",0,0,"3",,terminal_output +985,2490241,"TERMINAL",0,0,"4",,terminal_output +986,2491285,"TERMINAL",0,0,"5",,terminal_output +987,2492283,"TERMINAL",0,0,"6",,terminal_output +988,2493410,"TERMINAL",0,0,"7",,terminal_output +989,2494252,"TERMINAL",0,0,"8",,terminal_output +990,2495272,"TERMINAL",0,0,"9",,terminal_output +991,2496296,"TERMINAL",0,0,"50",,terminal_output +992,2497314,"TERMINAL",0,0,"1",,terminal_output +993,2498438,"TERMINAL",0,0,"2",,terminal_output +994,2499454,"TERMINAL",0,0,"3",,terminal_output +995,2500379,"TERMINAL",0,0,"4",,terminal_output +996,2501399,"TERMINAL",0,0,"5",,terminal_output +997,2502521,"TERMINAL",0,0,"6",,terminal_output +998,2503447,"TERMINAL",0,0,"7",,terminal_output +999,2504491,"TERMINAL",0,0,"8",,terminal_output +1000,2505497,"TERMINAL",0,0,"9",,terminal_output +1001,2506516,"TERMINAL",0,0,"7:00:00",,terminal_output +1002,2507521,"TERMINAL",0,0,"1",,terminal_output +1003,2508542,"TERMINAL",0,0,"2",,terminal_output +1004,2509564,"TERMINAL",0,0,"366",,terminal_output +1005,2510583,"TERMINAL",0,0,"4",,terminal_output +1006,2511604,"TERMINAL",0,0,"5",,terminal_output +1007,2512626,"TERMINAL",0,0,"6",,terminal_output +1008,2513645,"TERMINAL",0,0,"7",,terminal_output +1009,2514713,"TERMINAL",0,0,"8",,terminal_output +1010,2515690,"TERMINAL",0,0,"9",,terminal_output +1011,2516759,"TERMINAL",0,0,"10",,terminal_output +1012,2517732,"TERMINAL",0,0,"1",,terminal_output +1013,2518750,"TERMINAL",0,0,"2",,terminal_output +1014,2519773,"TERMINAL",0,0,"3",,terminal_output +1015,2520789,"TERMINAL",0,0,"4 2",,terminal_output +1016,2521812,"TERMINAL",0,0,"5",,terminal_output +1017,2522848,"TERMINAL",0,0,"6",,terminal_output +1018,2523849,"TERMINAL",0,0,"8",,terminal_output +1019,2524996,"TERMINAL",0,0,"9",,terminal_output +1020,2525977,"TERMINAL",0,0,"20",,terminal_output +1021,2526933,"TERMINAL",0,0,"1",,terminal_output +1022,2527932,"TERMINAL",0,0,"2",,terminal_output +1023,2528955,"TERMINAL",0,0,"3",,terminal_output +1024,2530069,"TERMINAL",0,0,"4",,terminal_output +1025,2531094,"TERMINAL",0,0,"5",,terminal_output +1026,2532020,"TERMINAL",0,0,"6",,terminal_output +1027,2533142,"TERMINAL",0,0,"7",,terminal_output +1028,2534163,"TERMINAL",0,0,"8",,terminal_output +1029,2535086,"TERMINAL",0,0,"9",,terminal_output +1030,2536098,"TERMINAL",0,0,"30",,terminal_output +1031,2537119,"TERMINAL",0,0,"1",,terminal_output +1032,2538140,"TERMINAL",0,0,"2",,terminal_output +1033,2539162,"TERMINAL",0,0,"3",,terminal_output +1034,2540186,"TERMINAL",0,0,"4",,terminal_output +1035,2541204,"TERMINAL",0,0,"5",,terminal_output +1036,2542226,"TERMINAL",0,0,"6",,terminal_output +1037,2543246,"TERMINAL",0,0,"70",,terminal_output +1038,2544269,"TERMINAL",0,0,"8",,terminal_output +1039,2545295,"TERMINAL",0,0,"9",,terminal_output +1040,2546310,"TERMINAL",0,0,"40",,terminal_output +1041,2547480,"TERMINAL",0,0,"11",,terminal_output +1042,2548506,"TERMINAL",0,0,"2",,terminal_output +1043,2549478,"TERMINAL",0,0,"3",,terminal_output +1044,2550458,"TERMINAL",0,0,"4",,terminal_output +1045,2551517,"TERMINAL",0,0,"5",,terminal_output +1046,2552496,"TERMINAL",0,0,"6",,terminal_output +1047,2553521,"TERMINAL",0,0,"7",,terminal_output +1048,2554539,"TERMINAL",0,0,"8",,terminal_output +1049,2555556,"TERMINAL",0,0,"9",,terminal_output +1050,2556578,"TERMINAL",0,0,"50",,terminal_output +1051,2557599,"TERMINAL",0,0,"1",,terminal_output +1052,2558617,"TERMINAL",0,0,"2",,terminal_output +1053,2559653,"TERMINAL",0,0,"3",,terminal_output +1054,2560796,"TERMINAL",0,0,"4",,terminal_output +1055,2561706,"TERMINAL",0,0,"5",,terminal_output +1056,2562700,"TERMINAL",0,0,"6",,terminal_output +1057,2563758,"TERMINAL",0,0,"7",,terminal_output +1058,2564746,"TERMINAL",0,0,"8",,terminal_output +1059,2565803,"TERMINAL",0,0,"9",,terminal_output +1060,2566833,"TERMINAL",0,0,"1:00",,terminal_output +1061,2567849,"TERMINAL",0,0,"1",,terminal_output +1062,2568827,"TERMINAL",0,0,"2",,terminal_output +1063,2569852,"TERMINAL",0,0,"4",,terminal_output +1064,2570927,"TERMINAL",0,0,"5",,terminal_output +1065,2571951,"TERMINAL",0,0,"6",,terminal_output +1066,2572912,"TERMINAL",0,0,"7",,terminal_output +1067,2573931,"TERMINAL",0,0,"8",,terminal_output +1068,2574950,"TERMINAL",0,0,"9",,terminal_output +1069,2576054,"TERMINAL",0,0,"10",,terminal_output +1070,2577009,"TERMINAL",0,0,"1",,terminal_output +1071,2578016,"TERMINAL",0,0,"2",,terminal_output +1072,2579045,"TERMINAL",0,0,"3",,terminal_output +1073,2580052,"TERMINAL",0,0,"4",,terminal_output +1074,2581169,"TERMINAL",0,0,"5",,terminal_output +1075,2582190,"TERMINAL",0,0,"6",,terminal_output +1076,2583118,"TERMINAL",0,0,"7",,terminal_output +1077,2584136,"TERMINAL",0,0,"8",,terminal_output +1078,2585160,"TERMINAL",0,0,"9",,terminal_output +1079,2586300,"TERMINAL",0,0,"20",,terminal_output +1080,2587213,"TERMINAL",0,0,"1",,terminal_output +1081,2588210,"TERMINAL",0,0,"2",,terminal_output +1082,2589229,"TERMINAL",0,0,"3",,terminal_output +1083,2590251,"TERMINAL",0,0,"4",,terminal_output +1084,2591286,"TERMINAL",0,0,"5",,terminal_output +1085,2592297,"TERMINAL",0,0,"6",,terminal_output +1086,2593350,"TERMINAL",0,0,"7",,terminal_output +1087,2594341,"TERMINAL",0,0,"8",,terminal_output +1088,2595364,"TERMINAL",0,0,"9",,terminal_output +1089,2596384,"TERMINAL",0,0,"302",,terminal_output +1090,2597448,"TERMINAL",0,0,"1",,terminal_output +1091,2598483,"TERMINAL",0,0,"2",,terminal_output +1092,2599447,"TERMINAL",0,0,"3",,terminal_output +1093,2600527,"TERMINAL",0,0,"4",,terminal_output +1094,2601519,"TERMINAL",0,0,"5",,terminal_output +1095,2602571,"TERMINAL",0,0,"61",,terminal_output +1096,2603540,"TERMINAL",0,0,"7",,terminal_output +1097,2604554,"TERMINAL",0,0,"8",,terminal_output +1098,2605585,"TERMINAL",0,0,"9",,terminal_output +1099,2606632,"TERMINAL",0,0,"40",,terminal_output +1100,2607653,"TERMINAL",0,0,"1",,terminal_output +1101,2608675,"TERMINAL",0,0,"2",,terminal_output +1102,2609739,"TERMINAL",0,0,"3",,terminal_output +1103,2610762,"TERMINAL",0,0,"4",,terminal_output +1104,2611786,"TERMINAL",0,0,"5",,terminal_output +1105,2612809,"TERMINAL",0,0,"6",,terminal_output +1106,2613776,"TERMINAL",0,0,"7",,terminal_output +1107,2614803,"TERMINAL",0,0,"8",,terminal_output +1108,2615825,"TERMINAL",0,0,"9",,terminal_output +1109,2616904,"TERMINAL",0,0,"51",,terminal_output +1110,2617869,"TERMINAL",0,0,"2",,terminal_output +1111,2618883,"TERMINAL",0,0,"3",,terminal_output +1112,2619926,"TERMINAL",0,0,"4",,terminal_output +1113,2620921,"TERMINAL",0,0,"5",,terminal_output +1114,2622025,"TERMINAL",0,0,"6",,terminal_output +1115,2623045,"TERMINAL",0,0,"7",,terminal_output +1116,2624005,"TERMINAL",0,0,"8",,terminal_output +1117,2625001,"TERMINAL",0,0,"9",,terminal_output +1118,2626119,"TERMINAL",0,0,"2:00",,terminal_output +1119,2627043,"TERMINAL",0,0,"1",,terminal_output +1120,2628060,"TERMINAL",0,0,"2",,terminal_output +1121,2629196,"TERMINAL",0,0,"3",,terminal_output +1122,2630221,"TERMINAL",0,0,"466",,terminal_output +1123,2631272,"TERMINAL",0,0,"5",,terminal_output +1124,2632268,"TERMINAL",0,0,"6",,terminal_output +1125,2633257,"TERMINAL",0,0,"7",,terminal_output +1126,2634316,"TERMINAL",0,0,"8",,terminal_output +1127,2635297,"TERMINAL",0,0,"92",,terminal_output +1128,2636421,"TERMINAL",0,0,"10",,terminal_output +1129,2637389,"TERMINAL",0,0,"1",,terminal_output +1130,2638454,"TERMINAL",0,0,"2",,terminal_output +1131,2639434,"TERMINAL",0,0,"3",,terminal_output +1132,2640440,"TERMINAL",0,0,"4",,terminal_output +1133,2641481,"TERMINAL",0,0,"5 2",,terminal_output +1134,2642508,"TERMINAL",0,0,"6",,terminal_output +1135,2643535,"TERMINAL",0,0,"7",,terminal_output +1136,2644552,"TERMINAL",0,0,"8",,terminal_output +1137,2645506,"TERMINAL",0,0,"9",,terminal_output +1138,2646908,"TERMINAL",0,0,"20",,terminal_output +1139,2647554,"TERMINAL",0,0,"1",,terminal_output +1140,2648680,"TERMINAL",0,0,"2",,terminal_output +1141,2649593,"TERMINAL",0,0,"3",,terminal_output +1142,2650621,"TERMINAL",0,0,"4",,terminal_output +1143,2651726,"TERMINAL",0,0,"5",,terminal_output +1144,2652663,"TERMINAL",0,0,"6",,terminal_output +1145,2653710,"TERMINAL",0,0,"7",,terminal_output +1146,2654757,"TERMINAL",0,0,"8",,terminal_output +1147,2655705,"TERMINAL",0,0,"9",,terminal_output +1148,2656739,"TERMINAL",0,0,"30",,terminal_output +1149,2657769,"TERMINAL",0,0,"1",,terminal_output +1150,2658888,"TERMINAL",0,0,"2",,terminal_output +1151,2659812,"TERMINAL",0,0,"3",,terminal_output +1152,2660939,"TERMINAL",0,0,"4",,terminal_output +1153,2662065,"TERMINAL",0,0,"5",,terminal_output +1154,2662889,"TERMINAL",0,0,"7",,terminal_output +1155,2664007,"TERMINAL",0,0,"8",,terminal_output +1156,2664931,"TERMINAL",0,0,"9",,terminal_output +1157,2665941,"TERMINAL",0,0,"40",,terminal_output +1158,2666983,"TERMINAL",0,0,"1",,terminal_output +1159,2667985,"TERMINAL",0,0,"2",,terminal_output +1160,2669029,"TERMINAL",0,0,"3",,terminal_output +1161,2670050,"TERMINAL",0,0,"4",,terminal_output +1162,2671074,"TERMINAL",0,0,"5",,terminal_output +1163,2672100,"TERMINAL",0,0,"6",,terminal_output +1164,2673123,"TERMINAL",0,0,"7",,terminal_output +1165,2674149,"TERMINAL",0,0,"8",,terminal_output +1166,2675274,"TERMINAL",0,0,"9",,terminal_output +1167,2676172,"TERMINAL",0,0,"50",,terminal_output +1168,2677222,"TERMINAL",0,0,"1",,terminal_output +1169,2678242,"TERMINAL",0,0,"2",,terminal_output +1170,2679217,"TERMINAL",0,0,"3",,terminal_output +1171,2680236,"TERMINAL",0,0,"4",,terminal_output +1172,2681273,"TERMINAL",0,0,"5",,terminal_output +1173,2682295,"TERMINAL",0,0,"6",,terminal_output +1174,2683365,"TERMINAL",0,0,"7",,terminal_output +1175,2684320,"TERMINAL",0,0,"8",,terminal_output +1176,2685417,"TERMINAL",0,0,"9",,terminal_output +1177,2686360,"TERMINAL",0,0,"3:00",,terminal_output +1178,2687382,"TERMINAL",0,0,"1",,terminal_output +1179,2688496,"TERMINAL",0,0,"2",,terminal_output +1180,2689432,"TERMINAL",0,0,"3",,terminal_output +1181,2690442,"TERMINAL",0,0,"4",,terminal_output +1182,2691547,"TERMINAL",0,0,"5",,terminal_output +1183,2692579,"TERMINAL",0,0,"6",,terminal_output +1184,2693506,"TERMINAL",0,0,"7",,terminal_output +1185,2694526,"TERMINAL",0,0,"8",,terminal_output +1186,2695555,"TERMINAL",0,0,"9",,terminal_output +1187,2696565,"TERMINAL",0,0,"10",,terminal_output +1188,2697583,"TERMINAL",0,0,"1",,terminal_output +1189,2698605,"TERMINAL",0,0,"2",,terminal_output +1190,2699622,"TERMINAL",0,0,"3",,terminal_output +1191,2700646,"TERMINAL",0,0,"4",,terminal_output +1192,2701695,"TERMINAL",0,0,"5",,terminal_output +1193,2702685,"TERMINAL",0,0,"6",,terminal_output +1194,2703788,"TERMINAL",0,0,"7",,terminal_output +1195,2704789,"TERMINAL",0,0,"8",,terminal_output +1196,2705742,"TERMINAL",0,0,"9",,terminal_output +1197,2706813,"TERMINAL",0,0,"20",,terminal_output +1198,2707786,"TERMINAL",0,0,"1",,terminal_output +1199,2708859,"TERMINAL",0,0,"2",,terminal_output +1200,2709825,"TERMINAL",0,0,"3",,terminal_output +1201,2710908,"TERMINAL",0,0,"5",,terminal_output +1202,2711863,"TERMINAL",0,0,"6",,terminal_output +1203,2712884,"TERMINAL",0,0,"73",,terminal_output +1204,2713449,"TERMINAL",0,0,"srun",,terminal_focus +1205,2713909,"TERMINAL",0,0,"81",,terminal_output +1206,2714930,"TERMINAL",0,0,"9",,terminal_output +1207,2715951,"TERMINAL",0,0,"30",,terminal_output +1208,2716216,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n# tf_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/open_ai_minecraft_tfrecord\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/train_lam_action_space_scaling_20/3318547/lam_1751657975_200000/\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n # --lam_checkpoint=$lam_ckpt_dir\n",shellscript,tab +1209,2716971,"TERMINAL",0,0,"1",,terminal_output +1210,2717273,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1372,0,"",shellscript,selection_mouse +1211,2717914,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1371,0,"",shellscript,selection_mouse +1212,2717989,"TERMINAL",0,0,"2",,terminal_output +1213,2719102,"TERMINAL",0,0,"3",,terminal_output +1214,2719145,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1411,0,"",shellscript,selection_mouse +1215,2719847,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1410,0,"",shellscript,selection_mouse +1216,2720035,"TERMINAL",0,0,"4",,terminal_output +1217,2721152,"TERMINAL",0,0,"5",,terminal_output +1218,2721783,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",84,0,"",shellscript,selection_mouse +1219,2722172,"TERMINAL",0,0,"6",,terminal_output +1220,2722947,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",83,1,"",shellscript,content +1221,2723080,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",82,1,"",shellscript,content +1222,2723100,"TERMINAL",0,0,"71",,terminal_output +1223,2724125,"TERMINAL",0,0,"8",,terminal_output +1224,2724162,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",82,0,"0",shellscript,content +1225,2724162,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",83,0,"",shellscript,selection_keyboard +1226,2724268,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",83,0,"0",shellscript,content +1227,2724268,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",84,0,"",shellscript,selection_keyboard +1228,2724544,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",85,0,"",shellscript,selection_command +1229,2724722,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",86,0,"",shellscript,selection_command +1230,2725135,"TERMINAL",0,0,"9",,terminal_output +1231,2726000,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",85,1,"",shellscript,content +1232,2726009,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",85,0,"3",shellscript,content +1233,2726010,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",86,0,"",shellscript,selection_keyboard +1234,2726155,"TERMINAL",0,0,"40",,terminal_output +1235,2727184,"TERMINAL",0,0,"1",,terminal_output +1236,2728213,"TERMINAL",0,0,"2",,terminal_output +1237,2728454,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",85,1,"",shellscript,content +1238,2729169,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",85,0,"0",shellscript,content +1239,2729170,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",86,0,"",shellscript,selection_keyboard +1240,2729228,"TERMINAL",0,0,"3",,terminal_output +1241,2729355,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",85,0,"",shellscript,selection_command +1242,2729866,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",84,0,"",shellscript,selection_command +1243,2730015,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",83,1,"",shellscript,content +1244,2730078,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",83,0,"1",shellscript,content +1245,2730078,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",84,0,"",shellscript,selection_keyboard +1246,2730238,"TERMINAL",0,0,"4",,terminal_output +1247,2731277,"TERMINAL",0,0,"5",,terminal_output +1248,2732276,"TERMINAL",0,0,"6",,terminal_output +1249,2732697,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1406,0,"",shellscript,selection_mouse +1250,2733147,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1405,1,"",shellscript,content +1251,2733271,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1404,1,"",shellscript,content +1252,2733330,"TERMINAL",0,0,"7",,terminal_output +1253,2733562,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1403,1,"",shellscript,content +1254,2733689,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1403,0,"4",shellscript,content +1255,2733690,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1404,0,"",shellscript,selection_keyboard +1256,2734319,"TERMINAL",0,0,"8",,terminal_output +1257,2734570,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1403,1,"",shellscript,content +1258,2734627,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1403,0,"3",shellscript,content +1259,2734627,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1404,0,"",shellscript,selection_keyboard +1260,2735384,"TERMINAL",0,0,"9",,terminal_output +1261,2736072,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1406,1,"",shellscript,content +1262,2736088,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1406,0,"5",shellscript,content +1263,2736088,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1407,0,"",shellscript,selection_keyboard +1264,2736358,"TERMINAL",0,0,"50",,terminal_output +1265,2737380,"TERMINAL",0,0,"1",,terminal_output +1266,2737453,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1453,0,"",shellscript,selection_mouse +1267,2737708,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1438,15,"0 \\n --log \",shellscript,selection_mouse +1268,2737724,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1437,16,"00 \\n --log \",shellscript,selection_mouse +1269,2737800,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1436,17,"000 \\n --log \",shellscript,selection_mouse +1270,2738092,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1436,0,"",shellscript,selection_mouse +1271,2738390,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1478,0,"",shellscript,selection_mouse +1272,2738431,"TERMINAL",0,0,"2",,terminal_output +1273,2738959,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1502,0,"",shellscript,selection_mouse +1274,2739478,"TERMINAL",0,0,"3",,terminal_output +1275,2739534,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1453,0,"",shellscript,selection_mouse +1276,2740025,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1549,0,"",shellscript,selection_mouse +1277,2740438,"TERMINAL",0,0,"4",,terminal_output +1278,2741458,"TERMINAL",0,0,"5",,terminal_output +1279,2742174,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=""debug""\nslurm_job_id=""debug-mihir""\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/train_lam_action_space_scaling_20/3318547/lam_1751657975_200000/\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n # --lam_checkpoint=$lam_ckpt_dir\n ",shellscript,tab +1280,2742530,"TERMINAL",0,0,"6",,terminal_output +1281,2743513,"TERMINAL",0,0,"7",,terminal_output +1282,2744602,"TERMINAL",0,0,"8",,terminal_output +1283,2745623,"TERMINAL",0,0,"9",,terminal_output +1284,2746562,"TERMINAL",0,0,"4:00",,terminal_output +1285,2747622,"TERMINAL",0,0,"1",,terminal_output +1286,2748612,"TERMINAL",0,0,"2",,terminal_output +1287,2748869,"TERMINAL",0,0,"srun",,terminal_focus +1288,2749621,"TERMINAL",0,0,"3",,terminal_output +1289,2750366,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"",shellscript,tab +1290,2750648,"TERMINAL",0,0,"4",,terminal_output +1291,2751765,"TERMINAL",0,0,"52",,terminal_output +1292,2752230,"train_tokenizer.py",0,0,"from dataclasses import dataclass, field\nimport os\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n min_lr: float = 0.0\n max_lr: float = 3e-4\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 8\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n\n\nargs = tyro.cli(Args)\n\n\ndef tokenizer_loss_fn(params, state, inputs):\n # --- Compute loss ---\n outputs = state.apply_fn(\n params,\n inputs,\n training=True,\n rngs={""params"": inputs[""rng""], ""dropout"": inputs[""dropout_rng""]},\n )\n mse = jnp.square(inputs[""videos""] - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = inputs[""videos""].clip(0, 1).reshape(-1, *inputs[""videos""].shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = pix.psnr(gt, recon).mean() # type: ignore\n ssim = pix.ssim(gt, recon).mean() # type: ignore\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n grad_fn = jax.value_and_grad(tokenizer_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n\n # --- Initialize model ---\n tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_height, args.image_width, args.image_channels)\n inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=jnp.float32,\n ),\n )\n init_params = tokenizer.init(_rng, inputs)\n\n param_counts = count_parameters_by_component(init_params)\n\n if args.log and jax.process_index() == 0:\n wandb.init(\n entity=args.entity,\n project=args.project,\n name=args.name,\n tags=args.tags,\n group=""debug"",\n config=args,\n )\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=tokenizer.apply, params=init_params, tx=tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(\n mesh, PartitionSpec(""data"", None, None, None, None)\n )\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add('model_state', ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler)\n handler_registry.add('model_state', ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler)\n handler_registry.add('dataloader_state', grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add('dataloader_state', grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler) # type: ignore\n \n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n \n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n \n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_train_state = jax.tree_util.tree_map(ocp.utils.to_shape_dtype_struct, train_state)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_train_state),\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator),\n )\n )\n train_state = restored[""model_state""]\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader = (jax.make_array_from_process_local_data(videos_sharding, elem) for elem in grain_iterator) # type: ignore\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _rng_dropout = jax.random.split(rng, 3)\n\n inputs = dict(videos=videos, rng=_rng, dropout_rng=_rng_dropout)\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.StandardSave(train_state),\n dataloader_state=grain.checkpoint.CheckpointSave(grain_iterator),\n )\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()",python,tab +1293,2752342,"train_tokenizer.py",284,10832,"\nimport orbax\nfrom orbax.checkpoint import PyTreeCheckpointer\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n checkpoint: str = """"\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n min_lr: float = 0.0\n max_lr: float = 3e-4\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 8\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_gradients: bool = False\n\n\nargs = tyro.cli(Args)\n\n\ndef tokenizer_loss_fn(params, state, inputs):\n # --- Compute loss ---\n outputs = state.apply_fn(\n params,\n inputs,\n training=True,\n rngs={""params"": inputs[""rng""], ""dropout"": inputs[""dropout_rng""]},\n )\n mse = jnp.square(inputs[""videos""] - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = inputs[""videos""].clip(0, 1).reshape(-1, *inputs[""videos""].shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = pix.psnr(gt, recon).mean()\n ssim = pix.ssim(gt, recon).mean()\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n grad_fn = jax.value_and_grad(tokenizer_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n\n # --- Initialize model ---\n tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_height, args.image_width, args.image_channels)\n inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=jnp.float32,\n ),\n )\n init_params = tokenizer.init(_rng, inputs)\n\n param_counts = count_parameters_by_component(init_params)\n\n if args.log and jax.process_index() == 0:\n wandb.init(\n entity=args.entity,\n project=args.project,\n name=args.name,\n tags=args.tags,\n group=""debug"",\n config=args,\n )\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=tokenizer.apply, params=init_params, tx=tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(\n mesh, PartitionSpec(""data"", None, None, None, None)\n )\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Load checkpoint ---\n step = 0\n if args.checkpoint:\n restore_target = {""model"": train_state}\n restore_args = orbax_utils.restore_args_from_target(restore_target)\n train_state.params[""params""].update(\n PyTreeCheckpointer()\n .restore(args.checkpoint, item=restore_target, restore_args=restore_args)[\n ""model""\n ]\n .params[""params""]\n )\n # Assume checkpoint is of the form tokenizer__\n step += int(args.checkpoint.split(""_"")[-1])\n\n # --- TRAIN LOOP ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n dataloader = get_dataloader(\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n array_record_files,\n args.seq_len,\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n dataloader = (jax.make_array_from_process_local_data(videos_sharding, elem) for elem in dataloader) # type: ignore\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _rng_dropout = jax.random.split(rng, 3)\n\n inputs = dict(videos=videos, rng=_rng, dropout_rng=_rng_dropout)\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n if step % args.log_checkpoint_interval == 0:\n ckpt = {""model"": train_state}\n orbax_checkpointer = orbax.checkpoint.PyTreeCheckpointer()\n save_args = orbax_utils.save_args_from_target(ckpt)\n orbax_checkpointer.save(\n os.path.join(os.getcwd(), args.ckpt_dir, f""tokenizer_{step}""),\n ckpt,\n save_args=save_args,\n )\n if step >= args.num_steps:\n break\n",python,content +1294,2752701,"TERMINAL",0,0,"6",,terminal_output +1295,2753159,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +1296,2753745,"TERMINAL",0,0,"srun",,terminal_focus +1297,2753760,"TERMINAL",0,0,"7",,terminal_output +1298,2754736,"TERMINAL",0,0,"8",,terminal_output +1299,2755759,"TERMINAL",0,0,"9",,terminal_output +1300,2756790,"TERMINAL",0,0,"10",,terminal_output +1301,2757347,"slurm/jobs/mihir/horeka/sbatch_dir.sh",0,0,"#!/usr/bin/env bash\n\n# Check if directory argument is provided\nif [ $# -lt 1 ]; then\n echo ""Usage: $0 ""\n exit 1\nfi\n\nDIR=""$1""\n\n# Check if the directory exists\nif [ ! -d ""$DIR"" ]; then\n echo ""Error: Directory '$DIR' does not exist.""\n exit 1\nfi\n\necho ""The following files will be sbatched from directory: $DIR""\nls ""$DIR""\n\nread -p ""Proceed to sbatch all files in $DIR? [y/N] "" confirm\nconfirm=${confirm,,} # tolower\n\nif [[ ""$confirm"" != ""y"" ]]; then\n echo ""Aborted.""\n exit 0\nfi\n\nfor file in ""$DIR""/*; do\n if [ -f ""$file"" ]; then\n echo ""Submitting $file""\n sbatch ""$file""\n fi\ndone\n",shellscript,tab +1302,2757784,"TERMINAL",0,0,"1",,terminal_output +1303,2758576,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"",shellscript,tab +1304,2758813,"TERMINAL",0,0,"2",,terminal_output +1305,2759857,"TERMINAL",0,0,"3",,terminal_output +1306,2760850,"TERMINAL",0,0,"5",,terminal_output +1307,2761867,"TERMINAL",0,0,"6",,terminal_output +1308,2762884,"TERMINAL",0,0,"7",,terminal_output +1309,2763708,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0523:~/Projects/jafar[?2004h]633;A(jafar) ]633;Ajafar[tum_cte0515@hkn0523 jafar]$ ]633;B]633;B",,terminal_output +1310,2764078,"TERMINAL",0,0,"s",,terminal_output +1311,2764194,"TERMINAL",0,0,"[?25lb[?25h",,terminal_output +1312,2764272,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1313,2764412,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +1314,2764467,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +1315,2764539,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +1316,2764703,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +1317,2764950,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",,terminal_output +1318,2766824,"TERMINAL",0,0,"\rslurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch\r\n[?2004l\rSubmitted batch job 3335345\r\n]0;tum_cte0515@hkn0523:~/Projects/jafar[?2004h]633;A(jafar) ]633;Ajafar[tum_cte0515@hkn0523 jafar]$ ]633;B]633;B",,terminal_output +1319,2769458,"TERMINAL",0,0,"q",,terminal_output +1320,2769647,"TERMINAL",0,0,"[?25lu[?25h[?25le[?25h",,terminal_output +1321,2769773,"TERMINAL",0,0,"[?25lu[?25h[?25le[?25h",,terminal_output +1322,2769914,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: squeue --mehkn0523.localdomain: Thu Jul 10 17:04:24 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3335345 accelerat train_dy tum_cte0 PD\t0:00\t 2 (None)3335335 accelerat interact tum_cte0 R13:02\t 2 hkn[0425,0428]3333584 accelerat interact tum_cte0 R 18:44:18\t 1 hkn0523",,terminal_output +1323,2770905,"TERMINAL",0,0,"539",,terminal_output +1324,2771913,"TERMINAL",0,0,"6420",,terminal_output +1325,2772929,"TERMINAL",0,0,"751",,terminal_output +1326,2773986,"TERMINAL",0,0,"862",,terminal_output +1327,2775011,"TERMINAL",0,0,"973",,terminal_output +1328,2775440,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0523:~/Projects/jafar[?2004h]633;A(jafar) ]633;Ajafar[tum_cte0515@hkn0523 jafar]$ ]633;B]633;B",,terminal_output +1329,2778309,"TERMINAL",0,0,"s",,terminal_output +1330,2778543,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +1331,2778658,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1332,2778784,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +1333,2778967,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1334,2780207,"TERMINAL",0,0,"[?2004l\r\r\nexit\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +1335,2782441,"TERMINAL",0,0,"scancel 3333584",,terminal_command +1336,2782506,"TERMINAL",0,0,"]633;E;2025-07-10 17:04:36 scancel 3333584;16c08be6-3885-420f-ac53-f5272aed6e54]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +1337,2783886,"TERMINAL",0,0,"queue",,terminal_command +1338,2783947,"TERMINAL",0,0,"]633;E;2025-07-10 17:04:38 queue;16c08be6-3885-420f-ac53-f5272aed6e54]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Jul 10 17:04:38 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3333584 accelerat interact tum_cte0 CG 18:44:30\t 1 hkn05233335345 accelerat train_dy tum_cte0 R\t0:01\t 2 hkn[0430-0431]3335335 accelerat interact tum_cte0 R13:16\t 2 hkn[0425,0428]",,terminal_output +1339,2785049,"TERMINAL",0,0,"927",,terminal_output +1340,2786014,"TERMINAL",0,0,"4038",,terminal_output +1341,2787051,"TERMINAL",0,0,"149",,terminal_output +1342,2787676,"TERMINAL",0,0,"srun",,terminal_focus +1343,2788343,"TERMINAL",0,0,"2520",,terminal_output +1344,2789361,"TERMINAL",0,0,"361",,terminal_output +1345,2790366,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"",shellscript,tab +1346,2790440,"TERMINAL",0,0,"472",,terminal_output +1347,2791471,"TERMINAL",0,0,"583",,terminal_output +1348,2792534,"TERMINAL",0,0,"694",,terminal_output +1349,2793543,"TERMINAL",0,0,"7105",,terminal_output +1350,2794555,"TERMINAL",0,0,"816",,terminal_output +1351,2795600,"TERMINAL",0,0,"927",,terminal_output +1352,2796624,"TERMINAL",0,0,"5038",,terminal_output +1353,2797666,"TERMINAL",0,0,"149",,terminal_output +1354,2798702,"TERMINAL",0,0,"2530",,terminal_output +1355,2799745,"TERMINAL",0,0,"361",,terminal_output +1356,2800812,"TERMINAL",0,0,"472",,terminal_output +1357,2801839,"TERMINAL",0,0,"594",,terminal_output +1358,2802908,"TERMINAL",0,0,"7205",,terminal_output +1359,2803989,"TERMINAL",0,0,"816",,terminal_output +1360,2805013,"TERMINAL",0,0,"927",,terminal_output +1361,2805993,"TERMINAL",0,0,"5:0038",,terminal_output +1362,2807019,"TERMINAL",0,0,"149",,terminal_output +1363,2808084,"TERMINAL",0,0,"2540",,terminal_output +1364,2809112,"TERMINAL",0,0,"361",,terminal_output +1365,2810134,"TERMINAL",0,0,"45345train_dy 0:27\t 2 hkn[0430-0431]3584interactCG 18:44:30\t 1 hkn05232",,terminal_output +1366,2811297,"TERMINAL",0,0,"53",,terminal_output +1367,2812286,"TERMINAL",0,0,"64",,terminal_output +1368,2813231,"TERMINAL",0,0,"75",,terminal_output +1369,2814336,"TERMINAL",0,0,"86",,terminal_output +1370,2815358,"TERMINAL",0,0,"97",,terminal_output +1371,2816349,"TERMINAL",0,0,"108",,terminal_output +1372,2817405,"TERMINAL",0,0,"19",,terminal_output +1373,2818423,"TERMINAL",0,0,"250",,terminal_output +1374,2819461,"TERMINAL",0,0,"31",,terminal_output +1375,2820579,"TERMINAL",0,0,"\r42",,terminal_output +1376,2821522,"TERMINAL",0,0,"53",,terminal_output +1377,2822627,"TERMINAL",0,0,"64",,terminal_output +1378,2823604,"TERMINAL",0,0,"\r75335 R 13:55\t 2 hkn[0425,0428]",,terminal_output +1379,2824646,"TERMINAL",0,0,"86",,terminal_output +1380,2825676,"TERMINAL",0,0,"97",,terminal_output +1381,2826829,"TERMINAL",0,0,"208",,terminal_output +1382,2827751,"TERMINAL",0,0,"19",,terminal_output +1383,2828876,"TERMINAL",0,0,"24:00",,terminal_output +1384,2829817,"TERMINAL",0,0,"32",,terminal_output +1385,2830925,"TERMINAL",0,0,"53",,terminal_output +1386,2831948,"TERMINAL",0,0,"64",,terminal_output +1387,2832923,"TERMINAL",0,0,"75",,terminal_output +1388,2833994,"TERMINAL",0,0,"86",,terminal_output +1389,2835001,"TERMINAL",0,0,"97",,terminal_output +1390,2836049,"TERMINAL",0,0,"308",,terminal_output +1391,2837073,"TERMINAL",0,0,"19",,terminal_output +1392,2838195,"TERMINAL",0,0,"210",,terminal_output +1393,2839164,"TERMINAL",0,0,"31",,terminal_output +1394,2840196,"TERMINAL",0,0,"42",,terminal_output +1395,2840881,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1786,0,"",shellscript,selection_mouse +1396,2841294,"TERMINAL",0,0,"53",,terminal_output +1397,2842286,"TERMINAL",0,0,"64",,terminal_output +1398,2842806,"TERMINAL",0,0,"watch",,terminal_focus +1399,2843323,"TERMINAL",0,0,"75",,terminal_output +1400,2844440,"TERMINAL",0,0,"86",,terminal_output +1401,2845473,"TERMINAL",0,0,"97",,terminal_output +1402,2846411,"TERMINAL",0,0,"408",,terminal_output +1403,2847520,"TERMINAL",0,0,"srun",,terminal_focus +1404,2847534,"TERMINAL",0,0,"19",,terminal_output +1405,2848503,"TERMINAL",0,0,"220",,terminal_output +1406,2848977,"TERMINAL",0,0,"watch",,terminal_focus +1407,2849534,"TERMINAL",0,0,"31",,terminal_output +1408,2850552,"TERMINAL",0,0,"42",,terminal_output +1409,2851939,"slurm/jobs/mihir/horeka/sbatch_dir.sh",0,0,"",shellscript,tab +1410,2852431,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +1411,2852942,"train_tokenizer.py",0,0,"",python,tab +1412,2856012,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +1413,2857006,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"",shellscript,tab +1414,2863247,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",169,0,"",shellscript,selection_mouse +1415,2864028,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",82,0,"",shellscript,selection_mouse +1416,2864171,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",82,2,"01",shellscript,selection_mouse +1417,2864362,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",82,3,"01:",shellscript,selection_mouse +1418,2864369,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",82,5,"01:00",shellscript,selection_mouse +1419,2864386,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",82,6,"01:00:",shellscript,selection_mouse +1420,2864441,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",82,8,"01:00:00",shellscript,selection_mouse +1421,2864615,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",90,0,"",shellscript,selection_mouse +1422,2865541,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",147,0,"",shellscript,selection_mouse +1423,2865900,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",147,1,"5",shellscript,selection_mouse +1424,2866047,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",123,26,"#SBATCH --cpus-per-task=5\n",shellscript,selection_mouse +1425,2866723,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",169,0,"",shellscript,selection_mouse +1426,2873897,"TERMINAL",0,0,"cd $ws_dir",,terminal_command +1427,2874479,"TERMINAL",0,0,"ls",,terminal_command +1428,2874492,"TERMINAL",0,0,"]633;E;2025-07-10 17:06:08 ls;16c08be6-3885-420f-ac53-f5272aed6e54]633;Ccheckpoints count_items.sh data data_new huggingface logs scripts\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared]633;D;0",,terminal_output +1429,2877829,"TERMINAL",0,0,"cd logs/logs_mihir/",,terminal_command +1430,2878493,"TERMINAL",0,0,"ls",,terminal_command +1431,2878546,"TERMINAL",0,0,"]633;E;2025-07-10 17:06:12 ls;16c08be6-3885-420f-ac53-f5272aed6e54]633;Ctrain_dyn_yolorun_3333026.log train_lam_action_space_scaling_20_3331285.log train_lam_model_size_scaling_38M_3317115.log train_tokenizer_model_size_scaling_140M_3316019.log\r\ntrain_dyn_yolorun_3333448.log train_lam_action_space_scaling_50_3320180.log train_lam_model_size_scaling_38M_3317231.log train_tokenizer_model_size_scaling_200M_3313563.log\r\ntrain_dyn_yolorun_3335345.log train_lam_action_space_scaling_50_3329789.log train_tokenizer_batch_size_scaling_16_node_3321526.log train_tokenizer_model_size_scaling_200M_3316020.log\r\ntrain_lam_action_space_scaling_10_3320179.log train_lam_action_space_scaling_50_3329804.log train_tokenizer_batch_size_scaling_1_node_3318551.log train_tokenizer_model_size_scaling_227M_3317234.log\r\ntrain_lam_action_space_scaling_10_3321529.log train_lam_action_space_scaling_50_3331286.log train_tokenizer_batch_size_scaling_2_node_3318552.log train_tokenizer_model_size_scaling_227M_3318555.log\r\ntrain_lam_action_space_scaling_10_3329786.log train_lam_action_space_scaling_6_3318549.log train_tokenizer_batch_size_scaling_2_node_3330806.log train_tokenizer_model_size_scaling_227M_3320173.log\r\ntrain_lam_action_space_scaling_10_3329801.log train_lam_action_space_scaling_6_3320178.log train_tokenizer_batch_size_scaling_2_node_3330848.log train_tokenizer_model_size_scaling_227M_3321523.log\r\ntrain_lam_action_space_scaling_10_3331283.log train_lam_action_space_scaling_6_3321528.log train_tokenizer_batch_size_scaling_2_node_3331282.log train_tokenizer_model_size_scaling_37M_3313565.log\r\ntrain_lam_action_space_scaling_12_3318546.log train_lam_action_space_scaling_6_3329790.log train_tokenizer_batch_size_scaling_4_node_3318553.log train_tokenizer_model_size_scaling_37M_3316022.log\r\ntrain_lam_action_space_scaling_12_3320177.log train_lam_action_space_scaling_6_3329805.log train_tokenizer_batch_size_scaling_4_node_3320175.log train_tokenizer_model_size_scaling_37M_3317232.log\r\ntrain_lam_action_space_scaling_12_3321527.log train_lam_action_space_scaling_6_3331287.log train_tokenizer_batch_size_scaling_4_node_3321524.log train_tokenizer_model_size_scaling_37M_3317239.log\r\ntrain_lam_action_space_scaling_12_3329787.log train_lam_action_space_scaling_8_3318550.log train_tokenizer_batch_size_scaling_8_node_3320176.log train_tokenizer_model_size_scaling_37M_3318556.log\r\ntrain_lam_action_space_scaling_12_3329802.log train_lam_action_space_scaling_8_3329791.log train_tokenizer_batch_size_scaling_8_node_3321525.log train_tokenizer_model_size_scaling_74M_3318557.log\r\ntrain_lam_action_space_scaling_12_3331284.log train_lam_action_space_scaling_8_3329806.log train_tokenizer_minecraft_overfit_sample_3309656.log train_tokenizer_model_size_scaling_74M_3320174.log\r\ntrain_lam_action_space_scaling_20_3318547.log train_lam_action_space_scaling_8_3331288.log train_tokenizer_model_size_scaling_127M_3317233.log train_tokenizer_model_size_scaling_74M_3321522.log\r\ntrain_lam_action_space_scaling_20_3329788.log train_lam_minecraft_overfit_sample_3309655.log train_tokenizer_model_size_scaling_127M_3318554.log train_tokenizer_model_size_scaling_80M_3313564.log\r\ntrain_lam_action_space_scaling_20_3329803.log train_lam_model_size_scaling_38M_3317098.log train_tokenizer_model_size_scaling_140M_3313562.log train_tokenizer_model_size_scaling_80M_3316026.log\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir]633;D;0",,terminal_output +1432,2883217,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3335345.log",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=01:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n# tf_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/open_ai_minecraft_tfrecord\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/train_lam_action_space_scaling_20/3318547/lam_1751657975_200000/\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n # --lam_checkpoint=$lam_ckpt_dir\nSLURM_STEP_NUM_TASKS=1\nSLURM_JOB_USER=tum_cte0515\nSLURM_TASKS_PER_NODE=4(x2)\nSLURM_JOB_UID=999226\nSLURM_STEP_GPUS=0\nSLURM_CPU_BIND=quiet,mask_cpu:0x00000003FC00000007F00000003FC00000007F\nSLURM_TASK_PID=3141369\nSLURM_JOB_GPUS=0,1,2,3\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\nSLURM_CPU_BIND_VERBOSE=quiet\nSLURMD_NODENAME=hkn0430\nSLURM_JOB_START_TIME=1752159877\nSLURM_STEP_NODELIST=hkn0523\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1752163477\nSLURM_CPUS_ON_NODE=24\nSLURM_JOB_CPUS_PER_NODE=24(x2)\nSLURM_GPUS_ON_NODE=4\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=accelerated\nSLURM_TRES_PER_TASK=cpu=5\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=2\nSLURM_STEPID=0\nSLURM_CPU_BIND_LIST=0x00000003FC00000007F00000003FC00000007F\nSLURM_JOBID=3335345\nSLURM_PTY_PORT=43975\nSLURM_JOB_QOS=normal\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\nSLURM_PTY_WIN_ROW=48\nSLURM_CPU_BIND_TYPE=mask_cpu:\nSLURMD_DEBUG=2\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=5\nSLURM_NTASKS=8\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e13.hkn0430\nSLURM_DISTRIBUTION=cyclic\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SRUN_COMM_HOST=10.0.7.201\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_PTY_WIN_COL=228\nSLURM_NODELIST=hkn[0430-0431]\nSLURM_SRUN_COMM_PORT=42223\nSLURM_STEP_ID=0\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=8\nSLURM_NNODES=2\nSLURM_SUBMIT_HOST=hkn0523.localdomain\nSLURM_JOB_ID=3335345\nSLURM_NODEID=0\nSLURMD_TRES_FREQ=gpu:high,memory=high\nSLURM_STEP_NUM_NODES=1\nSLURM_STEP_TASKS_PER_NODE=1\nSLURM_MPI_TYPE=pmi2\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_dyn_yolorun\nSLURM_NTASKS_PER_NODE=4\nSLURM_STEP_LAUNCHER_PORT=42223\nSLURM_JOB_GID=502226\nSLURM_JOB_NODELIST=hkn[0430-0431]\nSLURM_STEP_NUM_TASKS=1\nSLURM_JOB_USER=tum_cte0515\nSLURM_TASKS_PER_NODE=4(x2)\nSLURM_JOB_UID=999226\nSLURM_STEP_GPUS=0\nSLURM_CPU_BIND=quiet,mask_cpu:0x00000003FC00000007F00000003FC00000007F\nSLURM_TASK_PID=3141369\nSLURM_JOB_GPUS=0,1,2,3\nSLURM_LOCALID=0\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\nSLURM_CPU_BIND_VERBOSE=quiet\nSLURMD_NODENAME=hkn0430\nSLURM_JOB_START_TIME=1752159877\nSLURM_STEP_NODELIST=hkn0523\nSLURM_CLUSTER_NAME=hk\nSLURM_JOB_END_TIME=1752163477\nSLURM_CPUS_ON_NODE=24\nSLURM_JOB_CPUS_PER_NODE=24(x2)\nSLURM_GPUS_ON_NODE=4\nSLURM_GTIDS=0\nSLURM_JOB_PARTITION=accelerated\nSLURM_TRES_PER_TASK=cpu=5\nSLURM_OOM_KILL_STEP=0\nSLURM_JOB_NUM_NODES=2\nSLURM_STEPID=0\nSLURM_CPU_BIND_LIST=0x00000003FC00000007F00000003FC00000007F\nSLURM_JOBID=3335345\nSLURM_PTY_PORT=43975\nSLURM_JOB_QOS=normal\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\nSLURM_PTY_WIN_ROW=48\nSLURM_CPU_BIND_TYPE=mask_cpu:\nSLURMD_DEBUG=2\nSLURM_PROCID=0\nSLURM_CPUS_PER_TASK=5\nSLURM_NTASKS=8\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e13.hkn0430\nSLURM_DISTRIBUTION=cyclic\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\nSLURM_SRUN_COMM_HOST=10.0.7.201\nSLURM_SCRIPT_CONTEXT=prolog_task\nSLURM_PTY_WIN_COL=228\nSLURM_NODELIST=hkn[0430-0431]\nSLURM_SRUN_COMM_PORT=42223\nSLURM_STEP_ID=0\nSLURM_JOB_ACCOUNT=hk-project-p0023960\nSLURM_PRIO_PROCESS=0\nSLURM_NPROCS=8\nSLURM_NNODES=2\nSLURM_SUBMIT_HOST=hkn0523.localdomain\nSLURM_JOB_ID=3335345\nSLURM_NODEID=0\nSLURMD_TRES_FREQ=gpu:high,memory=high\nSLURM_STEP_NUM_NODES=1\nSLURM_STEP_TASKS_PER_NODE=1\nSLURM_MPI_TYPE=pmi2\nSLURM_CONF=/etc/slurm/slurm.conf\nSLURM_JOB_NAME=train_dyn_yolorun\nSLURM_NTASKS_PER_NODE=4\nSLURM_STEP_LAUNCHER_PORT=42223\nSLURM_JOB_GID=502226\nSLURM_JOB_NODELIST=hkn[0430-0431]\nsrun: error: CPU binding outside of job step allocation, allocated CPUs are: 0x000000003C00000003F000000003C00000003F.\nsrun: error: CPU binding outside of job step allocation, allocated CPUs are: 0x000000003C00000003F000000003C00000003F.\nsrun: error: Task launch for StepId=3335345.0 failed on node hkn0430: Unable to satisfy cpu bind request\nsrun: error: Task launch for StepId=3335345.0 failed on node hkn0431: Unable to satisfy cpu bind request\nsrun: error: Application launch failed: Unable to satisfy cpu bind request\nsrun: Job step aborted\n\n============================= JOB FEEDBACK =============================\n\nJob ID: 3335345\nCluster: hk\nUser/Group: tum_cte0515/hk-project-p0023960\nAccount: hk-project-p0023960\nState: FAILED (exit code 192)\nPartition: accelerated\nNodes: 2\nCores per node: 24\nNodelist: hkn[0430-0431]\nCPU Utilized: 00:00:00\nCPU Efficiency: 0.00% of 00:21:36 core-walltime\nJob Wall-clock time: 00:00:27\nStarttime: Thu Jul 10 17:04:37 2025\nEndtime: Thu Jul 10 17:05:04 2025\nMemory Utilized: 10.38 MB\nMemory Efficiency: 0.00% of 0.00 MB\nEnergy Consumed: 13356 Joule / 3.71 Watthours\nAverage node power draw: 494.666666666667 Watt\n",log,tab +1433,2883830,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3335345.log",1303,0,"",log,selection_mouse +1434,2884425,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/train_dyn_yolorun_3335345.log",6380,0,"",log,selection_command +1435,2907779,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"",shellscript,tab +1436,2910928,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",66,0,"",shellscript,selection_mouse +1437,2911068,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",65,1,"4",shellscript,selection_mouse +1438,2911148,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",60,6,"node=4",shellscript,selection_mouse +1439,2911210,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",56,10,"per-node=4",shellscript,selection_mouse +1440,2911214,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",49,17,"ntasks-per-node=4",shellscript,selection_mouse +1441,2911314,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",48,18,"-ntasks-per-node=4",shellscript,selection_mouse +1442,2911373,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",47,19,"--ntasks-per-node=4",shellscript,selection_mouse +1443,2911429,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",46,20," --ntasks-per-node=4",shellscript,selection_mouse +1444,2911860,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",46,0,"",shellscript,selection_mouse +1445,2911861,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",40,6,"SBATCH",shellscript,selection_mouse +1446,2912035,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",40,7,"SBATCH ",shellscript,selection_mouse +1447,2912036,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",40,8,"SBATCH -",shellscript,selection_mouse +1448,2912046,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",40,15,"SBATCH --ntasks",shellscript,selection_mouse +1449,2912103,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",40,19,"SBATCH --ntasks-per",shellscript,selection_mouse +1450,2912115,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",40,24,"SBATCH --ntasks-per-node",shellscript,selection_mouse +1451,2912173,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",40,25,"SBATCH --ntasks-per-node=",shellscript,selection_mouse +1452,2912174,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",40,26,"SBATCH --ntasks-per-node=4",shellscript,selection_mouse +1453,2912485,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",66,0,"",shellscript,selection_mouse +1454,2912778,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",65,1,"4",shellscript,selection_mouse +1455,2912981,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",60,6,"node=4",shellscript,selection_mouse +1456,2913040,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",56,10,"per-node=4",shellscript,selection_mouse +1457,2913040,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",55,11,"-per-node=4",shellscript,selection_mouse +1458,2913050,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",49,17,"ntasks-per-node=4",shellscript,selection_mouse +1459,2913465,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",51,0,"",shellscript,selection_mouse +1460,2913801,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",136,0,"",shellscript,selection_mouse +1461,2913949,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",133,4,"cpus",shellscript,selection_mouse +1462,2914180,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",133,5,"cpus-",shellscript,selection_mouse +1463,2914195,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",133,8,"cpus-per",shellscript,selection_mouse +1464,2914221,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",133,34,"cpus-per-task=5\n#SBATCH --gres=gpu",shellscript,selection_mouse +1465,2914236,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",133,36,"cpus-per-task=5\n#SBATCH --gres=gpu:4",shellscript,selection_mouse +1466,2914428,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",169,0,"",shellscript,selection_mouse +1467,2914940,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",397,0,"",shellscript,selection_mouse +1468,2915079,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",386,17,"train_dyn_yolorun",shellscript,selection_mouse +1469,2915293,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",291,112,"work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun",shellscript,selection_mouse +1470,2915307,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",188,215,"hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun",shellscript,selection_mouse +1471,2915367,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",186,217,"=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun",shellscript,selection_mouse +1472,2915367,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",159,244,"gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun",shellscript,selection_mouse +1473,2915368,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",133,270,"cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun",shellscript,selection_mouse +1474,2915430,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",101,302,"partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun",shellscript,selection_mouse +1475,2915773,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",102,0,"",shellscript,selection_mouse +1476,2926976,"TERMINAL",0,0,"cd",,terminal_command +1477,2926989,"TERMINAL",0,0,"]633;E;2025-07-10 17:07:01 cd;16c08be6-3885-420f-ac53-f5272aed6e54]633;C]0;tum_cte0515@hkn1993:~]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515",,terminal_output +1478,2931981,"TERMINAL",0,0,"cd Projects/jafar",,terminal_command +1479,2932602,"TERMINAL",0,0,"ls",,terminal_command +1480,2932616,"TERMINAL",0,0,"]633;E;2025-07-10 17:07:06 ls;16c08be6-3885-420f-ac53-f5272aed6e54]633;Cdata frame.png genie.py local-logs overfit_dir read_tf_record.py sample.py scripts_horeka tests train_tokenizer_bak.py wandb\r\ndebug frames gifs logs __pycache__ requirements-franz.txt sample.py_bak slurm train_dynamics.py train_tokenizer.py\r\nframe-knoms.png generate_dataset.py LICENSE models README.md requirements.txt scripts_cremers slurm-3309772.out train_lam.py utils\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +1481,2940672,"TERMINAL",0,0,"queue^C",,terminal_command +1482,2940687,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]633;E;;16c08be6-3885-420f-ac53-f5272aed6e54]633;C]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D",,terminal_output +1483,2942924,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"",shellscript,tab +1484,2947815,"TERMINAL",0,0,"sbatch slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",,terminal_command +1485,2947826,"TERMINAL",0,0,"]633;E;2025-07-10 17:07:21 sbatch slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch;16c08be6-3885-420f-ac53-f5272aed6e54]633;CSubmitted batch job 3335362\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +1486,2949743,"TERMINAL",0,0,"queue",,terminal_command +1487,2949797,"TERMINAL",0,0,"]633;E;2025-07-10 17:07:23 queue;16c08be6-3885-420f-ac53-f5272aed6e54]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Jul 10 17:07:23 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3335362 accelerat train_dy tum_cte0 PD\t0:00\t 2 (Priority)3335335 accelerat interact tum_cte0 R16:01\t 2 hkn[0425,0428]",,terminal_output +1488,2950834,"TERMINAL",0,0,"43",,terminal_output +1489,2952659,"TERMINAL",0,0,"idling",,terminal_command +1490,2952712,"TERMINAL",0,0,"]633;E;2025-07-10 17:07:26 idling;16c08be6-3885-420f-ac53-f5272aed6e54]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1993.localdomain: Thu Jul 10 17:07:26 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly:\t 4 nodes idle\rPartition dev_accelerated:\t 0 nodes idle\rPartition accelerated:\t 2 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 6 nodes idle",,terminal_output +1491,2954767,"TERMINAL",0,0,"queue",,terminal_command +1492,2954835,"TERMINAL",0,0,"]633;E;2025-07-10 17:07:28 queue;16c08be6-3885-420f-ac53-f5272aed6e54]633;C[?1049h(B[?7hEvery 1.0s: squeue --mehkn1993.localdomain: Thu Jul 10 17:07:28 2025JOBID PARTITION NAME USER ST\tTIME NODES NODELIST(REASON)3335362 accelerat train_dy tum_cte0 PD\t0:00\t 2 (Priority)3335335 accelerat interact tum_cte0 R16:07\t 2 hkn[0425,0428]",,terminal_output +1493,2955900,"TERMINAL",0,0,"308",,terminal_output +1494,2958924,"TERMINAL",0,0,"111",,terminal_output +1495,2961898,"TERMINAL",0,0,"44",,terminal_output +1496,2964967,"TERMINAL",0,0,"77",,terminal_output +1497,2967998,"TERMINAL",0,0,"4020",,terminal_output +1498,2971095,"TERMINAL",0,0,"33",,terminal_output +1499,2974898,"TERMINAL",0,0,"67",,terminal_output +1500,2977561,"TERMINAL",0,0,"509",,terminal_output +1501,2978585,"TERMINAL",0,0,"230",,terminal_output +1502,2979595,"TERMINAL",0,0,"31",,terminal_output +1503,2980640,"TERMINAL",0,0,"42",,terminal_output +1504,2981682,"TERMINAL",0,0,"53",,terminal_output +1505,2982750,"TERMINAL",0,0,"64",,terminal_output +1506,2983808,"TERMINAL",0,0,"75",,terminal_output +1507,2984934,"TERMINAL",0,0,"87",,terminal_output +1508,2985958,"TERMINAL",0,0,"8:008",,terminal_output +1509,2986943,"TERMINAL",0,0,"19",,terminal_output +1510,2987983,"TERMINAL",0,0,"240",,terminal_output +1511,2989131,"TERMINAL",0,0,"31",,terminal_output +1512,2990076,"TERMINAL",0,0,"42",,terminal_output +1513,2991119,"TERMINAL",0,0,"53",,terminal_output +1514,2992165,"TERMINAL",0,0,"64",,terminal_output +1515,2993215,"TERMINAL",0,0,"75",,terminal_output +1516,2994268,"TERMINAL",0,0,"86",,terminal_output +1517,2995382,"TERMINAL",0,0,"97",,terminal_output +1518,2996363,"TERMINAL",0,0,"108",,terminal_output +1519,2997429,"TERMINAL",0,0,"19",,terminal_output +1520,2998463,"TERMINAL",0,0,"250",,terminal_output +1521,2999508,"TERMINAL",0,0,"31",,terminal_output +1522,3000612,"TERMINAL",0,0,"42",,terminal_output +1523,3001602,"TERMINAL",0,0,"53",,terminal_output +1524,3002663,"TERMINAL",0,0,"64",,terminal_output +1525,3003695,"TERMINAL",0,0,"75",,terminal_output +1526,3004741,"TERMINAL",0,0,"86",,terminal_output +1527,3005814,"TERMINAL",0,0,"97",,terminal_output +1528,3006934,"TERMINAL",0,0,"209",,terminal_output +1529,3007995,"TERMINAL",0,0,"27:00",,terminal_output +1530,3009009,"TERMINAL",0,0,"31",,terminal_output +1531,3010024,"TERMINAL",0,0,"42",,terminal_output +1532,3011079,"TERMINAL",0,0,"53",,terminal_output +1533,3012172,"TERMINAL",0,0,"64",,terminal_output +1534,3013201,"TERMINAL",0,0,"75",,terminal_output +1535,3014168,"TERMINAL",0,0,"86",,terminal_output +1536,3015212,"TERMINAL",0,0,"97",,terminal_output +1537,3016292,"TERMINAL",0,0,"308",,terminal_output +1538,3017305,"TERMINAL",0,0,"19",,terminal_output +1539,3018421,"TERMINAL",0,0,"210",,terminal_output +1540,3019443,"TERMINAL",0,0,"31",,terminal_output +1541,3020468,"TERMINAL",0,0,"42",,terminal_output +1542,3021640,"TERMINAL",0,0,"53",,terminal_output +1543,3022618,"TERMINAL",0,0,"64",,terminal_output +1544,3023641,"TERMINAL",0,0,"75",,terminal_output +1545,3024667,"TERMINAL",0,0,"86",,terminal_output +1546,3025690,"TERMINAL",0,0,"97",,terminal_output +1547,3026735,"TERMINAL",0,0,"408",,terminal_output +1548,3027788,"TERMINAL",0,0,"19",,terminal_output +1549,3028847,"TERMINAL",0,0,"221",,terminal_output +1550,3029988,"TERMINAL",0,0,"4 R1hkn[0430-0431]2",,terminal_output +1551,3031011,"TERMINAL",0,0,"523",,terminal_output +1552,3031983,"TERMINAL",0,0,"634",,terminal_output +1553,3033062,"TERMINAL",0,0,"745",,terminal_output +1554,3034089,"TERMINAL",0,0,"856",,terminal_output +1555,3035218,"TERMINAL",0,0,"967",,terminal_output +1556,3036168,"TERMINAL",0,0,"5078",,terminal_output +1557,3037216,"TERMINAL",0,0,"189",,terminal_output +1558,3038284,"TERMINAL",0,0,"2930",,terminal_output +1559,3039312,"TERMINAL",0,0,"3101",,terminal_output +1560,3040357,"TERMINAL",0,0,"412",,terminal_output +1561,3041406,"TERMINAL",0,0,"523",,terminal_output +1562,3042483,"TERMINAL",0,0,"634",,terminal_output +1563,3043506,"TERMINAL",0,0,"745",,terminal_output +1564,3044542,"TERMINAL",0,0,"856",,terminal_output +1565,3045668,"TERMINAL",0,0,"967",,terminal_output +1566,3046681,"TERMINAL",0,0,"9:0078",,terminal_output +1567,3047703,"TERMINAL",0,0,"189",,terminal_output +1568,3048734,"TERMINAL",0,0,"2940",,terminal_output +1569,3049783,"TERMINAL",0,0,"3201",,terminal_output +1570,3050828,"TERMINAL",0,0,"423",,terminal_output +1571,3051905,"TERMINAL",0,0,"634",,terminal_output +1572,3052936,"TERMINAL",0,0,"745",,terminal_output +1573,3053971,"TERMINAL",0,0,"856",,terminal_output +1574,3055010,"TERMINAL",0,0,"967",,terminal_output +1575,3056061,"TERMINAL",0,0,"1078",,terminal_output +1576,3057111,"TERMINAL",0,0,"189",,terminal_output +1577,3058150,"TERMINAL",0,0,"2950",,terminal_output +1578,3059277,"TERMINAL",0,0,"3301",,terminal_output +1579,3060299,"TERMINAL",0,0,"412",,terminal_output +1580,3061296,"TERMINAL",0,0,"523",,terminal_output +1581,3062376,"slurm/jobs/mihir/horeka/sbatch_dir.sh",0,0,"",shellscript,tab +1582,3062472,"TERMINAL",0,0,"634",,terminal_output +1583,3063473,"TERMINAL",0,0,"745",,terminal_output +1584,3064500,"TERMINAL",0,0,"856",,terminal_output +1585,3065522,"TERMINAL",0,0,"967",,terminal_output +1586,3066587,"TERMINAL",0,0,"2078",,terminal_output +1587,3067697,"TERMINAL",0,0,"189",,terminal_output +1588,3068675,"TERMINAL",0,0,"298:00",,terminal_output +1589,3069730,"TERMINAL",0,0,"3401",,terminal_output +1590,3071052,"TERMINAL",0,0,"412",,terminal_output +1591,3071862,"TERMINAL",0,0,"523",,terminal_output +1592,3073308,"TERMINAL",0,0,"745",,terminal_output +1593,3073907,"TERMINAL",0,0,"856",,terminal_output +1594,3075460,"TERMINAL",0,0,"967",,terminal_output +1595,3076299,"TERMINAL",0,0,"3078",,terminal_output +1596,3077607,"TERMINAL",0,0,"189",,terminal_output +1597,3078097,"TERMINAL",0,0,"2910",,terminal_output +1598,3079179,"TERMINAL",0,0,"3501",,terminal_output +1599,3080241,"TERMINAL",0,0,"412",,terminal_output +1600,3081286,"TERMINAL",0,0,"523",,terminal_output +1601,3082322,"TERMINAL",0,0,"634",,terminal_output +1602,3083443,"TERMINAL",0,0,"745",,terminal_output +1603,3084468,"TERMINAL",0,0,"856",,terminal_output +1604,3085697,"TERMINAL",0,0,"967",,terminal_output +1605,3086517,"TERMINAL",0,0,"4078",,terminal_output +1606,3087641,"TERMINAL",0,0,"189",,terminal_output +1607,3088637,"TERMINAL",0,0,"2920",,terminal_output +1608,3089650,"TERMINAL",0,0,"31:001",,terminal_output +1609,3090689,"TERMINAL",0,0,"412",,terminal_output +1610,3091735,"TERMINAL",0,0,"523",,terminal_output +1611,3092836,"TERMINAL",0,0,"634",,terminal_output +1612,3093891,"TERMINAL",0,0,"756",,terminal_output +1613,3095145,"TERMINAL",0,0,"967",,terminal_output +1614,3095979,"TERMINAL",0,0,"5078",,terminal_output +1615,3097369,"TERMINAL",0,0,"189",,terminal_output +1616,3098071,"TERMINAL",0,0,"2930",,terminal_output +1617,3099112,"TERMINAL",0,0,"3101",,terminal_output +1618,3100235,"TERMINAL",0,0,"412",,terminal_output +1619,3101259,"TERMINAL",0,0,"523",,terminal_output +1620,3102284,"TERMINAL",0,0,"634",,terminal_output +1621,3103311,"TERMINAL",0,0,"745",,terminal_output +1622,3104333,"TERMINAL",0,0,"856",,terminal_output +1623,3105663,"TERMINAL",0,0,"967",,terminal_output +1624,3106451,"TERMINAL",0,0,"10:0078",,terminal_output +1625,3107465,"TERMINAL",0,0,"189",,terminal_output +1626,3108660,"TERMINAL",0,0,"2940",,terminal_output +1627,3109555,"TERMINAL",0,0,"3201",,terminal_output +1628,3109996,"TERMINAL",0,0,"Step 0, loss: 18.776222229003906\r\nStep 1, loss: 18.819902420043945\r\nStep 2, loss: 18.791563034057617\r\nStep 3, loss: 18.668718338012695\r\nStep 4, loss: 18.80548667907715\r\nStep 5, loss: 18.81717300415039\r\nStep 6, loss: 18.450780868530273\r\nStep 7, loss: 18.754093170166016\r\nStep 8, loss: 18.730485916137695\r\nStep 9, loss: 18.4219913482666\r\nStep 10, loss: 18.53896713256836\r\nStep 11, loss: 18.433452606201172\r\nStep 12, loss: 18.339122772216797\r\nStep 13, loss: 18.459253311157227\r\nStep 14, loss: 18.409881591796875\r\nStep 15, loss: 18.35429573059082\r\nStep 16, loss: 18.237462997436523\r\nStep 17, loss: 18.09574317932129\r\nStep 18, loss: 17.840925216674805\r\nStep 19, loss: 18.041223526000977\r\nStep 20, loss: 17.888032913208008\r\nStep 21, loss: 17.768489837646484\r\nStep 22, loss: 17.63291358947754\r\nStep 23, loss: 17.79960060119629\r\nStep 24, loss: 17.668100357055664\r\nStep 25, loss: 17.608400344848633\r\nStep 26, loss: 17.400070190429688\r\nStep 27, loss: 17.27320671081543\r\nStep 28, loss: 17.558292388916016\r\nStep 29, loss: 17.401700973510742\r\nStep 30, loss: 17.43343734741211\r\nStep 31, loss: 16.896129608154297\r\nStep 32, loss: 16.79721450805664\r\nStep 33, loss: 16.77494239807129\r\nStep 34, loss: 17.045551300048828\r\nStep 35, loss: 16.608722686767578\r\nStep 36, loss: 16.647741317749023\r\nStep 37, loss: 16.336868286132812\r\nStep 38, loss: 16.651290893554688\r\nStep 39, loss: 16.270305633544922\r\nStep 40, loss: 16.28514289855957\r\nStep 41, loss: 15.932032585144043\r\nStep 42, loss: 16.085798263549805\r\nStep 43, loss: 15.853764533996582\r\nStep 44, loss: 16.237834930419922\r\nStep 45, loss: 15.47584056854248\r\nStep 46, loss: 15.609787940979004\r\nStep 47, loss: 15.486068725585938\r\nStep 48, loss: 15.258798599243164\r\nStep 49, loss: 15.545014381408691\r\nStep 50, loss: 16.061813354492188\r\nStep 51, loss: 15.366859436035156\r\nStep 52, loss: 14.827864646911621\r\nStep 53, loss: 14.727177619934082\r\nStep 54, loss: 14.848566055297852\r\nStep 55, loss: 14.447760581970215\r\nStep 56, loss: 15.252120018005371\r\nStep 57, loss: 14.218976020812988\r\nStep 58, loss: 14.730356216430664\r\nStep 59, loss: 14.582534790039062\r\nStep 60, loss: 13.967232704162598\r\nStep 61, loss: 13.71136474609375\r\nStep 62, loss: 13.708770751953125\r\nStep 63, loss: 14.0950345993042\r\nStep 64, loss: 13.713918685913086\r\nStep 65, loss: 14.868953704833984\r\nStep 66, loss: 14.468207359313965\r\nStep 67, loss: 14.585230827331543\r\nStep 68, loss: 14.341510772705078\r\nStep 69, loss: 14.391881942749023\r\nStep 70, loss: 13.262972831726074\r\nStep 71, loss: 14.027960777282715\r\nStep 72, loss: 12.899559020996094\r\nStep 73, loss: 13.606071472167969\r\nStep 74, loss: 13.505942344665527\r\nStep 75, loss: 12.73733139038086\r\nStep 76, loss: 13.731353759765625\r\nStep 77, loss: 12.609628677368164\r\nStep 78, loss: 13.05951976776123\r\nStep 79, loss: 12.82889461517334\r\nStep 80, loss: 12.381841659545898\r\nStep 81, loss: 12.184906005859375\r\nStep 82, loss: 12.261787414550781\r\nStep 83, loss: 12.400224685668945\r\nStep 84, loss: 12.114161491394043\r\nStep 85, loss: 12.412161827087402\r\nStep 86, loss: 11.783485412597656\r\nStep 87, loss: 11.766879081726074\r\nStep 88, loss: 11.893105506896973\r\nStep 89, loss: 11.720884323120117\r\nStep 90, loss: 11.85186767578125\r\nStep 91, loss: 11.742947578430176\r\nStep 92, loss: 11.327075004577637\r\nStep 93, loss: 11.285242080688477\r\nStep 94, loss: 11.404678344726562\r\nStep 95, loss: 11.971321105957031\r\nStep 96, loss: 11.94084644317627\r\nStep 97, loss: 11.925477027893066\r\nStep 98, loss: 11.581640243530273\r\nStep 99, loss: 10.887056350708008\r\nStep 100, loss: 11.63808822631836\r\nStep 101, loss: 10.839936256408691\r\nStep 102, loss: 11.229722023010254\r\nStep 103, loss: 10.856170654296875\r\nStep 104, loss: 11.429742813110352\r\nStep 105, loss: 10.258073806762695\r\nStep 106, loss: 10.562714576721191\r\nStep 107, loss: 10.214714050292969\r\nStep 108, loss: 10.124898910522461\r\nStep 109, loss: 9.705013275146484\r\nStep 110, loss: 9.921828269958496\r\nStep 111, loss: 9.90079116821289\r\nStep 112, loss: 10.455109596252441\r\nStep 113, loss: 9.936418533325195\r\nStep 114, loss: 9.936537742614746\r\nStep 115, loss: 9.773576736450195\r\nStep 116, loss: 9.676142692565918\r\nStep 117, loss: 9.945345878601074\r\nStep 118, loss: 9.923107147216797\r\nStep 119, loss: 9.550374984741211\r\nStep 120, loss: 10.050488471984863\r\nStep 121, loss: 9.169127464294434\r\nStep 122, loss: 9.096636772155762\r\nStep 123, loss: 9.86887264251709\r\nStep 124, loss: 9.788836479187012\r\nStep 125, loss: 9.638827323913574\r\nStep 126, loss: 9.093913078308105\r\nStep 127, loss: 9.269700050354004\r\nStep 128, loss: 9.940503120422363\r\nStep 129, loss: 8.495964050292969\r\nStep 130, loss: 9.110383033752441\r\nStep 131, loss: 9.079448699951172\r\nStep 132, loss: 9.014200210571289\r\nStep 133, loss: 8.72136402130127\r\nStep 134, loss: 9.308770179748535\r\nStep 135, loss: 8.615967750549316\r\nStep 136, loss: 8.500845909118652\r\nStep 137, loss: 8.710356712341309\r\nStep 138, loss: 8.530448913574219\r\nStep 139, loss: 9.046561241149902\r\nStep 140, loss: 8.546600341796875\r\nStep 141, loss: 9.229374885559082\r\nStep 142, loss: 8.218626022338867\r\nStep 143, loss: 8.23629093170166\r\nStep 144, loss: 8.220922470092773\r\nStep 145, loss: 8.11863899230957\r\nStep 146, loss: 8.13656997680664\r\nStep 147, loss: 8.036163330078125\r\nStep 148, loss: 7.980085372924805\r\nStep 149, loss: 7.8308844566345215\r\nStep 150, loss: 7.997012615203857\r\nStep 151, loss: 8.312012672424316\r\nStep 152, loss: 8.257838249206543\r\nStep 153, loss: 7.67844820022583\r\nStep 154, loss: 8.097171783447266\r\nStep 155, loss: 7.956625461578369\r\nStep 156, loss: 7.661460876464844\r\nStep 157, loss: 7.60241174697876\r\nStep 158, loss: 7.5734171867370605\r\nStep 159, loss: 7.4999213218688965\r\nStep 160, loss: 7.889724254608154\r\nStep 161, loss: 7.616578578948975\r\nStep 162, loss: 7.669461250305176\r\nStep 163, loss: 7.583798408508301\r\nStep 164, loss: 7.573363780975342\r\nStep 165, loss: 7.425256252288818\r\nStep 166, loss: 7.534543991088867\r\nStep 167, loss: 7.315162658691406\r\nStep 168, loss: 7.1738691329956055\r\nStep 169, loss: 7.142663955688477\r\nStep 170, loss: 7.435649871826172\r\nStep 171, loss: 7.096683502197266\r\nStep 172, loss: 7.248353481292725\r\nStep 173, loss: 7.270962715148926\r\nStep 174, loss: 7.3473687171936035\r\nStep 175, loss: 6.999227523803711\r\nStep 176, loss: 7.4188151359558105\r\nStep 177, loss: 7.223378658294678\r\nStep 178, loss: 7.144739627838135\r\nStep 179, loss: 6.7800211906433105\r\nStep 180, loss: 7.156866073608398\r\nStep 181, loss: 6.900200843811035\r\nStep 182, loss: 7.0645365715026855\r\nStep 183, loss: 7.188953399658203\r\nStep 184, loss: 6.891933917999268\r\nStep 185, loss: 6.902243614196777\r\nStep 186, loss: 6.5922651290893555\r\nStep 187, loss: 7.096951007843018\r\nStep 188, loss: 6.9083991050720215\r\nStep 189, loss: 6.90083646774292\r\nStep 190, loss: 6.980688095092773\r\nStep 191, loss: 6.729495048522949\r\nStep 192, loss: 6.864687442779541\r\nStep 193, loss: 6.7847514152526855\r\nStep 194, loss: 6.816504955291748\r\nStep 195, loss: 6.671535968780518\r\nStep 196, loss: 6.762739181518555\r\nStep 197, loss: 6.497260570526123\r\nStep 198, loss: 7.014808654785156\r\nStep 199, loss: 6.6613945960998535\r\nStep 200, loss: 6.8338422775268555\r\nStep 201, loss: 6.816775798797607\r\nStep 202, loss: 6.6137614250183105\r\nStep 203, loss: 6.697750568389893\r\nStep 204, loss: 6.821736812591553\r\nStep 205, loss: 6.9376959800720215\r\nStep 206, loss: 6.731281757354736\r\nStep 207, loss: 6.579183578491211\r\nStep 208, loss: 6.68570613861084\r\nStep 209, loss: 6.527091026306152\r\nStep 210, loss: 6.523392200469971\r\nStep 211, loss: 6.566160678863525\r\nStep 212, loss: 6.912597179412842\r\nStep 213, loss: 6.662230968475342\r\nStep 214, loss: 6.962000846862793\r\nStep 215, loss: 6.555893898010254\r\nStep 216, loss: 6.651054382324219\r\nStep 217, loss: 6.657430648803711\r\nStep 218, loss: 6.917163372039795\r\nStep 219, loss: 6.651887893676758\r\nStep 220, loss: 6.725634574890137\r\nStep 221, loss: 6.670478343963623\r\nStep 222, loss: 6.46140193939209\r\nStep 223, loss: 6.792141437530518\r\nStep 224, loss: 6.598993301391602\r\nStep 225, loss: 6.382656097412109\r\nStep 226, loss: 6.67178201675415\r\nStep 227, loss: 6.364635467529297\r\nStep 228, loss: 6.6085052490234375\r\nStep 229, loss: 6.653966903686523\r\nStep 230, loss: 6.554793834686279\r\nStep 231, loss: 6.588866710662842\r\nStep 232, loss: 6.585617542266846\r\nStep 233, loss: 6.507537364959717\r\nStep 234, loss: 6.644289493560791\r\nStep 235, loss: 6.5886993408203125\r\nStep 236, loss: 6.520603656768799\r\nStep 237, loss: 6.404054164886475\r\nStep 238, loss: 6.414518356323242\r\nStep 239, loss: 6.4341325759887695\r\nStep 240, loss: 6.528247833251953\r\nStep 0, loss: 18.776222229003906\r\nStep 1, loss: 18.819902420043945\r\nStep 2, loss: 18.791563034057617\r\nStep 3, loss: 18.668718338012695\r\nStep 4, loss: 18.80548667907715\r\nStep 5, loss: 18.81717300415039\r\nStep 6, loss: 18.450780868530273\r\nStep 7, loss: 18.754093170166016\r\nStep 8, loss: 18.730485916137695\r\nStep 9, loss: 18.4219913482666\r\nStep 10, loss: 18.53896713256836\r\nStep 11, loss: 18.433452606201172\r\nStep 12, loss: 18.339122772216797\r\nStep 13, loss: 18.459253311157227\r\nStep 14, loss: 18.409881591796875\r\nStep 15, loss: 18.35429573059082\r\nStep 16, loss: 18.237462997436523\r\nStep 17, loss: 18.09574317932129\r\nStep 18, loss: 17.840925216674805\r\nStep 19, loss: 18.041223526000977\r\nStep 20, loss: 17.888032913208008\r\nStep 21, loss: 17.768489837646484\r\nStep 22, loss: 17.63291358947754\r\nStep 23, loss: 17.79960060119629\r\nStep 24, loss: 17.668100357055664\r\nStep 25, loss: 17.608400344848633\r\nStep 26, loss: 17.400070190429688\r\nStep 27, loss: 17.27320671081543\r\nStep 28, loss: 17.558292388916016\r\nStep 29, loss: 17.401700973510742\r\nStep 30, loss: 17.43343734741211\r\nStep 31, loss: 16.896129608154297\r\nStep 32, loss: 16.79721450805664\r\nStep 33, loss: 16.77494239807129\r\nStep 34, loss: 17.045551300048828\r\nStep 35, loss: 16.608722686767578\r\nStep 36, loss: 16.647741317749023\r\nStep 37, loss: 16.336868286132812\r\nStep 38, loss: 16.651290893554688\r\nStep 39, loss: 16.270305633544922\r\nStep 40, loss: 16.28514289855957\r\nStep 41, loss: 15.932032585144043\r\nStep 42, loss: 16.085798263549805\r\nStep 43, loss: 15.853764533996582\r\nStep 44, loss: 16.237834930419922\r\nStep 45, loss: 15.47584056854248\r\nStep 46, loss: 15.609787940979004\r\nStep 47, loss: 15.486068725585938\r\nStep 48, loss: 15.258798599243164\r\nStep 49, loss: 15.545014381408691\r\nStep 50, loss: 16.061813354492188\r\nStep 51, loss: 15.366859436035156\r\nStep 52, loss: 14.827864646911621\r\nStep 53, loss: 14.727177619934082\r\nStep 54, loss: 14.848566055297852\r\nStep 55, loss: 14.447760581970215\r\nStep 56, loss: 15.252120018005371\r\nStep 57, loss: 14.218976020812988\r\nStep 58, loss: 14.730356216430664\r\nStep 59, loss: 14.582534790039062\r\nStep 60, loss: 13.967232704162598\r\nStep 61, loss: 13.71136474609375\r\nStep 62, loss: 13.708770751953125\r\nStep 63, loss: 14.0950345993042\r\nStep 64, loss: 13.713918685913086\r\nStep 65, loss: 14.868953704833984\r\nStep 66, loss: 14.468207359313965\r\nStep 67, loss: 14.585230827331543\r\nStep 68, loss: 14.341510772705078\r\nStep 69, loss: 14.391881942749023\r\nStep 70, loss: 13.262972831726074\r\nStep 71, loss: 14.027960777282715\r\nStep 72, loss: 12.899559020996094\r\nStep 73, loss: 13.606071472167969\r\nStep 74, loss: 13.505942344665527\r\nStep 75, loss: 12.73733139038086\r\nStep 76, loss: 13.731353759765625\r\nStep 77, loss: 12.609628677368164\r\nStep 78, loss: 13.05951976776123\r\nStep 79, loss: 12.82889461517334\r\nStep 80, loss: 12.381841659545898\r\nStep 81, loss: 12.184906005859375\r\nStep 82, loss: 12.261787414550781\r\nStep 83, loss: 12.400224685668945\r\nStep 84, loss: 12.114161491394043\r\nStep 85, loss: 12.412161827087402\r\nStep 86, loss: 11.783485412597656\r\nStep 87, loss: 11.766879081726074\r\nStep 88, loss: 11.893105506896973\r\nStep 89, loss: 11.720884323120117\r\nStep 90, loss: 11.85186767578125\r\nStep 91, loss: 11.742947578430176\r\nStep 92, loss: 11.327075004577637\r\nStep 93, loss: 11.285242080688477\r\nStep 94, loss: 11.404678344726562\r\nStep 95, loss: 11.971321105957031\r\nStep 96, loss: 11.94084644317627\r\nStep 97, loss: 11.925477027893066\r\nStep 98, loss: 11.581640243530273\r\nStep 99, loss: 10.887056350708008\r\nStep 100, loss: 11.63808822631836\r\nStep 101, loss: 10.839936256408691\r\nStep 102, loss: 11.229722023010254\r\nStep 103, loss: 10.856170654296875\r\nStep 104, loss: 11.429742813110352\r\nStep 105, loss: 10.258073806762695\r\nStep 106, loss: 10.562714576721191\r\nStep 107, loss: 10.214714050292969\r\nStep 108, loss: 10.124898910522461\r\nStep 109, loss: 9.705013275146484\r\nStep 110, loss: 9.921828269958496\r\nStep 111, loss: 9.90079116821289\r\nStep 112, loss: 10.455109596252441\r\nStep 113, loss: 9.936418533325195\r\nStep 114, loss: 9.936537742614746\r\nStep 115, loss: 9.773576736450195\r\nStep 116, loss: 9.676142692565918\r\nStep 117, loss: 9.945345878601074\r\nStep 118, loss: 9.923107147216797\r\nStep 0, loss: 18.776222229003906\r\nStep 1, loss: 18.819902420043945\r\nStep 2, loss: 18.791563034057617\r\nStep 3, loss: 18.668718338012695\r\nStep 4, loss: 18.80548667907715\r\nStep 5, loss: 18.81717300415039\r\nStep 6, loss: 18.450780868530273\r\nStep 7, loss: 18.754093170166016\r\nStep 8, loss: 18.730485916137695\r\nStep 9, loss: 18.4219913482666\r\nStep 10, loss: 18.53896713256836\r\nStep 11, loss: 18.433452606201172\r\nStep 12, loss: 18.339122772216797\r\nStep 13, loss: 18.459253311157227\r\nStep 14, loss: 18.409881591796875\r\nStep 15, loss: 18.35429573059082\r\nStep 16, loss: 18.237462997436523\r\nStep 17, loss: 18.09574317932129\r\nStep 18, loss: 17.840925216674805\r\nStep 19, loss: 18.041223526000977\r\nStep 20, loss: 17.888032913208008\r\nStep 21, loss: 17.768489837646484\r\nStep 22, loss: 17.63291358947754\r\nStep 23, loss: 17.79960060119629\r\nStep 24, loss: 17.668100357055664\r\nStep 25, loss: 17.608400344848633\r\nStep 26, loss: 17.400070190429688\r\nStep 27, loss: 17.27320671081543\r\nStep 28, loss: 17.558292388916016\r\nStep 29, loss: 17.401700973510742\r\nStep 119, loss: 9.550374984741211\r\nStep 120, loss: 10.050488471984863\r\nStep 30, loss: 17.43343734741211\r\nStep 31, loss: 16.896129608154297\r\nStep 32, loss: 16.79721450805664\r\nStep 33, loss: 16.77494239807129\r\nStep 34, loss: 17.045551300048828\r\nStep 35, loss: 16.608722686767578\r\nStep 36, loss: 16.647741317749023\r\nStep 37, loss: 16.336868286132812\r\nStep 38, loss: 16.651290893554688\r\nStep 39, loss: 16.270305633544922\r\nStep 40, loss: 16.28514289855957\r\nStep 41, loss: 15.932032585144043\r\nStep 42, loss: 16.085798263549805\r\nStep 43, loss: 15.853764533996582\r\nStep 44, loss: 16.237834930419922\r\nStep 45, loss: 15.47584056854248\r\nStep 46, loss: 15.609787940979004\r\nStep 47, loss: 15.486068725585938\r\nStep 48, loss: 15.258798599243164\r\nStep 49, loss: 15.545014381408691\r\nStep 50, loss: 16.061813354492188\r\nStep 51, loss: 15.366859436035156\r\nStep 52, loss: 14.827864646911621\r\nStep 53, loss: 14.727177619934082\r\nStep 54, loss: 14.848566055297852\r\nStep 55, loss: 14.447760581970215\r\nStep 56, loss: 15.252120018005371\r\nStep 57, loss: 14.218976020812988\r\nStep 58, loss: 14.730356216430664\r\nStep 59, loss: 14.582534790039062\r\nStep 121, loss: 9.169127464294434\r\nStep 122, loss: 9.096636772155762\r\nStep 123, loss: 9.86887264251709\r\nStep 124, loss: 9.788836479187012\r\nStep 125, loss: 9.638827323913574\r\nStep 126, loss: 9.093913078308105\r\nStep 127, loss: 9.269700050354004\r\nStep 128, loss: 9.940503120422363\r\nStep 129, loss: 8.495964050292969\r\nStep 130, loss: 9.110383033752441\r\nStep 131, loss: 9.079448699951172\r\nStep 132, loss: 9.014200210571289\r\nStep 133, loss: 8.72136402130127\r\nStep 134, loss: 9.308770179748535\r\nStep 135, loss: 8.615967750549316\r\nStep 136, loss: 8.500845909118652\r\nStep 137, loss: 8.710356712341309\r\nStep 138, loss: 8.530448913574219\r\nStep 139, loss: 9.046561241149902\r\nStep 140, loss: 8.546600341796875\r\nStep 141, loss: 9.229374885559082\r\nStep 142, loss: 8.218626022338867\r\nStep 143, loss: 8.23629093170166\r\nStep 144, loss: 8.220922470092773\r\nStep 145, loss: 8.11863899230957\r\nStep 146, loss: 8.13656997680664\r\nStep 147, loss: 8.036163330078125\r\nStep 148, loss: 7.980085372924805\r\nStep 149, loss: 7.8308844566345215\r\nStep 150, loss: 7.997012615203857\r\nStep 60, loss: 13.967232704162598\r\nStep 61, loss: 13.71136474609375\r\nStep 62, loss: 13.708770751953125\r\nStep 63, loss: 14.0950345993042\r\nStep 64, loss: 13.713918685913086\r\nStep 65, loss: 14.868953704833984\r\nStep 66, loss: 14.468207359313965\r\nStep 67, loss: 14.585230827331543\r\nStep 68, loss: 14.341510772705078\r\nStep 69, loss: 14.391881942749023\r\nStep 70, loss: 13.262972831726074\r\nStep 71, loss: 14.027960777282715\r\nStep 72, loss: 12.899559020996094\r\nStep 73, loss: 13.606071472167969\r\nStep 74, loss: 13.505942344665527\r\nStep 75, loss: 12.73733139038086\r\nStep 76, loss: 13.731353759765625\r\nStep 77, loss: 12.609628677368164\r\nStep 78, loss: 13.05951976776123\r\nStep 79, loss: 12.82889461517334\r\nStep 80, loss: 12.381841659545898\r\nStep 81, loss: 12.184906005859375\r\nStep 82, loss: 12.261787414550781\r\nStep 83, loss: 12.400224685668945\r\nStep 84, loss: 12.114161491394043\r\nStep 85, loss: 12.412161827087402\r\nStep 86, loss: 11.783485412597656\r\nStep 87, loss: 11.766879081726074\r\nStep 88, loss: 11.893105506896973\r\nStep 89, loss: 11.720884323120117\r\nStep 151, loss: 8.312012672424316\r\nStep 152, loss: 8.257838249206543\r\nStep 153, loss: 7.67844820022583\r\nStep 154, loss: 8.097171783447266\r\nStep 155, loss: 7.956625461578369\r\nStep 156, loss: 7.661460876464844\r\nStep 157, loss: 7.60241174697876\r\nStep 158, loss: 7.5734171867370605\r\nStep 159, loss: 7.4999213218688965\r\nStep 160, loss: 7.889724254608154\r\nStep 161, loss: 7.616578578948975\r\nStep 162, loss: 7.669461250305176\r\nStep 163, loss: 7.583798408508301\r\nStep 164, loss: 7.573363780975342\r\nStep 165, loss: 7.425256252288818\r\nStep 166, loss: 7.534543991088867\r\nStep 167, loss: 7.315162658691406\r\nStep 168, loss: 7.1738691329956055\r\nStep 169, loss: 7.142663955688477\r\nStep 170, loss: 7.435649871826172\r\nStep 171, loss: 7.096683502197266\r\nStep 172, loss: 7.248353481292725\r\nStep 173, loss: 7.270962715148926\r\nStep 174, loss: 7.3473687171936035\r\nStep 175, loss: 6.999227523803711\r\nStep 176, loss: 7.4188151359558105\r\nStep 177, loss: 7.223378658294678\r\nStep 178, loss: 7.144739627838135\r\nStep 179, loss: 6.7800211906433105\r\nStep 90, loss: 11.85186767578125\r\nStep 91, loss: 11.742947578430176\r\nStep 92, loss: 11.327075004577637\r\nStep 93, loss: 11.285242080688477\r\nStep 94, loss: 11.404678344726562\r\nStep 95, loss: 11.971321105957031\r\nStep 96, loss: 11.94084644317627\r\nStep 97, loss: 11.925477027893066\r\nStep 98, loss: 11.581640243530273\r\nStep 99, loss: 10.887056350708008\r\nStep 100, loss: 11.63808822631836\r\nStep 101, loss: 10.839936256408691\r\nStep 102, loss: 11.229722023010254\r\nStep 103, loss: 10.856170654296875\r\nStep 104, loss: 11.429742813110352\r\nStep 105, loss: 10.258073806762695\r\nStep 106, loss: 10.562714576721191\r\nStep 107, loss: 10.214714050292969\r\nStep 108, loss: 10.124898910522461\r\nStep 109, loss: 9.705013275146484\r\nStep 110, loss: 9.921828269958496\r\nStep 111, loss: 9.90079116821289\r\nStep 112, loss: 10.455109596252441\r\nStep 113, loss: 9.936418533325195\r\nStep 114, loss: 9.936537742614746\r\nStep 115, loss: 9.773576736450195\r\nStep 116, loss: 9.676142692565918\r\nStep 117, loss: 9.945345878601074\r\nStep 118, loss: 9.923107147216797\r\nStep 180, loss: 7.156866073608398\r\nStep 181, loss: 6.900200843811035\r\nStep 182, loss: 7.0645365715026855\r\nStep 183, loss: 7.188953399658203\r\nStep 184, loss: 6.891933917999268\r\nStep 185, loss: 6.902243614196777\r\nStep 186, loss: 6.5922651290893555\r\nStep 187, loss: 7.096951007843018\r\nStep 188, loss: 6.9083991050720215\r\nStep 189, loss: 6.90083646774292\r\nStep 190, loss: 6.980688095092773\r\nStep 191, loss: 6.729495048522949\r\nStep 192, loss: 6.864687442779541\r\nStep 193, loss: 6.7847514152526855\r\nStep 194, loss: 6.816504955291748\r\nStep 195, loss: 6.671535968780518\r\nStep 196, loss: 6.762739181518555\r\nStep 197, loss: 6.497260570526123\r\nStep 198, loss: 7.014808654785156\r\nStep 199, loss: 6.6613945960998535\r\nStep 200, loss: 6.8338422775268555\r\nStep 201, loss: 6.816775798797607\r\nStep 202, loss: 6.6137614250183105\r\nStep 203, loss: 6.697750568389893\r\nStep 204, loss: 6.821736812591553\r\nStep 205, loss: 6.9376959800720215\r\nStep 206, loss: 6.731281757354736\r\nStep 207, loss: 6.579183578491211\r\nStep 208, loss: 6.68570613861084\r\nStep 119, loss: 9.550374984741211\r\nStep 120, loss: 10.050488471984863\r\nStep 209, loss: 6.527091026306152\r\nStep 210, loss: 6.523392200469971\r\nStep 211, loss: 6.566160678863525\r\nStep 212, loss: 6.912597179412842\r\nStep 213, loss: 6.662230968475342\r\nStep 214, loss: 6.962000846862793\r\nStep 215, loss: 6.555893898010254\r\nStep 216, loss: 6.651054382324219\r\nStep 217, loss: 6.657430648803711\r\nStep 218, loss: 6.917163372039795\r\nStep 219, loss: 6.651887893676758\r\nStep 220, loss: 6.725634574890137\r\nStep 221, loss: 6.670478343963623\r\nStep 222, loss: 6.46140193939209\r\nStep 223, loss: 6.792141437530518\r\nStep 224, loss: 6.598993301391602\r\nStep 225, loss: 6.382656097412109\r\nStep 226, loss: 6.67178201675415\r\nStep 227, loss: 6.364635467529297\r\nStep 228, loss: 6.6085052490234375\r\nStep 229, loss: 6.653966903686523\r\nStep 230, loss: 6.554793834686279\r\nStep 231, loss: 6.588866710662842\r\nStep 232, loss: 6.585617542266846\r\nStep 233, loss: 6.507537364959717\r\nStep 234, loss: 6.644289493560791\r\nStep 235, loss: 6.5886993408203125\r\nStep 236, loss: 6.520603656768799\r\nStep 237, loss: 6.404054164886475\r\nStep 238, loss: 6.414518356323242\r\nStep 121, loss: 9.169127464294434\r\nStep 122, loss: 9.096636772155762\r\nStep 123, loss: 9.86887264251709\r\nStep 124, loss: 9.788836479187012\r\nStep 125, loss: 9.638827323913574\r\nStep 126, loss: 9.093913078308105\r\nStep 127, loss: 9.269700050354004\r\nStep 128, loss: 9.940503120422363\r\nStep 129, loss: 8.495964050292969\r\nStep 130, loss: 9.110383033752441\r\nStep 131, loss: 9.079448699951172\r\nStep 132, loss: 9.014200210571289\r\nStep 133, loss: 8.72136402130127\r\nStep 134, loss: 9.308770179748535\r\nStep 135, loss: 8.615967750549316\r\nStep 136, loss: 8.500845909118652\r\nStep 137, loss: 8.710356712341309\r\nStep 138, loss: 8.530448913574219\r\nStep 139, loss: 9.046561241149902\r\nStep 140, loss: 8.546600341796875\r\nStep 141, loss: 9.229374885559082\r\nStep 142, loss: 8.218626022338867\r\nStep 143, loss: 8.23629093170166\r\nStep 144, loss: 8.220922470092773\r\nStep 145, loss: 8.11863899230957\r\nStep 146, loss: 8.13656997680664\r\nStep 147, loss: 8.036163330078125\r\nStep 148, loss: 7.980085372924805\r\nStep 149, loss: 7.8308844566345215\r\nStep 150, loss: 7.997012615203857\r\nStep 239, loss: 6.4341325759887695\r\nStep 240, loss: 6.528247833251953\r\nStep 151, loss: 8.312012672424316\r\nStep 152, loss: 8.257838249206543\r\nStep 153, loss: 7.67844820022583\r\nStep 154, loss: 8.097171783447266\r\nStep 155, loss: 7.956625461578369\r\nStep 156, loss: 7.661460876464844\r\nStep 157, loss: 7.60241174697876\r\nStep 158, loss: 7.5734171867370605\r\nStep 159, loss: 7.4999213218688965\r\nStep 160, loss: 7.889724254608154\r\nStep 161, loss: 7.616578578948975\r\nStep 162, loss: 7.669461250305176\r\nStep 163, loss: 7.583798408508301\r\nStep 164, loss: 7.573363780975342\r\nStep 165, loss: 7.425256252288818\r\nStep 166, loss: 7.534543991088867\r\nStep 167, loss: 7.315162658691406\r\nStep 168, loss: 7.1738691329956055\r\nStep 169, loss: 7.142663955688477\r\nStep 170, loss: 7.435649871826172\r\nStep 171, loss: 7.096683502197266\r\nStep 172, loss: 7.248353481292725\r\nStep 173, loss: 7.270962715148926\r\nStep 174, loss: 7.3473687171936035\r\nStep 175, loss: 6.999227523803711\r\nStep 176, loss: 7.4188151359558105\r\nStep 177, loss: 7.223378658294678\r\nStep 178, loss: 7.144739627838135\r\nStep 179, loss: 6.7800211906433105\r\nStep 0, loss: 18.776222229003906\r\nStep 1, loss: 18.819902420043945\r\nStep 2, loss: 18.791563034057617\r\nStep 3, loss: 18.668718338012695\r\nStep 4, loss: 18.80548667907715\r\nStep 5, loss: 18.81717300415039\r\nStep 6, loss: 18.450780868530273\r\nStep 7, loss: 18.754093170166016\r\nStep 8, loss: 18.730485916137695\r\nStep 9, loss: 18.4219913482666\r\nStep 10, loss: 18.53896713256836\r\nStep 11, loss: 18.433452606201172\r\nStep 12, loss: 18.339122772216797\r\nStep 13, loss: 18.459253311157227\r\nStep 14, loss: 18.409881591796875\r\nStep 15, loss: 18.35429573059082\r\nStep 16, loss: 18.237462997436523\r\nStep 17, loss: 18.09574317932129\r\nStep 18, loss: 17.840925216674805\r\nStep 19, loss: 18.041223526000977\r\nStep 20, loss: 17.888032913208008\r\nStep 21, loss: 17.768489837646484\r\nStep 22, loss: 17.63291358947754\r\nStep 23, loss: 17.79960060119629\r\nStep 24, loss: 17.668100357055664\r\nStep 25, loss: 17.608400344848633\r\nStep 26, loss: 17.400070190429688\r\nStep 27, loss: 17.27320671081543\r\nStep 28, loss: 17.558292388916016\r\nStep 29, loss: 17.401700973510742\r\nStep 180, loss: 7.156866073608398\r\nStep 181, loss: 6.900200843811035\r\nStep 182, loss: 7.0645365715026855\r\nStep 183, loss: 7.188953399658203\r\nStep 184, loss: 6.891933917999268\r\nStep 185, loss: 6.902243614196777\r\nStep 186, loss: 6.5922651290893555\r\nStep 187, loss: 7.096951007843018\r\nStep 188, loss: 6.9083991050720215\r\nStep 189, loss: 6.90083646774292\r\nStep 190, loss: 6.980688095092773\r\nStep 191, loss: 6.729495048522949\r\nStep 192, loss: 6.864687442779541\r\nStep 193, loss: 6.7847514152526855\r\nStep 194, loss: 6.816504955291748\r\nStep 195, loss: 6.671535968780518\r\nStep 196, loss: 6.762739181518555\r\nStep 197, loss: 6.497260570526123\r\nStep 198, loss: 7.014808654785156\r\nStep 199, loss: 6.6613945960998535\r\nStep 200, loss: 6.8338422775268555\r\nStep 201, loss: 6.816775798797607\r\nStep 202, loss: 6.6137614250183105\r\nStep 203, loss: 6.697750568389893\r\nStep 204, loss: 6.821736812591553\r\nStep 205, loss: 6.9376959800720215\r\nStep 206, loss: 6.731281757354736\r\nStep 207, loss: 6.579183578491211\r\nStep 208, loss: 6.68570613861084\r\nStep 30, loss: 17.43343734741211\r\nStep 31, loss: 16.896129608154297\r\nStep 32, loss: 16.79721450805664\r\nStep 33, loss: 16.77494239807129\r\nStep 34, loss: 17.045551300048828\r\nStep 35, loss: 16.608722686767578\r\nStep 36, loss: 16.647741317749023\r\nStep 37, loss: 16.336868286132812\r\nStep 38, loss: 16.651290893554688\r\nStep 39, loss: 16.270305633544922\r\nStep 40, loss: 16.28514289855957\r\nStep 41, loss: 15.932032585144043\r\nStep 42, loss: 16.085798263549805\r\nStep 43, loss: 15.853764533996582\r\nStep 44, loss: 16.237834930419922\r\nStep 45, loss: 15.47584056854248\r\nStep 46, loss: 15.609787940979004\r\nStep 47, loss: 15.486068725585938\r\nStep 48, loss: 15.258798599243164\r\nStep 49, loss: 15.545014381408691\r\nStep 50, loss: 16.061813354492188\r\nStep 51, loss: 15.366859436035156\r\nStep 52, loss: 14.827864646911621\r\nStep 53, loss: 14.727177619934082\r\nStep 54, loss: 14.848566055297852\r\nStep 55, loss: 14.447760581970215\r\nStep 56, loss: 15.252120018005371\r\nStep 57, loss: 14.218976020812988\r\nStep 58, loss: 14.730356216430664\r\nStep 59, loss: 14.582534790039062\r\nStep 209, loss: 6.527091026306152\r\nStep 210, loss: 6.523392200469971\r\nStep 211, loss: 6.566160678863525\r\nStep 212, loss: 6.912597179412842\r\nStep 213, loss: 6.662230968475342\r\nStep 214, loss: 6.962000846862793\r\nStep 215, loss: 6.555893898010254\r\nStep 216, loss: 6.651054382324219\r\nStep 217, loss: 6.657430648803711\r\nStep 218, loss: 6.917163372039795\r\nStep 219, loss: 6.651887893676758\r\nStep 220, loss: 6.725634574890137\r\nStep 221, loss: 6.670478343963623\r\nStep 222, loss: 6.46140193939209\r\nStep 223, loss: 6.792141437530518\r\nStep 224, loss: 6.598993301391602\r\nStep 225, loss: 6.382656097412109\r\nStep 226, loss: 6.67178201675415\r\nStep 227, loss: 6.364635467529297\r\nStep 228, loss: 6.6085052490234375\r\nStep 229, loss: 6.653966903686523\r\nStep 230, loss: 6.554793834686279\r\nStep 231, loss: 6.588866710662842\r\nStep 232, loss: 6.585617542266846\r\nStep 233, loss: 6.507537364959717\r\nStep 234, loss: 6.644289493560791\r\nStep 235, loss: 6.5886993408203125\r\nStep 236, loss: 6.520603656768799\r\nStep 237, loss: 6.404054164886475\r\nStep 238, loss: 6.414518356323242\r\nStep 60, loss: 13.967232704162598\r\nStep 61, loss: 13.71136474609375\r\nStep 62, loss: 13.708770751953125\r\nStep 63, loss: 14.0950345993042\r\nStep 64, loss: 13.713918685913086\r\nStep 65, loss: 14.868953704833984\r\nStep 66, loss: 14.468207359313965\r\nStep 67, loss: 14.585230827331543\r\nStep 68, loss: 14.341510772705078\r\nStep 69, loss: 14.391881942749023\r\nStep 70, loss: 13.262972831726074\r\nStep 71, loss: 14.027960777282715\r\nStep 72, loss: 12.899559020996094\r\nStep 73, loss: 13.606071472167969\r\nStep 74, loss: 13.505942344665527\r\nStep 75, loss: 12.73733139038086\r\nStep 76, loss: 13.731353759765625\r\nStep 77, loss: 12.609628677368164\r\nStep 78, loss: 13.05951976776123\r\nStep 79, loss: 12.82889461517334\r\nStep 80, loss: 12.381841659545898\r\nStep 81, loss: 12.184906005859375\r\nStep 82, loss: 12.261787414550781\r\nStep 83, loss: 12.400224685668945\r\nStep 84, loss: 12.114161491394043\r\nStep 85, loss: 12.412161827087402\r\nStep 86, loss: 11.783485412597656\r\nStep 87, loss: 11.766879081726074\r\nStep 88, loss: 11.893105506896973\r\nStep 89, loss: 11.720884323120117\r\nStep 239, loss: 6.4341325759887695\r\nStep 240, loss: 6.528247833251953\r\nStep 90, loss: 11.85186767578125\r\nStep 91, loss: 11.742947578430176\r\nStep 92, loss: 11.327075004577637\r\nStep 93, loss: 11.285242080688477\r\nStep 94, loss: 11.404678344726562\r\nStep 95, loss: 11.971321105957031\r\nStep 96, loss: 11.94084644317627\r\nStep 97, loss: 11.925477027893066\r\nStep 98, loss: 11.581640243530273\r\nStep 99, loss: 10.887056350708008\r\nStep 100, loss: 11.63808822631836\r\nStep 101, loss: 10.839936256408691\r\nStep 102, loss: 11.229722023010254\r\nStep 103, loss: 10.856170654296875\r\nStep 104, loss: 11.429742813110352\r\nStep 105, loss: 10.258073806762695\r\nStep 106, loss: 10.562714576721191\r\nStep 107, loss: 10.214714050292969\r\nStep 108, loss: 10.124898910522461\r\nStep 109, loss: 9.705013275146484\r\nStep 110, loss: 9.921828269958496\r\nStep 111, loss: 9.90079116821289\r\nStep 112, loss: 10.455109596252441\r\nStep 113, loss: 9.936418533325195\r\nStep 114, loss: 9.936537742614746\r\nStep 115, loss: 9.773576736450195\r\nStep 116, loss: 9.676142692565918\r\nStep 117, loss: 9.945345878601074\r\nStep 118, loss: 9.923107147216797\r\nStep 119, loss: 9.550374984741211\r\nStep 120, loss: 10.050488471984863\r\nStep 0, loss: 18.776222229003906\r\nStep 1, loss: 18.819902420043945\r\nStep 2, loss: 18.791563034057617\r\nStep 3, loss: 18.668718338012695\r\nStep 4, loss: 18.80548667907715\r\nStep 5, loss: 18.81717300415039\r\nStep 6, loss: 18.450780868530273\r\nStep 7, loss: 18.754093170166016\r\nStep 8, loss: 18.730485916137695\r\nStep 9, loss: 18.4219913482666\r\nStep 10, loss: 18.53896713256836\r\nStep 11, loss: 18.433452606201172\r\nStep 12, loss: 18.339122772216797\r\nStep 13, loss: 18.459253311157227\r\nStep 14, loss: 18.409881591796875\r\nStep 15, loss: 18.35429573059082\r\nStep 16, loss: 18.237462997436523\r\nStep 17, loss: 18.09574317932129\r\nStep 18, loss: 17.840925216674805\r\nStep 19, loss: 18.041223526000977\r\nStep 20, loss: 17.888032913208008\r\nStep 21, loss: 17.768489837646484\r\nStep 22, loss: 17.63291358947754\r\nStep 23, loss: 17.79960060119629\r\nStep 24, loss: 17.668100357055664\r\nStep 25, loss: 17.608400344848633\r\nStep 26, loss: 17.400070190429688\r\nStep 27, loss: 17.27320671081543\r\nStep 28, loss: 17.558292388916016\r\nStep 29, loss: 17.401700973510742\r\nStep 30, loss: 17.43343734741211\r\nStep 31, loss: 16.896129608154297\r\nStep 32, loss: 16.79721450805664\r\nStep 33, loss: 16.77494239807129\r\nStep 34, loss: 17.045551300048828\r\nStep 35, loss: 16.608722686767578\r\nStep 36, loss: 16.647741317749023\r\nStep 37, loss: 16.336868286132812\r\nStep 38, loss: 16.651290893554688\r\nStep 39, loss: 16.270305633544922\r\nStep 40, loss: 16.28514289855957\r\nStep 41, loss: 15.932032585144043\r\nStep 42, loss: 16.085798263549805\r\nStep 43, loss: 15.853764533996582\r\nStep 44, loss: 16.237834930419922\r\nStep 45, loss: 15.47584056854248\r\nStep 46, loss: 15.609787940979004\r\nStep 47, loss: 15.486068725585938\r\nStep 48, loss: 15.258798599243164\r\nStep 49, loss: 15.545014381408691\r\nStep 50, loss: 16.061813354492188\r\nStep 51, loss: 15.366859436035156\r\nStep 52, loss: 14.827864646911621\r\nStep 53, loss: 14.727177619934082\r\nStep 54, loss: 14.848566055297852\r\nStep 55, loss: 14.447760581970215\r\nStep 56, loss: 15.252120018005371\r\nStep 57, loss: 14.218976020812988\r\nStep 58, loss: 14.730356216430664\r\nStep 59, loss: 14.582534790039062\r\nStep 60, loss: 13.967232704162598\r\nStep 61, loss: 13.71136474609375\r\nStep 62, loss: 13.708770751953125\r\nStep 63, loss: 14.0950345993042\r\nStep 64, loss: 13.713918685913086\r\nStep 65, loss: 14.868953704833984\r\nStep 66, loss: 14.468207359313965\r\nStep 67, loss: 14.585230827331543\r\nStep 68, loss: 14.341510772705078\r\nStep 69, loss: 14.391881942749023\r\nStep 70, loss: 13.262972831726074\r\nStep 71, loss: 14.027960777282715\r\nStep 72, loss: 12.899559020996094\r\nStep 73, loss: 13.606071472167969\r\nStep 74, loss: 13.505942344665527\r\nStep 75, loss: 12.73733139038086\r\nStep 76, loss: 13.731353759765625\r\nStep 77, loss: 12.609628677368164\r\nStep 78, loss: 13.05951976776123\r\nStep 79, loss: 12.82889461517334\r\nStep 80, loss: 12.381841659545898\r\nStep 81, loss: 12.184906005859375\r\nStep 82, loss: 12.261787414550781\r\nStep 83, loss: 12.400224685668945\r\nStep 84, loss: 12.114161491394043\r\nStep 85, loss: 12.412161827087402\r\nStep 86, loss: 11.783485412597656\r\nStep 87, loss: 11.766879081726074\r\nStep 88, loss: 11.893105506896973\r\nStep 89, loss: 11.720884323120117\r\nStep 90, loss: 11.85186767578125\r\nStep 91, loss: 11.742947578430176\r\nStep 92, loss: 11.327075004577637\r\nStep 93, loss: 11.285242080688477\r\nStep 94, loss: 11.404678344726562\r\nStep 95, loss: 11.971321105957031\r\nStep 96, loss: 11.94084644317627\r\nStep 97, loss: 11.925477027893066\r\nStep 98, loss: 11.581640243530273\r\nStep 99, loss: 10.887056350708008\r\nStep 100, loss: 11.63808822631836\r\nStep 101, loss: 10.839936256408691\r\nStep 102, loss: 11.229722023010254\r\nStep 103, loss: 10.856170654296875\r\nStep 104, loss: 11.429742813110352\r\nStep 105, loss: 10.258073806762695\r\nStep 106, loss: 10.562714576721191\r\nStep 107, loss: 10.214714050292969\r\nStep 108, loss: 10.124898910522461\r\nStep 109, loss: 9.705013275146484\r\nStep 110, loss: 9.921828269958496\r\nStep 111, loss: 9.90079116821289\r\nStep 112, loss: 10.455109596252441\r\nStep 113, loss: 9.936418533325195\r\nStep 114, loss: 9.936537742614746\r\nStep 115, loss: 9.773576736450195\r\nStep 116, loss: 9.676142692565918\r\nStep 117, loss: 9.945345878601074\r\nStep 118, loss: 9.923107147216797\r\nStep 119, loss: 9.550374984741211\r\nStep 120, loss: 10.050488471984863\r\nStep 121, loss: 9.169127464294434\r\nStep 122, loss: 9.096636772155762\r\nStep 123, loss: 9.86887264251709\r\nStep 124, loss: 9.788836479187012\r\nStep 125, loss: 9.638827323913574\r\nStep 126, loss: 9.093913078308105\r\nStep 127, loss: 9.269700050354004\r\nStep 128, loss: 9.940503120422363\r\nStep 129, loss: 8.495964050292969\r\nStep 130, loss: 9.110383033752441\r\nStep 131, loss: 9.079448699951172\r\nStep 132, loss: 9.014200210571289\r\nStep 133, loss: 8.72136402130127\r\nStep 134, loss: 9.308770179748535\r\nStep 135, loss: 8.615967750549316\r\nStep 136, loss: 8.500845909118652\r\nStep 137, loss: 8.710356712341309\r\nStep 138, loss: 8.530448913574219\r\nStep 139, loss: 9.046561241149902\r\nStep 140, loss: 8.546600341796875\r\nStep 141, loss: 9.229374885559082\r\nStep 142, loss: 8.218626022338867\r\nStep 143, loss: 8.23629093170166\r\nStep 144, loss: 8.220922470092773\r\nStep 145, loss: 8.11863899230957\r\nStep 146, loss: 8.13656997680664\r\nStep 147, loss: 8.036163330078125\r\nStep 148, loss: 7.980085372924805\r\nStep 149, loss: 7.8308844566345215\r\nStep 150, loss: 7.997012615203857\r\nStep 151, loss: 8.312012672424316\r\nStep 152, loss: 8.257838249206543\r\nStep 153, loss: 7.67844820022583\r\nStep 154, loss: 8.097171783447266\r\nStep 155, loss: 7.956625461578369\r\nStep 156, loss: 7.661460876464844\r\nStep 157, loss: 7.60241174697876\r\nStep 158, loss: 7.5734171867370605\r\nStep 159, loss: 7.4999213218688965\r\nStep 160, loss: 7.889724254608154\r\nStep 161, loss: 7.616578578948975\r\nStep 162, loss: 7.669461250305176\r\nStep 163, loss: 7.583798408508301\r\nStep 164, loss: 7.573363780975342\r\nStep 165, loss: 7.425256252288818\r\nStep 166, loss: 7.534543991088867\r\nStep 167, loss: 7.315162658691406\r\nStep 168, loss: 7.1738691329956055\r\nStep 169, loss: 7.142663955688477\r\nStep 170, loss: 7.435649871826172\r\nStep 171, loss: 7.096683502197266\r\nStep 172, loss: 7.248353481292725\r\nStep 173, loss: 7.270962715148926\r\nStep 174, loss: 7.3473687171936035\r\nStep 175, loss: 6.999227523803711\r\nStep 176, loss: 7.4188151359558105\r\nStep 177, loss: 7.223378658294678\r\nStep 178, loss: 7.144739627838135\r\nStep 179, loss: 6.7800211906433105\r\nStep 180, loss: 7.156866073608398\r\nStep 181, loss: 6.900200843811035\r\nStep 182, loss: 7.0645365715026855\r\nStep 183, loss: 7.188953399658203\r\nStep 184, loss: 6.891933917999268\r\nStep 185, loss: 6.902243614196777\r\nStep 186, loss: 6.5922651290893555\r\nStep 187, loss: 7.096951007843018\r\nStep 188, loss: 6.9083991050720215\r\nStep 189, loss: 6.90083646774292\r\nStep 190, loss: 6.980688095092773\r\nStep 191, loss: 6.729495048522949\r\nStep 192, loss: 6.864687442779541\r\nStep 193, loss: 6.7847514152526855\r\nStep 194, loss: 6.816504955291748\r\nStep 195, loss: 6.671535968780518\r\nStep 196, loss: 6.762739181518555\r\nStep 197, loss: 6.497260570526123\r\nStep 198, loss: 7.014808654785156\r\nStep 199, loss: 6.6613945960998535\r\nStep 200, loss: 6.8338422775268555\r\nStep 201, loss: 6.816775798797607\r\nStep 202, loss: 6.6137614250183105\r\nStep 203, loss: 6.697750568389893\r\nStep 204, loss: 6.821736812591553\r\nStep 205, loss: 6.9376959800720215\r\nStep 206, loss: 6.731281757354736\r\nStep 207, loss: 6.579183578491211\r\nStep 208, loss: 6.68570613861084\r\nStep 209, loss: 6.527091026306152\r\nStep 210, loss: 6.523392200469971\r\nStep 211, loss: 6.566160678863525\r\nStep 212, loss: 6.912597179412842\r\nStep 213, loss: 6.662230968475342\r\nStep 214, loss: 6.962000846862793\r\nStep 215, loss: 6.555893898010254\r\nStep 216, loss: 6.651054382324219\r\nStep 217, loss: 6.657430648803711\r\nStep 218, loss: 6.917163372039795\r\nStep 219, loss: 6.651887893676758\r\nStep 220, loss: 6.725634574890137\r\nStep 221, loss: 6.670478343963623\r\nStep 222, loss: 6.46140193939209\r\nStep 223, loss: 6.792141437530518\r\nStep 224, loss: 6.598993301391602\r\nStep 225, loss: 6.382656097412109\r\nStep 226, loss: 6.67178201675415\r\nStep 227, loss: 6.364635467529297\r\nStep 228, loss: 6.6085052490234375\r\nStep 229, loss: 6.653966903686523\r\nStep 230, loss: 6.554793834686279\r\nStep 231, loss: 6.588866710662842\r\nStep 232, loss: 6.585617542266846\r\nStep 233, loss: 6.507537364959717\r\nStep 234, loss: 6.644289493560791\r\nStep 235, loss: 6.5886993408203125\r\nStep 236, loss: 6.520603656768799\r\nStep 237, loss: 6.404054164886475\r\nStep 238, loss: 6.414518356323242\r\nStep 239, loss: 6.4341325759887695\r\nStep 240, loss: 6.528247833251953\r\nStep 121, loss: 9.169127464294434\r\nStep 122, loss: 9.096636772155762\r\nStep 123, loss: 9.86887264251709\r\nStep 124, loss: 9.788836479187012\r\nStep 125, loss: 9.638827323913574\r\nStep 126, loss: 9.093913078308105\r\nStep 127, loss: 9.269700050354004\r\nStep 128, loss: 9.940503120422363\r\nStep 129, loss: 8.495964050292969\r\nStep 130, loss: 9.110383033752441\r\nStep 131, loss: 9.079448699951172\r\nStep 132, loss: 9.014200210571289\r\nStep 133, loss: 8.72136402130127\r\nStep 134, loss: 9.308770179748535\r\nStep 135, loss: 8.615967750549316\r\nStep 136, loss: 8.500845909118652\r\nStep 137, loss: 8.710356712341309\r\nStep 138, loss: 8.530448913574219\r\nStep 139, loss: 9.046561241149902\r\nStep 140, loss: 8.546600341796875\r\nStep 141, loss: 9.229374885559082\r\nStep 142, loss: 8.218626022338867\r\nStep 143, loss: 8.23629093170166\r\nStep 144, loss: 8.220922470092773\r\nStep 145, loss: 8.11863899230957\r\nStep 146, loss: 8.13656997680664\r\nStep 147, loss: 8.036163330078125\r\nStep 148, loss: 7.980085372924805\r\nStep 149, loss: 7.8308844566345215\r\nStep 150, loss: 7.997012615203857\r\nStep 151, loss: 8.312012672424316\r\nStep 152, loss: 8.257838249206543\r\nStep 153, loss: 7.67844820022583\r\nStep 154, loss: 8.097171783447266\r\nStep 155, loss: 7.956625461578369\r\nStep 156, loss: 7.661460876464844\r\nStep 157, loss: 7.60241174697876\r\nStep 158, loss: 7.5734171867370605\r\nStep 159, loss: 7.4999213218688965\r\nStep 160, loss: 7.889724254608154\r\nStep 161, loss: 7.616578578948975\r\nStep 162, loss: 7.669461250305176\r\nStep 163, loss: 7.583798408508301\r\nStep 164, loss: 7.573363780975342\r\nStep 165, loss: 7.425256252288818\r\nStep 166, loss: 7.534543991088867\r\nStep 167, loss: 7.315162658691406\r\nStep 168, loss: 7.1738691329956055\r\nStep 169, loss: 7.142663955688477\r\nStep 170, loss: 7.435649871826172\r\nStep 171, loss: 7.096683502197266\r\nStep 172, loss: 7.248353481292725\r\nStep 173, loss: 7.270962715148926\r\nStep 174, loss: 7.3473687171936035\r\nStep 175, loss: 6.999227523803711\r\nStep 176, loss: 7.4188151359558105\r\nStep 177, loss: 7.223378658294678\r\nStep 178, loss: 7.144739627838135\r\nStep 179, loss: 6.7800211906433105\r\nStep 180, loss: 7.156866073608398\r\nStep 181, loss: 6.900200843811035\r\nStep 182, loss: 7.0645365715026855\r\nStep 183, loss: 7.188953399658203\r\nStep 184, loss: 6.891933917999268\r\nStep 185, loss: 6.902243614196777\r\nStep 186, loss: 6.5922651290893555\r\nStep 187, loss: 7.096951007843018\r\nStep 188, loss: 6.9083991050720215\r\nStep 189, loss: 6.90083646774292\r\nStep 190, loss: 6.980688095092773\r\nStep 191, loss: 6.729495048522949\r\nStep 192, loss: 6.864687442779541\r\nStep 193, loss: 6.7847514152526855\r\nStep 194, loss: 6.816504955291748\r\nStep 195, loss: 6.671535968780518\r\nStep 196, loss: 6.762739181518555\r\nStep 197, loss: 6.497260570526123\r\nStep 198, loss: 7.014808654785156\r\nStep 199, loss: 6.6613945960998535\r\nStep 200, loss: 6.8338422775268555\r\nStep 201, loss: 6.816775798797607\r\nStep 202, loss: 6.6137614250183105\r\nStep 203, loss: 6.697750568389893\r\nStep 204, loss: 6.821736812591553\r\nStep 205, loss: 6.9376959800720215\r\nStep 206, loss: 6.731281757354736\r\nStep 207, loss: 6.579183578491211\r\nStep 208, loss: 6.68570613861084\r\nStep 209, loss: 6.527091026306152\r\nStep 210, loss: 6.523392200469971\r\nStep 211, loss: 6.566160678863525\r\nStep 212, loss: 6.912597179412842\r\nStep 213, loss: 6.662230968475342\r\nStep 214, loss: 6.962000846862793\r\nStep 215, loss: 6.555893898010254\r\nStep 216, loss: 6.651054382324219\r\nStep 217, loss: 6.657430648803711\r\nStep 218, loss: 6.917163372039795\r\nStep 219, loss: 6.651887893676758\r\nStep 220, loss: 6.725634574890137\r\nStep 221, loss: 6.670478343963623\r\nStep 222, loss: 6.46140193939209\r\nStep 223, loss: 6.792141437530518\r\nStep 224, loss: 6.598993301391602\r\nStep 225, loss: 6.382656097412109\r\nStep 226, loss: 6.67178201675415\r\nStep 227, loss: 6.364635467529297\r\nStep 228, loss: 6.6085052490234375\r\nStep 229, loss: 6.653966903686523\r\nStep 230, loss: 6.554793834686279\r\nStep 231, loss: 6.588866710662842\r\nStep 232, loss: 6.585617542266846\r\nStep 233, loss: 6.507537364959717\r\nStep 234, loss: 6.644289493560791\r\nStep 235, loss: 6.5886993408203125\r\nStep 236, loss: 6.520603656768799\r\nStep 237, loss: 6.404054164886475\r\nStep 238, loss: 6.414518356323242\r\nStep 239, loss: 6.4341325759887695\r\nStep 240, loss: 6.528247833251953\r\nStep 0, loss: 18.776222229003906\r\nStep 1, loss: 18.819902420043945\r\nStep 2, loss: 18.791563034057617\r\nStep 3, loss: 18.668718338012695\r\nStep 4, loss: 18.80548667907715\r\nStep 5, loss: 18.81717300415039\r\nStep 6, loss: 18.450780868530273\r\nStep 7, loss: 18.754093170166016\r\nStep 8, loss: 18.730485916137695\r\nStep 9, loss: 18.4219913482666\r\nStep 10, loss: 18.53896713256836\r\nStep 11, loss: 18.433452606201172\r\nStep 12, loss: 18.339122772216797\r\nStep 13, loss: 18.459253311157227\r\nStep 14, loss: 18.409881591796875\r\nStep 15, loss: 18.35429573059082\r\nStep 16, loss: 18.237462997436523\r\nStep 17, loss: 18.09574317932129\r\nStep 18, loss: 17.840925216674805\r\nStep 19, loss: 18.041223526000977\r\nStep 20, loss: 17.888032913208008\r\nStep 21, loss: 17.768489837646484\r\nStep 22, loss: 17.63291358947754\r\nStep 23, loss: 17.79960060119629\r\nStep 24, loss: 17.668100357055664\r\nStep 25, loss: 17.608400344848633\r\nStep 26, loss: 17.400070190429688\r\nStep 27, loss: 17.27320671081543\r\nStep 28, loss: 17.558292388916016\r\nStep 29, loss: 17.401700973510742\r\nStep 0, loss: 18.776222229003906\r\nStep 1, loss: 18.819902420043945\r\nStep 2, loss: 18.791563034057617\r\nStep 3, loss: 18.668718338012695\r\nStep 4, loss: 18.80548667907715\r\nStep 5, loss: 18.81717300415039\r\nStep 6, loss: 18.450780868530273\r\nStep 7, loss: 18.754093170166016\r\nStep 8, loss: 18.730485916137695\r\nStep 9, loss: 18.4219913482666\r\nStep 10, loss: 18.53896713256836\r\nStep 11, loss: 18.433452606201172\r\nStep 12, loss: 18.339122772216797\r\nStep 13, loss: 18.459253311157227\r\nStep 14, loss: 18.409881591796875\r\nStep 15, loss: 18.35429573059082\r\nStep 16, loss: 18.237462997436523\r\nStep 17, loss: 18.09574317932129\r\nStep 18, loss: 17.840925216674805\r\nStep 19, loss: 18.041223526000977\r\nStep 20, loss: 17.888032913208008\r\nStep 21, loss: 17.768489837646484\r\nStep 22, loss: 17.63291358947754\r\nStep 23, loss: 17.79960060119629\r\nStep 24, loss: 17.668100357055664\r\nStep 25, loss: 17.608400344848633\r\nStep 26, loss: 17.400070190429688\r\nStep 27, loss: 17.27320671081543\r\nStep 28, loss: 17.558292388916016\r\nStep 29, loss: 17.401700973510742\r\nStep 30, loss: 17.43343734741211\r\nStep 31, loss: 16.896129608154297\r\nStep 32, loss: 16.79721450805664\r\nStep 33, loss: 16.77494239807129\r\nStep 34, loss: 17.045551300048828\r\nStep 35, loss: 16.608722686767578\r\nStep 36, loss: 16.647741317749023\r\nStep 37, loss: 16.336868286132812\r\nStep 38, loss: 16.651290893554688\r\nStep 39, loss: 16.270305633544922\r\nStep 40, loss: 16.28514289855957\r\nStep 41, loss: 15.932032585144043\r\nStep 42, loss: 16.085798263549805\r\nStep 43, loss: 15.853764533996582\r\nStep 44, loss: 16.237834930419922\r\nStep 45, loss: 15.47584056854248\r\nStep 46, loss: 15.609787940979004\r\nStep 47, loss: 15.486068725585938\r\nStep 48, loss: 15.258798599243164\r\nStep 49, loss: 15.545014381408691\r\nStep 50, loss: 16.061813354492188\r\nStep 51, loss: 15.366859436035156\r\nStep 52, loss: 14.827864646911621\r\nStep 53, loss: 14.727177619934082\r\nStep 54, loss: 14.848566055297852\r\nStep 55, loss: 14.447760581970215\r\nStep 56, loss: 15.252120018005371\r\nStep 57, loss: 14.218976020812988\r\nStep 58, loss: 14.730356216430664\r\nStep 59, loss: 14.582534790039062\r\nStep 30, loss: 17.43343734741211\r\nStep 31, loss: 16.896129608154297\r\nStep 32, loss: 16.79721450805664\r\nStep 33, loss: 16.77494239807129\r\nStep 34, loss: 17.045551300048828\r\nStep 35, loss: 16.608722686767578\r\nStep 36, loss: 16.647741317749023\r\nStep 37, loss: 16.336868286132812\r\nStep 38, loss: 16.651290893554688\r\nStep 39, loss: 16.270305633544922\r\nStep 40, loss: 16.28514289855957\r\nStep 41, loss: 15.932032585144043\r\nStep 42, loss: 16.085798263549805\r\nStep 43, loss: 15.853764533996582\r\nStep 44, loss: 16.237834930419922\r\nStep 45, loss: 15.47584056854248\r\nStep 46, loss: 15.609787940979004\r\nStep 47, loss: 15.486068725585938\r\nStep 48, loss: 15.258798599243164\r\nStep 49, loss: 15.545014381408691\r\nStep 50, loss: 16.061813354492188\r\nStep 51, loss: 15.366859436035156\r\nStep 52, loss: 14.827864646911621\r\nStep 53, loss: 14.727177619934082\r\nStep 54, loss: 14.848566055297852\r\nStep 55, loss: 14.447760581970215\r\nStep 56, loss: 15.252120018005371\r\nStep 57, loss: 14.218976020812988\r\nStep 58, loss: 14.730356216430664\r\nStep 59, loss: 14.582534790039062\r\nStep 60, loss: 13.967232704162598\r\nStep 61, loss: 13.71136474609375\r\nStep 62, loss: 13.708770751953125\r\nStep 63, loss: 14.0950345993042\r\nStep 64, loss: 13.713918685913086\r\nStep 65, loss: 14.868953704833984\r\nStep 66, loss: 14.468207359313965\r\nStep 67, loss: 14.585230827331543\r\nStep 68, loss: 14.341510772705078\r\nStep 69, loss: 14.391881942749023\r\nStep 70, loss: 13.262972831726074\r\nStep 71, loss: 14.027960777282715\r\nStep 72, loss: 12.899559020996094\r\nStep 73, loss: 13.606071472167969\r\nStep 74, loss: 13.505942344665527\r\nStep 75, loss: 12.73733139038086\r\nStep 76, loss: 13.731353759765625\r\nStep 77, loss: 12.609628677368164\r\nStep 78, loss: 13.05951976776123\r\nStep 79, loss: 12.82889461517334\r\nStep 80, loss: 12.381841659545898\r\nStep 81, loss: 12.184906005859375\r\nStep 82, loss: 12.261787414550781\r\nStep 83, loss: 12.400224685668945\r\nStep 84, loss: 12.114161491394043\r\nStep 85, loss: 12.412161827087402\r\nStep 86, loss: 11.783485412597656\r\nStep 87, loss: 11.766879081726074\r\nStep 88, loss: 11.893105506896973\r\nStep 89, loss: 11.720884323120117\r\nStep 60, loss: 13.967232704162598\r\nStep 61, loss: 13.71136474609375\r\nStep 62, loss: 13.708770751953125\r\nStep 63, loss: 14.0950345993042\r\nStep 64, loss: 13.713918685913086\r\nStep 65, loss: 14.868953704833984\r\nStep 66, loss: 14.468207359313965\r\nStep 67, loss: 14.585230827331543\r\nStep 68, loss: 14.341510772705078\r\nStep 69, loss: 14.391881942749023\r\nStep 70, loss: 13.262972831726074\r\nStep 71, loss: 14.027960777282715\r\nStep 72, loss: 12.899559020996094\r\nStep 73, loss: 13.606071472167969\r\nStep 74, loss: 13.505942344665527\r\nStep 75, loss: 12.73733139038086\r\nStep 76, loss: 13.731353759765625\r\nStep 77, loss: 12.609628677368164\r\nStep 78, loss: 13.05951976776123\r\nStep 79, loss: 12.82889461517334\r\nStep 80, loss: 12.381841659545898\r\nStep 81, loss: 12.184906005859375\r\nStep 82, loss: 12.261787414550781\r\nStep 83, loss: 12.400224685668945\r\nStep 84, loss: 12.114161491394043\r\nStep 85, loss: 12.412161827087402\r\nStep 86, loss: 11.783485412597656\r\nStep 87, loss: 11.766879081726074\r\nStep 88, loss: 11.893105506896973\r\nStep 89, loss: 11.720884323120117\r\nStep 90, loss: 11.85186767578125\r\nStep 91, loss: 11.742947578430176\r\nStep 92, loss: 11.327075004577637\r\nStep 93, loss: 11.285242080688477\r\nStep 94, loss: 11.404678344726562\r\nStep 95, loss: 11.971321105957031\r\nStep 96, loss: 11.94084644317627\r\nStep 97, loss: 11.925477027893066\r\nStep 98, loss: 11.581640243530273\r\nStep 99, loss: 10.887056350708008\r\nStep 100, loss: 11.63808822631836\r\nStep 101, loss: 10.839936256408691\r\nStep 102, loss: 11.229722023010254\r\nStep 103, loss: 10.856170654296875\r\nStep 104, loss: 11.429742813110352\r\nStep 105, loss: 10.258073806762695\r\nStep 106, loss: 10.562714576721191\r\nStep 107, loss: 10.214714050292969\r\nStep 108, loss: 10.124898910522461\r\nStep 109, loss: 9.705013275146484\r\nStep 110, loss: 9.921828269958496\r\nStep 111, loss: 9.90079116821289\r\nStep 112, loss: 10.455109596252441\r\nStep 113, loss: 9.936418533325195\r\nStep 114, loss: 9.936537742614746\r\nStep 115, loss: 9.773576736450195\r\nStep 116, loss: 9.676142692565918\r\nStep 117, loss: 9.945345878601074\r\nStep 118, loss: 9.923107147216797\r\nStep 90, loss: 11.85186767578125\r\nStep 91, loss: 11.742947578430176\r\nStep 92, loss: 11.327075004577637\r\nStep 93, loss: 11.285242080688477\r\nStep 94, loss: 11.404678344726562\r\nStep 95, loss: 11.971321105957031\r\nStep 96, loss: 11.94084644317627\r\nStep 97, loss: 11.925477027893066\r\nStep 98, loss: 11.581640243530273\r\nStep 99, loss: 10.887056350708008\r\nStep 100, loss: 11.63808822631836\r\nStep 101, loss: 10.839936256408691\r\nStep 102, loss: 11.229722023010254\r\nStep 103, loss: 10.856170654296875\r\nStep 104, loss: 11.429742813110352\r\nStep 105, loss: 10.258073806762695\r\nStep 106, loss: 10.562714576721191\r\nStep 107, loss: 10.214714050292969\r\nStep 108, loss: 10.124898910522461\r\nStep 109, loss: 9.705013275146484\r\nStep 110, loss: 9.921828269958496\r\nStep 111, loss: 9.90079116821289\r\nStep 112, loss: 10.455109596252441\r\nStep 113, loss: 9.936418533325195\r\nStep 114, loss: 9.936537742614746\r\nStep 115, loss: 9.773576736450195\r\nStep 116, loss: 9.676142692565918\r\nStep 117, loss: 9.945345878601074\r\nStep 118, loss: 9.923107147216797\r\nStep 119, loss: 9.550374984741211\r\nStep 120, loss: 10.050488471984863\r\nStep 119, loss: 9.550374984741211\r\nStep 120, loss: 10.050488471984863\r\nStep 121, loss: 9.169127464294434\r\nStep 122, loss: 9.096636772155762\r\nStep 123, loss: 9.86887264251709\r\nStep 124, loss: 9.788836479187012\r\nStep 125, loss: 9.638827323913574\r\nStep 126, loss: 9.093913078308105\r\nStep 127, loss: 9.269700050354004\r\nStep 128, loss: 9.940503120422363\r\nStep 129, loss: 8.495964050292969\r\nStep 130, loss: 9.110383033752441\r\nStep 131, loss: 9.079448699951172\r\nStep 132, loss: 9.014200210571289\r\nStep 133, loss: 8.72136402130127\r\nStep 134, loss: 9.308770179748535\r\nStep 135, loss: 8.615967750549316\r\nStep 136, loss: 8.500845909118652\r\nStep 137, loss: 8.710356712341309\r\nStep 138, loss: 8.530448913574219\r\nStep 139, loss: 9.046561241149902\r\nStep 140, loss: 8.546600341796875\r\nStep 141, loss: 9.229374885559082\r\nStep 142, loss: 8.218626022338867\r\nStep 143, loss: 8.23629093170166\r\nStep 144, loss: 8.220922470092773\r\nStep 145, loss: 8.11863899230957\r\nStep 146, loss: 8.13656997680664\r\nStep 147, loss: 8.036163330078125\r\nStep 148, loss: 7.980085372924805\r\nStep 149, loss: 7.8308844566345215\r\nStep 150, loss: 7.997012615203857\r\nStep 121, loss: 9.169127464294434\r\nStep 122, loss: 9.096636772155762\r\nStep 123, loss: 9.86887264251709\r\nStep 124, loss: 9.788836479187012\r\nStep 125, loss: 9.638827323913574\r\nStep 126, loss: 9.093913078308105\r\nStep 127, loss: 9.269700050354004\r\nStep 128, loss: 9.940503120422363\r\nStep 129, loss: 8.495964050292969\r\nStep 130, loss: 9.110383033752441\r\nStep 131, loss: 9.079448699951172\r\nStep 132, loss: 9.014200210571289\r\nStep 133, loss: 8.72136402130127\r\nStep 134, loss: 9.308770179748535\r\nStep 135, loss: 8.615967750549316\r\nStep 136, loss: 8.500845909118652\r\nStep 137, loss: 8.710356712341309\r\nStep 138, loss: 8.530448913574219\r\nStep 139, loss: 9.046561241149902\r\nStep 140, loss: 8.546600341796875\r\nStep 141, loss: 9.229374885559082\r\nStep 142, loss: 8.218626022338867\r\nStep 143, loss: 8.23629093170166\r\nStep 144, loss: 8.220922470092773\r\nStep 145, loss: 8.11863899230957\r\nStep 146, loss: 8.13656997680664\r\nStep 147, loss: 8.036163330078125\r\nStep 148, loss: 7.980085372924805\r\nStep 149, loss: 7.8308844566345215\r\nStep 150, loss: 7.997012615203857\r\nStep 151, loss: 8.312012672424316\r\nStep 152, loss: 8.257838249206543\r\nStep 153, loss: 7.67844820022583\r\nStep 154, loss: 8.097171783447266\r\nStep 155, loss: 7.956625461578369\r\nStep 156, loss: 7.661460876464844\r\nStep 157, loss: 7.60241174697876\r\nStep 158, loss: 7.5734171867370605\r\nStep 159, loss: 7.4999213218688965\r\nStep 160, loss: 7.889724254608154\r\nStep 161, loss: 7.616578578948975\r\nStep 162, loss: 7.669461250305176\r\nStep 163, loss: 7.583798408508301\r\nStep 164, loss: 7.573363780975342\r\nStep 165, loss: 7.425256252288818\r\nStep 166, loss: 7.534543991088867\r\nStep 167, loss: 7.315162658691406\r\nStep 168, loss: 7.1738691329956055\r\nStep 169, loss: 7.142663955688477\r\nStep 170, loss: 7.435649871826172\r\nStep 171, loss: 7.096683502197266\r\nStep 172, loss: 7.248353481292725\r\nStep 173, loss: 7.270962715148926\r\nStep 174, loss: 7.3473687171936035\r\nStep 175, loss: 6.999227523803711\r\nStep 176, loss: 7.4188151359558105\r\nStep 177, loss: 7.223378658294678\r\nStep 178, loss: 7.144739627838135\r\nStep 179, loss: 6.7800211906433105\r\nStep 151, loss: 8.312012672424316\r\nStep 152, loss: 8.257838249206543\r\nStep 153, loss: 7.67844820022583\r\nStep 154, loss: 8.097171783447266\r\nStep 155, loss: 7.956625461578369\r\nStep 156, loss: 7.661460876464844\r\nStep 157, loss: 7.60241174697876\r\nStep 158, loss: 7.5734171867370605\r\nStep 159, loss: 7.4999213218688965\r\nStep 160, loss: 7.889724254608154\r\nStep 161, loss: 7.616578578948975\r\nStep 162, loss: 7.669461250305176\r\nStep 163, loss: 7.583798408508301\r\nStep 164, loss: 7.573363780975342\r\nStep 165, loss: 7.425256252288818\r\nStep 166, loss: 7.534543991088867\r\nStep 167, loss: 7.315162658691406\r\nStep 168, loss: 7.1738691329956055\r\nStep 169, loss: 7.142663955688477\r\nStep 170, loss: 7.435649871826172\r\nStep 171, loss: 7.096683502197266\r\nStep 172, loss: 7.248353481292725\r\nStep 173, loss: 7.270962715148926\r\nStep 174, loss: 7.3473687171936035\r\nStep 175, loss: 6.999227523803711\r\nStep 176, loss: 7.4188151359558105\r\nStep 177, loss: 7.223378658294678\r\nStep 178, loss: 7.144739627838135\r\nStep 179, loss: 6.7800211906433105\r\nStep 180, loss: 7.156866073608398\r\nStep 181, loss: 6.900200843811035\r\nStep 182, loss: 7.0645365715026855\r\nStep 183, loss: 7.188953399658203\r\nStep 184, loss: 6.891933917999268\r\nStep 185, loss: 6.902243614196777\r\nStep 186, loss: 6.5922651290893555\r\nStep 187, loss: 7.096951007843018\r\nStep 188, loss: 6.9083991050720215\r\nStep 189, loss: 6.90083646774292\r\nStep 190, loss: 6.980688095092773\r\nStep 191, loss: 6.729495048522949\r\nStep 192, loss: 6.864687442779541\r\nStep 193, loss: 6.7847514152526855\r\nStep 194, loss: 6.816504955291748\r\nStep 195, loss: 6.671535968780518\r\nStep 196, loss: 6.762739181518555\r\nStep 197, loss: 6.497260570526123\r\nStep 198, loss: 7.014808654785156\r\nStep 199, loss: 6.6613945960998535\r\nStep 200, loss: 6.8338422775268555\r\nStep 201, loss: 6.816775798797607\r\nStep 202, loss: 6.6137614250183105\r\nStep 203, loss: 6.697750568389893\r\nStep 204, loss: 6.821736812591553\r\nStep 205, loss: 6.9376959800720215\r\nStep 206, loss: 6.731281757354736\r\nStep 207, loss: 6.579183578491211\r\nStep 208, loss: 6.68570613861084\r\nStep 180, loss: 7.156866073608398\r\nStep 181, loss: 6.900200843811035\r\nStep 182, loss: 7.0645365715026855\r\nStep 183, loss: 7.188953399658203\r\nStep 184, loss: 6.891933917999268\r\nStep 185, loss: 6.902243614196777\r\nStep 186, loss: 6.5922651290893555\r\nStep 187, loss: 7.096951007843018\r\nStep 188, loss: 6.9083991050720215\r\nStep 189, loss: 6.90083646774292\r\nStep 190, loss: 6.980688095092773\r\nStep 191, loss: 6.729495048522949\r\nStep 192, loss: 6.864687442779541\r\nStep 193, loss: 6.7847514152526855\r\nStep 194, loss: 6.816504955291748\r\nStep 195, loss: 6.671535968780518\r\nStep 196, loss: 6.762739181518555\r\nStep 197, loss: 6.497260570526123\r\nStep 198, loss: 7.014808654785156\r\nStep 199, loss: 6.6613945960998535\r\nStep 200, loss: 6.8338422775268555\r\nStep 201, loss: 6.816775798797607\r\nStep 202, loss: 6.6137614250183105\r\nStep 203, loss: 6.697750568389893\r\nStep 204, loss: 6.821736812591553\r\nStep 205, loss: 6.9376959800720215\r\nStep 206, loss: 6.731281757354736\r\nStep 207, loss: 6.579183578491211\r\nStep 208, loss: 6.68570613861084\r\nStep 209, loss: 6.527091026306152\r\nStep 210, loss: 6.523392200469971\r\nStep 211, loss: 6.566160678863525\r\nStep 212, loss: 6.912597179412842\r\nStep 213, loss: 6.662230968475342\r\nStep 214, loss: 6.962000846862793\r\nStep 215, loss: 6.555893898010254\r\nStep 216, loss: 6.651054382324219\r\nStep 217, loss: 6.657430648803711\r\nStep 218, loss: 6.917163372039795\r\nStep 219, loss: 6.651887893676758\r\nStep 220, loss: 6.725634574890137\r\nStep 221, loss: 6.670478343963623\r\nStep 222, loss: 6.46140193939209\r\nStep 223, loss: 6.792141437530518\r\nStep 224, loss: 6.598993301391602\r\nStep 225, loss: 6.382656097412109\r\nStep 226, loss: 6.67178201675415\r\nStep 227, loss: 6.364635467529297\r\nStep 228, loss: 6.6085052490234375\r\nStep 229, loss: 6.653966903686523\r\nStep 230, loss: 6.554793834686279\r\nStep 231, loss: 6.588866710662842\r\nStep 232, loss: 6.585617542266846\r\nStep 233, loss: 6.507537364959717\r\nStep 234, loss: 6.644289493560791\r\nStep 235, loss: 6.5886993408203125\r\nStep 236, loss: 6.520603656768799\r\nStep 237, loss: 6.404054164886475\r\nStep 238, loss: 6.414518356323242\r\nStep 209, loss: 6.527091026306152\r\nStep 210, loss: 6.523392200469971\r\nStep 211, loss: 6.566160678863525\r\nStep 212, loss: 6.912597179412842\r\nStep 213, loss: 6.662230968475342\r\nStep 214, loss: 6.962000846862793\r\nStep 215, loss: 6.555893898010254\r\nStep 216, loss: 6.651054382324219\r\nStep 217, loss: 6.657430648803711\r\nStep 218, loss: 6.917163372039795\r\nStep 219, loss: 6.651887893676758\r\nStep 220, loss: 6.725634574890137\r\nStep 221, loss: 6.670478343963623\r\nStep 222, loss: 6.46140193939209\r\nStep 223, loss: 6.792141437530518\r\nStep 224, loss: 6.598993301391602\r\nStep 225, loss: 6.382656097412109\r\nStep 226, loss: 6.67178201675415\r\nStep 227, loss: 6.364635467529297\r\nStep 228, loss: 6.6085052490234375\r\nStep 229, loss: 6.653966903686523\r\nStep 230, loss: 6.554793834686279\r\nStep 231, loss: 6.588866710662842\r\nStep 232, loss: 6.585617542266846\r\nStep 233, loss: 6.507537364959717\r\nStep 234, loss: 6.644289493560791\r\nStep 235, loss: 6.5886993408203125\r\nStep 236, loss: 6.520603656768799\r\nStep 237, loss: 6.404054164886475\r\nStep 238, loss: 6.414518356323242\r\nStep 239, loss: 6.4341325759887695\r\nStep 240, loss: 6.528247833251953\r\nStep 239, loss: 6.4341325759887695\r\nStep 240, loss: 6.528247833251953\r\nStep 0, loss: 18.776222229003906\r\nStep 1, loss: 18.819902420043945\r\nStep 2, loss: 18.791563034057617\r\nStep 3, loss: 18.668718338012695\r\nStep 4, loss: 18.80548667907715\r\nStep 5, loss: 18.81717300415039\r\nStep 6, loss: 18.450780868530273\r\nStep 7, loss: 18.754093170166016\r\nStep 8, loss: 18.730485916137695\r\nStep 9, loss: 18.4219913482666\r\nStep 10, loss: 18.53896713256836\r\nStep 11, loss: 18.433452606201172\r\nStep 12, loss: 18.339122772216797\r\nStep 13, loss: 18.459253311157227\r\nStep 14, loss: 18.409881591796875\r\nStep 15, loss: 18.35429573059082\r\nStep 16, loss: 18.237462997436523\r\nStep 17, loss: 18.09574317932129\r\nStep 18, loss: 17.840925216674805\r\nStep 19, loss: 18.041223526000977\r\nStep 20, loss: 17.888032913208008\r\nStep 21, loss: 17.768489837646484\r\nStep 22, loss: 17.63291358947754\r\nStep 23, loss: 17.79960060119629\r\nStep 24, loss: 17.668100357055664\r\nStep 25, loss: 17.608400344848633\r\nStep 26, loss: 17.400070190429688\r\nStep 27, loss: 17.27320671081543\r\nStep 28, loss: 17.558292388916016\r\nStep 29, loss: 17.401700973510742\r\nStep 30, loss: 17.43343734741211\r\nStep 31, loss: 16.896129608154297\r\nStep 32, loss: 16.79721450805664\r\nStep 33, loss: 16.77494239807129\r\nStep 34, loss: 17.045551300048828\r\nStep 35, loss: 16.608722686767578\r\nStep 36, loss: 16.647741317749023\r\nStep 37, loss: 16.336868286132812\r\nStep 38, loss: 16.651290893554688\r\nStep 39, loss: 16.270305633544922\r\nStep 40, loss: 16.28514289855957\r\nStep 41, loss: 15.932032585144043\r\nStep 42, loss: 16.085798263549805\r\nStep 43, loss: 15.853764533996582\r\nStep 44, loss: 16.237834930419922\r\nStep 45, loss: 15.47584056854248\r\nStep 46, loss: 15.609787940979004\r\nStep 47, loss: 15.486068725585938\r\nStep 48, loss: 15.258798599243164\r\nStep 49, loss: 15.545014381408691\r\nStep 50, loss: 16.061813354492188\r\nStep 51, loss: 15.366859436035156\r\nStep 52, loss: 14.827864646911621\r\nStep 53, loss: 14.727177619934082\r\nStep 54, loss: 14.848566055297852\r\nStep 55, loss: 14.447760581970215\r\nStep 56, loss: 15.252120018005371\r\nStep 57, loss: 14.218976020812988\r\nStep 58, loss: 14.730356216430664\r\nStep 59, loss: 14.582534790039062\r\nStep 60, loss: 13.967232704162598\r\nStep 61, loss: 13.71136474609375\r\nStep 62, loss: 13.708770751953125\r\nStep 63, loss: 14.0950345993042\r\nStep 64, loss: 13.713918685913086\r\nStep 65, loss: 14.868953704833984\r\nStep 66, loss: 14.468207359313965\r\nStep 67, loss: 14.585230827331543\r\nStep 68, loss: 14.341510772705078\r\nStep 69, loss: 14.391881942749023\r\nStep 70, loss: 13.262972831726074\r\nStep 71, loss: 14.027960777282715\r\nStep 72, loss: 12.899559020996094\r\nStep 73, loss: 13.606071472167969\r\nStep 74, loss: 13.505942344665527\r\nStep 75, loss: 12.73733139038086\r\nStep 76, loss: 13.731353759765625\r\nStep 77, loss: 12.609628677368164\r\nStep 78, loss: 13.05951976776123\r\nStep 79, loss: 12.82889461517334\r\nStep 80, loss: 12.381841659545898\r\nStep 81, loss: 12.184906005859375\r\nStep 82, loss: 12.261787414550781\r\nStep 83, loss: 12.400224685668945\r\nStep 84, loss: 12.114161491394043\r\nStep 85, loss: 12.412161827087402\r\nStep 86, loss: 11.783485412597656\r\nStep 87, loss: 11.766879081726074\r\nStep 88, loss: 11.893105506896973\r\nStep 89, loss: 11.720884323120117\r\nStep 90, loss: 11.85186767578125\r\nStep 91, loss: 11.742947578430176\r\nStep 92, loss: 11.327075004577637\r\nStep 93, loss: 11.285242080688477\r\nStep 94, loss: 11.404678344726562\r\nStep 95, loss: 11.971321105957031\r\nStep 96, loss: 11.94084644317627\r\nStep 97, loss: 11.925477027893066\r\nStep 98, loss: 11.581640243530273\r\nStep 99, loss: 10.887056350708008\r\nStep 100, loss: 11.63808822631836\r\nStep 101, loss: 10.839936256408691\r\nStep 102, loss: 11.229722023010254\r\nStep 103, loss: 10.856170654296875\r\nStep 104, loss: 11.429742813110352\r\nStep 105, loss: 10.258073806762695\r\nStep 106, loss: 10.562714576721191\r\nStep 107, loss: 10.214714050292969\r\nStep 108, loss: 10.124898910522461\r\nStep 109, loss: 9.705013275146484\r\nStep 110, loss: 9.921828269958496\r\nStep 111, loss: 9.90079116821289\r\nStep 112, loss: 10.455109596252441\r\nStep 113, loss: 9.936418533325195\r\nStep 114, loss: 9.936537742614746\r\nStep 115, loss: 9.773576736450195\r\nStep 116, loss: 9.676142692565918\r\nStep 117, loss: 9.945345878601074\r\nStep 118, loss: 9.923107147216797\r\nStep 119, loss: 9.550374984741211\r\nStep 120, loss: 10.050488471984863\r\nStep 121, loss: 9.169127464294434\r\nStep 122, loss: 9.096636772155762\r\nStep 123, loss: 9.86887264251709\r\nStep 124, loss: 9.788836479187012\r\nStep 125, loss: 9.638827323913574\r\nStep 126, loss: 9.093913078308105\r\nStep 127, loss: 9.269700050354004\r\nStep 128, loss: 9.940503120422363\r\nStep 129, loss: 8.495964050292969\r\nStep 130, loss: 9.110383033752441\r\nStep 131, loss: 9.079448699951172\r\nStep 132, loss: 9.014200210571289\r\nStep 133, loss: 8.72136402130127\r\nStep 134, loss: 9.308770179748535\r\nStep 135, loss: 8.615967750549316\r\nStep 136, loss: 8.500845909118652\r\nStep 137, loss: 8.710356712341309\r\nStep 138, loss: 8.530448913574219\r\nStep 139, loss: 9.046561241149902\r\nStep 140, loss: 8.546600341796875\r\nStep 141, loss: 9.229374885559082\r\nStep 142, loss: 8.218626022338867\r\nStep 143, loss: 8.23629093170166\r\nStep 144, loss: 8.220922470092773\r\nStep 145, loss: 8.11863899230957\r\nStep 146, loss: 8.13656997680664\r\nStep 147, loss: 8.036163330078125\r\nStep 148, loss: 7.980085372924805\r\nStep 149, loss: 7.8308844566345215\r\nStep 150, loss: 7.997012615203857\r\nStep 151, loss: 8.312012672424316\r\nStep 152, loss: 8.257838249206543\r\nStep 153, loss: 7.67844820022583\r\nStep 154, loss: 8.097171783447266\r\nStep 155, loss: 7.956625461578369\r\nStep 156, loss: 7.661460876464844\r\nStep 157, loss: 7.60241174697876\r\nStep 158, loss: 7.5734171867370605\r\nStep 159, loss: 7.4999213218688965\r\nStep 160, loss: 7.889724254608154\r\nStep 161, loss: 7.616578578948975\r\nStep 162, loss: 7.669461250305176\r\nStep 163, loss: 7.583798408508301\r\nStep 164, loss: 7.573363780975342\r\nStep 165, loss: 7.425256252288818\r\nStep 166, loss: 7.534543991088867\r\nStep 167, loss: 7.315162658691406\r\nStep 168, loss: 7.1738691329956055\r\nStep 169, loss: 7.142663955688477\r\nStep 170, loss: 7.435649871826172\r\nStep 171, loss: 7.096683502197266\r\nStep 172, loss: 7.248353481292725\r\nStep 173, loss: 7.270962715148926\r\nStep 174, loss: 7.3473687171936035\r\nStep 175, loss: 6.999227523803711\r\nStep 176, loss: 7.4188151359558105\r\nStep 177, loss: 7.223378658294678\r\nStep 178, loss: 7.144739627838135\r\nStep 179, loss: 6.7800211906433105\r\nStep 180, loss: 7.156866073608398\r\nStep 181, loss: 6.900200843811035\r\nStep 182, loss: 7.0645365715026855\r\nStep 183, loss: 7.188953399658203\r\nStep 184, loss: 6.891933917999268\r\nStep 185, loss: 6.902243614196777\r\nStep 186, loss: 6.5922651290893555\r\nStep 187, loss: 7.096951007843018\r\nStep 188, loss: 6.9083991050720215\r\nStep 189, loss: 6.90083646774292\r\nStep 190, loss: 6.980688095092773\r\nStep 191, loss: 6.729495048522949\r\nStep 192, loss: 6.864687442779541\r\nStep 193, loss: 6.7847514152526855\r\nStep 194, loss: 6.816504955291748\r\nStep 195, loss: 6.671535968780518\r\nStep 196, loss: 6.762739181518555\r\nStep 197, loss: 6.497260570526123\r\nStep 198, loss: 7.014808654785156\r\nStep 199, loss: 6.6613945960998535\r\nStep 200, loss: 6.8338422775268555\r\nStep 201, loss: 6.816775798797607\r\nStep 202, loss: 6.6137614250183105\r\nStep 203, loss: 6.697750568389893\r\nStep 204, loss: 6.821736812591553\r\nStep 205, loss: 6.9376959800720215\r\nStep 206, loss: 6.731281757354736\r\nStep 207, loss: 6.579183578491211\r\nStep 208, loss: 6.68570613861084\r\nStep 209, loss: 6.527091026306152\r\nStep 210, loss: 6.523392200469971\r\nStep 211, loss: 6.566160678863525\r\nStep 212, loss: 6.912597179412842\r\nStep 213, loss: 6.662230968475342\r\nStep 214, loss: 6.962000846862793\r\nStep 215, loss: 6.555893898010254\r\nStep 216, loss: 6.651054382324219\r\nStep 217, loss: 6.657430648803711\r\nStep 218, loss: 6.917163372039795\r\nStep 219, loss: 6.651887893676758\r\nStep 220, loss: 6.725634574890137\r\nStep 221, loss: 6.670478343963623\r\nStep 222, loss: 6.46140193939209\r\nStep 223, loss: 6.792141437530518\r\nStep 224, loss: 6.598993301391602\r\nStep 225, loss: 6.382656097412109\r\nStep 226, loss: 6.67178201675415\r\nStep 227, loss: 6.364635467529297\r\nStep 228, loss: 6.6085052490234375\r\nStep 229, loss: 6.653966903686523\r\nStep 230, loss: 6.554793834686279\r\nStep 231, loss: 6.588866710662842\r\nStep 232, loss: 6.585617542266846\r\nStep 233, loss: 6.507537364959717\r\nStep 234, loss: 6.644289493560791\r\nStep 235, loss: 6.5886993408203125\r\nStep 236, loss: 6.520603656768799\r\nStep 237, loss: 6.404054164886475\r\nStep 238, loss: 6.414518356323242\r\nStep 239, loss: 6.4341325759887695\r\nStep 240, loss: 6.528247833251953\r\n",,terminal_output +1629,3110681,"TERMINAL",0,0,"412",,terminal_output +1630,3111697,"TERMINAL",0,0,"523",,terminal_output +1631,3112729,"TERMINAL",0,0,"634",,terminal_output +1632,3113766,"TERMINAL",0,0,"745",,terminal_output +1633,3114985,"TERMINAL",0,0,"856",,terminal_output +1634,3116208,"TERMINAL",0,0,"978",,terminal_output +1635,3117233,"TERMINAL",0,0,"1189",,terminal_output +1636,3118159,"TERMINAL",0,0,"2950",,terminal_output +1637,3119284,"TERMINAL",0,0,"3301",,terminal_output +1638,3120426,"TERMINAL",0,0,"412",,terminal_output +1639,3121126,"TERMINAL",0,0,"523",,terminal_output +1640,3122153,"TERMINAL",0,0,"634",,terminal_output +1641,3123176,"TERMINAL",0,0,"745",,terminal_output +1642,3124300,"TERMINAL",0,0,"856",,terminal_output +1643,3125736,"TERMINAL",0,0,"967",,terminal_output +1644,3126315,"TERMINAL",0,0,"2078",,terminal_output +1645,3127336,"TERMINAL",0,0,"189",,terminal_output +1646,3128401,"TERMINAL",0,0,"299:00",,terminal_output +1647,3129421,"TERMINAL",0,0,"3401",,terminal_output +1648,3130649,"TERMINAL",0,0,"412",,terminal_output +1649,3131674,"TERMINAL",0,0,"523",,terminal_output +1650,3132601,"TERMINAL",0,0,"634",,terminal_output +1651,3133621,"TERMINAL",0,0,"745",,terminal_output +1652,3134700,"TERMINAL",0,0,"856",,terminal_output +1653,3135751,"TERMINAL",0,0,"967",,terminal_output +1654,3136802,"TERMINAL",0,0,"3078",,terminal_output +1655,3137764,"TERMINAL",0,0,"189",,terminal_output +1656,3138834,"TERMINAL",0,0,"2910",,terminal_output +1657,3140172,"TERMINAL",0,0,"3512",,terminal_output +1658,3141195,"TERMINAL",0,0,"523",,terminal_output +1659,3142121,"TERMINAL",0,0,"634",,terminal_output +1660,3143024,"TERMINAL",0,0,"745",,terminal_output +1661,3144068,"TERMINAL",0,0,"856",,terminal_output +1662,3145093,"TERMINAL",0,0,"967",,terminal_output +1663,3146103,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=6 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-6-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 6 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +1664,3146193,"TERMINAL",0,0,"4078",,terminal_output +1665,3147136,"TERMINAL",0,0,"189",,terminal_output +1666,3148264,"TERMINAL",0,0,"2920",,terminal_output +1667,3149289,"TERMINAL",0,0,"32:001",,terminal_output +1668,3150318,"TERMINAL",0,0,"412",,terminal_output +1669,3151290,"TERMINAL",0,0,"523",,terminal_output +1670,3152334,"TERMINAL",0,0,"634",,terminal_output +1671,3153375,"TERMINAL",0,0,"745",,terminal_output +1672,3154516,"TERMINAL",0,0,"856",,terminal_output +1673,3155534,"TERMINAL",0,0,"967",,terminal_output +1674,3156562,"TERMINAL",0,0,"5078",,terminal_output +1675,3157585,"TERMINAL",0,0,"189",,terminal_output +1676,3158712,"TERMINAL",0,0,"2930",,terminal_output +1677,3159734,"TERMINAL",0,0,"3101",,terminal_output +1678,3160696,"TERMINAL",0,0,"412",,terminal_output +1679,3161738,"TERMINAL",0,0,"523",,terminal_output +1680,3162783,"TERMINAL",0,0,"634",,terminal_output +1681,3163818,"TERMINAL",0,0,"756",,terminal_output +1682,3164854,"TERMINAL",0,0,"967",,terminal_output +1683,3165979,"TERMINAL",0,0,"1:0078",,terminal_output +1684,3167003,"TERMINAL",0,0,"189",,terminal_output +1685,3167992,"TERMINAL",0,0,"2940",,terminal_output +1686,3169047,"TERMINAL",0,0,"3201",,terminal_output +1687,3170072,"TERMINAL",0,0,"412",,terminal_output +1688,3171200,"TERMINAL",0,0,"523",,terminal_output +1689,3172148,"TERMINAL",0,0,"634",,terminal_output +1690,3173188,"TERMINAL",0,0,"745",,terminal_output +1691,3174273,"TERMINAL",0,0,"856",,terminal_output +1692,3175288,"TERMINAL",0,0,"967",,terminal_output +1693,3176409,"TERMINAL",0,0,"1078",,terminal_output +1694,3177382,"TERMINAL",0,0,"189",,terminal_output +1695,3178417,"TERMINAL",0,0,"2950",,terminal_output +1696,3179456,"TERMINAL",0,0,"3301",,terminal_output +1697,3180499,"TERMINAL",0,0,"412",,terminal_output +1698,3181581,"TERMINAL",0,0,"523",,terminal_output +1699,3182596,"TERMINAL",0,0,"634",,terminal_output +1700,3183642,"TERMINAL",0,0,"745",,terminal_output +1701,3184704,"TERMINAL",0,0,"856",,terminal_output +1702,3185743,"TERMINAL",0,0,"967",,terminal_output +1703,3186789,"TERMINAL",0,0,"2078",,terminal_output +1704,3187840,"TERMINAL",0,0,"1920:00",,terminal_output +1705,3188915,"TERMINAL",0,0,"3401",,terminal_output +1706,3189918,"TERMINAL",0,0,"412",,terminal_output +1707,3191066,"TERMINAL",0,0,"523",,terminal_output +1708,3192046,"TERMINAL",0,0,"634",,terminal_output +1709,3193077,"TERMINAL",0,0,"745",,terminal_output +1710,3194159,"TERMINAL",0,0,"856",,terminal_output +1711,3195200,"TERMINAL",0,0,"967",,terminal_output +1712,3196167,"TERMINAL",0,0,"3078",,terminal_output +1713,3197222,"TERMINAL",0,0,"189",,terminal_output +1714,3198337,"TERMINAL",0,0,"2910",,terminal_output +1715,3199021,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=8 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-8-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 8 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +1716,3199313,"TERMINAL",0,0,"3501",,terminal_output +1717,3200350,"TERMINAL",0,0,"412",,terminal_output +1718,3201406,"TERMINAL",0,0,"523",,terminal_output +1719,3202286,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=12 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-12-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 12 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +1720,3202465,"TERMINAL",0,0,"634",,terminal_output +1721,3203555,"TERMINAL",0,0,"745",,terminal_output +1722,3204583,"TERMINAL",0,0,"856",,terminal_output +1723,3205577,"TERMINAL",0,0,"967",,terminal_output +1724,3206610,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-20-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 20 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +1725,3206675,"TERMINAL",0,0,"4078",,terminal_output +1726,3207703,"TERMINAL",0,0,"189",,terminal_output +1727,3208780,"TERMINAL",0,0,"2920",,terminal_output +1728,3209818,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=50 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-50-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 50 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +1729,3209824,"TERMINAL",0,0,"33:001",,terminal_output +1730,3210821,"TERMINAL",0,0,"423",,terminal_output +1731,3211955,"TERMINAL",0,0,"634",,terminal_output +1732,3212905,"TERMINAL",0,0,"745",,terminal_output +1733,3214004,"TERMINAL",0,0,"856",,terminal_output +1734,3215005,"TERMINAL",0,0,"967",,terminal_output +1735,3216157,"TERMINAL",0,0,"5078",,terminal_output +1736,3217177,"TERMINAL",0,0,"189",,terminal_output +1737,3218211,"TERMINAL",0,0,"2930",,terminal_output +1738,3219187,"TERMINAL",0,0,"3101",,terminal_output +1739,3220236,"TERMINAL",0,0,"412",,terminal_output +1740,3221299,"TERMINAL",0,0,"523",,terminal_output +1741,3222328,"TERMINAL",0,0,"634",,terminal_output +1742,3223372,"TERMINAL",0,0,"745",,terminal_output +1743,3224447,"TERMINAL",0,0,"856",,terminal_output +1744,3225462,"TERMINAL",0,0,"967",,terminal_output +1745,3226640,"TERMINAL",0,0,"2:0078",,terminal_output +1746,3227553,"TERMINAL",0,0,"189",,terminal_output +1747,3228650,"TERMINAL",0,0,"2940",,terminal_output +1748,3229653,"TERMINAL",0,0,"3201",,terminal_output +1749,3230716,"TERMINAL",0,0,"412",,terminal_output +1750,3231750,"TERMINAL",0,0,"523",,terminal_output +1751,3232809,"TERMINAL",0,0,"634",,terminal_output +1752,3233856,"TERMINAL",0,0,"756",,terminal_output +1753,3234903,"TERMINAL",0,0,"967",,terminal_output +1754,3235949,"TERMINAL",0,0,"1078",,terminal_output +1755,3237052,"TERMINAL",0,0,"189",,terminal_output +1756,3238051,"TERMINAL",0,0,"2950",,terminal_output +1757,3239100,"TERMINAL",0,0,"3301",,terminal_output +1758,3240219,"TERMINAL",0,0,"412",,terminal_output +1759,3241291,"TERMINAL",0,0,"523",,terminal_output +1760,3242266,"TERMINAL",0,0,"634",,terminal_output +1761,3243278,"TERMINAL",0,0,"745",,terminal_output +1762,3244320,"TERMINAL",0,0,"856",,terminal_output +1763,3245442,"TERMINAL",0,0,"967",,terminal_output +1764,3246468,"TERMINAL",0,0,"2078",,terminal_output +1765,3247493,"TERMINAL",0,0,"189",,terminal_output +1766,3248619,"TERMINAL",0,0,"291:00",,terminal_output +1767,3249647,"TERMINAL",0,0,"3401",,terminal_output +1768,3250618,"TERMINAL",0,0,"412",,terminal_output +1769,3251672,"TERMINAL",0,0,"523",,terminal_output +1770,3252732,"TERMINAL",0,0,"634",,terminal_output +1771,3253767,"TERMINAL",0,0,"745",,terminal_output +1772,3254825,"TERMINAL",0,0,"867",,terminal_output +1773,3255882,"TERMINAL",0,0,"3078",,terminal_output +1774,3256935,"TERMINAL",0,0,"189",,terminal_output +1775,3257971,"TERMINAL",0,0,"2910",,terminal_output +1776,3259061,"TERMINAL",0,0,"3501",,terminal_output +1777,3260063,"TERMINAL",0,0,"412",,terminal_output +1778,3261214,"TERMINAL",0,0,"523",,terminal_output +1779,3262235,"TERMINAL",0,0,"634",,terminal_output +1780,3263265,"TERMINAL",0,0,"745",,terminal_output +1781,3264284,"TERMINAL",0,0,"856",,terminal_output +1782,3264752,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",878,0,"",shellscript,selection_mouse +1783,3265310,"TERMINAL",0,0,"967",,terminal_output +1784,3266348,"TERMINAL",0,0,"4078",,terminal_output +1785,3267464,"TERMINAL",0,0,"189",,terminal_output +1786,3268446,"TERMINAL",0,0,"2920",,terminal_output +1787,3269507,"TERMINAL",0,0,"34:001",,terminal_output +1788,3270560,"TERMINAL",0,0,"412",,terminal_output +1789,3271602,"TERMINAL",0,0,"523",,terminal_output +1790,3272650,"TERMINAL",0,0,"634",,terminal_output +1791,3273696,"TERMINAL",0,0,"745",,terminal_output +1792,3274749,"TERMINAL",0,0,"856",,terminal_output +1793,3275801,"TERMINAL",0,0,"967",,terminal_output +1794,3276878,"TERMINAL",0,0,"5089",,terminal_output +1795,3277896,"TERMINAL",0,0,"2930",,terminal_output +1796,3279027,"TERMINAL",0,0,"3101",,terminal_output +1797,3279989,"TERMINAL",0,0,"412",,terminal_output +1798,3281037,"TERMINAL",0,0,"523",,terminal_output +1799,3282082,"TERMINAL",0,0,"634",,terminal_output +1800,3283169,"TERMINAL",0,0,"745",,terminal_output +1801,3284223,"TERMINAL",0,0,"856",,terminal_output +1802,3285276,"TERMINAL",0,0,"967",,terminal_output +1803,3286288,"TERMINAL",0,0,"3:0078",,terminal_output +1804,3287324,"TERMINAL",0,0,"189",,terminal_output +1805,3288452,"TERMINAL",0,0,"2940",,terminal_output +1806,3289477,"TERMINAL",0,0,"3201",,terminal_output +1807,3290483,"TERMINAL",0,0,"412",,terminal_output +1808,3291511,"TERMINAL",0,0,"523",,terminal_output +1809,3292563,"TERMINAL",0,0,"634",,terminal_output +1810,3293609,"TERMINAL",0,0,"745",,terminal_output +1811,3294666,"TERMINAL",0,0,"856",,terminal_output +1812,3295721,"TERMINAL",0,0,"967",,terminal_output +1813,3296759,"TERMINAL",0,0,"1078",,terminal_output +1814,3297807,"TERMINAL",0,0,"189",,terminal_output +1815,3298845,"TERMINAL",0,0,"23051",,terminal_output +1816,3299920,"TERMINAL",0,0,"412",,terminal_output +1817,3300946,"TERMINAL",0,0,"523",,terminal_output +1818,3301994,"TERMINAL",0,0,"634",,terminal_output +1819,3303093,"TERMINAL",0,0,"745",,terminal_output +1820,3304074,"TERMINAL",0,0,"856",,terminal_output +1821,3305159,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",0,0,"",shellscript,tab +1822,3305271,"TERMINAL",0,0,"967",,terminal_output +1823,3306172,"TERMINAL",0,0,"2078",,terminal_output +1824,3307219,"TERMINAL",0,0,"189",,terminal_output +1825,3308319,"TERMINAL",0,0,"292:00",,terminal_output +1826,3309388,"TERMINAL",0,0,"3401",,terminal_output +1827,3310475,"TERMINAL",0,0,"412",,terminal_output +1828,3311405,"TERMINAL",0,0,"523",,terminal_output +1829,3312513,"TERMINAL",0,0,"634",,terminal_output +1830,3313540,"TERMINAL",0,0,"745",,terminal_output +1831,3314541,"TERMINAL",0,0,"856",,terminal_output +1832,3315591,"TERMINAL",0,0,"967",,terminal_output +1833,3316639,"TERMINAL",0,0,"3078",,terminal_output +1834,3317690,"TERMINAL",0,0,"189",,terminal_output +1835,3318737,"TERMINAL",0,0,"2910",,terminal_output +1836,3319790,"TERMINAL",0,0,"3501",,terminal_output +1837,3320829,"TERMINAL",0,0,"423",,terminal_output +1838,3321879,"TERMINAL",0,0,"634",,terminal_output +1839,3322930,"TERMINAL",0,0,"745",,terminal_output +1840,3323983,"TERMINAL",0,0,"856",,terminal_output +1841,3325019,"TERMINAL",0,0,"967",,terminal_output +1842,3326071,"TERMINAL",0,0,"4078",,terminal_output +1843,3327115,"TERMINAL",0,0,"189",,terminal_output +1844,3328180,"TERMINAL",0,0,"2920",,terminal_output +1845,3329311,"TERMINAL",0,0,"35:001",,terminal_output +1846,3330331,"TERMINAL",0,0,"412",,terminal_output +1847,3331311,"TERMINAL",0,0,"523",,terminal_output +1848,3332354,"TERMINAL",0,0,"634",,terminal_output +1849,3333505,"TERMINAL",0,0,"745",,terminal_output +1850,3334452,"TERMINAL",0,0,"856",,terminal_output +1851,3335503,"TERMINAL",0,0,"967",,terminal_output +1852,3336583,"TERMINAL",0,0,"5078",,terminal_output +1853,3337596,"TERMINAL",0,0,"189",,terminal_output +1854,3338645,"TERMINAL",0,0,"2930",,terminal_output +1855,3339761,"TERMINAL",0,0,"3101",,terminal_output +1856,3340776,"TERMINAL",0,0,"412",,terminal_output +1857,3341791,"TERMINAL",0,0,"523",,terminal_output +1858,3342833,"TERMINAL",0,0,"645",,terminal_output +1859,3343892,"TERMINAL",0,0,"856",,terminal_output +1860,3344929,"TERMINAL",0,0,"967",,terminal_output +1861,3346000,"TERMINAL",0,0,"4:0078",,terminal_output +1862,3347026,"TERMINAL",0,0,"189",,terminal_output +1863,3348072,"TERMINAL",0,0,"2940",,terminal_output +1864,3349172,"TERMINAL",0,0,"3201",,terminal_output +1865,3350198,"TERMINAL",0,0,"412",,terminal_output +1866,3351295,"TERMINAL",0,0,"523",,terminal_output +1867,3352274,"TERMINAL",0,0,"634",,terminal_output +1868,3353384,"TERMINAL",0,0,"745",,terminal_output +1869,3354372,"TERMINAL",0,0,"856",,terminal_output +1870,3355525,"TERMINAL",0,0,"967",,terminal_output +1871,3356463,"TERMINAL",0,0,"1078",,terminal_output +1872,3357515,"TERMINAL",0,0,"189",,terminal_output +1873,3359651,"TERMINAL",0,0,"23051",,terminal_output +1874,3360745,"TERMINAL",0,0,"412",,terminal_output +1875,3361743,"TERMINAL",0,0,"523",,terminal_output +1876,3362796,"TERMINAL",0,0,"634",,terminal_output +1877,3363839,"TERMINAL",0,0,"756",,terminal_output +1878,3364890,"TERMINAL",0,0,"967",,terminal_output +1879,3365938,"TERMINAL",0,0,"2078",,terminal_output +1880,3366985,"TERMINAL",0,0,"189",,terminal_output +1881,3368052,"TERMINAL",0,0,"293:00",,terminal_output +1882,3368441,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1326,0,"",shellscript,selection_mouse +1883,3369054,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1134,0,"",shellscript,selection_mouse +1884,3369109,"TERMINAL",0,0,"3401",,terminal_output +1885,3370167,"TERMINAL",0,0,"412",,terminal_output +1886,3370437,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +1887,3371169,"TERMINAL",0,0,"523",,terminal_output +1888,3372214,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",802,0,"",shellscript,selection_mouse +1889,3372228,"TERMINAL",0,0,"634",,terminal_output +1890,3373341,"TERMINAL",0,0,"745",,terminal_output +1891,3373948,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",0,0,"",shellscript,tab +1892,3374315,"TERMINAL",0,0,"856",,terminal_output +1893,3375385,"TERMINAL",0,0,"967",,terminal_output +1894,3376413,"TERMINAL",0,0,"3078",,terminal_output +1895,3377537,"TERMINAL",0,0,"189",,terminal_output +1896,3378508,"TERMINAL",0,0,"2910",,terminal_output +1897,3379589,"TERMINAL",0,0,"3501",,terminal_output +1898,3380609,"TERMINAL",0,0,"412",,terminal_output +1899,3381647,"TERMINAL",0,0,"523",,terminal_output +1900,3382760,"TERMINAL",0,0,"634",,terminal_output +1901,3383788,"TERMINAL",0,0,"745",,terminal_output +1902,3384812,"TERMINAL",0,0,"856",,terminal_output +1903,3385839,"TERMINAL",0,0,"978",,terminal_output +1904,3386884,"TERMINAL",0,0,"4189",,terminal_output +1905,3387939,"TERMINAL",0,0,"2920",,terminal_output +1906,3388971,"TERMINAL",0,0,"36:001",,terminal_output +1907,3390025,"TERMINAL",0,0,"412",,terminal_output +1908,3391615,"TERMINAL",0,0,"523",,terminal_output +1909,3392664,"TERMINAL",0,0,"634",,terminal_output +1910,3393820,"TERMINAL",0,0,"745",,terminal_output +1911,3394847,"TERMINAL",0,0,"856",,terminal_output +1912,3395803,"TERMINAL",0,0,"967",,terminal_output +1913,3396859,"TERMINAL",0,0,"5089",,terminal_output +1914,3397937,"TERMINAL",0,0,"2930",,terminal_output +1915,3399040,"TERMINAL",0,0,"3101",,terminal_output +1916,3400173,"TERMINAL",0,0,"412",,terminal_output +1917,3401090,"TERMINAL",0,0,"523",,terminal_output +1918,3402221,"TERMINAL",0,0,"634",,terminal_output +1919,3403245,"TERMINAL",0,0,"745",,terminal_output +1920,3404265,"TERMINAL",0,0,"856",,terminal_output +1921,3405263,"TERMINAL",0,0,"967",,terminal_output +1922,3406312,"TERMINAL",0,0,"5:0078",,terminal_output +1923,3408382,"TERMINAL",0,0,"189",,terminal_output +1924,3408734,"TERMINAL",0,0,"2940",,terminal_output +1925,3409491,"TERMINAL",0,0,"3201",,terminal_output +1926,3410515,"TERMINAL",0,0,"412",,terminal_output +1927,3411579,"TERMINAL",0,0,"523",,terminal_output +1928,3412662,"TERMINAL",0,0,"634",,terminal_output +1929,3413685,"TERMINAL",0,0,"745",,terminal_output +1930,3414694,"TERMINAL",0,0,"856",,terminal_output +1931,3415844,"TERMINAL",0,0,"967",,terminal_output +1932,3416811,"TERMINAL",0,0,"1078",,terminal_output +1933,3417840,"TERMINAL",0,0,"1950",,terminal_output +1934,3418887,"TERMINAL",0,0,"3301",,terminal_output +1935,3419927,"TERMINAL",0,0,"412",,terminal_output +1936,3420978,"TERMINAL",0,0,"523",,terminal_output +1937,3422082,"TERMINAL",0,0,"634",,terminal_output +1938,3423107,"TERMINAL",0,0,"745",,terminal_output +1939,3424133,"TERMINAL",0,0,"856",,terminal_output +1940,3425174,"TERMINAL",0,0,"967",,terminal_output +1941,3426294,"TERMINAL",0,0,"2078",,terminal_output +1942,3427309,"TERMINAL",0,0,"189",,terminal_output +1943,3428291,"TERMINAL",0,0,"294:00",,terminal_output +1944,3429355,"TERMINAL",0,0,"3401",,terminal_output +1945,3430480,"TERMINAL",0,0,"412",,terminal_output +1946,3431503,"TERMINAL",0,0,"523",,terminal_output +1947,3432527,"TERMINAL",0,0,"634",,terminal_output +1948,3433511,"TERMINAL",0,0,"745",,terminal_output +1949,3434578,"TERMINAL",0,0,"856",,terminal_output +1950,3435604,"TERMINAL",0,0,"967",,terminal_output +1951,3436627,"TERMINAL",0,0,"3078",,terminal_output +1952,3437671,"TERMINAL",0,0,"189",,terminal_output +1953,3438720,"TERMINAL",0,0,"2910",,terminal_output +1954,3439787,"TERMINAL",0,0,"3501",,terminal_output +1955,3440829,"TERMINAL",0,0,"412",,terminal_output +1956,3441984,"TERMINAL",0,0,"534",,terminal_output +1957,3442909,"TERMINAL",0,0,"745",,terminal_output +1958,3443953,"TERMINAL",0,0,"856",,terminal_output +1959,3445023,"TERMINAL",0,0,"967",,terminal_output +1960,3446056,"TERMINAL",0,0,"4078",,terminal_output +1961,3447097,"TERMINAL",0,0,"189",,terminal_output +1962,3448150,"TERMINAL",0,0,"2920",,terminal_output +1963,3449217,"TERMINAL",0,0,"37:001",,terminal_output +1964,3450243,"TERMINAL",0,0,"412",,terminal_output +1965,3451294,"TERMINAL",0,0,"523",,terminal_output +1966,3452395,"TERMINAL",0,0,"634",,terminal_output +1967,3453389,"TERMINAL",0,0,"745",,terminal_output +1968,3454544,"TERMINAL",0,0,"856",,terminal_output +1969,3455482,"TERMINAL",0,0,"967",,terminal_output +1970,3456544,"TERMINAL",0,0,"5078",,terminal_output +1971,3457614,"TERMINAL",0,0,"189",,terminal_output +1972,3458640,"TERMINAL",0,0,"2930",,terminal_output +1973,3459672,"TERMINAL",0,0,"3101",,terminal_output +1974,3460791,"TERMINAL",0,0,"412",,terminal_output +1975,3461816,"TERMINAL",0,0,"523",,terminal_output +1976,3462945,"TERMINAL",0,0,"645",,terminal_output +1977,3463886,"TERMINAL",0,0,"856",,terminal_output +1978,3464934,"TERMINAL",0,0,"967",,terminal_output +1979,3466014,"TERMINAL",0,0,"6:0078",,terminal_output +1980,3467044,"TERMINAL",0,0,"189",,terminal_output +1981,3468164,"TERMINAL",0,0,"2940",,terminal_output +1982,3469187,"TERMINAL",0,0,"3201",,terminal_output +1983,3470170,"TERMINAL",0,0,"412",,terminal_output +1984,3471276,"TERMINAL",0,0,"523",,terminal_output +1985,3472364,"TERMINAL",0,0,"634",,terminal_output +1986,3473319,"TERMINAL",0,0,"745",,terminal_output +1987,3474414,"TERMINAL",0,0,"856",,terminal_output +1988,3475434,"TERMINAL",0,0,"967",,terminal_output +1989,3476464,"TERMINAL",0,0,"1078",,terminal_output +1990,3477513,"TERMINAL",0,0,"189",,terminal_output +1991,3478618,"TERMINAL",0,0,"2950",,terminal_output +1992,3479634,"TERMINAL",0,0,"3301",,terminal_output +1993,3480652,"TERMINAL",0,0,"412",,terminal_output +1994,3481716,"TERMINAL",0,0,"523",,terminal_output +1995,3482748,"TERMINAL",0,0,"634",,terminal_output +1996,3483812,"TERMINAL",0,0,"745",,terminal_output +1997,3484861,"TERMINAL",0,0,"867",,terminal_output +1998,3485912,"TERMINAL",0,0,"2078",,terminal_output +1999,3486949,"TERMINAL",0,0,"189",,terminal_output +2000,3487992,"TERMINAL",0,0,"295:00",,terminal_output +2001,3489024,"TERMINAL",0,0,"3401",,terminal_output +2002,3490191,"TERMINAL",0,0,"412",,terminal_output +2003,3491118,"TERMINAL",0,0,"523",,terminal_output +2004,3492161,"TERMINAL",0,0,"634",,terminal_output +2005,3493252,"TERMINAL",0,0,"745",,terminal_output +2006,3494263,"TERMINAL",0,0,"856",,terminal_output +2007,3495404,"TERMINAL",0,0,"967",,terminal_output +2008,3496355,"TERMINAL",0,0,"3078",,terminal_output +2009,3497395,"TERMINAL",0,0,"189",,terminal_output +2010,3498446,"TERMINAL",0,0,"2910",,terminal_output +2011,3499499,"TERMINAL",0,0,"3501",,terminal_output +2012,3500532,"TERMINAL",0,0,"412",,terminal_output +2013,3501588,"TERMINAL",0,0,"523",,terminal_output +2014,3502678,"TERMINAL",0,0,"634",,terminal_output +2015,3503698,"TERMINAL",0,0,"745",,terminal_output +2016,3504742,"TERMINAL",0,0,"856",,terminal_output +2017,3505767,"TERMINAL",0,0,"967",,terminal_output +2018,3506872,"TERMINAL",0,0,"4078",,terminal_output +2019,3507858,"TERMINAL",0,0,"1920",,terminal_output +2020,3508902,"TERMINAL",0,0,"38:001",,terminal_output +2021,3509949,"TERMINAL",0,0,"412",,terminal_output +2022,3510990,"TERMINAL",0,0,"523",,terminal_output +2023,3512193,"TERMINAL",0,0,"634",,terminal_output +2024,3513221,"TERMINAL",0,0,"745",,terminal_output +2025,3514303,"TERMINAL",0,0,"856",,terminal_output +2026,3515300,"TERMINAL",0,0,"967",,terminal_output +2027,3516347,"TERMINAL",0,0,"5078",,terminal_output +2028,3517388,"TERMINAL",0,0,"189",,terminal_output +2029,3518544,"TERMINAL",0,0,"2930",,terminal_output +2030,3519484,"TERMINAL",0,0,"3101",,terminal_output +2031,3520595,"TERMINAL",0,0,"412",,terminal_output +2032,3521581,"TERMINAL",0,0,"523",,terminal_output +2033,3522639,"TERMINAL",0,0,"634",,terminal_output +2034,3523607,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/train_lam_action_space_scaling_20/3318547/lam_1751657975_200000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=12 \\n --min_lr=0 \\n --max_lr=3e-4 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-batch-size-scaling-1-node-$slurm_job_id \\n --tags dynamics debug \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +2035,3523676,"TERMINAL",0,0,"745",,terminal_output +2036,3524838,"TERMINAL",0,0,"856",,terminal_output +2037,3525827,"TERMINAL",0,0,"978",,terminal_output +2038,3526875,"TERMINAL",0,0,"7:0189",,terminal_output +2039,3527919,"TERMINAL",0,0,"2940",,terminal_output +2040,3528970,"TERMINAL",0,0,"3201",,terminal_output +2041,3530021,"TERMINAL",0,0,"412",,terminal_output +2042,3531140,"TERMINAL",0,0,"523",,terminal_output +2043,3532108,"TERMINAL",0,0,"634",,terminal_output +2044,3533150,"TERMINAL",0,0,"745",,terminal_output +2045,3534204,"TERMINAL",0,0,"856",,terminal_output +2046,3535339,"TERMINAL",0,0,"967",,terminal_output +2047,3536300,"TERMINAL",0,0,"1078",,terminal_output +2048,3537444,"TERMINAL",0,0,"189",,terminal_output +2049,3538379,"TERMINAL",0,0,"2950",,terminal_output +2050,3539408,"TERMINAL",0,0,"3301",,terminal_output +2051,3540465,"TERMINAL",0,0,"412",,terminal_output +2052,3541599,"TERMINAL",0,0,"523",,terminal_output +2053,3542597,"TERMINAL",0,0,"634",,terminal_output +2054,3543599,"TERMINAL",0,0,"745",,terminal_output +2055,3544661,"TERMINAL",0,0,"856",,terminal_output +2056,3545673,"TERMINAL",0,0,"967",,terminal_output +2057,3546716,"TERMINAL",0,0,"2078",,terminal_output +2058,3547843,"TERMINAL",0,0,"189",,terminal_output +2059,3548823,"TERMINAL",0,0,"296:00",,terminal_output +2060,3549656,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics/sqrt_lr/train_dynamics_32_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=32\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=05:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_32_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1536 \\n --min_lr=3e-4 \\n --max_lr=3e-4 \\n --log_image_interval=250 \\n --log \\n --name=tokenizer-batch-size-scaling-32-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 32-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +2061,3549883,"TERMINAL",0,0,"3412",,terminal_output +2062,3550526,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics/sqrt_lr/train_dynamics_16_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=16\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_16_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=768 \\n --min_lr=0 \\n --max_lr=4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-16-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 16-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +2063,3550890,"TERMINAL",0,0,"523",,terminal_output +2064,3551933,"TERMINAL",0,0,"634",,terminal_output +2065,3552975,"TERMINAL",0,0,"745",,terminal_output +2066,3554015,"TERMINAL",0,0,"856",,terminal_output +2067,3555059,"TERMINAL",0,0,"967",,terminal_output +2068,3556129,"TERMINAL",0,0,"3078",,terminal_output +2069,3557153,"TERMINAL",0,0,"189",,terminal_output +2070,3558199,"TERMINAL",0,0,"2910",,terminal_output +2071,3559307,"TERMINAL",0,0,"3501",,terminal_output +2072,3560295,"TERMINAL",0,0,"412",,terminal_output +2073,3561348,"TERMINAL",0,0,"523",,terminal_output +2074,3562385,"TERMINAL",0,0,"634",,terminal_output +2075,3563498,"TERMINAL",0,0,"745",,terminal_output +2076,3564354,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/train_lam_action_space_scaling_20/3318547/lam_1751657975_200000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=12 \\n --min_lr=0 \\n --max_lr=3e-4 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-batch-size-scaling-1-node-$slurm_job_id \\n --tags dynamics debug \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +2077,3564476,"TERMINAL",0,0,"856",,terminal_output +2078,3565548,"TERMINAL",0,0,"967",,terminal_output +2079,3565705,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1315,0,"",shellscript,selection_mouse +2080,3565727,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1314,0,"",shellscript,selection_command +2081,3566573,"TERMINAL",0,0,"4078",,terminal_output +2082,3567613,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1295,0,"",shellscript,selection_mouse +2083,3567615,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1294,0,"",shellscript,selection_command +2084,3567672,"TERMINAL",0,0,"189",,terminal_output +2085,3568164,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1278,0,"",shellscript,selection_mouse +2086,3568168,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1277,0,"",shellscript,selection_command +2087,3568725,"TERMINAL",0,0,"2920",,terminal_output +2088,3569706,"TERMINAL",0,0,"39:001",,terminal_output +2089,3570773,"TERMINAL",0,0,"412",,terminal_output +2090,3571551,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1295,0,"",shellscript,selection_mouse +2091,3571555,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1294,0,"",shellscript,selection_command +2092,3571794,"TERMINAL",0,0,"523",,terminal_output +2093,3572142,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1276,0,"",shellscript,selection_mouse +2094,3572921,"TERMINAL",0,0,"645",,terminal_output +2095,3573883,"TERMINAL",0,0,"856",,terminal_output +2096,3574937,"TERMINAL",0,0,"967",,terminal_output +2097,3575398,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1275,1,"",shellscript,content +2098,3575550,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1274,1,"",shellscript,content +2099,3575986,"TERMINAL",0,0,"5078",,terminal_output +2100,3577039,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1274,0,"4",shellscript,content +2101,3577040,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1275,0,"",shellscript,selection_keyboard +2102,3577063,"TERMINAL",0,0,"189",,terminal_output +2103,3577106,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1275,0,"8",shellscript,content +2104,3577107,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1276,0,"",shellscript,selection_keyboard +2105,3577725,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1275,0,"",shellscript,selection_command +2106,3577849,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1242,0,"",shellscript,selection_command +2107,3578034,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1275,0,"",shellscript,selection_command +2108,3578087,"TERMINAL",0,0,"2930",,terminal_output +2109,3578178,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1294,0,"",shellscript,selection_command +2110,3578295,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1314,0,"",shellscript,selection_command +2111,3578773,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1313,0,"",shellscript,selection_command +2112,3578923,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1312,0,"",shellscript,selection_command +2113,3579036,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1311,0,"",shellscript,selection_command +2114,3579143,"TERMINAL",0,0,"3101",,terminal_output +2115,3579227,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1310,0,"",shellscript,selection_command +2116,3579372,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1309,0,"",shellscript,selection_command +2117,3580192,"TERMINAL",0,0,"412",,terminal_output +2118,3581249,"TERMINAL",0,0,"523",,terminal_output +2119,3582134,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +2120,3582287,"TERMINAL",0,0,"634",,terminal_output +2121,3583243,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1007,0,"",shellscript,selection_mouse +2122,3583343,"TERMINAL",0,0,"745",,terminal_output +2123,3584352,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1007,1,"5",shellscript,content +2124,3584411,"TERMINAL",0,0,"856",,terminal_output +2125,3585429,"TERMINAL",0,0,"967",,terminal_output +2126,3586542,"TERMINAL",0,0,"8:0078",,terminal_output +2127,3587527,"TERMINAL",0,0,"189",,terminal_output +2128,3588589,"TERMINAL",0,0,"2940",,terminal_output +2129,3589672,"TERMINAL",0,0,"3201",,terminal_output +2130,3590685,"TERMINAL",0,0,"412",,terminal_output +2131,3591501,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",84,0,"",shellscript,selection_mouse +2132,3591706,"TERMINAL",0,0,"523",,terminal_output +2133,3592787,"TERMINAL",0,0,"634",,terminal_output +2134,3593809,"TERMINAL",0,0,"745",,terminal_output +2135,3594872,"TERMINAL",0,0,"867",,terminal_output +2136,3595960,"TERMINAL",0,0,"1078",,terminal_output +2137,3596927,"TERMINAL",0,0,"189",,terminal_output +2138,3597966,"TERMINAL",0,0,"2950",,terminal_output +2139,3598616,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",397,0,"",shellscript,selection_mouse +2140,3599009,"TERMINAL",0,0,"3301",,terminal_output +2141,3599511,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1310,0,"",shellscript,selection_mouse +2142,3599648,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1305,5,"_dir\n",shellscript,selection_mouse +2143,3599668,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1060,250,"int_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2144,3599724,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",842,468,"kdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2145,3599731,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",21,1289,"#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2146,3599732,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,1310,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2147,3600074,"TERMINAL",0,0,"412",,terminal_output +2148,3601088,"TERMINAL",0,0,"523",,terminal_output +2149,3601318,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1310,0,"",shellscript,selection_mouse +2150,3601664,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1309,1,"\n",shellscript,selection_mouse +2151,3601671,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",990,320,"\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2152,3601689,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",715,595,"SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2153,3601714,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",439,871,"sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2154,3601744,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",124,1186,"SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2155,3601759,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",40,1270,"SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2156,3601788,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",22,1288,"SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2157,3602205,"TERMINAL",0,0,"634",,terminal_output +2158,3602276,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",92,1218,"SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2159,3602300,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",150,1160,"SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2160,3602355,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",428,882,"\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2161,3602355,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",498,812,"unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2162,3602356,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",686,624,"SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2163,3602368,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",728,582,"\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2164,3602384,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",851,459,"CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2165,3602441,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",866,444,"\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2166,3602451,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",883,427,"\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2167,3602507,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",866,444,"\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2168,3602508,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",851,459,"CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2169,3602519,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",715,595,"SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2170,3602537,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",460,850,"\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2171,3602556,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",278,1032,"-error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2172,3602573,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",92,1218,"SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2173,3602588,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",40,1270,"SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2174,3602603,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",22,1288,"SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2175,3602670,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",20,1290,"\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2176,3603180,"TERMINAL",0,0,"745",,terminal_output +2177,3603284,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",700,0,"",shellscript,selection_mouse +2178,3603301,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",699,0,"",shellscript,selection_command +2179,3603939,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",635,0,"",shellscript,selection_mouse +2180,3604084,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",631,30,"open_ai_minecraft_arrayrecords",shellscript,selection_mouse +2181,3604220,"TERMINAL",0,0,"856",,terminal_output +2182,3604676,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",650,0,"",shellscript,selection_mouse +2183,3604677,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",631,30,"open_ai_minecraft_arrayrecords",shellscript,selection_mouse +2184,3605280,"TERMINAL",0,0,"967",,terminal_output +2185,3606304,"TERMINAL",0,0,"2078",,terminal_output +2186,3607347,"TERMINAL",0,0,"189",,terminal_output +2187,3608456,"TERMINAL",0,0,"297:00",,terminal_output +2188,3609428,"TERMINAL",0,0,"3401",,terminal_output +2189,3609610,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2190,3610502,"TERMINAL",0,0,"412",,terminal_output +2191,3611740,"TERMINAL",0,0,"523",,terminal_output +2192,3612505,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,tab +2193,3612562,"TERMINAL",0,0,"634",,terminal_output +2194,3613605,"TERMINAL",0,0,"745",,terminal_output +2195,3614697,"TERMINAL",0,0,"856",,terminal_output +2196,3615723,"TERMINAL",0,0,"967",,terminal_output +2197,3616755,"TERMINAL",0,0,"3078",,terminal_output +2198,3617871,"TERMINAL",0,0,"189",,terminal_output +2199,3618818,"TERMINAL",0,0,"25011",,terminal_output +2200,3619801,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",0,0,"",shellscript,tab +2201,3619884,"TERMINAL",0,0,"412",,terminal_output +2202,3620913,"TERMINAL",0,0,"523",,terminal_output +2203,3621973,"TERMINAL",0,0,"634",,terminal_output +2204,3622968,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",0,0,"",shellscript,tab +2205,3623037,"TERMINAL",0,0,"745",,terminal_output +2206,3623871,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",0,0,"",shellscript,tab +2207,3624090,"TERMINAL",0,0,"856",,terminal_output +2208,3625024,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",669,0,"",shellscript,selection_mouse +2209,3625126,"TERMINAL",0,0,"967",,terminal_output +2210,3625321,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",669,1,"\n",shellscript,selection_mouse +2211,3625789,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",649,0,"",shellscript,selection_mouse +2212,3625937,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",630,38,"open_ai_minecraft_arrayrecords_chunked",shellscript,selection_mouse +2213,3626151,"TERMINAL",0,0,"4078",,terminal_output +2214,3627294,"TERMINAL",0,0,"189",,terminal_output +2215,3627974,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",623,0,"",shellscript,selection_mouse +2216,3628092,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",621,8,"data_new",shellscript,selection_mouse +2217,3628220,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",547,122,"array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n",shellscript,selection_mouse +2218,3628235,"TERMINAL",0,0,"2920",,terminal_output +2219,3629348,"TERMINAL",0,0,"310:001",,terminal_output +2220,3630301,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,tab +2221,3630378,"TERMINAL",0,0,"412",,terminal_output +2222,3631355,"TERMINAL",0,0,"523",,terminal_output +2223,3632618,"TERMINAL",0,0,"634",,terminal_output +2224,3633641,"TERMINAL",0,0,"745",,terminal_output +2225,3633910,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2226,3634678,"TERMINAL",0,0,"856",,terminal_output +2227,3635356,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",547,0,"",shellscript,selection_mouse +2228,3635508,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",547,127,"\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90",shellscript,selection_mouse +2229,3635539,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",547,117,"\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10",shellscript,selection_mouse +2230,3635564,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",547,96,"\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_mine",shellscript,selection_mouse +2231,3635580,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",547,81,"\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_n",shellscript,selection_mouse +2232,3635595,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",547,128,"\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n",shellscript,selection_mouse +2233,3635720,"TERMINAL",0,0,"967",,terminal_output +2234,3636072,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",675,0,"",shellscript,selection_mouse +2235,3636561,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",596,0,"",shellscript,selection_mouse +2236,3636762,"TERMINAL",0,0,"5078",,terminal_output +2237,3637444,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",597,0," --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=82 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-tf-records-$slurm_job_id \\n --tags dynamics yolo-run tf_records \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir",shellscript,content +2238,3637445,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",597,0,"",shellscript,selection_command +2239,3637815,"TERMINAL",0,0,"189",,terminal_output +2240,3638853,"TERMINAL",0,0,"21031",,terminal_output +2241,3638882,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",597,432,"",shellscript,content +2242,3638941,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",596,0,"",shellscript,selection_command +2243,3639277,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",674,0,"\n",shellscript,content +2244,3639480,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",675,0,"array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n",shellscript,content +2245,3639900,"TERMINAL",0,0,"412",,terminal_output +2246,3640156,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",675,0,"",shellscript,selection_command +2247,3640476,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",548,0,"",shellscript,selection_command +2248,3640809,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",548,127,"",shellscript,content +2249,3640948,"TERMINAL",0,0,"523",,terminal_output +2250,3641996,"TERMINAL",0,0,"634",,terminal_output +2251,3642582,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/train_lam_action_space_scaling_20/3318547/lam_1751657975_200000/\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --lam_checkpoint=$lam_ckpt_dir\n",shellscript,tab +2252,3643033,"TERMINAL",0,0,"745",,terminal_output +2253,3643319,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",583,0,"",shellscript,selection_mouse +2254,3643925,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",669,0,"\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",shellscript,content +2255,3643930,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",670,0,"",shellscript,selection_command +2256,3644084,"TERMINAL",0,0,"856",,terminal_output +2257,3644474,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",548,0,"",shellscript,selection_command +2258,3645135,"TERMINAL",0,0,"967",,terminal_output +2259,3645304,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",548,122,"",shellscript,content +2260,3646284,"TERMINAL",0,0,"9:0078",,terminal_output +2261,3646664,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=4\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_4_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=192 \\n --min_lr=0 \\n --max_lr=2.00e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-4-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 4-node sqrt-lr \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +2262,3647210,"TERMINAL",0,0,"189",,terminal_output +2263,3648007,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",623,0,"",shellscript,selection_mouse +2264,3648254,"TERMINAL",0,0,"2940",,terminal_output +2265,3648661,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",674,0,"\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",shellscript,content +2266,3648676,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",675,0,"",shellscript,selection_command +2267,3648906,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",548,0,"",shellscript,selection_command +2268,3649196,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",548,127,"",shellscript,content +2269,3649317,"TERMINAL",0,0,"3201",,terminal_output +2270,3649887,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",623,0,"",shellscript,selection_mouse +2271,3650146,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",622,8,"data_new",shellscript,selection_mouse +2272,3650356,"TERMINAL",0,0,"412",,terminal_output +2273,3650537,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",622,0,"",shellscript,selection_command +2274,3651379,"TERMINAL",0,0,"523",,terminal_output +2275,3652486,"TERMINAL",0,0,"634",,terminal_output +2276,3653458,"TERMINAL",0,0,"745",,terminal_output +2277,3653662,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,tab +2278,3654494,"TERMINAL",0,0,"856",,terminal_output +2279,3655559,"TERMINAL",0,0,"967",,terminal_output +2280,3656621,"TERMINAL",0,0,"1078",,terminal_output +2281,3657561,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",609,0,"",shellscript,selection_mouse +2282,3657629,"TERMINAL",0,0,"189",,terminal_output +2283,3658677,"TERMINAL",0,0,"2950",,terminal_output +2284,3659167,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",674,0,"\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",shellscript,content +2285,3659177,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",675,0,"",shellscript,selection_command +2286,3660059,"TERMINAL",0,0,"3312",,terminal_output +2287,3660410,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",547,0,"",shellscript,selection_mouse +2288,3660798,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",609,0,"",shellscript,selection_mouse +2289,3661081,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",548,127,"",shellscript,content +2290,3661141,"TERMINAL",0,0,"523",,terminal_output +2291,3661394,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",670,0,"",shellscript,selection_command +2292,3661620,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",548,0,"",shellscript,selection_command +2293,3662167,"TERMINAL",0,0,"634",,terminal_output +2294,3663241,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=16\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_16_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=768 \\n --min_lr=0 \\n --max_lr=4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-16-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 16-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +2295,3663335,"TERMINAL",0,0,"745",,terminal_output +2296,3664285,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",677,0,"",shellscript,selection_mouse +2297,3664300,"TERMINAL",0,0,"856",,terminal_output +2298,3665013,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",677,0,"\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",shellscript,content +2299,3665029,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",678,0,"",shellscript,selection_command +2300,3665225,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",677,0,"",shellscript,selection_command +2301,3665329,"TERMINAL",0,0,"967",,terminal_output +2302,3665550,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",677,1,"",shellscript,content +2303,3665859,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",550,0,"",shellscript,selection_command +2304,3666188,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",550,127,"",shellscript,content +2305,3666313,"TERMINAL",0,0,"2078",,terminal_output +2306,3666395,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",672,0,"",shellscript,selection_command +2307,3666807,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",550,0,"",shellscript,selection_command +2308,3666995,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",671,0,"\n",shellscript,content +2309,3667352,"TERMINAL",0,0,"189",,terminal_output +2310,3668389,"TERMINAL",0,0,"298:00",,terminal_output +2311,3668838,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,tab +2312,3669440,"TERMINAL",0,0,"3401",,terminal_output +2313,3669874,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"",shellscript,tab +2314,3670483,"TERMINAL",0,0,"412",,terminal_output +2315,3670793,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2316,3671358,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2317,3671605,"TERMINAL",0,0,"523",,terminal_output +2318,3672148,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2319,3672578,"TERMINAL",0,0,"634",,terminal_output +2320,3672870,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"",shellscript,tab +2321,3673631,"TERMINAL",0,0,"745",,terminal_output +2322,3674707,"TERMINAL",0,0,"856",,terminal_output +2323,3675306,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2324,3675729,"TERMINAL",0,0,"967",,terminal_output +2325,3675973,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"",shellscript,tab +2326,3676872,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,tab +2327,3676908,"TERMINAL",0,0,"3078",,terminal_output +2328,3677740,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",0,0,"",shellscript,tab +2329,3677850,"TERMINAL",0,0,"189",,terminal_output +2330,3678305,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,tab +2331,3678695,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"",shellscript,tab +2332,3678849,"TERMINAL",0,0,"25011",,terminal_output +2333,3679918,"TERMINAL",0,0,"412",,terminal_output +2334,3680805,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2335,3680972,"TERMINAL",0,0,"523",,terminal_output +2336,3681822,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2337,3681992,"TERMINAL",0,0,"634",,terminal_output +2338,3683054,"TERMINAL",0,0,"745",,terminal_output +2339,3684086,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2340,3684104,"TERMINAL",0,0,"856",,terminal_output +2341,3685153,"TERMINAL",0,0,"967",,terminal_output +2342,3685704,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2343,3686159,"TERMINAL",0,0,"4078",,terminal_output +2344,3687300,"TERMINAL",0,0,"189",,terminal_output +2345,3687444,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2346,3688324,"TERMINAL",0,0,"2920",,terminal_output +2347,3688468,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2348,3689164,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",671,0,"",shellscript,selection_mouse +2349,3689321,"TERMINAL",0,0,"31:001",,terminal_output +2350,3690067,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",671,1,"",shellscript,content +2351,3690379,"TERMINAL",0,0,"412",,terminal_output +2352,3691456,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2353,3691475,"TERMINAL",0,0,"523",,terminal_output +2354,3692392,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2355,3692471,"TERMINAL",0,0,"634",,terminal_output +2356,3693551,"TERMINAL",0,0,"745",,terminal_output +2357,3693683,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2358,3694572,"TERMINAL",0,0,"856",,terminal_output +2359,3695123,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2360,3695586,"TERMINAL",0,0,"967",,terminal_output +2361,3696031,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2362,3696638,"TERMINAL",0,0,"5078",,terminal_output +2363,3696814,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2364,3697351,"TERMINAL",0,0,"Step 241, loss: 6.375746250152588\r\nStep 242, loss: 6.384668350219727\r\nStep 243, loss: 6.2770094871521\r\nStep 244, loss: 6.596880912780762\r\nStep 245, loss: 6.688802242279053\r\nStep 246, loss: 6.219254493713379\r\nStep 247, loss: 6.591236591339111\r\nStep 248, loss: 6.286152362823486\r\nStep 249, loss: 6.393139839172363\r\nStep 250, loss: 6.294431686401367\r\nStep 251, loss: 6.2147321701049805\r\nStep 252, loss: 6.365173816680908\r\nStep 253, loss: 6.267394065856934\r\nStep 254, loss: 6.570861339569092\r\nStep 255, loss: 6.34361457824707\r\nStep 256, loss: 6.068459510803223\r\nStep 257, loss: 6.029990196228027\r\nStep 258, loss: 6.469965934753418\r\nStep 259, loss: 6.202692985534668\r\nStep 260, loss: 6.448121070861816\r\nStep 261, loss: 6.35492467880249\r\nStep 262, loss: 6.2771992683410645\r\nStep 263, loss: 6.350648403167725\r\nStep 264, loss: 6.209310054779053\r\nStep 265, loss: 6.30444860458374\r\nStep 266, loss: 6.190724849700928\r\nStep 267, loss: 6.2750067710876465\r\nStep 268, loss: 6.282001972198486\r\nStep 269, loss: 6.261999130249023\r\nStep 270, loss: 6.238827228546143\r\nStep 271, loss: 6.399940013885498\r\nStep 272, loss: 6.577332496643066\r\nStep 273, loss: 6.290578365325928\r\nStep 274, loss: 6.126301288604736\r\nStep 275, loss: 6.4832916259765625\r\nStep 276, loss: 6.190239906311035\r\nStep 277, loss: 6.054000377655029\r\nStep 278, loss: 6.173544406890869\r\nStep 279, loss: 5.889945030212402\r\nStep 280, loss: 6.498538494110107\r\nStep 281, loss: 6.207729816436768\r\nStep 282, loss: 6.37006139755249\r\nStep 283, loss: 6.058416366577148\r\nStep 284, loss: 6.195253372192383\r\nStep 285, loss: 6.223158359527588\r\nStep 286, loss: 5.904426574707031\r\nStep 287, loss: 6.371645450592041\r\nStep 288, loss: 6.0909423828125\r\nStep 289, loss: 6.016209125518799\r\nStep 290, loss: 6.319799900054932\r\nStep 291, loss: 6.296480178833008\r\nStep 292, loss: 6.148037910461426\r\nStep 293, loss: 5.923506259918213\r\nStep 294, loss: 5.763485431671143\r\nStep 295, loss: 6.133178234100342\r\nStep 296, loss: 6.073211193084717\r\nStep 297, loss: 6.169839859008789\r\nStep 298, loss: 6.141224384307861\r\nStep 299, loss: 6.168487071990967\r\nStep 300, loss: 6.119058132171631\r\nStep 301, loss: 6.03639554977417\r\nStep 302, loss: 6.216285228729248\r\nStep 303, loss: 6.022148609161377\r\nStep 304, loss: 5.897799968719482\r\nStep 305, loss: 5.894784927368164\r\nStep 306, loss: 6.132763385772705\r\nStep 307, loss: 5.888872146606445\r\nStep 308, loss: 6.012820243835449\r\nStep 309, loss: 6.126757621765137\r\nStep 310, loss: 6.202884197235107\r\nStep 311, loss: 6.018051624298096\r\nStep 312, loss: 6.108768939971924\r\nStep 313, loss: 5.92648458480835\r\nStep 314, loss: 6.071450710296631\r\nStep 315, loss: 5.71877384185791\r\nStep 316, loss: 5.99001407623291\r\nStep 317, loss: 5.833607196807861\r\nStep 318, loss: 6.032952785491943\r\nStep 319, loss: 6.038209438323975\r\nStep 320, loss: 6.10830545425415\r\nStep 321, loss: 5.989357948303223\r\nStep 322, loss: 6.006596088409424\r\nStep 323, loss: 6.012803077697754\r\nStep 324, loss: 6.0025787353515625\r\nStep 325, loss: 5.972760200500488\r\nStep 326, loss: 6.077414512634277\r\nStep 327, loss: 5.926943302154541\r\nStep 328, loss: 6.139301300048828\r\nStep 329, loss: 5.791858196258545\r\nStep 330, loss: 5.903587818145752\r\nStep 331, loss: 6.382968902587891\r\nStep 332, loss: 6.010067462921143\r\nStep 333, loss: 6.022067070007324\r\nStep 334, loss: 5.914710998535156\r\nStep 335, loss: 6.026205539703369\r\nStep 336, loss: 6.101585865020752\r\nStep 337, loss: 5.8053202629089355\r\nStep 338, loss: 5.739813327789307\r\nStep 339, loss: 5.901891231536865\r\nStep 340, loss: 5.769777774810791\r\nStep 341, loss: 5.84769868850708\r\nStep 342, loss: 5.975956916809082\r\nStep 343, loss: 5.790035247802734\r\nStep 344, loss: 5.943611145019531\r\nStep 345, loss: 5.850645065307617\r\nStep 346, loss: 5.7808990478515625\r\nStep 347, loss: 5.690489292144775\r\nStep 348, loss: 5.986003398895264\r\nStep 349, loss: 5.836080551147461\r\nStep 350, loss: 5.887240886688232\r\nStep 351, loss: 5.904907703399658\r\nStep 352, loss: 5.790065288543701\r\nStep 353, loss: 5.890659809112549\r\nStep 354, loss: 5.948431015014648\r\nStep 355, loss: 5.6976847648620605\r\nStep 356, loss: 5.886941909790039\r\nStep 357, loss: 5.837642192840576\r\nStep 358, loss: 5.8052215576171875\r\nStep 359, loss: 5.958202838897705\r\nStep 360, loss: 6.015711784362793\r\nStep 241, loss: 6.375746250152588\r\nStep 242, loss: 6.384668350219727\r\nStep 243, loss: 6.2770094871521\r\nStep 244, loss: 6.596880912780762\r\nStep 245, loss: 6.688802242279053\r\nStep 246, loss: 6.219254493713379\r\nStep 247, loss: 6.591236591339111\r\nStep 248, loss: 6.286152362823486\r\nStep 249, loss: 6.393139839172363\r\nStep 250, loss: 6.294431686401367\r\nStep 251, loss: 6.2147321701049805\r\nStep 252, loss: 6.365173816680908\r\nStep 253, loss: 6.267394065856934\r\nStep 254, loss: 6.570861339569092\r\nStep 255, loss: 6.34361457824707\r\nStep 256, loss: 6.068459510803223\r\nStep 257, loss: 6.029990196228027\r\nStep 258, loss: 6.469965934753418\r\nStep 259, loss: 6.202692985534668\r\nStep 260, loss: 6.448121070861816\r\nStep 261, loss: 6.35492467880249\r\nStep 262, loss: 6.2771992683410645\r\nStep 263, loss: 6.350648403167725\r\nStep 264, loss: 6.209310054779053\r\nStep 265, loss: 6.30444860458374\r\nStep 266, loss: 6.190724849700928\r\nStep 267, loss: 6.2750067710876465\r\nStep 268, loss: 6.282001972198486\r\nStep 269, loss: 6.261999130249023\r\nStep 270, loss: 6.238827228546143\r\nStep 241, loss: 6.375746250152588\r\nStep 242, loss: 6.384668350219727\r\nStep 243, loss: 6.2770094871521\r\nStep 244, loss: 6.596880912780762\r\nStep 245, loss: 6.688802242279053\r\nStep 246, loss: 6.219254493713379\r\nStep 247, loss: 6.591236591339111\r\nStep 248, loss: 6.286152362823486\r\nStep 249, loss: 6.393139839172363\r\nStep 250, loss: 6.294431686401367\r\nStep 251, loss: 6.2147321701049805\r\nStep 252, loss: 6.365173816680908\r\nStep 253, loss: 6.267394065856934\r\nStep 254, loss: 6.570861339569092\r\nStep 255, loss: 6.34361457824707\r\nStep 256, loss: 6.068459510803223\r\nStep 257, loss: 6.029990196228027\r\nStep 258, loss: 6.469965934753418\r\nStep 259, loss: 6.202692985534668\r\nStep 260, loss: 6.448121070861816\r\nStep 261, loss: 6.35492467880249\r\nStep 262, loss: 6.2771992683410645\r\nStep 263, loss: 6.350648403167725\r\nStep 264, loss: 6.209310054779053\r\nStep 265, loss: 6.30444860458374\r\nStep 266, loss: 6.190724849700928\r\nStep 267, loss: 6.2750067710876465\r\nStep 268, loss: 6.282001972198486\r\nStep 269, loss: 6.261999130249023\r\nStep 270, loss: 6.238827228546143\r\nStep 271, loss: 6.399940013885498\r\nStep 272, loss: 6.577332496643066\r\nStep 273, loss: 6.290578365325928\r\nStep 274, loss: 6.126301288604736\r\nStep 275, loss: 6.4832916259765625\r\nStep 276, loss: 6.190239906311035\r\nStep 277, loss: 6.054000377655029\r\nStep 278, loss: 6.173544406890869\r\nStep 279, loss: 5.889945030212402\r\nStep 280, loss: 6.498538494110107\r\nStep 281, loss: 6.207729816436768\r\nStep 282, loss: 6.37006139755249\r\nStep 283, loss: 6.058416366577148\r\nStep 284, loss: 6.195253372192383\r\nStep 285, loss: 6.223158359527588\r\nStep 286, loss: 5.904426574707031\r\nStep 287, loss: 6.371645450592041\r\nStep 288, loss: 6.0909423828125\r\nStep 289, loss: 6.016209125518799\r\nStep 290, loss: 6.319799900054932\r\nStep 291, loss: 6.296480178833008\r\nStep 292, loss: 6.148037910461426\r\nStep 293, loss: 5.923506259918213\r\nStep 294, loss: 5.763485431671143\r\nStep 295, loss: 6.133178234100342\r\nStep 296, loss: 6.073211193084717\r\nStep 297, loss: 6.169839859008789\r\nStep 298, loss: 6.141224384307861\r\nStep 299, loss: 6.168487071990967\r\nStep 300, loss: 6.119058132171631\r\nStep 271, loss: 6.399940013885498\r\nStep 272, loss: 6.577332496643066\r\nStep 273, loss: 6.290578365325928\r\nStep 274, loss: 6.126301288604736\r\nStep 275, loss: 6.4832916259765625\r\nStep 276, loss: 6.190239906311035\r\nStep 277, loss: 6.054000377655029\r\nStep 278, loss: 6.173544406890869\r\nStep 279, loss: 5.889945030212402\r\nStep 280, loss: 6.498538494110107\r\nStep 281, loss: 6.207729816436768\r\nStep 282, loss: 6.37006139755249\r\nStep 283, loss: 6.058416366577148\r\nStep 284, loss: 6.195253372192383\r\nStep 285, loss: 6.223158359527588\r\nStep 286, loss: 5.904426574707031\r\nStep 287, loss: 6.371645450592041\r\nStep 288, loss: 6.0909423828125\r\nStep 289, loss: 6.016209125518799\r\nStep 290, loss: 6.319799900054932\r\nStep 291, loss: 6.296480178833008\r\nStep 292, loss: 6.148037910461426\r\nStep 293, loss: 5.923506259918213\r\nStep 294, loss: 5.763485431671143\r\nStep 295, loss: 6.133178234100342\r\nStep 296, loss: 6.073211193084717\r\nStep 297, loss: 6.169839859008789\r\nStep 298, loss: 6.141224384307861\r\nStep 299, loss: 6.168487071990967\r\nStep 300, loss: 6.119058132171631\r\nStep 301, loss: 6.03639554977417\r\nStep 302, loss: 6.216285228729248\r\nStep 303, loss: 6.022148609161377\r\nStep 304, loss: 5.897799968719482\r\nStep 305, loss: 5.894784927368164\r\nStep 306, loss: 6.132763385772705\r\nStep 307, loss: 5.888872146606445\r\nStep 308, loss: 6.012820243835449\r\nStep 309, loss: 6.126757621765137\r\nStep 310, loss: 6.202884197235107\r\nStep 311, loss: 6.018051624298096\r\nStep 312, loss: 6.108768939971924\r\nStep 313, loss: 5.92648458480835\r\nStep 314, loss: 6.071450710296631\r\nStep 315, loss: 5.71877384185791\r\nStep 316, loss: 5.99001407623291\r\nStep 317, loss: 5.833607196807861\r\nStep 318, loss: 6.032952785491943\r\nStep 319, loss: 6.038209438323975\r\nStep 320, loss: 6.10830545425415\r\nStep 321, loss: 5.989357948303223\r\nStep 322, loss: 6.006596088409424\r\nStep 323, loss: 6.012803077697754\r\nStep 324, loss: 6.0025787353515625\r\nStep 325, loss: 5.972760200500488\r\nStep 326, loss: 6.077414512634277\r\nStep 327, loss: 5.926943302154541\r\nStep 328, loss: 6.139301300048828\r\nStep 329, loss: 5.791858196258545\r\nStep 330, loss: 5.903587818145752\r\nStep 301, loss: 6.03639554977417\r\nStep 302, loss: 6.216285228729248\r\nStep 303, loss: 6.022148609161377\r\nStep 304, loss: 5.897799968719482\r\nStep 305, loss: 5.894784927368164\r\nStep 306, loss: 6.132763385772705\r\nStep 307, loss: 5.888872146606445\r\nStep 308, loss: 6.012820243835449\r\nStep 309, loss: 6.126757621765137\r\nStep 310, loss: 6.202884197235107\r\nStep 311, loss: 6.018051624298096\r\nStep 312, loss: 6.108768939971924\r\nStep 313, loss: 5.92648458480835\r\nStep 314, loss: 6.071450710296631\r\nStep 315, loss: 5.71877384185791\r\nStep 316, loss: 5.99001407623291\r\nStep 317, loss: 5.833607196807861\r\nStep 318, loss: 6.032952785491943\r\nStep 319, loss: 6.038209438323975\r\nStep 320, loss: 6.10830545425415\r\nStep 321, loss: 5.989357948303223\r\nStep 322, loss: 6.006596088409424\r\nStep 323, loss: 6.012803077697754\r\nStep 324, loss: 6.0025787353515625\r\nStep 325, loss: 5.972760200500488\r\nStep 326, loss: 6.077414512634277\r\nStep 327, loss: 5.926943302154541\r\nStep 328, loss: 6.139301300048828\r\nStep 329, loss: 5.791858196258545\r\nStep 330, loss: 5.903587818145752\r\nStep 331, loss: 6.382968902587891\r\nStep 332, loss: 6.010067462921143\r\nStep 333, loss: 6.022067070007324\r\nStep 334, loss: 5.914710998535156\r\nStep 335, loss: 6.026205539703369\r\nStep 336, loss: 6.101585865020752\r\nStep 337, loss: 5.8053202629089355\r\nStep 338, loss: 5.739813327789307\r\nStep 339, loss: 5.901891231536865\r\nStep 340, loss: 5.769777774810791\r\nStep 341, loss: 5.84769868850708\r\nStep 342, loss: 5.975956916809082\r\nStep 343, loss: 5.790035247802734\r\nStep 344, loss: 5.943611145019531\r\nStep 345, loss: 5.850645065307617\r\nStep 346, loss: 5.7808990478515625\r\nStep 347, loss: 5.690489292144775\r\nStep 348, loss: 5.986003398895264\r\nStep 349, loss: 5.836080551147461\r\nStep 350, loss: 5.887240886688232\r\nStep 351, loss: 5.904907703399658\r\nStep 352, loss: 5.790065288543701\r\nStep 353, loss: 5.890659809112549\r\nStep 354, loss: 5.948431015014648\r\nStep 355, loss: 5.6976847648620605\r\nStep 356, loss: 5.886941909790039\r\nStep 357, loss: 5.837642192840576\r\nStep 358, loss: 5.8052215576171875\r\nStep 359, loss: 5.958202838897705\r\nStep 360, loss: 6.015711784362793\r\nStep 331, loss: 6.382968902587891\r\nStep 332, loss: 6.010067462921143\r\nStep 333, loss: 6.022067070007324\r\nStep 334, loss: 5.914710998535156\r\nStep 335, loss: 6.026205539703369\r\nStep 336, loss: 6.101585865020752\r\nStep 337, loss: 5.8053202629089355\r\nStep 338, loss: 5.739813327789307\r\nStep 339, loss: 5.901891231536865\r\nStep 340, loss: 5.769777774810791\r\nStep 341, loss: 5.84769868850708\r\nStep 342, loss: 5.975956916809082\r\nStep 343, loss: 5.790035247802734\r\nStep 344, loss: 5.943611145019531\r\nStep 345, loss: 5.850645065307617\r\nStep 346, loss: 5.7808990478515625\r\nStep 347, loss: 5.690489292144775\r\nStep 348, loss: 5.986003398895264\r\nStep 349, loss: 5.836080551147461\r\nStep 350, loss: 5.887240886688232\r\nStep 351, loss: 5.904907703399658\r\nStep 352, loss: 5.790065288543701\r\nStep 353, loss: 5.890659809112549\r\nStep 354, loss: 5.948431015014648\r\nStep 355, loss: 5.6976847648620605\r\nStep 356, loss: 5.886941909790039\r\nStep 357, loss: 5.837642192840576\r\nStep 358, loss: 5.8052215576171875\r\nStep 359, loss: 5.958202838897705\r\nStep 360, loss: 6.015711784362793\r\nStep 361, loss: 5.924943923950195\r\nStep 362, loss: 5.905316352844238\r\nStep 363, loss: 5.8577880859375\r\nStep 364, loss: 5.833385944366455\r\nStep 365, loss: 6.09619140625\r\nStep 366, loss: 5.879598140716553\r\nStep 367, loss: 5.800159931182861\r\nStep 368, loss: 5.879344463348389\r\nStep 369, loss: 5.841940879821777\r\nStep 370, loss: 6.016635894775391\r\nStep 371, loss: 5.777203559875488\r\nStep 372, loss: 5.687002658843994\r\nStep 373, loss: 5.890249252319336\r\nStep 374, loss: 5.735091209411621\r\nStep 375, loss: 5.94771146774292\r\nStep 376, loss: 5.657384872436523\r\nStep 377, loss: 5.763212203979492\r\nStep 378, loss: 6.015496730804443\r\nStep 379, loss: 5.793003082275391\r\nStep 380, loss: 5.845806121826172\r\nStep 381, loss: 5.729050159454346\r\nStep 382, loss: 5.7567267417907715\r\nStep 383, loss: 5.785036087036133\r\nStep 384, loss: 5.758927345275879\r\nStep 385, loss: 5.695383071899414\r\nStep 386, loss: 5.753874778747559\r\nStep 387, loss: 5.656781196594238\r\nStep 388, loss: 5.64355993270874\r\nStep 389, loss: 5.938015460968018\r\nStep 390, loss: 5.836848735809326\r\nStep 361, loss: 5.924943923950195\r\nStep 362, loss: 5.905316352844238\r\nStep 363, loss: 5.8577880859375\r\nStep 364, loss: 5.833385944366455\r\nStep 365, loss: 6.09619140625\r\nStep 366, loss: 5.879598140716553\r\nStep 367, loss: 5.800159931182861\r\nStep 368, loss: 5.879344463348389\r\nStep 369, loss: 5.841940879821777\r\nStep 370, loss: 6.016635894775391\r\nStep 371, loss: 5.777203559875488\r\nStep 372, loss: 5.687002658843994\r\nStep 373, loss: 5.890249252319336\r\nStep 374, loss: 5.735091209411621\r\nStep 375, loss: 5.94771146774292\r\nStep 376, loss: 5.657384872436523\r\nStep 377, loss: 5.763212203979492\r\nStep 378, loss: 6.015496730804443\r\nStep 379, loss: 5.793003082275391\r\nStep 380, loss: 5.845806121826172\r\nStep 381, loss: 5.729050159454346\r\nStep 382, loss: 5.7567267417907715\r\nStep 383, loss: 5.785036087036133\r\nStep 384, loss: 5.758927345275879\r\nStep 385, loss: 5.695383071899414\r\nStep 386, loss: 5.753874778747559\r\nStep 387, loss: 5.656781196594238\r\nStep 388, loss: 5.64355993270874\r\nStep 389, loss: 5.938015460968018\r\nStep 390, loss: 5.836848735809326\r\nStep 391, loss: 5.777060031890869\r\nStep 392, loss: 5.80929708480835\r\nStep 393, loss: 5.858407974243164\r\nStep 394, loss: 5.820347785949707\r\nStep 395, loss: 5.752736568450928\r\nStep 396, loss: 5.977144241333008\r\nStep 397, loss: 5.999248504638672\r\nStep 398, loss: 5.756777763366699\r\nStep 399, loss: 5.889488220214844\r\nStep 400, loss: 5.789821624755859\r\nStep 401, loss: 5.788527488708496\r\nStep 402, loss: 5.627648830413818\r\nStep 403, loss: 5.792341709136963\r\nStep 404, loss: 5.86679220199585\r\nStep 405, loss: 5.916896820068359\r\nStep 406, loss: 5.835608005523682\r\nStep 407, loss: 6.11745023727417\r\nStep 408, loss: 5.773310661315918\r\nStep 409, loss: 5.883132457733154\r\nStep 410, loss: 5.6630048751831055\r\nStep 411, loss: 5.871366500854492\r\nStep 412, loss: 5.903974533081055\r\nStep 413, loss: 5.535669803619385\r\nStep 414, loss: 5.646414756774902\r\nStep 415, loss: 5.724021911621094\r\nStep 416, loss: 5.677767753601074\r\nStep 417, loss: 5.652090072631836\r\nStep 418, loss: 5.597585678100586\r\nStep 419, loss: 5.762701988220215\r\nStep 420, loss: 5.502464771270752\r\nStep 391, loss: 5.777060031890869\r\nStep 392, loss: 5.80929708480835\r\nStep 393, loss: 5.858407974243164\r\nStep 394, loss: 5.820347785949707\r\nStep 395, loss: 5.752736568450928\r\nStep 396, loss: 5.977144241333008\r\nStep 397, loss: 5.999248504638672\r\nStep 398, loss: 5.756777763366699\r\nStep 399, loss: 5.889488220214844\r\nStep 400, loss: 5.789821624755859\r\nStep 401, loss: 5.788527488708496\r\nStep 402, loss: 5.627648830413818\r\nStep 403, loss: 5.792341709136963\r\nStep 404, loss: 5.86679220199585\r\nStep 405, loss: 5.916896820068359\r\nStep 406, loss: 5.835608005523682\r\nStep 407, loss: 6.11745023727417\r\nStep 408, loss: 5.773310661315918\r\nStep 409, loss: 5.883132457733154\r\nStep 410, loss: 5.6630048751831055\r\nStep 411, loss: 5.871366500854492\r\nStep 412, loss: 5.903974533081055\r\nStep 413, loss: 5.535669803619385\r\nStep 414, loss: 5.646414756774902\r\nStep 415, loss: 5.724021911621094\r\nStep 416, loss: 5.677767753601074\r\nStep 417, loss: 5.652090072631836\r\nStep 418, loss: 5.597585678100586\r\nStep 419, loss: 5.762701988220215\r\nStep 420, loss: 5.502464771270752\r\nStep 421, loss: 5.753362655639648\r\nStep 422, loss: 5.559510707855225\r\nStep 423, loss: 5.701245307922363\r\nStep 424, loss: 5.822718620300293\r\nStep 425, loss: 5.51902437210083\r\nStep 426, loss: 5.723858833312988\r\nStep 427, loss: 5.752884387969971\r\nStep 428, loss: 5.7527642250061035\r\nStep 429, loss: 5.536493301391602\r\nStep 430, loss: 5.853747367858887\r\nStep 431, loss: 5.712923526763916\r\nStep 432, loss: 5.617330551147461\r\nStep 433, loss: 5.601428985595703\r\nStep 434, loss: 5.9634552001953125\r\nStep 435, loss: 5.687324523925781\r\nStep 436, loss: 5.664196968078613\r\nStep 437, loss: 5.776155471801758\r\nStep 438, loss: 5.555196285247803\r\nStep 439, loss: 5.63498592376709\r\nStep 440, loss: 5.8197760581970215\r\nStep 441, loss: 5.577805995941162\r\nStep 442, loss: 5.509472846984863\r\nStep 443, loss: 5.6083526611328125\r\nStep 444, loss: 5.517510890960693\r\nStep 445, loss: 5.543288707733154\r\nStep 446, loss: 5.672813415527344\r\nStep 447, loss: 5.723514080047607\r\nStep 448, loss: 5.612575531005859\r\nStep 449, loss: 5.857496738433838\r\nStep 450, loss: 5.798945903778076\r\nStep 421, loss: 5.753362655639648\r\nStep 422, loss: 5.559510707855225\r\nStep 423, loss: 5.701245307922363\r\nStep 424, loss: 5.822718620300293\r\nStep 425, loss: 5.51902437210083\r\nStep 426, loss: 5.723858833312988\r\nStep 427, loss: 5.752884387969971\r\nStep 428, loss: 5.7527642250061035\r\nStep 429, loss: 5.536493301391602\r\nStep 430, loss: 5.853747367858887\r\nStep 431, loss: 5.712923526763916\r\nStep 432, loss: 5.617330551147461\r\nStep 433, loss: 5.601428985595703\r\nStep 434, loss: 5.9634552001953125\r\nStep 435, loss: 5.687324523925781\r\nStep 436, loss: 5.664196968078613\r\nStep 437, loss: 5.776155471801758\r\nStep 438, loss: 5.555196285247803\r\nStep 439, loss: 5.63498592376709\r\nStep 440, loss: 5.8197760581970215\r\nStep 441, loss: 5.577805995941162\r\nStep 442, loss: 5.509472846984863\r\nStep 443, loss: 5.6083526611328125\r\nStep 444, loss: 5.517510890960693\r\nStep 445, loss: 5.543288707733154\r\nStep 446, loss: 5.672813415527344\r\nStep 447, loss: 5.723514080047607\r\nStep 448, loss: 5.612575531005859\r\nStep 449, loss: 5.857496738433838\r\nStep 450, loss: 5.798945903778076\r\nStep 451, loss: 5.441054821014404\r\nStep 452, loss: 5.596620082855225\r\nStep 453, loss: 5.544404983520508\r\nStep 454, loss: 5.754658222198486\r\nStep 455, loss: 5.781772613525391\r\nStep 456, loss: 5.656846523284912\r\nStep 457, loss: 5.571268081665039\r\nStep 458, loss: 5.702635765075684\r\nStep 459, loss: 5.473593235015869\r\nStep 460, loss: 5.5884690284729\r\nStep 461, loss: 5.590051651000977\r\nStep 462, loss: 5.594494342803955\r\nStep 463, loss: 5.555944919586182\r\nStep 464, loss: 5.735914707183838\r\nStep 465, loss: 5.679328441619873\r\nStep 466, loss: 5.410174369812012\r\nStep 467, loss: 5.712713718414307\r\nStep 468, loss: 5.571666717529297\r\nStep 469, loss: 5.591692924499512\r\nStep 470, loss: 5.54882287979126\r\nStep 471, loss: 5.527029991149902\r\nStep 472, loss: 5.44807767868042\r\nStep 473, loss: 5.462024211883545\r\nStep 474, loss: 5.600245952606201\r\nStep 475, loss: 5.820129871368408\r\nStep 476, loss: 5.609768390655518\r\nStep 477, loss: 5.626163959503174\r\nStep 478, loss: 5.776749610900879\r\nStep 479, loss: 5.380319595336914\r\nStep 480, loss: 5.529082298278809\r\nStep 451, loss: 5.441054821014404\r\nStep 452, loss: 5.596620082855225\r\nStep 453, loss: 5.544404983520508\r\nStep 454, loss: 5.754658222198486\r\nStep 455, loss: 5.781772613525391\r\nStep 456, loss: 5.656846523284912\r\nStep 457, loss: 5.571268081665039\r\nStep 458, loss: 5.702635765075684\r\nStep 459, loss: 5.473593235015869\r\nStep 460, loss: 5.5884690284729\r\nStep 461, loss: 5.590051651000977\r\nStep 462, loss: 5.594494342803955\r\nStep 463, loss: 5.555944919586182\r\nStep 464, loss: 5.735914707183838\r\nStep 465, loss: 5.679328441619873\r\nStep 466, loss: 5.410174369812012\r\nStep 467, loss: 5.712713718414307\r\nStep 468, loss: 5.571666717529297\r\nStep 469, loss: 5.591692924499512\r\nStep 470, loss: 5.54882287979126\r\nStep 471, loss: 5.527029991149902\r\nStep 472, loss: 5.44807767868042\r\nStep 473, loss: 5.462024211883545\r\nStep 474, loss: 5.600245952606201\r\nStep 475, loss: 5.820129871368408\r\nStep 476, loss: 5.609768390655518\r\nStep 477, loss: 5.626163959503174\r\nStep 478, loss: 5.776749610900879\r\nStep 479, loss: 5.380319595336914\r\nStep 480, loss: 5.529082298278809\r\nStep 481, loss: 5.674862861633301\r\nStep 241, loss: 6.375746250152588\r\nStep 242, loss: 6.384668350219727\r\nStep 243, loss: 6.2770094871521\r\nStep 244, loss: 6.596880912780762\r\nStep 245, loss: 6.688802242279053\r\nStep 246, loss: 6.219254493713379\r\nStep 247, loss: 6.591236591339111\r\nStep 248, loss: 6.286152362823486\r\nStep 249, loss: 6.393139839172363\r\nStep 250, loss: 6.294431686401367\r\nStep 251, loss: 6.2147321701049805\r\nStep 252, loss: 6.365173816680908\r\nStep 253, loss: 6.267394065856934\r\nStep 254, loss: 6.570861339569092\r\nStep 255, loss: 6.34361457824707\r\nStep 256, loss: 6.068459510803223\r\nStep 257, loss: 6.029990196228027\r\nStep 258, loss: 6.469965934753418\r\nStep 259, loss: 6.202692985534668\r\nStep 260, loss: 6.448121070861816\r\nStep 261, loss: 6.35492467880249\r\nStep 262, loss: 6.2771992683410645\r\nStep 263, loss: 6.350648403167725\r\nStep 264, loss: 6.209310054779053\r\nStep 265, loss: 6.30444860458374\r\nStep 266, loss: 6.190724849700928\r\nStep 267, loss: 6.2750067710876465\r\nStep 268, loss: 6.282001972198486\r\nStep 269, loss: 6.261999130249023\r\nStep 270, loss: 6.238827228546143\r\nStep 241, loss: 6.375746250152588\r\nStep 242, loss: 6.384668350219727\r\nStep 243, loss: 6.2770094871521\r\nStep 244, loss: 6.596880912780762\r\nStep 245, loss: 6.688802242279053\r\nStep 246, loss: 6.219254493713379\r\nStep 247, loss: 6.591236591339111\r\nStep 248, loss: 6.286152362823486\r\nStep 249, loss: 6.393139839172363\r\nStep 250, loss: 6.294431686401367\r\nStep 251, loss: 6.2147321701049805\r\nStep 252, loss: 6.365173816680908\r\nStep 253, loss: 6.267394065856934\r\nStep 254, loss: 6.570861339569092\r\nStep 255, loss: 6.34361457824707\r\nStep 256, loss: 6.068459510803223\r\nStep 257, loss: 6.029990196228027\r\nStep 258, loss: 6.469965934753418\r\nStep 259, loss: 6.202692985534668\r\nStep 260, loss: 6.448121070861816\r\nStep 261, loss: 6.35492467880249\r\nStep 262, loss: 6.2771992683410645\r\nStep 263, loss: 6.350648403167725\r\nStep 264, loss: 6.209310054779053\r\nStep 265, loss: 6.30444860458374\r\nStep 266, loss: 6.190724849700928\r\nStep 267, loss: 6.2750067710876465\r\nStep 268, loss: 6.282001972198486\r\nStep 269, loss: 6.261999130249023\r\nStep 270, loss: 6.238827228546143\r\nStep 271, loss: 6.399940013885498\r\nStep 272, loss: 6.577332496643066\r\nStep 273, loss: 6.290578365325928\r\nStep 274, loss: 6.126301288604736\r\nStep 275, loss: 6.4832916259765625\r\nStep 276, loss: 6.190239906311035\r\nStep 277, loss: 6.054000377655029\r\nStep 278, loss: 6.173544406890869\r\nStep 279, loss: 5.889945030212402\r\nStep 280, loss: 6.498538494110107\r\nStep 281, loss: 6.207729816436768\r\nStep 282, loss: 6.37006139755249\r\nStep 283, loss: 6.058416366577148\r\nStep 284, loss: 6.195253372192383\r\nStep 285, loss: 6.223158359527588\r\nStep 286, loss: 5.904426574707031\r\nStep 287, loss: 6.371645450592041\r\nStep 288, loss: 6.0909423828125\r\nStep 289, loss: 6.016209125518799\r\nStep 290, loss: 6.319799900054932\r\nStep 291, loss: 6.296480178833008\r\nStep 292, loss: 6.148037910461426\r\nStep 293, loss: 5.923506259918213\r\nStep 294, loss: 5.763485431671143\r\nStep 295, loss: 6.133178234100342\r\nStep 296, loss: 6.073211193084717\r\nStep 297, loss: 6.169839859008789\r\nStep 298, loss: 6.141224384307861\r\nStep 299, loss: 6.168487071990967\r\nStep 300, loss: 6.119058132171631\r\nStep 271, loss: 6.399940013885498\r\nStep 272, loss: 6.577332496643066\r\nStep 273, loss: 6.290578365325928\r\nStep 274, loss: 6.126301288604736\r\nStep 275, loss: 6.4832916259765625\r\nStep 276, loss: 6.190239906311035\r\nStep 277, loss: 6.054000377655029\r\nStep 278, loss: 6.173544406890869\r\nStep 279, loss: 5.889945030212402\r\nStep 280, loss: 6.498538494110107\r\nStep 281, loss: 6.207729816436768\r\nStep 282, loss: 6.37006139755249\r\nStep 283, loss: 6.058416366577148\r\nStep 284, loss: 6.195253372192383\r\nStep 285, loss: 6.223158359527588\r\nStep 286, loss: 5.904426574707031\r\nStep 287, loss: 6.371645450592041\r\nStep 288, loss: 6.0909423828125\r\nStep 289, loss: 6.016209125518799\r\nStep 290, loss: 6.319799900054932\r\nStep 291, loss: 6.296480178833008\r\nStep 292, loss: 6.148037910461426\r\nStep 293, loss: 5.923506259918213\r\nStep 294, loss: 5.763485431671143\r\nStep 295, loss: 6.133178234100342\r\nStep 296, loss: 6.073211193084717\r\nStep 297, loss: 6.169839859008789\r\nStep 298, loss: 6.141224384307861\r\nStep 299, loss: 6.168487071990967\r\nStep 300, loss: 6.119058132171631\r\nStep 301, loss: 6.03639554977417\r\nStep 302, loss: 6.216285228729248\r\nStep 303, loss: 6.022148609161377\r\nStep 304, loss: 5.897799968719482\r\nStep 305, loss: 5.894784927368164\r\nStep 306, loss: 6.132763385772705\r\nStep 307, loss: 5.888872146606445\r\nStep 308, loss: 6.012820243835449\r\nStep 309, loss: 6.126757621765137\r\nStep 310, loss: 6.202884197235107\r\nStep 311, loss: 6.018051624298096\r\nStep 312, loss: 6.108768939971924\r\nStep 313, loss: 5.92648458480835\r\nStep 314, loss: 6.071450710296631\r\nStep 315, loss: 5.71877384185791\r\nStep 316, loss: 5.99001407623291\r\nStep 317, loss: 5.833607196807861\r\nStep 318, loss: 6.032952785491943\r\nStep 319, loss: 6.038209438323975\r\nStep 320, loss: 6.10830545425415\r\nStep 321, loss: 5.989357948303223\r\nStep 322, loss: 6.006596088409424\r\nStep 323, loss: 6.012803077697754\r\nStep 324, loss: 6.0025787353515625\r\nStep 325, loss: 5.972760200500488\r\nStep 326, loss: 6.077414512634277\r\nStep 327, loss: 5.926943302154541\r\nStep 328, loss: 6.139301300048828\r\nStep 329, loss: 5.791858196258545\r\nStep 330, loss: 5.903587818145752\r\nStep 301, loss: 6.03639554977417\r\nStep 302, loss: 6.216285228729248\r\nStep 303, loss: 6.022148609161377\r\nStep 304, loss: 5.897799968719482\r\nStep 305, loss: 5.894784927368164\r\nStep 306, loss: 6.132763385772705\r\nStep 307, loss: 5.888872146606445\r\nStep 308, loss: 6.012820243835449\r\nStep 309, loss: 6.126757621765137\r\nStep 310, loss: 6.202884197235107\r\nStep 311, loss: 6.018051624298096\r\nStep 312, loss: 6.108768939971924\r\nStep 313, loss: 5.92648458480835\r\nStep 314, loss: 6.071450710296631\r\nStep 315, loss: 5.71877384185791\r\nStep 316, loss: 5.99001407623291\r\nStep 317, loss: 5.833607196807861\r\nStep 318, loss: 6.032952785491943\r\nStep 319, loss: 6.038209438323975\r\nStep 320, loss: 6.10830545425415\r\nStep 321, loss: 5.989357948303223\r\nStep 322, loss: 6.006596088409424\r\nStep 323, loss: 6.012803077697754\r\nStep 324, loss: 6.0025787353515625\r\nStep 325, loss: 5.972760200500488\r\nStep 326, loss: 6.077414512634277\r\nStep 327, loss: 5.926943302154541\r\nStep 328, loss: 6.139301300048828\r\nStep 329, loss: 5.791858196258545\r\nStep 330, loss: 5.903587818145752\r\nStep 331, loss: 6.382968902587891\r\nStep 332, loss: 6.010067462921143\r\nStep 333, loss: 6.022067070007324\r\nStep 334, loss: 5.914710998535156\r\nStep 335, loss: 6.026205539703369\r\nStep 336, loss: 6.101585865020752\r\nStep 337, loss: 5.8053202629089355\r\nStep 338, loss: 5.739813327789307\r\nStep 339, loss: 5.901891231536865\r\nStep 340, loss: 5.769777774810791\r\nStep 341, loss: 5.84769868850708\r\nStep 342, loss: 5.975956916809082\r\nStep 343, loss: 5.790035247802734\r\nStep 344, loss: 5.943611145019531\r\nStep 345, loss: 5.850645065307617\r\nStep 346, loss: 5.7808990478515625\r\nStep 347, loss: 5.690489292144775\r\nStep 348, loss: 5.986003398895264\r\nStep 349, loss: 5.836080551147461\r\nStep 350, loss: 5.887240886688232\r\nStep 351, loss: 5.904907703399658\r\nStep 352, loss: 5.790065288543701\r\nStep 353, loss: 5.890659809112549\r\nStep 354, loss: 5.948431015014648\r\nStep 355, loss: 5.6976847648620605\r\nStep 356, loss: 5.886941909790039\r\nStep 357, loss: 5.837642192840576\r\nStep 358, loss: 5.8052215576171875\r\nStep 359, loss: 5.958202838897705\r\nStep 360, loss: 6.015711784362793\r\nStep 331, loss: 6.382968902587891\r\nStep 332, loss: 6.010067462921143\r\nStep 333, loss: 6.022067070007324\r\nStep 334, loss: 5.914710998535156\r\nStep 335, loss: 6.026205539703369\r\nStep 336, loss: 6.101585865020752\r\nStep 337, loss: 5.8053202629089355\r\nStep 338, loss: 5.739813327789307\r\nStep 339, loss: 5.901891231536865\r\nStep 340, loss: 5.769777774810791\r\nStep 341, loss: 5.84769868850708\r\nStep 342, loss: 5.975956916809082\r\nStep 343, loss: 5.790035247802734\r\nStep 344, loss: 5.943611145019531\r\nStep 345, loss: 5.850645065307617\r\nStep 346, loss: 5.7808990478515625\r\nStep 347, loss: 5.690489292144775\r\nStep 348, loss: 5.986003398895264\r\nStep 349, loss: 5.836080551147461\r\nStep 350, loss: 5.887240886688232\r\nStep 351, loss: 5.904907703399658\r\nStep 352, loss: 5.790065288543701\r\nStep 353, loss: 5.890659809112549\r\nStep 354, loss: 5.948431015014648\r\nStep 355, loss: 5.6976847648620605\r\nStep 356, loss: 5.886941909790039\r\nStep 357, loss: 5.837642192840576\r\nStep 358, loss: 5.8052215576171875\r\nStep 359, loss: 5.958202838897705\r\nStep 360, loss: 6.015711784362793\r\nStep 361, loss: 5.924943923950195\r\nStep 362, loss: 5.905316352844238\r\nStep 363, loss: 5.8577880859375\r\nStep 364, loss: 5.833385944366455\r\nStep 365, loss: 6.09619140625\r\nStep 366, loss: 5.879598140716553\r\nStep 367, loss: 5.800159931182861\r\nStep 368, loss: 5.879344463348389\r\nStep 369, loss: 5.841940879821777\r\nStep 370, loss: 6.016635894775391\r\nStep 371, loss: 5.777203559875488\r\nStep 372, loss: 5.687002658843994\r\nStep 373, loss: 5.890249252319336\r\nStep 374, loss: 5.735091209411621\r\nStep 375, loss: 5.94771146774292\r\nStep 376, loss: 5.657384872436523\r\nStep 377, loss: 5.763212203979492\r\nStep 378, loss: 6.015496730804443\r\nStep 379, loss: 5.793003082275391\r\nStep 380, loss: 5.845806121826172\r\nStep 381, loss: 5.729050159454346\r\nStep 382, loss: 5.7567267417907715\r\nStep 383, loss: 5.785036087036133\r\nStep 384, loss: 5.758927345275879\r\nStep 385, loss: 5.695383071899414\r\nStep 386, loss: 5.753874778747559\r\nStep 387, loss: 5.656781196594238\r\nStep 388, loss: 5.64355993270874\r\nStep 389, loss: 5.938015460968018\r\nStep 390, loss: 5.836848735809326\r\nStep 241, loss: 6.375746250152588\r\nStep 242, loss: 6.384668350219727\r\nStep 243, loss: 6.2770094871521\r\nStep 244, loss: 6.596880912780762\r\nStep 245, loss: 6.688802242279053\r\nStep 246, loss: 6.219254493713379\r\nStep 247, loss: 6.591236591339111\r\nStep 248, loss: 6.286152362823486\r\nStep 249, loss: 6.393139839172363\r\nStep 250, loss: 6.294431686401367\r\nStep 251, loss: 6.2147321701049805\r\nStep 252, loss: 6.365173816680908\r\nStep 253, loss: 6.267394065856934\r\nStep 254, loss: 6.570861339569092\r\nStep 255, loss: 6.34361457824707\r\nStep 256, loss: 6.068459510803223\r\nStep 257, loss: 6.029990196228027\r\nStep 258, loss: 6.469965934753418\r\nStep 259, loss: 6.202692985534668\r\nStep 260, loss: 6.448121070861816\r\nStep 261, loss: 6.35492467880249\r\nStep 262, loss: 6.2771992683410645\r\nStep 263, loss: 6.350648403167725\r\nStep 264, loss: 6.209310054779053\r\nStep 265, loss: 6.30444860458374\r\nStep 266, loss: 6.190724849700928\r\nStep 267, loss: 6.2750067710876465\r\nStep 268, loss: 6.282001972198486\r\nStep 269, loss: 6.261999130249023\r\nStep 270, loss: 6.238827228546143\r\nStep 391, loss: 5.777060031890869\r\nStep 392, loss: 5.80929708480835\r\nStep 393, loss: 5.858407974243164\r\nStep 394, loss: 5.820347785949707\r\nStep 395, loss: 5.752736568450928\r\nStep 396, loss: 5.977144241333008\r\nStep 397, loss: 5.999248504638672\r\nStep 398, loss: 5.756777763366699\r\nStep 399, loss: 5.889488220214844\r\nStep 400, loss: 5.789821624755859\r\nStep 401, loss: 5.788527488708496\r\nStep 402, loss: 5.627648830413818\r\nStep 403, loss: 5.792341709136963\r\nStep 404, loss: 5.86679220199585\r\nStep 405, loss: 5.916896820068359\r\nStep 406, loss: 5.835608005523682\r\nStep 407, loss: 6.11745023727417\r\nStep 408, loss: 5.773310661315918\r\nStep 409, loss: 5.883132457733154\r\nStep 410, loss: 5.6630048751831055\r\nStep 411, loss: 5.871366500854492\r\nStep 412, loss: 5.903974533081055\r\nStep 413, loss: 5.535669803619385\r\nStep 414, loss: 5.646414756774902\r\nStep 415, loss: 5.724021911621094\r\nStep 416, loss: 5.677767753601074\r\nStep 417, loss: 5.652090072631836\r\nStep 418, loss: 5.597585678100586\r\nStep 419, loss: 5.762701988220215\r\nStep 420, loss: 5.502464771270752\r\nStep 271, loss: 6.399940013885498\r\nStep 272, loss: 6.577332496643066\r\nStep 273, loss: 6.290578365325928\r\nStep 274, loss: 6.126301288604736\r\nStep 275, loss: 6.4832916259765625\r\nStep 276, loss: 6.190239906311035\r\nStep 277, loss: 6.054000377655029\r\nStep 278, loss: 6.173544406890869\r\nStep 279, loss: 5.889945030212402\r\nStep 280, loss: 6.498538494110107\r\nStep 281, loss: 6.207729816436768\r\nStep 282, loss: 6.37006139755249\r\nStep 283, loss: 6.058416366577148\r\nStep 284, loss: 6.195253372192383\r\nStep 285, loss: 6.223158359527588\r\nStep 286, loss: 5.904426574707031\r\nStep 287, loss: 6.371645450592041\r\nStep 288, loss: 6.0909423828125\r\nStep 289, loss: 6.016209125518799\r\nStep 290, loss: 6.319799900054932\r\nStep 291, loss: 6.296480178833008\r\nStep 292, loss: 6.148037910461426\r\nStep 293, loss: 5.923506259918213\r\nStep 294, loss: 5.763485431671143\r\nStep 295, loss: 6.133178234100342\r\nStep 296, loss: 6.073211193084717\r\nStep 297, loss: 6.169839859008789\r\nStep 298, loss: 6.141224384307861\r\nStep 299, loss: 6.168487071990967\r\nStep 300, loss: 6.119058132171631\r\nStep 421, loss: 5.753362655639648\r\nStep 422, loss: 5.559510707855225\r\nStep 423, loss: 5.701245307922363\r\nStep 424, loss: 5.822718620300293\r\nStep 425, loss: 5.51902437210083\r\nStep 426, loss: 5.723858833312988\r\nStep 427, loss: 5.752884387969971\r\nStep 428, loss: 5.7527642250061035\r\nStep 429, loss: 5.536493301391602\r\nStep 430, loss: 5.853747367858887\r\nStep 431, loss: 5.712923526763916\r\nStep 432, loss: 5.617330551147461\r\nStep 433, loss: 5.601428985595703\r\nStep 434, loss: 5.9634552001953125\r\nStep 435, loss: 5.687324523925781\r\nStep 436, loss: 5.664196968078613\r\nStep 437, loss: 5.776155471801758\r\nStep 438, loss: 5.555196285247803\r\nStep 439, loss: 5.63498592376709\r\nStep 440, loss: 5.8197760581970215\r\nStep 441, loss: 5.577805995941162\r\nStep 442, loss: 5.509472846984863\r\nStep 443, loss: 5.6083526611328125\r\nStep 444, loss: 5.517510890960693\r\nStep 445, loss: 5.543288707733154\r\nStep 446, loss: 5.672813415527344\r\nStep 447, loss: 5.723514080047607\r\nStep 448, loss: 5.612575531005859\r\nStep 449, loss: 5.857496738433838\r\nStep 450, loss: 5.798945903778076\r\nStep 301, loss: 6.03639554977417\r\nStep 302, loss: 6.216285228729248\r\nStep 303, loss: 6.022148609161377\r\nStep 304, loss: 5.897799968719482\r\nStep 305, loss: 5.894784927368164\r\nStep 306, loss: 6.132763385772705\r\nStep 307, loss: 5.888872146606445\r\nStep 308, loss: 6.012820243835449\r\nStep 309, loss: 6.126757621765137\r\nStep 310, loss: 6.202884197235107\r\nStep 311, loss: 6.018051624298096\r\nStep 312, loss: 6.108768939971924\r\nStep 313, loss: 5.92648458480835\r\nStep 314, loss: 6.071450710296631\r\nStep 315, loss: 5.71877384185791\r\nStep 316, loss: 5.99001407623291\r\nStep 317, loss: 5.833607196807861\r\nStep 318, loss: 6.032952785491943\r\nStep 319, loss: 6.038209438323975\r\nStep 320, loss: 6.10830545425415\r\nStep 321, loss: 5.989357948303223\r\nStep 322, loss: 6.006596088409424\r\nStep 323, loss: 6.012803077697754\r\nStep 324, loss: 6.0025787353515625\r\nStep 325, loss: 5.972760200500488\r\nStep 326, loss: 6.077414512634277\r\nStep 327, loss: 5.926943302154541\r\nStep 328, loss: 6.139301300048828\r\nStep 329, loss: 5.791858196258545\r\nStep 330, loss: 5.903587818145752\r\nStep 451, loss: 5.441054821014404\r\nStep 452, loss: 5.596620082855225\r\nStep 453, loss: 5.544404983520508\r\nStep 454, loss: 5.754658222198486\r\nStep 455, loss: 5.781772613525391\r\nStep 456, loss: 5.656846523284912\r\nStep 457, loss: 5.571268081665039\r\nStep 458, loss: 5.702635765075684\r\nStep 459, loss: 5.473593235015869\r\nStep 460, loss: 5.5884690284729\r\nStep 461, loss: 5.590051651000977\r\nStep 462, loss: 5.594494342803955\r\nStep 463, loss: 5.555944919586182\r\nStep 464, loss: 5.735914707183838\r\nStep 465, loss: 5.679328441619873\r\nStep 466, loss: 5.410174369812012\r\nStep 467, loss: 5.712713718414307\r\nStep 468, loss: 5.571666717529297\r\nStep 469, loss: 5.591692924499512\r\nStep 470, loss: 5.54882287979126\r\nStep 471, loss: 5.527029991149902\r\nStep 472, loss: 5.44807767868042\r\nStep 473, loss: 5.462024211883545\r\nStep 474, loss: 5.600245952606201\r\nStep 475, loss: 5.820129871368408\r\nStep 476, loss: 5.609768390655518\r\nStep 477, loss: 5.626163959503174\r\nStep 478, loss: 5.776749610900879\r\nStep 479, loss: 5.380319595336914\r\nStep 480, loss: 5.529082298278809\r\nStep 331, loss: 6.382968902587891\r\nStep 332, loss: 6.010067462921143\r\nStep 333, loss: 6.022067070007324\r\nStep 334, loss: 5.914710998535156\r\nStep 335, loss: 6.026205539703369\r\nStep 336, loss: 6.101585865020752\r\nStep 337, loss: 5.8053202629089355\r\nStep 338, loss: 5.739813327789307\r\nStep 339, loss: 5.901891231536865\r\nStep 340, loss: 5.769777774810791\r\nStep 341, loss: 5.84769868850708\r\nStep 342, loss: 5.975956916809082\r\nStep 343, loss: 5.790035247802734\r\nStep 344, loss: 5.943611145019531\r\nStep 345, loss: 5.850645065307617\r\nStep 346, loss: 5.7808990478515625\r\nStep 347, loss: 5.690489292144775\r\nStep 348, loss: 5.986003398895264\r\nStep 349, loss: 5.836080551147461\r\nStep 350, loss: 5.887240886688232\r\nStep 351, loss: 5.904907703399658\r\nStep 352, loss: 5.790065288543701\r\nStep 353, loss: 5.890659809112549\r\nStep 354, loss: 5.948431015014648\r\nStep 355, loss: 5.6976847648620605\r\nStep 356, loss: 5.886941909790039\r\nStep 357, loss: 5.837642192840576\r\nStep 358, loss: 5.8052215576171875\r\nStep 359, loss: 5.958202838897705\r\nStep 360, loss: 6.015711784362793\r\nStep 481, loss: 5.674862861633301\r\nStep 361, loss: 5.924943923950195\r\nStep 362, loss: 5.905316352844238\r\nStep 363, loss: 5.8577880859375\r\nStep 364, loss: 5.833385944366455\r\nStep 365, loss: 6.09619140625\r\nStep 366, loss: 5.879598140716553\r\nStep 367, loss: 5.800159931182861\r\nStep 368, loss: 5.879344463348389\r\nStep 369, loss: 5.841940879821777\r\nStep 370, loss: 6.016635894775391\r\nStep 371, loss: 5.777203559875488\r\nStep 372, loss: 5.687002658843994\r\nStep 373, loss: 5.890249252319336\r\nStep 374, loss: 5.735091209411621\r\nStep 375, loss: 5.94771146774292\r\nStep 376, loss: 5.657384872436523\r\nStep 377, loss: 5.763212203979492\r\nStep 378, loss: 6.015496730804443\r\nStep 379, loss: 5.793003082275391\r\nStep 380, loss: 5.845806121826172\r\nStep 381, loss: 5.729050159454346\r\nStep 382, loss: 5.7567267417907715\r\nStep 383, loss: 5.785036087036133\r\nStep 384, loss: 5.758927345275879\r\nStep 385, loss: 5.695383071899414\r\nStep 386, loss: 5.753874778747559\r\nStep 387, loss: 5.656781196594238\r\nStep 388, loss: 5.64355993270874\r\nStep 389, loss: 5.938015460968018\r\nStep 390, loss: 5.836848735809326\r\nStep 361, loss: 5.924943923950195\r\nStep 362, loss: 5.905316352844238\r\nStep 363, loss: 5.8577880859375\r\nStep 364, loss: 5.833385944366455\r\nStep 365, loss: 6.09619140625\r\nStep 366, loss: 5.879598140716553\r\nStep 367, loss: 5.800159931182861\r\nStep 368, loss: 5.879344463348389\r\nStep 369, loss: 5.841940879821777\r\nStep 370, loss: 6.016635894775391\r\nStep 371, loss: 5.777203559875488\r\nStep 372, loss: 5.687002658843994\r\nStep 373, loss: 5.890249252319336\r\nStep 374, loss: 5.735091209411621\r\nStep 375, loss: 5.94771146774292\r\nStep 376, loss: 5.657384872436523\r\nStep 377, loss: 5.763212203979492\r\nStep 378, loss: 6.015496730804443\r\nStep 379, loss: 5.793003082275391\r\nStep 380, loss: 5.845806121826172\r\nStep 381, loss: 5.729050159454346\r\nStep 382, loss: 5.7567267417907715\r\nStep 383, loss: 5.785036087036133\r\nStep 384, loss: 5.758927345275879\r\nStep 385, loss: 5.695383071899414\r\nStep 386, loss: 5.753874778747559\r\nStep 387, loss: 5.656781196594238\r\nStep 388, loss: 5.64355993270874\r\nStep 389, loss: 5.938015460968018\r\nStep 390, loss: 5.836848735809326\r\nStep 391, loss: 5.777060031890869\r\nStep 392, loss: 5.80929708480835\r\nStep 393, loss: 5.858407974243164\r\nStep 394, loss: 5.820347785949707\r\nStep 395, loss: 5.752736568450928\r\nStep 396, loss: 5.977144241333008\r\nStep 397, loss: 5.999248504638672\r\nStep 398, loss: 5.756777763366699\r\nStep 399, loss: 5.889488220214844\r\nStep 400, loss: 5.789821624755859\r\nStep 401, loss: 5.788527488708496\r\nStep 402, loss: 5.627648830413818\r\nStep 403, loss: 5.792341709136963\r\nStep 404, loss: 5.86679220199585\r\nStep 405, loss: 5.916896820068359\r\nStep 406, loss: 5.835608005523682\r\nStep 407, loss: 6.11745023727417\r\nStep 408, loss: 5.773310661315918\r\nStep 409, loss: 5.883132457733154\r\nStep 410, loss: 5.6630048751831055\r\nStep 411, loss: 5.871366500854492\r\nStep 412, loss: 5.903974533081055\r\nStep 413, loss: 5.535669803619385\r\nStep 414, loss: 5.646414756774902\r\nStep 415, loss: 5.724021911621094\r\nStep 416, loss: 5.677767753601074\r\nStep 417, loss: 5.652090072631836\r\nStep 418, loss: 5.597585678100586\r\nStep 419, loss: 5.762701988220215\r\nStep 420, loss: 5.502464771270752\r\nStep 391, loss: 5.777060031890869\r\nStep 392, loss: 5.80929708480835\r\nStep 393, loss: 5.858407974243164\r\nStep 394, loss: 5.820347785949707\r\nStep 395, loss: 5.752736568450928\r\nStep 396, loss: 5.977144241333008\r\nStep 397, loss: 5.999248504638672\r\nStep 398, loss: 5.756777763366699\r\nStep 399, loss: 5.889488220214844\r\nStep 400, loss: 5.789821624755859\r\nStep 401, loss: 5.788527488708496\r\nStep 402, loss: 5.627648830413818\r\nStep 403, loss: 5.792341709136963\r\nStep 404, loss: 5.86679220199585\r\nStep 405, loss: 5.916896820068359\r\nStep 406, loss: 5.835608005523682\r\nStep 407, loss: 6.11745023727417\r\nStep 408, loss: 5.773310661315918\r\nStep 409, loss: 5.883132457733154\r\nStep 410, loss: 5.6630048751831055\r\nStep 411, loss: 5.871366500854492\r\nStep 412, loss: 5.903974533081055\r\nStep 413, loss: 5.535669803619385\r\nStep 414, loss: 5.646414756774902\r\nStep 415, loss: 5.724021911621094\r\nStep 416, loss: 5.677767753601074\r\nStep 417, loss: 5.652090072631836\r\nStep 418, loss: 5.597585678100586\r\nStep 419, loss: 5.762701988220215\r\nStep 420, loss: 5.502464771270752\r\nStep 421, loss: 5.753362655639648\r\nStep 422, loss: 5.559510707855225\r\nStep 423, loss: 5.701245307922363\r\nStep 424, loss: 5.822718620300293\r\nStep 425, loss: 5.51902437210083\r\nStep 426, loss: 5.723858833312988\r\nStep 427, loss: 5.752884387969971\r\nStep 428, loss: 5.7527642250061035\r\nStep 429, loss: 5.536493301391602\r\nStep 430, loss: 5.853747367858887\r\nStep 431, loss: 5.712923526763916\r\nStep 432, loss: 5.617330551147461\r\nStep 433, loss: 5.601428985595703\r\nStep 434, loss: 5.9634552001953125\r\nStep 435, loss: 5.687324523925781\r\nStep 436, loss: 5.664196968078613\r\nStep 437, loss: 5.776155471801758\r\nStep 438, loss: 5.555196285247803\r\nStep 439, loss: 5.63498592376709\r\nStep 440, loss: 5.8197760581970215\r\nStep 441, loss: 5.577805995941162\r\nStep 442, loss: 5.509472846984863\r\nStep 443, loss: 5.6083526611328125\r\nStep 444, loss: 5.517510890960693\r\nStep 445, loss: 5.543288707733154\r\nStep 446, loss: 5.672813415527344\r\nStep 447, loss: 5.723514080047607\r\nStep 448, loss: 5.612575531005859\r\nStep 449, loss: 5.857496738433838\r\nStep 450, loss: 5.798945903778076\r\nStep 421, loss: 5.753362655639648\r\nStep 422, loss: 5.559510707855225\r\nStep 423, loss: 5.701245307922363\r\nStep 424, loss: 5.822718620300293\r\nStep 425, loss: 5.51902437210083\r\nStep 426, loss: 5.723858833312988\r\nStep 427, loss: 5.752884387969971\r\nStep 428, loss: 5.7527642250061035\r\nStep 429, loss: 5.536493301391602\r\nStep 430, loss: 5.853747367858887\r\nStep 431, loss: 5.712923526763916\r\nStep 432, loss: 5.617330551147461\r\nStep 433, loss: 5.601428985595703\r\nStep 434, loss: 5.9634552001953125\r\nStep 435, loss: 5.687324523925781\r\nStep 436, loss: 5.664196968078613\r\nStep 437, loss: 5.776155471801758\r\nStep 438, loss: 5.555196285247803\r\nStep 439, loss: 5.63498592376709\r\nStep 440, loss: 5.8197760581970215\r\nStep 441, loss: 5.577805995941162\r\nStep 442, loss: 5.509472846984863\r\nStep 443, loss: 5.6083526611328125\r\nStep 444, loss: 5.517510890960693\r\nStep 445, loss: 5.543288707733154\r\nStep 446, loss: 5.672813415527344\r\nStep 447, loss: 5.723514080047607\r\nStep 448, loss: 5.612575531005859\r\nStep 449, loss: 5.857496738433838\r\nStep 450, loss: 5.798945903778076\r\nStep 451, loss: 5.441054821014404\r\nStep 452, loss: 5.596620082855225\r\nStep 453, loss: 5.544404983520508\r\nStep 454, loss: 5.754658222198486\r\nStep 455, loss: 5.781772613525391\r\nStep 456, loss: 5.656846523284912\r\nStep 457, loss: 5.571268081665039\r\nStep 458, loss: 5.702635765075684\r\nStep 459, loss: 5.473593235015869\r\nStep 460, loss: 5.5884690284729\r\nStep 461, loss: 5.590051651000977\r\nStep 462, loss: 5.594494342803955\r\nStep 463, loss: 5.555944919586182\r\nStep 464, loss: 5.735914707183838\r\nStep 465, loss: 5.679328441619873\r\nStep 466, loss: 5.410174369812012\r\nStep 467, loss: 5.712713718414307\r\nStep 468, loss: 5.571666717529297\r\nStep 469, loss: 5.591692924499512\r\nStep 470, loss: 5.54882287979126\r\nStep 471, loss: 5.527029991149902\r\nStep 472, loss: 5.44807767868042\r\nStep 473, loss: 5.462024211883545\r\nStep 474, loss: 5.600245952606201\r\nStep 475, loss: 5.820129871368408\r\nStep 476, loss: 5.609768390655518\r\nStep 477, loss: 5.626163959503174\r\nStep 478, loss: 5.776749610900879\r\nStep 479, loss: 5.380319595336914\r\nStep 480, loss: 5.529082298278809\r\nStep 451, loss: 5.441054821014404\r\nStep 452, loss: 5.596620082855225\r\nStep 453, loss: 5.544404983520508\r\nStep 454, loss: 5.754658222198486\r\nStep 455, loss: 5.781772613525391\r\nStep 456, loss: 5.656846523284912\r\nStep 457, loss: 5.571268081665039\r\nStep 458, loss: 5.702635765075684\r\nStep 459, loss: 5.473593235015869\r\nStep 460, loss: 5.5884690284729\r\nStep 461, loss: 5.590051651000977\r\nStep 462, loss: 5.594494342803955\r\nStep 463, loss: 5.555944919586182\r\nStep 464, loss: 5.735914707183838\r\nStep 465, loss: 5.679328441619873\r\nStep 466, loss: 5.410174369812012\r\nStep 467, loss: 5.712713718414307\r\nStep 468, loss: 5.571666717529297\r\nStep 469, loss: 5.591692924499512\r\nStep 470, loss: 5.54882287979126\r\nStep 471, loss: 5.527029991149902\r\nStep 472, loss: 5.44807767868042\r\nStep 473, loss: 5.462024211883545\r\nStep 474, loss: 5.600245952606201\r\nStep 475, loss: 5.820129871368408\r\nStep 476, loss: 5.609768390655518\r\nStep 477, loss: 5.626163959503174\r\nStep 478, loss: 5.776749610900879\r\nStep 479, loss: 5.380319595336914\r\nStep 480, loss: 5.529082298278809\r\nStep 361, loss: 5.924943923950195\r\nStep 362, loss: 5.905316352844238\r\nStep 363, loss: 5.8577880859375\r\nStep 364, loss: 5.833385944366455\r\nStep 365, loss: 6.09619140625\r\nStep 366, loss: 5.879598140716553\r\nStep 367, loss: 5.800159931182861\r\nStep 368, loss: 5.879344463348389\r\nStep 369, loss: 5.841940879821777\r\nStep 370, loss: 6.016635894775391\r\nStep 371, loss: 5.777203559875488\r\nStep 372, loss: 5.687002658843994\r\nStep 373, loss: 5.890249252319336\r\nStep 374, loss: 5.735091209411621\r\nStep 375, loss: 5.94771146774292\r\nStep 376, loss: 5.657384872436523\r\nStep 377, loss: 5.763212203979492\r\nStep 378, loss: 6.015496730804443\r\nStep 379, loss: 5.793003082275391\r\nStep 380, loss: 5.845806121826172\r\nStep 381, loss: 5.729050159454346\r\nStep 382, loss: 5.7567267417907715\r\nStep 383, loss: 5.785036087036133\r\nStep 384, loss: 5.758927345275879\r\nStep 385, loss: 5.695383071899414\r\nStep 386, loss: 5.753874778747559\r\nStep 387, loss: 5.656781196594238\r\nStep 388, loss: 5.64355993270874\r\nStep 389, loss: 5.938015460968018\r\nStep 390, loss: 5.836848735809326\r\nStep 481, loss: 5.674862861633301\r\nStep 391, loss: 5.777060031890869\r\nStep 392, loss: 5.80929708480835\r\nStep 393, loss: 5.858407974243164\r\nStep 394, loss: 5.820347785949707\r\nStep 395, loss: 5.752736568450928\r\nStep 396, loss: 5.977144241333008\r\nStep 397, loss: 5.999248504638672\r\nStep 398, loss: 5.756777763366699\r\nStep 399, loss: 5.889488220214844\r\nStep 400, loss: 5.789821624755859\r\nStep 401, loss: 5.788527488708496\r\nStep 402, loss: 5.627648830413818\r\nStep 403, loss: 5.792341709136963\r\nStep 404, loss: 5.86679220199585\r\nStep 405, loss: 5.916896820068359\r\nStep 406, loss: 5.835608005523682\r\nStep 407, loss: 6.11745023727417\r\nStep 408, loss: 5.773310661315918\r\nStep 409, loss: 5.883132457733154\r\nStep 410, loss: 5.6630048751831055\r\nStep 411, loss: 5.871366500854492\r\nStep 412, loss: 5.903974533081055\r\nStep 413, loss: 5.535669803619385\r\nStep 414, loss: 5.646414756774902\r\nStep 415, loss: 5.724021911621094\r\nStep 416, loss: 5.677767753601074\r\nStep 417, loss: 5.652090072631836\r\nStep 418, loss: 5.597585678100586\r\nStep 419, loss: 5.762701988220215\r\nStep 420, loss: 5.502464771270752\r\nStep 481, loss: 5.674862861633301\r\nStep 421, loss: 5.753362655639648\r\nStep 422, loss: 5.559510707855225\r\nStep 423, loss: 5.701245307922363\r\nStep 424, loss: 5.822718620300293\r\nStep 425, loss: 5.51902437210083\r\nStep 426, loss: 5.723858833312988\r\nStep 427, loss: 5.752884387969971\r\nStep 428, loss: 5.7527642250061035\r\nStep 429, loss: 5.536493301391602\r\nStep 430, loss: 5.853747367858887\r\nStep 431, loss: 5.712923526763916\r\nStep 432, loss: 5.617330551147461\r\nStep 433, loss: 5.601428985595703\r\nStep 434, loss: 5.9634552001953125\r\nStep 435, loss: 5.687324523925781\r\nStep 436, loss: 5.664196968078613\r\nStep 437, loss: 5.776155471801758\r\nStep 438, loss: 5.555196285247803\r\nStep 439, loss: 5.63498592376709\r\nStep 440, loss: 5.8197760581970215\r\nStep 441, loss: 5.577805995941162\r\nStep 442, loss: 5.509472846984863\r\nStep 443, loss: 5.6083526611328125\r\nStep 444, loss: 5.517510890960693\r\nStep 445, loss: 5.543288707733154\r\nStep 446, loss: 5.672813415527344\r\nStep 447, loss: 5.723514080047607\r\nStep 448, loss: 5.612575531005859\r\nStep 449, loss: 5.857496738433838\r\nStep 450, loss: 5.798945903778076\r\nStep 241, loss: 6.375746250152588\r\nStep 242, loss: 6.384668350219727\r\nStep 243, loss: 6.2770094871521\r\nStep 244, loss: 6.596880912780762\r\nStep 245, loss: 6.688802242279053\r\nStep 246, loss: 6.219254493713379\r\nStep 247, loss: 6.591236591339111\r\nStep 248, loss: 6.286152362823486\r\nStep 249, loss: 6.393139839172363\r\nStep 250, loss: 6.294431686401367\r\nStep 251, loss: 6.2147321701049805\r\nStep 252, loss: 6.365173816680908\r\nStep 253, loss: 6.267394065856934\r\nStep 254, loss: 6.570861339569092\r\nStep 255, loss: 6.34361457824707\r\nStep 256, loss: 6.068459510803223\r\nStep 257, loss: 6.029990196228027\r\nStep 258, loss: 6.469965934753418\r\nStep 259, loss: 6.202692985534668\r\nStep 260, loss: 6.448121070861816\r\nStep 261, loss: 6.35492467880249\r\nStep 262, loss: 6.2771992683410645\r\nStep 263, loss: 6.350648403167725\r\nStep 264, loss: 6.209310054779053\r\nStep 265, loss: 6.30444860458374\r\nStep 266, loss: 6.190724849700928\r\nStep 267, loss: 6.2750067710876465\r\nStep 268, loss: 6.282001972198486\r\nStep 269, loss: 6.261999130249023\r\nStep 270, loss: 6.238827228546143\r\nStep 451, loss: 5.441054821014404\r\nStep 452, loss: 5.596620082855225\r\nStep 453, loss: 5.544404983520508\r\nStep 454, loss: 5.754658222198486\r\nStep 455, loss: 5.781772613525391\r\nStep 456, loss: 5.656846523284912\r\nStep 457, loss: 5.571268081665039\r\nStep 458, loss: 5.702635765075684\r\nStep 459, loss: 5.473593235015869\r\nStep 460, loss: 5.5884690284729\r\nStep 461, loss: 5.590051651000977\r\nStep 462, loss: 5.594494342803955\r\nStep 463, loss: 5.555944919586182\r\nStep 464, loss: 5.735914707183838\r\nStep 465, loss: 5.679328441619873\r\nStep 466, loss: 5.410174369812012\r\nStep 467, loss: 5.712713718414307\r\nStep 468, loss: 5.571666717529297\r\nStep 469, loss: 5.591692924499512\r\nStep 470, loss: 5.54882287979126\r\nStep 471, loss: 5.527029991149902\r\nStep 472, loss: 5.44807767868042\r\nStep 473, loss: 5.462024211883545\r\nStep 474, loss: 5.600245952606201\r\nStep 475, loss: 5.820129871368408\r\nStep 476, loss: 5.609768390655518\r\nStep 477, loss: 5.626163959503174\r\nStep 478, loss: 5.776749610900879\r\nStep 479, loss: 5.380319595336914\r\nStep 480, loss: 5.529082298278809\r\nStep 271, loss: 6.399940013885498\r\nStep 272, loss: 6.577332496643066\r\nStep 273, loss: 6.290578365325928\r\nStep 274, loss: 6.126301288604736\r\nStep 275, loss: 6.4832916259765625\r\nStep 276, loss: 6.190239906311035\r\nStep 277, loss: 6.054000377655029\r\nStep 278, loss: 6.173544406890869\r\nStep 279, loss: 5.889945030212402\r\nStep 280, loss: 6.498538494110107\r\nStep 281, loss: 6.207729816436768\r\nStep 282, loss: 6.37006139755249\r\nStep 283, loss: 6.058416366577148\r\nStep 284, loss: 6.195253372192383\r\nStep 285, loss: 6.223158359527588\r\nStep 286, loss: 5.904426574707031\r\nStep 287, loss: 6.371645450592041\r\nStep 288, loss: 6.0909423828125\r\nStep 289, loss: 6.016209125518799\r\nStep 290, loss: 6.319799900054932\r\nStep 291, loss: 6.296480178833008\r\nStep 292, loss: 6.148037910461426\r\nStep 293, loss: 5.923506259918213\r\nStep 294, loss: 5.763485431671143\r\nStep 295, loss: 6.133178234100342\r\nStep 296, loss: 6.073211193084717\r\nStep 297, loss: 6.169839859008789\r\nStep 298, loss: 6.141224384307861\r\nStep 299, loss: 6.168487071990967\r\nStep 300, loss: 6.119058132171631\r\nStep 481, loss: 5.674862861633301\r\nStep 301, loss: 6.03639554977417\r\nStep 302, loss: 6.216285228729248\r\nStep 303, loss: 6.022148609161377\r\nStep 304, loss: 5.897799968719482\r\nStep 305, loss: 5.894784927368164\r\nStep 306, loss: 6.132763385772705\r\nStep 307, loss: 5.888872146606445\r\nStep 308, loss: 6.012820243835449\r\nStep 309, loss: 6.126757621765137\r\nStep 310, loss: 6.202884197235107\r\nStep 311, loss: 6.018051624298096\r\nStep 312, loss: 6.108768939971924\r\nStep 313, loss: 5.92648458480835\r\nStep 314, loss: 6.071450710296631\r\nStep 315, loss: 5.71877384185791\r\nStep 316, loss: 5.99001407623291\r\nStep 317, loss: 5.833607196807861\r\nStep 318, loss: 6.032952785491943\r\nStep 319, loss: 6.038209438323975\r\nStep 320, loss: 6.10830545425415\r\nStep 321, loss: 5.989357948303223\r\nStep 322, loss: 6.006596088409424\r\nStep 323, loss: 6.012803077697754\r\nStep 324, loss: 6.0025787353515625\r\nStep 325, loss: 5.972760200500488\r\nStep 326, loss: 6.077414512634277\r\nStep 327, loss: 5.926943302154541\r\nStep 328, loss: 6.139301300048828\r\nStep 329, loss: 5.791858196258545\r\nStep 330, loss: 5.903587818145752\r\nStep 481, loss: 5.674862861633301\r\nStep 331, loss: 6.382968902587891\r\nStep 332, loss: 6.010067462921143\r\nStep 333, loss: 6.022067070007324\r\nStep 334, loss: 5.914710998535156\r\nStep 335, loss: 6.026205539703369\r\nStep 336, loss: 6.101585865020752\r\nStep 337, loss: 5.8053202629089355\r\nStep 338, loss: 5.739813327789307\r\nStep 339, loss: 5.901891231536865\r\nStep 340, loss: 5.769777774810791\r\nStep 341, loss: 5.84769868850708\r\nStep 342, loss: 5.975956916809082\r\nStep 343, loss: 5.790035247802734\r\nStep 344, loss: 5.943611145019531\r\nStep 345, loss: 5.850645065307617\r\nStep 346, loss: 5.7808990478515625\r\nStep 347, loss: 5.690489292144775\r\nStep 348, loss: 5.986003398895264\r\nStep 349, loss: 5.836080551147461\r\nStep 350, loss: 5.887240886688232\r\nStep 351, loss: 5.904907703399658\r\nStep 352, loss: 5.790065288543701\r\nStep 353, loss: 5.890659809112549\r\nStep 354, loss: 5.948431015014648\r\nStep 355, loss: 5.6976847648620605\r\nStep 356, loss: 5.886941909790039\r\nStep 357, loss: 5.837642192840576\r\nStep 358, loss: 5.8052215576171875\r\nStep 359, loss: 5.958202838897705\r\nStep 360, loss: 6.015711784362793\r\nStep 241, loss: 6.375746250152588\r\nStep 242, loss: 6.384668350219727\r\nStep 243, loss: 6.2770094871521\r\nStep 244, loss: 6.596880912780762\r\nStep 245, loss: 6.688802242279053\r\nStep 246, loss: 6.219254493713379\r\nStep 247, loss: 6.591236591339111\r\nStep 248, loss: 6.286152362823486\r\nStep 249, loss: 6.393139839172363\r\nStep 250, loss: 6.294431686401367\r\nStep 251, loss: 6.2147321701049805\r\nStep 252, loss: 6.365173816680908\r\nStep 253, loss: 6.267394065856934\r\nStep 254, loss: 6.570861339569092\r\nStep 255, loss: 6.34361457824707\r\nStep 256, loss: 6.068459510803223\r\nStep 257, loss: 6.029990196228027\r\nStep 258, loss: 6.469965934753418\r\nStep 259, loss: 6.202692985534668\r\nStep 260, loss: 6.448121070861816\r\nStep 261, loss: 6.35492467880249\r\nStep 262, loss: 6.2771992683410645\r\nStep 263, loss: 6.350648403167725\r\nStep 264, loss: 6.209310054779053\r\nStep 265, loss: 6.30444860458374\r\nStep 266, loss: 6.190724849700928\r\nStep 267, loss: 6.2750067710876465\r\nStep 268, loss: 6.282001972198486\r\nStep 269, loss: 6.261999130249023\r\nStep 270, loss: 6.238827228546143\r\nStep 361, loss: 5.924943923950195\r\nStep 362, loss: 5.905316352844238\r\nStep 363, loss: 5.8577880859375\r\nStep 364, loss: 5.833385944366455\r\nStep 365, loss: 6.09619140625\r\nStep 366, loss: 5.879598140716553\r\nStep 367, loss: 5.800159931182861\r\nStep 368, loss: 5.879344463348389\r\nStep 369, loss: 5.841940879821777\r\nStep 370, loss: 6.016635894775391\r\nStep 371, loss: 5.777203559875488\r\nStep 372, loss: 5.687002658843994\r\nStep 373, loss: 5.890249252319336\r\nStep 374, loss: 5.735091209411621\r\nStep 375, loss: 5.94771146774292\r\nStep 376, loss: 5.657384872436523\r\nStep 377, loss: 5.763212203979492\r\nStep 378, loss: 6.015496730804443\r\nStep 379, loss: 5.793003082275391\r\nStep 380, loss: 5.845806121826172\r\nStep 381, loss: 5.729050159454346\r\nStep 382, loss: 5.7567267417907715\r\nStep 383, loss: 5.785036087036133\r\nStep 384, loss: 5.758927345275879\r\nStep 385, loss: 5.695383071899414\r\nStep 386, loss: 5.753874778747559\r\nStep 387, loss: 5.656781196594238\r\nStep 388, loss: 5.64355993270874\r\nStep 389, loss: 5.938015460968018\r\nStep 390, loss: 5.836848735809326\r\nStep 271, loss: 6.399940013885498\r\nStep 272, loss: 6.577332496643066\r\nStep 273, loss: 6.290578365325928\r\nStep 274, loss: 6.126301288604736\r\nStep 275, loss: 6.4832916259765625\r\nStep 276, loss: 6.190239906311035\r\nStep 277, loss: 6.054000377655029\r\nStep 278, loss: 6.173544406890869\r\nStep 279, loss: 5.889945030212402\r\nStep 280, loss: 6.498538494110107\r\nStep 281, loss: 6.207729816436768\r\nStep 282, loss: 6.37006139755249\r\nStep 283, loss: 6.058416366577148\r\nStep 284, loss: 6.195253372192383\r\nStep 285, loss: 6.223158359527588\r\nStep 286, loss: 5.904426574707031\r\nStep 287, loss: 6.371645450592041\r\nStep 288, loss: 6.0909423828125\r\nStep 289, loss: 6.016209125518799\r\nStep 290, loss: 6.319799900054932\r\nStep 291, loss: 6.296480178833008\r\nStep 292, loss: 6.148037910461426\r\nStep 293, loss: 5.923506259918213\r\nStep 294, loss: 5.763485431671143\r\nStep 295, loss: 6.133178234100342\r\nStep 296, loss: 6.073211193084717\r\nStep 297, loss: 6.169839859008789\r\nStep 298, loss: 6.141224384307861\r\nStep 299, loss: 6.168487071990967\r\nStep 300, loss: 6.119058132171631\r\nStep 391, loss: 5.777060031890869\r\nStep 392, loss: 5.80929708480835\r\nStep 393, loss: 5.858407974243164\r\nStep 394, loss: 5.820347785949707\r\nStep 395, loss: 5.752736568450928\r\nStep 396, loss: 5.977144241333008\r\nStep 397, loss: 5.999248504638672\r\nStep 398, loss: 5.756777763366699\r\nStep 399, loss: 5.889488220214844\r\nStep 400, loss: 5.789821624755859\r\nStep 401, loss: 5.788527488708496\r\nStep 402, loss: 5.627648830413818\r\nStep 403, loss: 5.792341709136963\r\nStep 404, loss: 5.86679220199585\r\nStep 405, loss: 5.916896820068359\r\nStep 406, loss: 5.835608005523682\r\nStep 407, loss: 6.11745023727417\r\nStep 408, loss: 5.773310661315918\r\nStep 409, loss: 5.883132457733154\r\nStep 410, loss: 5.6630048751831055\r\nStep 411, loss: 5.871366500854492\r\nStep 412, loss: 5.903974533081055\r\nStep 413, loss: 5.535669803619385\r\nStep 414, loss: 5.646414756774902\r\nStep 415, loss: 5.724021911621094\r\nStep 416, loss: 5.677767753601074\r\nStep 417, loss: 5.652090072631836\r\nStep 418, loss: 5.597585678100586\r\nStep 419, loss: 5.762701988220215\r\nStep 420, loss: 5.502464771270752\r\nStep 301, loss: 6.03639554977417\r\nStep 302, loss: 6.216285228729248\r\nStep 303, loss: 6.022148609161377\r\nStep 304, loss: 5.897799968719482\r\nStep 305, loss: 5.894784927368164\r\nStep 306, loss: 6.132763385772705\r\nStep 307, loss: 5.888872146606445\r\nStep 308, loss: 6.012820243835449\r\nStep 309, loss: 6.126757621765137\r\nStep 310, loss: 6.202884197235107\r\nStep 311, loss: 6.018051624298096\r\nStep 312, loss: 6.108768939971924\r\nStep 313, loss: 5.92648458480835\r\nStep 314, loss: 6.071450710296631\r\nStep 315, loss: 5.71877384185791\r\nStep 316, loss: 5.99001407623291\r\nStep 317, loss: 5.833607196807861\r\nStep 318, loss: 6.032952785491943\r\nStep 319, loss: 6.038209438323975\r\nStep 320, loss: 6.10830545425415\r\nStep 321, loss: 5.989357948303223\r\nStep 322, loss: 6.006596088409424\r\nStep 323, loss: 6.012803077697754\r\nStep 324, loss: 6.0025787353515625\r\nStep 325, loss: 5.972760200500488\r\nStep 326, loss: 6.077414512634277\r\nStep 327, loss: 5.926943302154541\r\nStep 328, loss: 6.139301300048828\r\nStep 329, loss: 5.791858196258545\r\nStep 330, loss: 5.903587818145752\r\nStep 421, loss: 5.753362655639648\r\nStep 422, loss: 5.559510707855225\r\nStep 423, loss: 5.701245307922363\r\nStep 424, loss: 5.822718620300293\r\nStep 425, loss: 5.51902437210083\r\nStep 426, loss: 5.723858833312988\r\nStep 427, loss: 5.752884387969971\r\nStep 428, loss: 5.7527642250061035\r\nStep 429, loss: 5.536493301391602\r\nStep 430, loss: 5.853747367858887\r\nStep 431, loss: 5.712923526763916\r\nStep 432, loss: 5.617330551147461\r\nStep 433, loss: 5.601428985595703\r\nStep 434, loss: 5.9634552001953125\r\nStep 435, loss: 5.687324523925781\r\nStep 436, loss: 5.664196968078613\r\nStep 437, loss: 5.776155471801758\r\nStep 438, loss: 5.555196285247803\r\nStep 439, loss: 5.63498592376709\r\nStep 440, loss: 5.8197760581970215\r\nStep 441, loss: 5.577805995941162\r\nStep 442, loss: 5.509472846984863\r\nStep 443, loss: 5.6083526611328125\r\nStep 444, loss: 5.517510890960693\r\nStep 445, loss: 5.543288707733154\r\nStep 446, loss: 5.672813415527344\r\nStep 447, loss: 5.723514080047607\r\nStep 448, loss: 5.612575531005859\r\nStep 449, loss: 5.857496738433838\r\nStep 450, loss: 5.798945903778076\r\nStep 331, loss: 6.382968902587891\r\nStep 332, loss: 6.010067462921143\r\nStep 333, loss: 6.022067070007324\r\nStep 334, loss: 5.914710998535156\r\nStep 335, loss: 6.026205539703369\r\nStep 336, loss: 6.101585865020752\r\nStep 337, loss: 5.8053202629089355\r\nStep 338, loss: 5.739813327789307\r\nStep 339, loss: 5.901891231536865\r\nStep 340, loss: 5.769777774810791\r\nStep 341, loss: 5.84769868850708\r\nStep 342, loss: 5.975956916809082\r\nStep 343, loss: 5.790035247802734\r\nStep 344, loss: 5.943611145019531\r\nStep 345, loss: 5.850645065307617\r\nStep 346, loss: 5.7808990478515625\r\nStep 347, loss: 5.690489292144775\r\nStep 348, loss: 5.986003398895264\r\nStep 349, loss: 5.836080551147461\r\nStep 350, loss: 5.887240886688232\r\nStep 351, loss: 5.904907703399658\r\nStep 352, loss: 5.790065288543701\r\nStep 353, loss: 5.890659809112549\r\nStep 354, loss: 5.948431015014648\r\nStep 355, loss: 5.6976847648620605\r\nStep 356, loss: 5.886941909790039\r\nStep 357, loss: 5.837642192840576\r\nStep 358, loss: 5.8052215576171875\r\nStep 359, loss: 5.958202838897705\r\nStep 360, loss: 6.015711784362793\r\nStep 451, loss: 5.441054821014404\r\nStep 452, loss: 5.596620082855225\r\nStep 453, loss: 5.544404983520508\r\nStep 454, loss: 5.754658222198486\r\nStep 455, loss: 5.781772613525391\r\nStep 456, loss: 5.656846523284912\r\nStep 457, loss: 5.571268081665039\r\nStep 458, loss: 5.702635765075684\r\nStep 459, loss: 5.473593235015869\r\nStep 460, loss: 5.5884690284729\r\nStep 461, loss: 5.590051651000977\r\nStep 462, loss: 5.594494342803955\r\nStep 463, loss: 5.555944919586182\r\nStep 464, loss: 5.735914707183838\r\nStep 465, loss: 5.679328441619873\r\nStep 466, loss: 5.410174369812012\r\nStep 467, loss: 5.712713718414307\r\nStep 468, loss: 5.571666717529297\r\nStep 469, loss: 5.591692924499512\r\nStep 470, loss: 5.54882287979126\r\nStep 471, loss: 5.527029991149902\r\nStep 472, loss: 5.44807767868042\r\nStep 473, loss: 5.462024211883545\r\nStep 474, loss: 5.600245952606201\r\nStep 475, loss: 5.820129871368408\r\nStep 476, loss: 5.609768390655518\r\nStep 477, loss: 5.626163959503174\r\nStep 478, loss: 5.776749610900879\r\nStep 479, loss: 5.380319595336914\r\nStep 480, loss: 5.529082298278809\r\nStep 361, loss: 5.924943923950195\r\nStep 362, loss: 5.905316352844238\r\nStep 363, loss: 5.8577880859375\r\nStep 364, loss: 5.833385944366455\r\nStep 365, loss: 6.09619140625\r\nStep 366, loss: 5.879598140716553\r\nStep 367, loss: 5.800159931182861\r\nStep 368, loss: 5.879344463348389\r\nStep 369, loss: 5.841940879821777\r\nStep 370, loss: 6.016635894775391\r\nStep 371, loss: 5.777203559875488\r\nStep 372, loss: 5.687002658843994\r\nStep 373, loss: 5.890249252319336\r\nStep 374, loss: 5.735091209411621\r\nStep 375, loss: 5.94771146774292\r\nStep 376, loss: 5.657384872436523\r\nStep 377, loss: 5.763212203979492\r\nStep 378, loss: 6.015496730804443\r\nStep 379, loss: 5.793003082275391\r\nStep 380, loss: 5.845806121826172\r\nStep 381, loss: 5.729050159454346\r\nStep 382, loss: 5.7567267417907715\r\nStep 383, loss: 5.785036087036133\r\nStep 384, loss: 5.758927345275879\r\nStep 385, loss: 5.695383071899414\r\nStep 386, loss: 5.753874778747559\r\nStep 387, loss: 5.656781196594238\r\nStep 388, loss: 5.64355993270874\r\nStep 389, loss: 5.938015460968018\r\nStep 390, loss: 5.836848735809326\r\nStep 481, loss: 5.674862861633301\r\nStep 391, loss: 5.777060031890869\r\nStep 392, loss: 5.80929708480835\r\nStep 393, loss: 5.858407974243164\r\nStep 394, loss: 5.820347785949707\r\nStep 395, loss: 5.752736568450928\r\nStep 396, loss: 5.977144241333008\r\nStep 397, loss: 5.999248504638672\r\nStep 398, loss: 5.756777763366699\r\nStep 399, loss: 5.889488220214844\r\nStep 400, loss: 5.789821624755859\r\nStep 401, loss: 5.788527488708496\r\nStep 402, loss: 5.627648830413818\r\nStep 403, loss: 5.792341709136963\r\nStep 404, loss: 5.86679220199585\r\nStep 405, loss: 5.916896820068359\r\nStep 406, loss: 5.835608005523682\r\nStep 407, loss: 6.11745023727417\r\nStep 408, loss: 5.773310661315918\r\nStep 409, loss: 5.883132457733154\r\nStep 410, loss: 5.6630048751831055\r\nStep 411, loss: 5.871366500854492\r\nStep 412, loss: 5.903974533081055\r\nStep 413, loss: 5.535669803619385\r\nStep 414, loss: 5.646414756774902\r\nStep 415, loss: 5.724021911621094\r\nStep 416, loss: 5.677767753601074\r\nStep 417, loss: 5.652090072631836\r\nStep 418, loss: 5.597585678100586\r\nStep 419, loss: 5.762701988220215\r\nStep 420, loss: 5.502464771270752\r\nStep 421, loss: 5.753362655639648\r\nStep 422, loss: 5.559510707855225\r\nStep 423, loss: 5.701245307922363\r\nStep 424, loss: 5.822718620300293\r\nStep 425, loss: 5.51902437210083\r\nStep 426, loss: 5.723858833312988\r\nStep 427, loss: 5.752884387969971\r\nStep 428, loss: 5.7527642250061035\r\nStep 429, loss: 5.536493301391602\r\nStep 430, loss: 5.853747367858887\r\nStep 431, loss: 5.712923526763916\r\nStep 432, loss: 5.617330551147461\r\nStep 433, loss: 5.601428985595703\r\nStep 434, loss: 5.9634552001953125\r\nStep 435, loss: 5.687324523925781\r\nStep 436, loss: 5.664196968078613\r\nStep 437, loss: 5.776155471801758\r\nStep 438, loss: 5.555196285247803\r\nStep 439, loss: 5.63498592376709\r\nStep 440, loss: 5.8197760581970215\r\nStep 441, loss: 5.577805995941162\r\nStep 442, loss: 5.509472846984863\r\nStep 443, loss: 5.6083526611328125\r\nStep 444, loss: 5.517510890960693\r\nStep 445, loss: 5.543288707733154\r\nStep 446, loss: 5.672813415527344\r\nStep 447, loss: 5.723514080047607\r\nStep 448, loss: 5.612575531005859\r\nStep 449, loss: 5.857496738433838\r\nStep 450, loss: 5.798945903778076\r\nStep 451, loss: 5.441054821014404\r\nStep 452, loss: 5.596620082855225\r\nStep 453, loss: 5.544404983520508\r\nStep 454, loss: 5.754658222198486\r\nStep 455, loss: 5.781772613525391\r\nStep 456, loss: 5.656846523284912\r\nStep 457, loss: 5.571268081665039\r\nStep 458, loss: 5.702635765075684\r\nStep 459, loss: 5.473593235015869\r\nStep 460, loss: 5.5884690284729\r\nStep 461, loss: 5.590051651000977\r\nStep 462, loss: 5.594494342803955\r\nStep 463, loss: 5.555944919586182\r\nStep 464, loss: 5.735914707183838\r\nStep 465, loss: 5.679328441619873\r\nStep 466, loss: 5.410174369812012\r\nStep 467, loss: 5.712713718414307\r\nStep 468, loss: 5.571666717529297\r\nStep 469, loss: 5.591692924499512\r\nStep 470, loss: 5.54882287979126\r\nStep 471, loss: 5.527029991149902\r\nStep 472, loss: 5.44807767868042\r\nStep 473, loss: 5.462024211883545\r\nStep 474, loss: 5.600245952606201\r\nStep 475, loss: 5.820129871368408\r\nStep 476, loss: 5.609768390655518\r\nStep 477, loss: 5.626163959503174\r\nStep 478, loss: 5.776749610900879\r\nStep 479, loss: 5.380319595336914\r\nStep 480, loss: 5.529082298278809\r\nStep 481, loss: 5.674862861633301\r\n",,terminal_output +2365,3697748,"TERMINAL",0,0,"189",,terminal_output +2366,3698399,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2367,3698731,"TERMINAL",0,0,"2930",,terminal_output +2368,3699084,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",875,0,"",shellscript,selection_mouse +2369,3699797,"TERMINAL",0,0,"3101",,terminal_output +2370,3699913,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",862,17,"",shellscript,content +2371,3700372,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",862,1,"",shellscript,content +2372,3700817,"TERMINAL",0,0,"412",,terminal_output +2373,3701949,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2374,3701970,"TERMINAL",0,0,"634",,terminal_output +2375,3702906,"TERMINAL",0,0,"745",,terminal_output +2376,3703164,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2377,3703948,"TERMINAL",0,0,"856",,terminal_output +2378,3704996,"TERMINAL",0,0,"967",,terminal_output +2379,3706061,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2380,3706079,"TERMINAL",0,0,"20:0078",,terminal_output +2381,3707084,"TERMINAL",0,0,"189",,terminal_output +2382,3708132,"TERMINAL",0,0,"2940",,terminal_output +2383,3709041,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",864,0,"",shellscript,selection_mouse +2384,3709176,"TERMINAL",0,0,"3201",,terminal_output +2385,3709801,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",863,0,"",shellscript,selection_command +2386,3709972,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,0,"",shellscript,selection_command +2387,3710240,"TERMINAL",0,0,"412",,terminal_output +2388,3710369,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,18,"",shellscript,content +2389,3710723,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1001,0,"",shellscript,selection_command +2390,3711296,"TERMINAL",0,0,"523",,terminal_output +2391,3712320,"TERMINAL",0,0,"634",,terminal_output +2392,3713332,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1002,0,"",shellscript,selection_command +2393,3713386,"TERMINAL",0,0,"745",,terminal_output +2394,3713825,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1003,0,"",shellscript,selection_command +2395,3713866,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1004,0,"",shellscript,selection_command +2396,3713929,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1005,0,"",shellscript,selection_command +2397,3713929,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1006,0,"",shellscript,selection_command +2398,3713946,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1007,0,"",shellscript,selection_command +2399,3714009,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1008,0,"",shellscript,selection_command +2400,3714024,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1009,0,"",shellscript,selection_command +2401,3714040,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1010,0,"",shellscript,selection_command +2402,3714099,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1011,0,"",shellscript,selection_command +2403,3714099,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1012,0,"",shellscript,selection_command +2404,3714236,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1013,0,"",shellscript,selection_command +2405,3714418,"TERMINAL",0,0,"856",,terminal_output +2406,3715104,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1012,0,"",shellscript,selection_command +2407,3715461,"TERMINAL",0,0,"967",,terminal_output +2408,3716304,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1013,0,"tokenizer_ckpt_dir",shellscript,content +2409,3716307,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1030,0,"",shellscript,selection_command +2410,3716512,"TERMINAL",0,0,"1078",,terminal_output +2411,3717558,"TERMINAL",0,0,"189",,terminal_output +2412,3717705,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1029,0,"",shellscript,selection_command +2413,3718196,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1028,0,"",shellscript,selection_command +2414,3718260,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1027,0,"",shellscript,selection_command +2415,3718261,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1026,0,"",shellscript,selection_command +2416,3718327,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1025,0,"",shellscript,selection_command +2417,3718328,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1024,0,"",shellscript,selection_command +2418,3718344,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1023,0,"",shellscript,selection_command +2419,3718405,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1022,0,"",shellscript,selection_command +2420,3718406,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1021,0,"",shellscript,selection_command +2421,3718469,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1020,0,"",shellscript,selection_command +2422,3718469,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1019,0,"",shellscript,selection_command +2423,3718562,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1018,0,"",shellscript,selection_command +2424,3718632,"TERMINAL",0,0,"2950",,terminal_output +2425,3718687,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1017,0,"",shellscript,selection_command +2426,3718848,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1016,0,"",shellscript,selection_command +2427,3719019,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1015,0,"",shellscript,selection_command +2428,3719115,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1014,0,"",shellscript,selection_command +2429,3719249,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1013,0,"",shellscript,selection_command +2430,3719365,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1012,0,"",shellscript,selection_command +2431,3719650,"TERMINAL",0,0,"3301",,terminal_output +2432,3719962,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1012,1,"r",shellscript,selection_command +2433,3720201,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1011,2,"ir",shellscript,selection_command +2434,3720693,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1010,3,"dir",shellscript,selection_command +2435,3720725,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1009,4,"_dir",shellscript,selection_command +2436,3720787,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1008,5,"t_dir",shellscript,selection_command +2437,3720789,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1007,6,"pt_dir",shellscript,selection_command +2438,3720790,"TERMINAL",0,0,"412",,terminal_output +2439,3720809,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1006,7,"kpt_dir",shellscript,selection_command +2440,3720848,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1005,8,"ckpt_dir",shellscript,selection_command +2441,3720878,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1004,9,"_ckpt_dir",shellscript,selection_command +2442,3720936,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1003,10,"m_ckpt_dir",shellscript,selection_command +2443,3720936,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1002,11,"am_ckpt_dir",shellscript,selection_command +2444,3720958,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1001,12,"lam_ckpt_dir",shellscript,selection_command +2445,3721355,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1001,12,"",shellscript,content +2446,3721533,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,0,"",shellscript,selection_command +2447,3721730,"TERMINAL",0,0,"523",,terminal_output +2448,3722770,"TERMINAL",0,0,"634",,terminal_output +2449,3723119,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",863,0,"lam_ckpt_dir",shellscript,content +2450,3723125,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",874,0,"",shellscript,selection_command +2451,3723855,"TERMINAL",0,0,"756",,terminal_output +2452,3724367,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",875,0,"",shellscript,selection_command +2453,3724887,"TERMINAL",0,0,"967",,terminal_output +2454,3725220,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",875,0,"=",shellscript,content +2455,3725221,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",876,0,"",shellscript,selection_keyboard +2456,3725502,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",875,0,"",shellscript,selection_command +2457,3725912,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",863,0,"",shellscript,selection_command +2458,3725916,"TERMINAL",0,0,"2078",,terminal_output +2459,3726125,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,0,"",shellscript,selection_command +2460,3726803,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,1,"",shellscript,content +2461,3726957,"TERMINAL",0,0,"189",,terminal_output +2462,3727999,"TERMINAL",0,0,"299:00",,terminal_output +2463,3729037,"TERMINAL",0,0,"3401",,terminal_output +2464,3730086,"TERMINAL",0,0,"412",,terminal_output +2465,3731128,"TERMINAL",0,0,"523",,terminal_output +2466,3732257,"TERMINAL",0,0,"634",,terminal_output +2467,3733264,"TERMINAL",0,0,"745",,terminal_output +2468,3733292,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,151,"",shellscript,content +2469,3733393,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1017,0,"",shellscript,selection_command +2470,3733895,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1018,0,"",shellscript,selection_command +2471,3733910,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1035,0,"",shellscript,selection_command +2472,3733951,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1036,0,"",shellscript,selection_command +2473,3734018,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1068,0,"",shellscript,selection_command +2474,3734018,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1101,0,"",shellscript,selection_command +2475,3734053,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1123,0,"",shellscript,selection_command +2476,3734065,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1140,0,"",shellscript,selection_command +2477,3734119,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1160,0,"",shellscript,selection_command +2478,3734155,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1192,0,"",shellscript,selection_command +2479,3734156,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1204,0,"",shellscript,selection_command +2480,3734207,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1241,0,"",shellscript,selection_command +2481,3734223,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1303,0,"",shellscript,selection_command +2482,3734279,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1331,0,"",shellscript,selection_command +2483,3734279,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1357,0,"",shellscript,selection_command +2484,3734280,"TERMINAL",0,0,"856",,terminal_output +2485,3734308,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1379,0,"",shellscript,selection_command +2486,3734338,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1428,0,"",shellscript,selection_command +2487,3734374,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1465,0,"",shellscript,selection_command +2488,3734893,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1428,0,"",shellscript,selection_command +2489,3735319,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1428,37,"",shellscript,content +2490,3735333,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1432,0,"",shellscript,selection_command +2491,3735333,"TERMINAL",0,0,"967",,terminal_output +2492,3736396,"TERMINAL",0,0,"3078",,terminal_output +2493,3737378,"TERMINAL",0,0,"189",,terminal_output +2494,3738514,"TERMINAL",0,0,"2910",,terminal_output +2495,3739464,"TERMINAL",0,0,"3501",,terminal_output +2496,3739565,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",980,0,"",shellscript,selection_mouse +2497,3739683,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",949,42,"train_tokenizer_batch_size_scaling_16_node",shellscript,selection_mouse +2498,3739832,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,155,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n",shellscript,selection_mouse +2499,3740386,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",980,0,"",shellscript,selection_mouse +2500,3740386,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",949,42,"train_tokenizer_batch_size_scaling_16_node",shellscript,selection_mouse +2501,3740504,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,155,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n",shellscript,selection_mouse +2502,3740565,"TERMINAL",0,0,"412",,terminal_output +2503,3741239,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1016,0,"",shellscript,selection_mouse +2504,3741248,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1015,0,"",shellscript,selection_command +2505,3741560,"TERMINAL",0,0,"523",,terminal_output +2506,3741601,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1016,0,"",shellscript,selection_mouse +2507,3741616,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1015,0,"",shellscript,selection_command +2508,3741755,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1015,1,"/",shellscript,selection_mouse +2509,3741755,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1015,2,"/\n",shellscript,selection_mouse +2510,3741755,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1016,0,"",shellscript,selection_command +2511,3741768,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1016,1,"\n",shellscript,selection_mouse +2512,3741829,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",949,67,"train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/",shellscript,selection_mouse +2513,3741831,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",937,79,"checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/",shellscript,selection_mouse +2514,3741851,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",936,80,"/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/",shellscript,selection_mouse +2515,3741895,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",922,94,"jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/",shellscript,selection_mouse +2516,3741910,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1016,1,"\n",shellscript,selection_mouse +2517,3741969,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",892,124,"workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/",shellscript,selection_mouse +2518,3742029,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",891,125,"/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/",shellscript,selection_mouse +2519,3742030,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",861,155,"\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/",shellscript,selection_mouse +2520,3742299,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,154,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/",shellscript,selection_mouse +2521,3742699,"TERMINAL",0,0,"634",,terminal_output +2522,3742835,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",877,0,"",shellscript,selection_mouse +2523,3742836,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,18,"tokenizer_ckpt_dir",shellscript,selection_mouse +2524,3743006,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,155,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n",shellscript,selection_mouse +2525,3743099,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,129,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node",shellscript,selection_mouse +2526,3743116,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,137,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526",shellscript,selection_mouse +2527,3743172,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,153,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000",shellscript,selection_mouse +2528,3743182,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,155,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n",shellscript,selection_mouse +2529,3743352,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,153,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000",shellscript,selection_mouse +2530,3743649,"TERMINAL",0,0,"745",,terminal_output +2531,3743775,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1013,0,"",shellscript,selection_mouse +2532,3743776,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1000,15,"tokenizer_22000",shellscript,selection_mouse +2533,3743933,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",992,23,"3321526/tokenizer_22000",shellscript,selection_mouse +2534,3743951,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",861,154,"\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000",shellscript,selection_mouse +2535,3744012,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",922,93,"jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000",shellscript,selection_mouse +2536,3744013,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1000,17,"tokenizer_22000/\n",shellscript,selection_mouse +2537,3744018,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1000,34,"tokenizer_22000/\n\nenv | grep SLURM",shellscript,selection_mouse +2538,3744182,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1000,17,"tokenizer_22000/\n",shellscript,selection_mouse +2539,3744242,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",882,133,"hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000",shellscript,selection_mouse +2540,3744303,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",881,134,"/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000",shellscript,selection_mouse +2541,3744317,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",862,153,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000",shellscript,selection_mouse +2542,3744399,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",861,154,"\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000",shellscript,selection_mouse +2543,3744575,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",861,0,"",shellscript,selection_mouse +2544,3744698,"TERMINAL",0,0,"856",,terminal_output +2545,3745770,"TERMINAL",0,0,"967",,terminal_output +2546,3746423,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2547,3746802,"TERMINAL",0,0,"4078",,terminal_output +2548,3747514,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",871,0,"",shellscript,selection_mouse +2549,3747838,"TERMINAL",0,0,"1920",,terminal_output +2550,3748292,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",862,151,"",shellscript,content +2551,3748371,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1017,0,"",shellscript,selection_command +2552,3748867,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1018,0,"",shellscript,selection_command +2553,3748898,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1035,0,"",shellscript,selection_command +2554,3748928,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1036,0,"",shellscript,selection_command +2555,3748929,"TERMINAL",0,0,"32:001",,terminal_output +2556,3749002,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1068,0,"",shellscript,selection_command +2557,3749003,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1101,0,"",shellscript,selection_command +2558,3749007,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1123,0,"",shellscript,selection_command +2559,3749066,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1140,0,"",shellscript,selection_command +2560,3749078,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1162,0,"",shellscript,selection_command +2561,3749135,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1194,0,"",shellscript,selection_command +2562,3749136,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1206,0,"",shellscript,selection_command +2563,3749195,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1236,0,"",shellscript,selection_command +2564,3749196,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1273,0,"",shellscript,selection_command +2565,3749255,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1317,0,"",shellscript,selection_command +2566,3749256,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1348,0,"",shellscript,selection_command +2567,3749313,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1374,0,"",shellscript,selection_command +2568,3749392,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1396,0,"",shellscript,selection_command +2569,3749548,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1445,0,"",shellscript,selection_command +2570,3749708,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1481,0,"",shellscript,selection_command +2571,3749912,"TERMINAL",0,0,"412",,terminal_output +2572,3750224,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1481,35,"",shellscript,content +2573,3750375,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1445,0,"",shellscript,selection_command +2574,3750575,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1396,0,"",shellscript,selection_command +2575,3750698,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1374,0,"",shellscript,selection_command +2576,3750834,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1348,0,"",shellscript,selection_command +2577,3750972,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1317,0,"",shellscript,selection_command +2578,3750972,"TERMINAL",0,0,"523",,terminal_output +2579,3751115,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1273,0,"",shellscript,selection_command +2580,3751252,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1236,0,"",shellscript,selection_command +2581,3752027,"TERMINAL",0,0,"634",,terminal_output +2582,3753045,"TERMINAL",0,0,"745",,terminal_output +2583,3754089,"TERMINAL",0,0,"856",,terminal_output +2584,3755131,"TERMINAL",0,0,"967",,terminal_output +2585,3755790,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"",shellscript,tab +2586,3756193,"TERMINAL",0,0,"5078",,terminal_output +2587,3756961,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1300,0,"",shellscript,selection_mouse +2588,3757085,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1299,1,"\n",shellscript,selection_mouse +2589,3757105,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",985,315,"\n --max_lr=2.00e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-4-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 4-node sqrt-lr \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2590,3757165,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",533,767,"/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=192 \\n --min_lr=0 \\n --max_lr=2.00e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-4-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 4-node sqrt-lr \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2591,3757170,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",39,1261,"#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_4_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=192 \\n --min_lr=0 \\n --max_lr=2.00e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-4-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 4-node sqrt-lr \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2592,3757171,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,1300,"#!/usr/bin/env bash\n\n#SBATCH --nodes=4\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_4_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=192 \\n --min_lr=0 \\n --max_lr=2.00e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-4-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 4-node sqrt-lr \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2593,3757224,"TERMINAL",0,0,"189",,terminal_output +2594,3758042,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"",shellscript,selection_command +2595,3758307,"TERMINAL",0,0,"2930",,terminal_output +2596,3759357,"TERMINAL",0,0,"3101",,terminal_output +2597,3760418,"TERMINAL",0,0,"412",,terminal_output +2598,3761400,"TERMINAL",0,0,"523",,terminal_output +2599,3762435,"TERMINAL",0,0,"634",,terminal_output +2600,3763525,"TERMINAL",0,0,"745",,terminal_output +2601,3764592,"TERMINAL",0,0,"856",,terminal_output +2602,3765574,"TERMINAL",0,0,"967",,terminal_output +2603,3766612,"TERMINAL",0,0,"1:0078",,terminal_output +2604,3767692,"TERMINAL",0,0,"189",,terminal_output +2605,3768705,"TERMINAL",0,0,"2940",,terminal_output +2606,3769749,"TERMINAL",0,0,"3201",,terminal_output +2607,3770786,"TERMINAL",0,0,"412",,terminal_output +2608,3771833,"TERMINAL",0,0,"534",,terminal_output +2609,3772913,"TERMINAL",0,0,"745",,terminal_output +2610,3773934,"TERMINAL",0,0,"856",,terminal_output +2611,3774992,"TERMINAL",0,0,"967",,terminal_output +2612,3775997,"TERMINAL",0,0,"1078",,terminal_output +2613,3777045,"TERMINAL",0,0,"189",,terminal_output +2614,3778106,"TERMINAL",0,0,"2950",,terminal_output +2615,3778373,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",862,0,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n",shellscript,content +2616,3778394,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1017,17,"",shellscript,content +2617,3778395,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1018,0,"env | grep SLURM\n\n",shellscript,content +2618,3778422,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1036,0,"srun python train_dynamics.py \\n",shellscript,content +2619,3778480,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1068,33,"",shellscript,content +2620,3778639,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1141,0," --max_lr=2e-4 \\n",shellscript,content +2621,3778741,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1161,23,"",shellscript,content +2622,3778842,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1193,37,"",shellscript,content +2623,3778859,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1205,0," --log_checkpoint_interval=1000 \\n --log \\n",shellscript,content +2624,3778986,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1242,0," --name=dynamics-batch-size-scaling-4-node-$slurm_job_id \\n",shellscript,content +2625,3778986,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1304,0," --tags dynamics debug \\n",shellscript,content +2626,3779049,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1332,0," --entity instant-uv \\n",shellscript,content +2627,3779112,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1358,166,"",shellscript,content +2628,3779170,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1380,0," --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n",shellscript,content +2629,3779170,"TERMINAL",0,0,"3301",,terminal_output +2630,3780188,"TERMINAL",0,0,"412",,terminal_output +2631,3781216,"TERMINAL",0,0,"523",,terminal_output +2632,3782257,"TERMINAL",0,0,"634",,terminal_output +2633,3783358,"TERMINAL",0,0,"745",,terminal_output +2634,3784380,"TERMINAL",0,0,"856",,terminal_output +2635,3785397,"TERMINAL",0,0,"967",,terminal_output +2636,3786441,"TERMINAL",0,0,"2078",,terminal_output +2637,3787165,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,tab +2638,3787511,"TERMINAL",0,0,"189",,terminal_output +2639,3787638,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1248,0,"",shellscript,selection_mouse +2640,3787655,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1247,0,"",shellscript,selection_command +2641,3788137,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1305,0,"",shellscript,selection_mouse +2642,3788303,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1304,1,"\n",shellscript,selection_mouse +2643,3788322,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1086,219,"\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2644,3788340,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",875,430,"URM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2645,3788395,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",492,813,"odule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2646,3788399,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",20,1285,"\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2647,3788399,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,1305,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2648,3788526,"TERMINAL",0,0,"2930:00",,terminal_output +2649,3789400,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",98,0,"",shellscript,selection_mouse +2650,3789579,"TERMINAL",0,0,"3401",,terminal_output +2651,3790584,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,selection_mouse +2652,3790652,"TERMINAL",0,0,"412",,terminal_output +2653,3790725,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,20,"#!/usr/bin/env bash\n",shellscript,selection_mouse +2654,3790740,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,95,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBA",shellscript,selection_mouse +2655,3790755,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,428,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n",shellscript,selection_mouse +2656,3790813,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,906,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenize",shellscript,selection_mouse +2657,3790813,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,1037,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \",shellscript,selection_mouse +2658,3790814,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,1192,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-sc",shellscript,selection_mouse +2659,3790826,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,1270,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \",shellscript,selection_mouse +2660,3790842,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,1304,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +2661,3790903,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,1305,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-8-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 8-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2662,3791661,"TERMINAL",0,0,"523",,terminal_output +2663,3792435,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,selection_command +2664,3792715,"TERMINAL",0,0,"634",,terminal_output +2665,3793770,"TERMINAL",0,0,"745",,terminal_output +2666,3794818,"TERMINAL",0,0,"856",,terminal_output +2667,3795834,"TERMINAL",0,0,"978",,terminal_output +2668,3796881,"TERMINAL",0,0,"3189",,terminal_output +2669,3797942,"TERMINAL",0,0,"2910",,terminal_output +2670,3799061,"TERMINAL",0,0,"3501",,terminal_output +2671,3800037,"TERMINAL",0,0,"412",,terminal_output +2672,3801087,"TERMINAL",0,0,"523",,terminal_output +2673,3802142,"TERMINAL",0,0,"634",,terminal_output +2674,3803189,"TERMINAL",0,0,"745",,terminal_output +2675,3804236,"TERMINAL",0,0,"856",,terminal_output +2676,3805278,"TERMINAL",0,0,"967",,terminal_output +2677,3806326,"TERMINAL",0,0,"4078",,terminal_output +2678,3807390,"TERMINAL",0,0,"189",,terminal_output +2679,3807877,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",862,0,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n",shellscript,content +2680,3807898,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1017,17,"",shellscript,content +2681,3807915,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1018,0,"env | grep SLURM\n\n",shellscript,content +2682,3807945,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1036,0,"srun python train_dynamics.py \\n",shellscript,content +2683,3808021,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1068,33,"",shellscript,content +2684,3808198,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1141,0," --max_lr=2.8e-4 \\n",shellscript,content +2685,3808230,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1163,20,"",shellscript,content +2686,3808290,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1195,37,"",shellscript,content +2687,3808299,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1207,0," --log_checkpoint_interval=1000 \\n --log \\n",shellscript,content +2688,3808431,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1244,0," --name=dynamics-batch-size-scaling-8-node-$slurm_job_id \\n",shellscript,content +2689,3808446,"TERMINAL",0,0,"2920",,terminal_output +2690,3808462,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1306,0," --tags dynamics debug \\n",shellscript,content +2691,3808523,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1334,0," --entity instant-uv \\n",shellscript,content +2692,3808524,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1360,174,"",shellscript,content +2693,3808544,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1382,0," --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n",shellscript,content +2694,3809479,"TERMINAL",0,0,"33:001",,terminal_output +2695,3810559,"TERMINAL",0,0,"412",,terminal_output +2696,3811660,"TERMINAL",0,0,"523",,terminal_output +2697,3812744,"TERMINAL",0,0,"634",,terminal_output +2698,3813767,"TERMINAL",0,0,"745",,terminal_output +2699,3814731,"TERMINAL",0,0,"856",,terminal_output +2700,3815819,"TERMINAL",0,0,"967",,terminal_output +2701,3816845,"TERMINAL",0,0,"5078",,terminal_output +2702,3817877,"TERMINAL",0,0,"2930",,terminal_output +2703,3818913,"TERMINAL",0,0,"3101",,terminal_output +2704,3820013,"TERMINAL",0,0,"412",,terminal_output +2705,3821001,"TERMINAL",0,0,"523",,terminal_output +2706,3822048,"TERMINAL",0,0,"634",,terminal_output +2707,3823095,"TERMINAL",0,0,"745",,terminal_output +2708,3824134,"TERMINAL",0,0,"856",,terminal_output +2709,3825167,"TERMINAL",0,0,"967",,terminal_output +2710,3826276,"TERMINAL",0,0,"2:0078",,terminal_output +2711,3827346,"TERMINAL",0,0,"189",,terminal_output +2712,3828410,"TERMINAL",0,0,"2940",,terminal_output +2713,3829544,"TERMINAL",0,0,"3201",,terminal_output +2714,3829825,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",0,0,"",shellscript,tab +2715,3830427,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1309,0,"",shellscript,selection_mouse +2716,3830477,"TERMINAL",0,0,"412",,terminal_output +2717,3830623,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1056,253,"oint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-16-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 16-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2718,3830653,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",672,637,"\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=768 \\n --min_lr=0 \\n --max_lr=4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-16-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 16-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2719,3830669,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",68,1241,"#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_16_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=768 \\n --min_lr=0 \\n --max_lr=4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-16-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 16-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2720,3830686,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",0,1309,"#!/usr/bin/env bash\n\n#SBATCH --nodes=16\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_16_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=768 \\n --min_lr=0 \\n --max_lr=4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-16-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 16-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +2721,3831636,"TERMINAL",0,0,"523",,terminal_output +2722,3832207,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",0,0,"",shellscript,selection_command +2723,3832568,"TERMINAL",0,0,"634",,terminal_output +2724,3833613,"TERMINAL",0,0,"745",,terminal_output +2725,3834762,"TERMINAL",0,0,"856",,terminal_output +2726,3835693,"TERMINAL",0,0,"967",,terminal_output +2727,3836755,"TERMINAL",0,0,"1078",,terminal_output +2728,3837847,"TERMINAL",0,0,"189",,terminal_output +2729,3838823,"TERMINAL",0,0,"23051",,terminal_output +2730,3839861,"TERMINAL",0,0,"412",,terminal_output +2731,3840899,"TERMINAL",0,0,"523",,terminal_output +2732,3841929,"TERMINAL",0,0,"634",,terminal_output +2733,3843025,"TERMINAL",0,0,"745",,terminal_output +2734,3843690,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",864,0,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n",shellscript,content +2735,3843707,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1019,17,"",shellscript,content +2736,3843708,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1020,0,"env | grep SLURM\n\n",shellscript,content +2737,3843765,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1038,0,"srun python train_dynamics.py \\n",shellscript,content +2738,3843780,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1070,33,"",shellscript,content +2739,3843981,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1195,37,"",shellscript,content +2740,3844004,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1207,0," --log_checkpoint_interval=1000 \\n --log \\n",shellscript,content +2741,3844089,"TERMINAL",0,0,"856",,terminal_output +2742,3844152,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1244,0," --name=dynamics-batch-size-scaling-16-node-$slurm_job_id \\n",shellscript,content +2743,3844213,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1307,0," --tags dynamics debug \\n",shellscript,content +2744,3844272,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1335,0," --entity instant-uv \\n",shellscript,content +2745,3844273,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1361,176,"",shellscript,content +2746,3844347,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1383,0," --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n",shellscript,content +2747,3845070,"TERMINAL",0,0,"967",,terminal_output +2748,3846120,"TERMINAL",0,0,"2078",,terminal_output +2749,3847186,"TERMINAL",0,0,"189",,terminal_output +2750,3848232,"TERMINAL",0,0,"291:00",,terminal_output +2751,3849304,"TERMINAL",0,0,"3401",,terminal_output +2752,3850460,"TERMINAL",0,0,"412",,terminal_output +2753,3851364,"TERMINAL",0,0,"523",,terminal_output +2754,3852415,"TERMINAL",0,0,"634",,terminal_output +2755,3853560,"TERMINAL",0,0,"745",,terminal_output +2756,3854525,"TERMINAL",0,0,"856",,terminal_output +2757,3854701,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,tab +2758,3855651,"TERMINAL",0,0,"967",,terminal_output +2759,3856592,"TERMINAL",0,0,"3078",,terminal_output +2760,3857699,"TERMINAL",0,0,"189",,terminal_output +2761,3858720,"TERMINAL",0,0,"2910",,terminal_output +2762,3859747,"TERMINAL",0,0,"3501",,terminal_output +2763,3860761,"TERMINAL",0,0,"412",,terminal_output +2764,3861811,"TERMINAL",0,0,"523",,terminal_output +2765,3862921,"TERMINAL",0,0,"745",,terminal_output +2766,3863944,"TERMINAL",0,0,"856",,terminal_output +2767,3864957,"TERMINAL",0,0,"967",,terminal_output +2768,3866013,"TERMINAL",0,0,"4078",,terminal_output +2769,3867059,"TERMINAL",0,0,"189",,terminal_output +2770,3868099,"TERMINAL",0,0,"2920",,terminal_output +2771,3869152,"TERMINAL",0,0,"34:001",,terminal_output +2772,3870194,"TERMINAL",0,0,"412",,terminal_output +2773,3871309,"TERMINAL",0,0,"523",,terminal_output +2774,3872311,"TERMINAL",0,0,"634",,terminal_output +2775,3873486,"TERMINAL",0,0,"745",,terminal_output +2776,3874713,"TERMINAL",0,0,"856",,terminal_output +2777,3875583,"TERMINAL",0,0,"967",,terminal_output +2778,3876626,"TERMINAL",0,0,"5078",,terminal_output +2779,3877774,"TERMINAL",0,0,"189",,terminal_output +2780,3878717,"TERMINAL",0,0,"2930",,terminal_output +2781,3879814,"TERMINAL",0,0,"3101",,terminal_output +2782,3880838,"TERMINAL",0,0,"423",,terminal_output +2783,3882007,"TERMINAL",0,0,"634",,terminal_output +2784,3882987,"TERMINAL",0,0,"745",,terminal_output +2785,3884731,"TERMINAL",0,0,"856",,terminal_output +2786,3885752,"TERMINAL",0,0,"967",,terminal_output +2787,3886752,"TERMINAL",0,0,"3:0078",,terminal_output +2788,3887804,"TERMINAL",0,0,"189",,terminal_output +2789,3888949,"TERMINAL",0,0,"22041",,terminal_output +2790,3889954,"TERMINAL",0,0,"412",,terminal_output +2791,3890938,"TERMINAL",0,0,"523",,terminal_output +2792,3892006,"TERMINAL",0,0,"634",,terminal_output +2793,3893063,"TERMINAL",0,0,"745",,terminal_output +2794,3894076,"TERMINAL",0,0,"856",,terminal_output +2795,3895125,"TERMINAL",0,0,"967",,terminal_output +2796,3896199,"TERMINAL",0,0,"1078",,terminal_output +2797,3897329,"TERMINAL",0,0,"189",,terminal_output +2798,3898274,"TERMINAL",0,0,"2950",,terminal_output +2799,3899329,"TERMINAL",0,0,"3301",,terminal_output +2800,3900389,"TERMINAL",0,0,"412",,terminal_output +2801,3901450,"TERMINAL",0,0,"523",,terminal_output +2802,3902552,"TERMINAL",0,0,"634",,terminal_output +2803,3903517,"TERMINAL",0,0,"745",,terminal_output +2804,3904563,"TERMINAL",0,0,"856",,terminal_output +2805,3905613,"TERMINAL",0,0,"967",,terminal_output +2806,3906748,"TERMINAL",0,0,"2078",,terminal_output +2807,3907709,"TERMINAL",0,0,"189",,terminal_output +2808,3908798,"TERMINAL",0,0,"292:00",,terminal_output +2809,3909819,"TERMINAL",0,0,"3401",,terminal_output +2810,3910877,"TERMINAL",0,0,"423",,terminal_output +2811,3911910,"TERMINAL",0,0,"634",,terminal_output +2812,3913001,"TERMINAL",0,0,"745",,terminal_output +2813,3914054,"TERMINAL",0,0,"856",,terminal_output +2814,3915078,"TERMINAL",0,0,"967",,terminal_output +2815,3916088,"TERMINAL",0,0,"3078",,terminal_output +2816,3917193,"TERMINAL",0,0,"189",,terminal_output +2817,3918172,"TERMINAL",0,0,"2910",,terminal_output +2818,3919221,"TERMINAL",0,0,"3501",,terminal_output +2819,3920271,"TERMINAL",0,0,"412",,terminal_output +2820,3921325,"TERMINAL",0,0,"523",,terminal_output +2821,3922417,"TERMINAL",0,0,"634",,terminal_output +2822,3923410,"TERMINAL",0,0,"745",,terminal_output +2823,3924466,"TERMINAL",0,0,"856",,terminal_output +2824,3925512,"TERMINAL",0,0,"967",,terminal_output +2825,3926628,"TERMINAL",0,0,"4078",,terminal_output +2826,3927642,"TERMINAL",0,0,"189",,terminal_output +2827,3928666,"TERMINAL",0,0,"2920",,terminal_output +2828,3929669,"TERMINAL",0,0,"35:001",,terminal_output +2829,3930715,"TERMINAL",0,0,"412",,terminal_output +2830,3931752,"TERMINAL",0,0,"523",,terminal_output +2831,3932858,"TERMINAL",0,0,"634",,terminal_output +2832,3933833,"TERMINAL",0,0,"756",,terminal_output +2833,3934904,"TERMINAL",0,0,"967",,terminal_output +2834,3935928,"TERMINAL",0,0,"5078",,terminal_output +2835,3937055,"TERMINAL",0,0,"189",,terminal_output +2836,3938386,"TERMINAL",0,0,"2930",,terminal_output +2837,3939324,"TERMINAL",0,0,"3101",,terminal_output +2838,3940369,"TERMINAL",0,0,"412",,terminal_output +2839,3941420,"TERMINAL",0,0,"523",,terminal_output +2840,3942519,"TERMINAL",0,0,"634",,terminal_output +2841,3943610,"TERMINAL",0,0,"745",,terminal_output +2842,3944640,"TERMINAL",0,0,"856",,terminal_output +2843,3945621,"TERMINAL",0,0,"967",,terminal_output +2844,3946663,"TERMINAL",0,0,"4:0078",,terminal_output +2845,3947713,"TERMINAL",0,0,"189",,terminal_output +2846,3948835,"TERMINAL",0,0,"2940",,terminal_output +2847,3949861,"TERMINAL",0,0,"3201",,terminal_output +2848,3950841,"TERMINAL",0,0,"423",,terminal_output +2849,3951884,"TERMINAL",0,0,"634",,terminal_output +2850,3952779,"slurm/utils/alfred/sqrt_lr_scaling.py",0,0,"",python,tab +2851,3952951,"TERMINAL",0,0,"745",,terminal_output +2852,3953927,"slurm/utils/alfred/sqrt_lr_scaling.py",0,0,"import math as mt\n\nbatch_size_og = 128\nlr_old = 3e-4\n\nbatch_sizes = [2,4,8, 12, 24, 48, 96, 192, 384, 768]\n\nfor batch_size_new in batch_sizes:\n lr_new = mt.sqrt(batch_size_new/batch_size_og) * lr_old\n print(f""Batch size: {batch_size_new}, Learning rate: {lr_new}"")\n",python,content +2853,3953984,"TERMINAL",0,0,"856",,terminal_output +2854,3955008,"TERMINAL",0,0,"967",,terminal_output +2855,3955491,"slurm/utils/alfred/sqrt_lr_scaling.py",52,0,"",python,selection_mouse +2856,3956113,"slurm/utils/alfred/sqrt_lr_scaling.py",36,0,"",python,selection_mouse +2857,3956120,"TERMINAL",0,0,"1078",,terminal_output +2858,3956331,"slurm/utils/alfred/sqrt_lr_scaling.py",35,3,"128",python,selection_mouse +2859,3956481,"slurm/utils/alfred/sqrt_lr_scaling.py",19,20,"batch_size_og = 128\n",python,selection_mouse +2860,3957113,"slurm/utils/alfred/sqrt_lr_scaling.py",36,0,"",python,selection_mouse +2861,3957113,"slurm/utils/alfred/sqrt_lr_scaling.py",35,3,"128",python,selection_mouse +2862,3957128,"TERMINAL",0,0,"189",,terminal_output +2863,3958103,"TERMINAL",0,0,"2950",,terminal_output +2864,3959192,"TERMINAL",0,0,"3301",,terminal_output +2865,3959309,"slurm/utils/alfred/sqrt_lr_scaling.py",52,0,"",python,selection_mouse +2866,3959997,"slurm/utils/alfred/sqrt_lr_scaling.py",51,1,"",python,content +2867,3960164,"slurm/utils/alfred/sqrt_lr_scaling.py",51,0,"5",python,content +2868,3960166,"slurm/utils/alfred/sqrt_lr_scaling.py",52,0,"",python,selection_keyboard +2869,3960235,"TERMINAL",0,0,"412",,terminal_output +2870,3960533,"slurm/utils/alfred/sqrt_lr_scaling.py",32,0,"",python,selection_command +2871,3961287,"TERMINAL",0,0,"523",,terminal_output +2872,3961687,"slurm/utils/alfred/sqrt_lr_scaling.py",37,1,"",python,content +2873,3961822,"slurm/utils/alfred/sqrt_lr_scaling.py",36,1,"",python,content +2874,3961959,"slurm/utils/alfred/sqrt_lr_scaling.py",35,1,"",python,content +2875,3962896,"slurm/utils/alfred/sqrt_lr_scaling.py",35,0,"8",python,content +2876,3962897,"slurm/utils/alfred/sqrt_lr_scaling.py",36,0,"",python,selection_keyboard +2877,3962948,"TERMINAL",0,0,"645",,terminal_output +2878,3963324,"slurm/utils/alfred/sqrt_lr_scaling.py",36,0,"*",python,content +2879,3963325,"slurm/utils/alfred/sqrt_lr_scaling.py",37,0,"",python,selection_keyboard +2880,3963990,"TERMINAL",0,0,"856",,terminal_output +2881,3965034,"TERMINAL",0,0,"967",,terminal_output +2882,3965158,"slurm/utils/alfred/sqrt_lr_scaling.py",37,0,"4",python,content +2883,3965159,"slurm/utils/alfred/sqrt_lr_scaling.py",38,0,"",python,selection_keyboard +2884,3965173,"slurm/utils/alfred/sqrt_lr_scaling.py",38,0,"8",python,content +2885,3965174,"slurm/utils/alfred/sqrt_lr_scaling.py",39,0,"",python,selection_keyboard +2886,3965907,"slurm/utils/alfred/sqrt_lr_scaling.py",54,0,"",python,selection_mouse +2887,3966089,"TERMINAL",0,0,"2078",,terminal_output +2888,3966397,"slurm/utils/alfred/sqrt_lr_scaling.py",83,0,"",python,selection_mouse +2889,3966944,"slurm/utils/alfred/sqrt_lr_scaling.py",272,0,"",python,selection_mouse +2890,3967108,"TERMINAL",0,0,"189",,terminal_output +2891,3968150,"TERMINAL",0,0,"293:00",,terminal_output +2892,3968275,"TERMINAL",0,0,"bash",,terminal_focus +2893,3969182,"TERMINAL",0,0,"3401",,terminal_output +2894,3970231,"TERMINAL",0,0,"412",,terminal_output +2895,3971310,"TERMINAL",0,0,"523",,terminal_output +2896,3972260,"TERMINAL",0,0,"watch",,terminal_focus +2897,3972351,"TERMINAL",0,0,"634",,terminal_output +2898,3973201,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +2899,3975671,"TERMINAL",0,0,"python slurm/utils/alfred/sqrt_lr_scaling.py",,terminal_command +2900,3975721,"TERMINAL",0,0,"]633;E;2025-07-10 17:24:29 python slurm/utils/alfred/sqrt_lr_scaling.py;16c08be6-3885-420f-ac53-f5272aed6e54]633;C",,terminal_output +2901,3975924,"TERMINAL",0,0,"Batch size: 2, Learning rate: 2.1650635094610967e-06\r\nBatch size: 4, Learning rate: 3.0618621784789727e-06\r\nBatch size: 8, Learning rate: 4.330127018922193e-06\r\nBatch size: 12, Learning rate: 5.303300858899107e-06\r\nBatch size: 24, Learning rate: 7.5e-06\r\nBatch size: 48, Learning rate: 1.0606601717798214e-05\r\nBatch size: 96, Learning rate: 1.5e-05\r\nBatch size: 192, Learning rate: 2.1213203435596428e-05\r\nBatch size: 384, Learning rate: 3e-05\r\nBatch size: 768, Learning rate: 4.2426406871192855e-05\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +2902,3988437,"TERMINAL",0,0,"bash",,terminal_focus +2903,3991819,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2904,3998710,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1120,0,"",shellscript,selection_mouse +2905,4002578,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,tab +2906,4004273,"TERMINAL",0,0,"bash",,terminal_focus +2907,4045519,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,tab +2908,4045519,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1158,0,"",shellscript,selection_mouse +2909,4046339,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1157,0,"",shellscript,selection_command +2910,4046768,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1156,1,"",shellscript,content +2911,4046864,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1155,1,"",shellscript,content +2912,4047175,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1154,1,"",shellscript,content +2913,4048024,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1154,0,"3",shellscript,content +2914,4048024,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1155,0,"",shellscript,selection_keyboard +2915,4049555,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1157,1,"",shellscript,content +2916,4049621,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1157,0,"5",shellscript,content +2917,4049622,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1158,0,"",shellscript,selection_keyboard +2918,4055833,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",0,0,"",shellscript,tab +2919,4059210,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1157,0,"",shellscript,selection_mouse +2920,4060545,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1157,0,".",shellscript,content +2921,4060547,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1158,0,"",shellscript,selection_keyboard +2922,4060714,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1158,0,"2",shellscript,content +2923,4060715,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1159,0,"",shellscript,selection_keyboard +2924,4061873,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1161,0,"",shellscript,selection_command +2925,4063097,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1161,1,"",shellscript,content +2926,4063108,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1161,0,"5",shellscript,content +2927,4063108,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1162,0,"",shellscript,selection_keyboard +2928,4066939,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"",shellscript,tab +2929,4068755,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1156,0,"",shellscript,selection_mouse +2930,4073740,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1155,0,"",shellscript,selection_mouse +2931,4074718,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1155,0,".",shellscript,content +2932,4074719,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1156,0,"",shellscript,selection_keyboard +2933,4074991,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1156,0,"1",shellscript,content +2934,4074992,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1157,0,"",shellscript,selection_keyboard +2935,4076475,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1159,1,"",shellscript,content +2936,4076538,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1159,0,"5",shellscript,content +2937,4076538,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1160,0,"",shellscript,selection_keyboard +2938,4081093,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2939,4082218,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"",shellscript,tab +2940,4114548,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2941,4122689,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1156,0,"",shellscript,selection_mouse +2942,4123573,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1155,1,"",shellscript,content +2943,4123719,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1155,0,"5",shellscript,content +2944,4123720,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1156,0,"",shellscript,selection_keyboard +2945,4124590,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1158,0,"",shellscript,selection_command +2946,4124950,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1159,0,"",shellscript,selection_command +2947,4125527,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1158,1,"",shellscript,content +2948,4125591,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1158,0,"5",shellscript,content +2949,4125592,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1159,0,"",shellscript,selection_keyboard +2950,4133776,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2951,4136612,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1317,0,"",shellscript,selection_mouse +2952,4137510,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1155,0,"",shellscript,selection_mouse +2953,4138101,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1153,0,"",shellscript,selection_mouse +2954,4143185,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1153,1,"1",shellscript,content +2955,4143757,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1154,0,"",shellscript,selection_command +2956,4143983,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1155,0,"",shellscript,selection_command +2957,4144118,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1156,0,"",shellscript,selection_command +2958,4149945,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1156,1,"5",shellscript,content +2959,4152301,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2960,4154025,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"",shellscript,tab +2961,4156354,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,tab +2962,4157644,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",0,0,"",shellscript,tab +2963,4163084,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1354,0,"",shellscript,selection_mouse +2964,4164792,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"",shellscript,tab +2965,4165772,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"",shellscript,tab +2966,4166259,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2967,4166865,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2968,4253369,"slurm/jobs/mihir/horeka/yolo-runs/tester.sh",0,0,"",shellscript,tab +2969,4258458,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"",shellscript,tab +2970,4260052,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",1781,0,"",shellscript,selection_mouse +2971,4262452,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",968,0,"",shellscript,selection_mouse +2972,4262615,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",961,12,"lam_ckpt_dir",shellscript,selection_mouse +2973,4282900,"TERMINAL",0,0,"Step 482, loss: 5.577463626861572\r\nStep 483, loss: 5.537669658660889\r\nStep 484, loss: 5.941127300262451\r\nStep 485, loss: 5.490207195281982\r\nStep 486, loss: 5.543642044067383\r\nStep 487, loss: 5.644903659820557\r\nStep 488, loss: 5.602255821228027\r\nStep 489, loss: 5.7087721824646\r\nStep 490, loss: 5.508888244628906\r\nStep 491, loss: 5.632482528686523\r\nStep 492, loss: 5.39115571975708\r\nStep 493, loss: 5.692109107971191\r\nStep 494, loss: 5.630179405212402\r\nStep 495, loss: 5.478193759918213\r\nStep 496, loss: 5.638604164123535\r\nStep 497, loss: 5.54039192199707\r\nStep 498, loss: 5.653471946716309\r\nStep 499, loss: 5.654304027557373\r\nStep 500, loss: 5.648343086242676\r\nStep 501, loss: 5.6619086265563965\r\nStep 502, loss: 5.326898574829102\r\nStep 503, loss: 5.6919264793396\r\nStep 504, loss: 5.549910068511963\r\nStep 505, loss: 5.593076229095459\r\nStep 506, loss: 5.713366508483887\r\nStep 507, loss: 5.6414794921875\r\nStep 508, loss: 5.393517017364502\r\nStep 509, loss: 5.7733964920043945\r\nStep 510, loss: 5.574671745300293\r\nStep 511, loss: 5.704086780548096\r\nStep 482, loss: 5.577463626861572\r\nStep 483, loss: 5.537669658660889\r\nStep 484, loss: 5.941127300262451\r\nStep 485, loss: 5.490207195281982\r\nStep 486, loss: 5.543642044067383\r\nStep 487, loss: 5.644903659820557\r\nStep 488, loss: 5.602255821228027\r\nStep 489, loss: 5.7087721824646\r\nStep 490, loss: 5.508888244628906\r\nStep 491, loss: 5.632482528686523\r\nStep 492, loss: 5.39115571975708\r\nStep 493, loss: 5.692109107971191\r\nStep 494, loss: 5.630179405212402\r\nStep 495, loss: 5.478193759918213\r\nStep 496, loss: 5.638604164123535\r\nStep 497, loss: 5.54039192199707\r\nStep 498, loss: 5.653471946716309\r\nStep 499, loss: 5.654304027557373\r\nStep 500, loss: 5.648343086242676\r\nStep 501, loss: 5.6619086265563965\r\nStep 502, loss: 5.326898574829102\r\nStep 503, loss: 5.6919264793396\r\nStep 504, loss: 5.549910068511963\r\nStep 505, loss: 5.593076229095459\r\nStep 506, loss: 5.713366508483887\r\nStep 507, loss: 5.6414794921875\r\nStep 508, loss: 5.393517017364502\r\nStep 509, loss: 5.7733964920043945\r\nStep 510, loss: 5.574671745300293\r\nStep 511, loss: 5.704086780548096\r\nStep 512, loss: 5.621325492858887\r\nStep 513, loss: 5.49443244934082\r\nStep 514, loss: 5.439146518707275\r\nStep 515, loss: 5.49471378326416\r\nStep 516, loss: 5.5571980476379395\r\nStep 517, loss: 5.517596244812012\r\nStep 518, loss: 5.4848527908325195\r\nStep 519, loss: 5.647706031799316\r\nStep 520, loss: 5.3533124923706055\r\nStep 521, loss: 5.372676372528076\r\nStep 522, loss: 5.403843879699707\r\nStep 523, loss: 5.5930867195129395\r\nStep 524, loss: 5.593081474304199\r\nStep 525, loss: 5.538381099700928\r\nStep 526, loss: 5.677945613861084\r\nStep 527, loss: 5.529627323150635\r\nStep 528, loss: 5.611823558807373\r\nStep 529, loss: 5.472219944000244\r\nStep 530, loss: 5.792049884796143\r\nStep 531, loss: 5.482363224029541\r\nStep 532, loss: 5.5036163330078125\r\nStep 533, loss: 5.576632499694824\r\nStep 534, loss: 5.6589226722717285\r\nStep 535, loss: 5.598089694976807\r\nStep 536, loss: 5.496463298797607\r\nStep 537, loss: 5.403911590576172\r\nStep 538, loss: 5.450916290283203\r\nStep 539, loss: 5.476700782775879\r\nStep 540, loss: 5.646368503570557\r\nStep 512, loss: 5.621325492858887\r\nStep 513, loss: 5.49443244934082\r\nStep 514, loss: 5.439146518707275\r\nStep 515, loss: 5.49471378326416\r\nStep 516, loss: 5.5571980476379395\r\nStep 517, loss: 5.517596244812012\r\nStep 518, loss: 5.4848527908325195\r\nStep 519, loss: 5.647706031799316\r\nStep 520, loss: 5.3533124923706055\r\nStep 521, loss: 5.372676372528076\r\nStep 522, loss: 5.403843879699707\r\nStep 523, loss: 5.5930867195129395\r\nStep 524, loss: 5.593081474304199\r\nStep 525, loss: 5.538381099700928\r\nStep 526, loss: 5.677945613861084\r\nStep 527, loss: 5.529627323150635\r\nStep 528, loss: 5.611823558807373\r\nStep 529, loss: 5.472219944000244\r\nStep 530, loss: 5.792049884796143\r\nStep 531, loss: 5.482363224029541\r\nStep 532, loss: 5.5036163330078125\r\nStep 533, loss: 5.576632499694824\r\nStep 534, loss: 5.6589226722717285\r\nStep 535, loss: 5.598089694976807\r\nStep 536, loss: 5.496463298797607\r\nStep 537, loss: 5.403911590576172\r\nStep 538, loss: 5.450916290283203\r\nStep 539, loss: 5.476700782775879\r\nStep 540, loss: 5.646368503570557\r\nStep 541, loss: 5.7761688232421875\r\nStep 542, loss: 5.410217761993408\r\nStep 543, loss: 5.411986827850342\r\nStep 544, loss: 5.435133457183838\r\nStep 545, loss: 5.382740020751953\r\nStep 546, loss: 5.442299842834473\r\nStep 547, loss: 5.272629261016846\r\nStep 548, loss: 5.307929039001465\r\nStep 549, loss: 5.5719523429870605\r\nStep 550, loss: 5.59519100189209\r\nStep 551, loss: 5.573168754577637\r\nStep 552, loss: 5.326801776885986\r\nStep 553, loss: 5.4029083251953125\r\nStep 554, loss: 5.825585842132568\r\nStep 555, loss: 5.606141090393066\r\nStep 556, loss: 5.3618693351745605\r\nStep 557, loss: 5.3861494064331055\r\nStep 558, loss: 5.679725646972656\r\nStep 559, loss: 5.418980121612549\r\nStep 560, loss: 5.386108875274658\r\nStep 561, loss: 5.720038414001465\r\nStep 562, loss: 5.490231990814209\r\nStep 563, loss: 5.3749494552612305\r\nStep 564, loss: 5.465063571929932\r\nStep 565, loss: 5.612521648406982\r\nStep 566, loss: 5.725881099700928\r\nStep 567, loss: 5.359390735626221\r\nStep 568, loss: 5.486212253570557\r\nStep 569, loss: 5.615264415740967\r\nStep 541, loss: 5.7761688232421875\r\nStep 542, loss: 5.410217761993408\r\nStep 543, loss: 5.411986827850342\r\nStep 544, loss: 5.435133457183838\r\nStep 545, loss: 5.382740020751953\r\nStep 546, loss: 5.442299842834473\r\nStep 547, loss: 5.272629261016846\r\nStep 548, loss: 5.307929039001465\r\nStep 549, loss: 5.5719523429870605\r\nStep 550, loss: 5.59519100189209\r\nStep 551, loss: 5.573168754577637\r\nStep 552, loss: 5.326801776885986\r\nStep 553, loss: 5.4029083251953125\r\nStep 554, loss: 5.825585842132568\r\nStep 555, loss: 5.606141090393066\r\nStep 556, loss: 5.3618693351745605\r\nStep 557, loss: 5.3861494064331055\r\nStep 558, loss: 5.679725646972656\r\nStep 559, loss: 5.418980121612549\r\nStep 560, loss: 5.386108875274658\r\nStep 561, loss: 5.720038414001465\r\nStep 562, loss: 5.490231990814209\r\nStep 563, loss: 5.3749494552612305\r\nStep 564, loss: 5.465063571929932\r\nStep 565, loss: 5.612521648406982\r\nStep 566, loss: 5.725881099700928\r\nStep 567, loss: 5.359390735626221\r\nStep 568, loss: 5.486212253570557\r\nStep 569, loss: 5.615264415740967\r\nStep 570, loss: 5.597212791442871\r\nStep 571, loss: 5.368577003479004\r\nStep 572, loss: 5.364856243133545\r\nStep 573, loss: 5.5607733726501465\r\nStep 574, loss: 5.502553462982178\r\nStep 575, loss: 5.4582037925720215\r\nStep 576, loss: 5.623786449432373\r\nStep 577, loss: 5.269536972045898\r\nStep 578, loss: 5.4617390632629395\r\nStep 579, loss: 5.435409069061279\r\nStep 580, loss: 5.52940034866333\r\nStep 581, loss: 5.613592147827148\r\nStep 582, loss: 5.687880516052246\r\nStep 583, loss: 5.469274044036865\r\nStep 584, loss: 5.551769256591797\r\nStep 585, loss: 5.4086995124816895\r\nStep 586, loss: 5.383148670196533\r\nStep 587, loss: 5.516772747039795\r\nStep 588, loss: 5.366397857666016\r\nStep 589, loss: 5.504728317260742\r\nStep 590, loss: 5.628626823425293\r\nStep 591, loss: 5.551731109619141\r\nStep 592, loss: 5.3733110427856445\r\nStep 593, loss: 5.490641117095947\r\nStep 594, loss: 5.366167068481445\r\nStep 595, loss: 5.499856472015381\r\nStep 596, loss: 5.299220085144043\r\nStep 597, loss: 5.158448219299316\r\nStep 598, loss: 5.433690547943115\r\nStep 570, loss: 5.597212791442871\r\nStep 571, loss: 5.368577003479004\r\nStep 572, loss: 5.364856243133545\r\nStep 573, loss: 5.5607733726501465\r\nStep 574, loss: 5.502553462982178\r\nStep 575, loss: 5.4582037925720215\r\nStep 576, loss: 5.623786449432373\r\nStep 577, loss: 5.269536972045898\r\nStep 578, loss: 5.4617390632629395\r\nStep 579, loss: 5.435409069061279\r\nStep 580, loss: 5.52940034866333\r\nStep 581, loss: 5.613592147827148\r\nStep 582, loss: 5.687880516052246\r\nStep 583, loss: 5.469274044036865\r\nStep 584, loss: 5.551769256591797\r\nStep 585, loss: 5.4086995124816895\r\nStep 586, loss: 5.383148670196533\r\nStep 587, loss: 5.516772747039795\r\nStep 588, loss: 5.366397857666016\r\nStep 589, loss: 5.504728317260742\r\nStep 590, loss: 5.628626823425293\r\nStep 591, loss: 5.551731109619141\r\nStep 592, loss: 5.3733110427856445\r\nStep 593, loss: 5.490641117095947\r\nStep 594, loss: 5.366167068481445\r\nStep 595, loss: 5.499856472015381\r\nStep 596, loss: 5.299220085144043\r\nStep 597, loss: 5.158448219299316\r\nStep 598, loss: 5.433690547943115\r\nStep 599, loss: 5.329139232635498\r\nStep 600, loss: 5.557008266448975\r\nStep 601, loss: 5.331687927246094\r\nStep 599, loss: 5.329139232635498\r\nStep 600, loss: 5.557008266448975\r\nStep 601, loss: 5.331687927246094\r\nStep 482, loss: 5.577463626861572\r\nStep 483, loss: 5.537669658660889\r\nStep 484, loss: 5.941127300262451\r\nStep 485, loss: 5.490207195281982\r\nStep 486, loss: 5.543642044067383\r\nStep 487, loss: 5.644903659820557\r\nStep 488, loss: 5.602255821228027\r\nStep 489, loss: 5.7087721824646\r\nStep 490, loss: 5.508888244628906\r\nStep 491, loss: 5.632482528686523\r\nStep 492, loss: 5.39115571975708\r\nStep 493, loss: 5.692109107971191\r\nStep 494, loss: 5.630179405212402\r\nStep 495, loss: 5.478193759918213\r\nStep 496, loss: 5.638604164123535\r\nStep 497, loss: 5.54039192199707\r\nStep 498, loss: 5.653471946716309\r\nStep 499, loss: 5.654304027557373\r\nStep 500, loss: 5.648343086242676\r\nStep 501, loss: 5.6619086265563965\r\nStep 502, loss: 5.326898574829102\r\nStep 503, loss: 5.6919264793396\r\nStep 504, loss: 5.549910068511963\r\nStep 505, loss: 5.593076229095459\r\nStep 506, loss: 5.713366508483887\r\nStep 507, loss: 5.6414794921875\r\nStep 508, loss: 5.393517017364502\r\nStep 509, loss: 5.7733964920043945\r\nStep 510, loss: 5.574671745300293\r\nStep 511, loss: 5.704086780548096\r\nStep 482, loss: 5.577463626861572\r\nStep 483, loss: 5.537669658660889\r\nStep 484, loss: 5.941127300262451\r\nStep 485, loss: 5.490207195281982\r\nStep 486, loss: 5.543642044067383\r\nStep 487, loss: 5.644903659820557\r\nStep 488, loss: 5.602255821228027\r\nStep 489, loss: 5.7087721824646\r\nStep 490, loss: 5.508888244628906\r\nStep 491, loss: 5.632482528686523\r\nStep 492, loss: 5.39115571975708\r\nStep 493, loss: 5.692109107971191\r\nStep 494, loss: 5.630179405212402\r\nStep 495, loss: 5.478193759918213\r\nStep 496, loss: 5.638604164123535\r\nStep 497, loss: 5.54039192199707\r\nStep 498, loss: 5.653471946716309\r\nStep 499, loss: 5.654304027557373\r\nStep 500, loss: 5.648343086242676\r\nStep 501, loss: 5.6619086265563965\r\nStep 502, loss: 5.326898574829102\r\nStep 503, loss: 5.6919264793396\r\nStep 504, loss: 5.549910068511963\r\nStep 505, loss: 5.593076229095459\r\nStep 506, loss: 5.713366508483887\r\nStep 507, loss: 5.6414794921875\r\nStep 508, loss: 5.393517017364502\r\nStep 509, loss: 5.7733964920043945\r\nStep 510, loss: 5.574671745300293\r\nStep 511, loss: 5.704086780548096\r\nStep 512, loss: 5.621325492858887\r\nStep 513, loss: 5.49443244934082\r\nStep 514, loss: 5.439146518707275\r\nStep 515, loss: 5.49471378326416\r\nStep 516, loss: 5.5571980476379395\r\nStep 517, loss: 5.517596244812012\r\nStep 518, loss: 5.4848527908325195\r\nStep 519, loss: 5.647706031799316\r\nStep 520, loss: 5.3533124923706055\r\nStep 521, loss: 5.372676372528076\r\nStep 522, loss: 5.403843879699707\r\nStep 523, loss: 5.5930867195129395\r\nStep 524, loss: 5.593081474304199\r\nStep 525, loss: 5.538381099700928\r\nStep 526, loss: 5.677945613861084\r\nStep 527, loss: 5.529627323150635\r\nStep 528, loss: 5.611823558807373\r\nStep 529, loss: 5.472219944000244\r\nStep 530, loss: 5.792049884796143\r\nStep 531, loss: 5.482363224029541\r\nStep 532, loss: 5.5036163330078125\r\nStep 533, loss: 5.576632499694824\r\nStep 534, loss: 5.6589226722717285\r\nStep 535, loss: 5.598089694976807\r\nStep 536, loss: 5.496463298797607\r\nStep 537, loss: 5.403911590576172\r\nStep 538, loss: 5.450916290283203\r\nStep 539, loss: 5.476700782775879\r\nStep 540, loss: 5.646368503570557\r\nStep 512, loss: 5.621325492858887\r\nStep 513, loss: 5.49443244934082\r\nStep 514, loss: 5.439146518707275\r\nStep 515, loss: 5.49471378326416\r\nStep 516, loss: 5.5571980476379395\r\nStep 517, loss: 5.517596244812012\r\nStep 518, loss: 5.4848527908325195\r\nStep 519, loss: 5.647706031799316\r\nStep 520, loss: 5.3533124923706055\r\nStep 521, loss: 5.372676372528076\r\nStep 522, loss: 5.403843879699707\r\nStep 523, loss: 5.5930867195129395\r\nStep 524, loss: 5.593081474304199\r\nStep 525, loss: 5.538381099700928\r\nStep 526, loss: 5.677945613861084\r\nStep 527, loss: 5.529627323150635\r\nStep 528, loss: 5.611823558807373\r\nStep 529, loss: 5.472219944000244\r\nStep 530, loss: 5.792049884796143\r\nStep 531, loss: 5.482363224029541\r\nStep 532, loss: 5.5036163330078125\r\nStep 533, loss: 5.576632499694824\r\nStep 534, loss: 5.6589226722717285\r\nStep 535, loss: 5.598089694976807\r\nStep 536, loss: 5.496463298797607\r\nStep 537, loss: 5.403911590576172\r\nStep 538, loss: 5.450916290283203\r\nStep 539, loss: 5.476700782775879\r\nStep 540, loss: 5.646368503570557\r\nStep 541, loss: 5.7761688232421875\r\nStep 542, loss: 5.410217761993408\r\nStep 543, loss: 5.411986827850342\r\nStep 544, loss: 5.435133457183838\r\nStep 545, loss: 5.382740020751953\r\nStep 546, loss: 5.442299842834473\r\nStep 547, loss: 5.272629261016846\r\nStep 548, loss: 5.307929039001465\r\nStep 549, loss: 5.5719523429870605\r\nStep 550, loss: 5.59519100189209\r\nStep 551, loss: 5.573168754577637\r\nStep 552, loss: 5.326801776885986\r\nStep 553, loss: 5.4029083251953125\r\nStep 554, loss: 5.825585842132568\r\nStep 555, loss: 5.606141090393066\r\nStep 556, loss: 5.3618693351745605\r\nStep 557, loss: 5.3861494064331055\r\nStep 558, loss: 5.679725646972656\r\nStep 559, loss: 5.418980121612549\r\nStep 560, loss: 5.386108875274658\r\nStep 561, loss: 5.720038414001465\r\nStep 562, loss: 5.490231990814209\r\nStep 563, loss: 5.3749494552612305\r\nStep 564, loss: 5.465063571929932\r\nStep 565, loss: 5.612521648406982\r\nStep 566, loss: 5.725881099700928\r\nStep 567, loss: 5.359390735626221\r\nStep 568, loss: 5.486212253570557\r\nStep 569, loss: 5.615264415740967\r\nStep 541, loss: 5.7761688232421875\r\nStep 542, loss: 5.410217761993408\r\nStep 543, loss: 5.411986827850342\r\nStep 544, loss: 5.435133457183838\r\nStep 545, loss: 5.382740020751953\r\nStep 546, loss: 5.442299842834473\r\nStep 547, loss: 5.272629261016846\r\nStep 548, loss: 5.307929039001465\r\nStep 549, loss: 5.5719523429870605\r\nStep 550, loss: 5.59519100189209\r\nStep 551, loss: 5.573168754577637\r\nStep 552, loss: 5.326801776885986\r\nStep 553, loss: 5.4029083251953125\r\nStep 554, loss: 5.825585842132568\r\nStep 555, loss: 5.606141090393066\r\nStep 556, loss: 5.3618693351745605\r\nStep 557, loss: 5.3861494064331055\r\nStep 558, loss: 5.679725646972656\r\nStep 559, loss: 5.418980121612549\r\nStep 560, loss: 5.386108875274658\r\nStep 561, loss: 5.720038414001465\r\nStep 562, loss: 5.490231990814209\r\nStep 563, loss: 5.3749494552612305\r\nStep 564, loss: 5.465063571929932\r\nStep 565, loss: 5.612521648406982\r\nStep 566, loss: 5.725881099700928\r\nStep 567, loss: 5.359390735626221\r\nStep 568, loss: 5.486212253570557\r\nStep 569, loss: 5.615264415740967\r\nStep 570, loss: 5.597212791442871\r\nStep 571, loss: 5.368577003479004\r\nStep 572, loss: 5.364856243133545\r\nStep 573, loss: 5.5607733726501465\r\nStep 574, loss: 5.502553462982178\r\nStep 575, loss: 5.4582037925720215\r\nStep 576, loss: 5.623786449432373\r\nStep 577, loss: 5.269536972045898\r\nStep 578, loss: 5.4617390632629395\r\nStep 579, loss: 5.435409069061279\r\nStep 580, loss: 5.52940034866333\r\nStep 581, loss: 5.613592147827148\r\nStep 582, loss: 5.687880516052246\r\nStep 583, loss: 5.469274044036865\r\nStep 584, loss: 5.551769256591797\r\nStep 585, loss: 5.4086995124816895\r\nStep 586, loss: 5.383148670196533\r\nStep 587, loss: 5.516772747039795\r\nStep 588, loss: 5.366397857666016\r\nStep 589, loss: 5.504728317260742\r\nStep 590, loss: 5.628626823425293\r\nStep 591, loss: 5.551731109619141\r\nStep 592, loss: 5.3733110427856445\r\nStep 593, loss: 5.490641117095947\r\nStep 594, loss: 5.366167068481445\r\nStep 595, loss: 5.499856472015381\r\nStep 596, loss: 5.299220085144043\r\nStep 597, loss: 5.158448219299316\r\nStep 598, loss: 5.433690547943115\r\nStep 570, loss: 5.597212791442871\r\nStep 571, loss: 5.368577003479004\r\nStep 572, loss: 5.364856243133545\r\nStep 573, loss: 5.5607733726501465\r\nStep 574, loss: 5.502553462982178\r\nStep 575, loss: 5.4582037925720215\r\nStep 576, loss: 5.623786449432373\r\nStep 577, loss: 5.269536972045898\r\nStep 578, loss: 5.4617390632629395\r\nStep 579, loss: 5.435409069061279\r\nStep 580, loss: 5.52940034866333\r\nStep 581, loss: 5.613592147827148\r\nStep 582, loss: 5.687880516052246\r\nStep 583, loss: 5.469274044036865\r\nStep 584, loss: 5.551769256591797\r\nStep 585, loss: 5.4086995124816895\r\nStep 586, loss: 5.383148670196533\r\nStep 587, loss: 5.516772747039795\r\nStep 588, loss: 5.366397857666016\r\nStep 589, loss: 5.504728317260742\r\nStep 590, loss: 5.628626823425293\r\nStep 591, loss: 5.551731109619141\r\nStep 592, loss: 5.3733110427856445\r\nStep 593, loss: 5.490641117095947\r\nStep 594, loss: 5.366167068481445\r\nStep 595, loss: 5.499856472015381\r\nStep 596, loss: 5.299220085144043\r\nStep 597, loss: 5.158448219299316\r\nStep 598, loss: 5.433690547943115\r\nStep 599, loss: 5.329139232635498\r\nStep 600, loss: 5.557008266448975\r\nStep 601, loss: 5.331687927246094\r\nStep 599, loss: 5.329139232635498\r\nStep 600, loss: 5.557008266448975\r\nStep 601, loss: 5.331687927246094\r\nStep 602, loss: 5.6626973152160645\r\nStep 603, loss: 5.384581089019775\r\nStep 604, loss: 5.374907970428467\r\nStep 605, loss: 5.254096984863281\r\nStep 606, loss: 5.541470527648926\r\nStep 607, loss: 5.296286106109619\r\nStep 608, loss: 5.154754638671875\r\nStep 609, loss: 5.358010292053223\r\nStep 610, loss: 5.432250022888184\r\nStep 611, loss: 5.315914630889893\r\nStep 612, loss: 5.464038848876953\r\nStep 613, loss: 5.403680324554443\r\nStep 614, loss: 5.4701828956604\r\nStep 615, loss: 5.374019622802734\r\nStep 616, loss: 5.368079662322998\r\nStep 617, loss: 5.493638515472412\r\nStep 618, loss: 5.4959821701049805\r\nStep 619, loss: 5.174816608428955\r\nStep 620, loss: 5.175816059112549\r\nStep 621, loss: 5.408751964569092\r\nStep 622, loss: 5.620441913604736\r\nStep 623, loss: 5.198512554168701\r\nStep 624, loss: 5.317488670349121\r\nStep 625, loss: 5.265618801116943\r\nStep 626, loss: 5.509090900421143\r\nStep 627, loss: 5.488482475280762\r\nStep 628, loss: 5.2329254150390625\r\nStep 629, loss: 5.395798206329346\r\nStep 630, loss: 5.485594749450684\r\nStep 631, loss: 5.229202747344971\r\nStep 602, loss: 5.6626973152160645\r\nStep 603, loss: 5.384581089019775\r\nStep 604, loss: 5.374907970428467\r\nStep 605, loss: 5.254096984863281\r\nStep 606, loss: 5.541470527648926\r\nStep 607, loss: 5.296286106109619\r\nStep 608, loss: 5.154754638671875\r\nStep 609, loss: 5.358010292053223\r\nStep 610, loss: 5.432250022888184\r\nStep 611, loss: 5.315914630889893\r\nStep 612, loss: 5.464038848876953\r\nStep 613, loss: 5.403680324554443\r\nStep 614, loss: 5.4701828956604\r\nStep 615, loss: 5.374019622802734\r\nStep 616, loss: 5.368079662322998\r\nStep 617, loss: 5.493638515472412\r\nStep 618, loss: 5.4959821701049805\r\nStep 619, loss: 5.174816608428955\r\nStep 620, loss: 5.175816059112549\r\nStep 621, loss: 5.408751964569092\r\nStep 622, loss: 5.620441913604736\r\nStep 623, loss: 5.198512554168701\r\nStep 624, loss: 5.317488670349121\r\nStep 625, loss: 5.265618801116943\r\nStep 626, loss: 5.509090900421143\r\nStep 627, loss: 5.488482475280762\r\nStep 628, loss: 5.2329254150390625\r\nStep 629, loss: 5.395798206329346\r\nStep 630, loss: 5.485594749450684\r\nStep 631, loss: 5.229202747344971\r\nStep 632, loss: 5.387337684631348\r\nStep 633, loss: 5.488489151000977\r\nStep 634, loss: 5.307542324066162\r\nStep 635, loss: 5.517372131347656\r\nStep 636, loss: 5.626887798309326\r\nStep 637, loss: 5.219997882843018\r\nStep 638, loss: 5.400931358337402\r\nStep 639, loss: 5.269510269165039\r\nStep 640, loss: 5.465200424194336\r\nStep 641, loss: 5.52509069442749\r\nStep 642, loss: 5.433084964752197\r\nStep 643, loss: 5.42751932144165\r\nStep 644, loss: 5.291433334350586\r\nStep 645, loss: 5.4225239753723145\r\nStep 646, loss: 5.282150745391846\r\nStep 647, loss: 5.644361972808838\r\nStep 648, loss: 5.237563610076904\r\nStep 649, loss: 5.553693771362305\r\nStep 650, loss: 5.0682220458984375\r\nStep 651, loss: 5.371950149536133\r\nStep 652, loss: 5.409918308258057\r\nStep 653, loss: 5.362102508544922\r\nStep 654, loss: 5.344768047332764\r\nStep 655, loss: 5.31875467300415\r\nStep 656, loss: 5.197376728057861\r\nStep 657, loss: 5.290582656860352\r\nStep 658, loss: 5.126527786254883\r\nStep 659, loss: 5.216240882873535\r\nStep 660, loss: 5.198899269104004\r\nStep 661, loss: 5.200804710388184\r\nStep 632, loss: 5.387337684631348\r\nStep 633, loss: 5.488489151000977\r\nStep 634, loss: 5.307542324066162\r\nStep 635, loss: 5.517372131347656\r\nStep 636, loss: 5.626887798309326\r\nStep 637, loss: 5.219997882843018\r\nStep 638, loss: 5.400931358337402\r\nStep 639, loss: 5.269510269165039\r\nStep 640, loss: 5.465200424194336\r\nStep 641, loss: 5.52509069442749\r\nStep 642, loss: 5.433084964752197\r\nStep 643, loss: 5.42751932144165\r\nStep 644, loss: 5.291433334350586\r\nStep 645, loss: 5.4225239753723145\r\nStep 646, loss: 5.282150745391846\r\nStep 647, loss: 5.644361972808838\r\nStep 648, loss: 5.237563610076904\r\nStep 649, loss: 5.553693771362305\r\nStep 650, loss: 5.0682220458984375\r\nStep 651, loss: 5.371950149536133\r\nStep 652, loss: 5.409918308258057\r\nStep 653, loss: 5.362102508544922\r\nStep 654, loss: 5.344768047332764\r\nStep 655, loss: 5.31875467300415\r\nStep 656, loss: 5.197376728057861\r\nStep 657, loss: 5.290582656860352\r\nStep 658, loss: 5.126527786254883\r\nStep 659, loss: 5.216240882873535\r\nStep 660, loss: 5.198899269104004\r\nStep 661, loss: 5.200804710388184\r\nStep 662, loss: 5.3883256912231445\r\nStep 663, loss: 5.612620830535889\r\nStep 664, loss: 5.255474090576172\r\nStep 665, loss: 5.145941734313965\r\nStep 666, loss: 5.431971549987793\r\nStep 667, loss: 5.3477983474731445\r\nStep 668, loss: 5.1120734214782715\r\nStep 669, loss: 5.181546211242676\r\nStep 670, loss: 5.1419501304626465\r\nStep 671, loss: 5.471360206604004\r\nStep 672, loss: 5.205795764923096\r\nStep 673, loss: 5.430771350860596\r\nStep 674, loss: 5.19976282119751\r\nStep 675, loss: 5.763006210327148\r\nStep 676, loss: 5.463747024536133\r\nStep 677, loss: 5.308181285858154\r\nStep 678, loss: 5.46526575088501\r\nStep 679, loss: 5.533337116241455\r\nStep 680, loss: 5.622751235961914\r\nStep 681, loss: 5.210545063018799\r\nStep 682, loss: 5.1722002029418945\r\nStep 683, loss: 5.419670581817627\r\nStep 684, loss: 5.494363307952881\r\nStep 685, loss: 5.412539005279541\r\nStep 686, loss: 5.154549598693848\r\nStep 687, loss: 5.436656475067139\r\nStep 688, loss: 5.432368278503418\r\nStep 689, loss: 5.210556507110596\r\nStep 690, loss: 5.408603668212891\r\nStep 691, loss: 5.160996913909912\r\nStep 662, loss: 5.3883256912231445\r\nStep 663, loss: 5.612620830535889\r\nStep 664, loss: 5.255474090576172\r\nStep 665, loss: 5.145941734313965\r\nStep 666, loss: 5.431971549987793\r\nStep 667, loss: 5.3477983474731445\r\nStep 668, loss: 5.1120734214782715\r\nStep 669, loss: 5.181546211242676\r\nStep 670, loss: 5.1419501304626465\r\nStep 671, loss: 5.471360206604004\r\nStep 672, loss: 5.205795764923096\r\nStep 673, loss: 5.430771350860596\r\nStep 674, loss: 5.19976282119751\r\nStep 675, loss: 5.763006210327148\r\nStep 676, loss: 5.463747024536133\r\nStep 677, loss: 5.308181285858154\r\nStep 678, loss: 5.46526575088501\r\nStep 679, loss: 5.533337116241455\r\nStep 680, loss: 5.622751235961914\r\nStep 681, loss: 5.210545063018799\r\nStep 682, loss: 5.1722002029418945\r\nStep 683, loss: 5.419670581817627\r\nStep 684, loss: 5.494363307952881\r\nStep 685, loss: 5.412539005279541\r\nStep 686, loss: 5.154549598693848\r\nStep 687, loss: 5.436656475067139\r\nStep 688, loss: 5.432368278503418\r\nStep 689, loss: 5.210556507110596\r\nStep 690, loss: 5.408603668212891\r\nStep 691, loss: 5.160996913909912\r\nStep 692, loss: 5.424027919769287\r\nStep 693, loss: 5.170679092407227\r\nStep 694, loss: 5.356256008148193\r\nStep 695, loss: 5.472571849822998\r\nStep 696, loss: 5.417373180389404\r\nStep 697, loss: 5.388498783111572\r\nStep 698, loss: 5.262965679168701\r\nStep 699, loss: 5.498363494873047\r\nStep 700, loss: 5.313055038452148\r\nStep 701, loss: 5.241844177246094\r\nStep 702, loss: 5.448908805847168\r\nStep 703, loss: 5.337677955627441\r\nStep 704, loss: 5.359001636505127\r\nStep 705, loss: 5.145549774169922\r\nStep 706, loss: 5.362729549407959\r\nStep 707, loss: 5.109035491943359\r\nStep 708, loss: 5.128541946411133\r\nStep 709, loss: 5.284796237945557\r\nStep 710, loss: 5.48291540145874\r\nStep 711, loss: 5.406717777252197\r\nStep 712, loss: 5.267970085144043\r\nStep 713, loss: 5.198863506317139\r\nStep 714, loss: 5.358706474304199\r\nStep 715, loss: 5.192930221557617\r\nStep 716, loss: 5.401691913604736\r\nStep 717, loss: 5.32148551940918\r\nStep 718, loss: 5.543673515319824\r\nStep 719, loss: 5.136351585388184\r\nStep 720, loss: 5.422923564910889\r\nStep 721, loss: 5.2997212409973145\r\nStep 692, loss: 5.424027919769287\r\nStep 693, loss: 5.170679092407227\r\nStep 694, loss: 5.356256008148193\r\nStep 695, loss: 5.472571849822998\r\nStep 696, loss: 5.417373180389404\r\nStep 697, loss: 5.388498783111572\r\nStep 698, loss: 5.262965679168701\r\nStep 699, loss: 5.498363494873047\r\nStep 700, loss: 5.313055038452148\r\nStep 701, loss: 5.241844177246094\r\nStep 702, loss: 5.448908805847168\r\nStep 703, loss: 5.337677955627441\r\nStep 704, loss: 5.359001636505127\r\nStep 705, loss: 5.145549774169922\r\nStep 706, loss: 5.362729549407959\r\nStep 707, loss: 5.109035491943359\r\nStep 708, loss: 5.128541946411133\r\nStep 709, loss: 5.284796237945557\r\nStep 710, loss: 5.48291540145874\r\nStep 711, loss: 5.406717777252197\r\nStep 712, loss: 5.267970085144043\r\nStep 713, loss: 5.198863506317139\r\nStep 714, loss: 5.358706474304199\r\nStep 715, loss: 5.192930221557617\r\nStep 716, loss: 5.401691913604736\r\nStep 717, loss: 5.32148551940918\r\nStep 718, loss: 5.543673515319824\r\nStep 719, loss: 5.136351585388184\r\nStep 720, loss: 5.422923564910889\r\nStep 721, loss: 5.2997212409973145\r\nStep 602, loss: 5.6626973152160645\r\nStep 603, loss: 5.384581089019775\r\nStep 604, loss: 5.374907970428467\r\nStep 605, loss: 5.254096984863281\r\nStep 606, loss: 5.541470527648926\r\nStep 607, loss: 5.296286106109619\r\nStep 608, loss: 5.154754638671875\r\nStep 609, loss: 5.358010292053223\r\nStep 610, loss: 5.432250022888184\r\nStep 611, loss: 5.315914630889893\r\nStep 612, loss: 5.464038848876953\r\nStep 613, loss: 5.403680324554443\r\nStep 614, loss: 5.4701828956604\r\nStep 615, loss: 5.374019622802734\r\nStep 616, loss: 5.368079662322998\r\nStep 617, loss: 5.493638515472412\r\nStep 618, loss: 5.4959821701049805\r\nStep 619, loss: 5.174816608428955\r\nStep 620, loss: 5.175816059112549\r\nStep 621, loss: 5.408751964569092\r\nStep 622, loss: 5.620441913604736\r\nStep 623, loss: 5.198512554168701\r\nStep 624, loss: 5.317488670349121\r\nStep 625, loss: 5.265618801116943\r\nStep 626, loss: 5.509090900421143\r\nStep 627, loss: 5.488482475280762\r\nStep 628, loss: 5.2329254150390625\r\nStep 629, loss: 5.395798206329346\r\nStep 630, loss: 5.485594749450684\r\nStep 631, loss: 5.229202747344971\r\nStep 602, loss: 5.6626973152160645\r\nStep 603, loss: 5.384581089019775\r\nStep 604, loss: 5.374907970428467\r\nStep 605, loss: 5.254096984863281\r\nStep 606, loss: 5.541470527648926\r\nStep 607, loss: 5.296286106109619\r\nStep 608, loss: 5.154754638671875\r\nStep 609, loss: 5.358010292053223\r\nStep 610, loss: 5.432250022888184\r\nStep 611, loss: 5.315914630889893\r\nStep 612, loss: 5.464038848876953\r\nStep 613, loss: 5.403680324554443\r\nStep 614, loss: 5.4701828956604\r\nStep 615, loss: 5.374019622802734\r\nStep 616, loss: 5.368079662322998\r\nStep 617, loss: 5.493638515472412\r\nStep 618, loss: 5.4959821701049805\r\nStep 619, loss: 5.174816608428955\r\nStep 620, loss: 5.175816059112549\r\nStep 621, loss: 5.408751964569092\r\nStep 622, loss: 5.620441913604736\r\nStep 623, loss: 5.198512554168701\r\nStep 624, loss: 5.317488670349121\r\nStep 625, loss: 5.265618801116943\r\nStep 626, loss: 5.509090900421143\r\nStep 627, loss: 5.488482475280762\r\nStep 628, loss: 5.2329254150390625\r\nStep 629, loss: 5.395798206329346\r\nStep 630, loss: 5.485594749450684\r\nStep 631, loss: 5.229202747344971\r\nStep 632, loss: 5.387337684631348\r\nStep 633, loss: 5.488489151000977\r\nStep 634, loss: 5.307542324066162\r\nStep 635, loss: 5.517372131347656\r\nStep 636, loss: 5.626887798309326\r\nStep 637, loss: 5.219997882843018\r\nStep 638, loss: 5.400931358337402\r\nStep 639, loss: 5.269510269165039\r\nStep 640, loss: 5.465200424194336\r\nStep 641, loss: 5.52509069442749\r\nStep 642, loss: 5.433084964752197\r\nStep 643, loss: 5.42751932144165\r\nStep 644, loss: 5.291433334350586\r\nStep 645, loss: 5.4225239753723145\r\nStep 646, loss: 5.282150745391846\r\nStep 647, loss: 5.644361972808838\r\nStep 648, loss: 5.237563610076904\r\nStep 649, loss: 5.553693771362305\r\nStep 650, loss: 5.0682220458984375\r\nStep 651, loss: 5.371950149536133\r\nStep 652, loss: 5.409918308258057\r\nStep 653, loss: 5.362102508544922\r\nStep 654, loss: 5.344768047332764\r\nStep 655, loss: 5.31875467300415\r\nStep 656, loss: 5.197376728057861\r\nStep 657, loss: 5.290582656860352\r\nStep 658, loss: 5.126527786254883\r\nStep 659, loss: 5.216240882873535\r\nStep 660, loss: 5.198899269104004\r\nStep 661, loss: 5.200804710388184\r\nStep 632, loss: 5.387337684631348\r\nStep 633, loss: 5.488489151000977\r\nStep 634, loss: 5.307542324066162\r\nStep 635, loss: 5.517372131347656\r\nStep 636, loss: 5.626887798309326\r\nStep 637, loss: 5.219997882843018\r\nStep 638, loss: 5.400931358337402\r\nStep 639, loss: 5.269510269165039\r\nStep 640, loss: 5.465200424194336\r\nStep 641, loss: 5.52509069442749\r\nStep 642, loss: 5.433084964752197\r\nStep 643, loss: 5.42751932144165\r\nStep 644, loss: 5.291433334350586\r\nStep 645, loss: 5.4225239753723145\r\nStep 646, loss: 5.282150745391846\r\nStep 647, loss: 5.644361972808838\r\nStep 648, loss: 5.237563610076904\r\nStep 649, loss: 5.553693771362305\r\nStep 650, loss: 5.0682220458984375\r\nStep 651, loss: 5.371950149536133\r\nStep 652, loss: 5.409918308258057\r\nStep 653, loss: 5.362102508544922\r\nStep 654, loss: 5.344768047332764\r\nStep 655, loss: 5.31875467300415\r\nStep 656, loss: 5.197376728057861\r\nStep 657, loss: 5.290582656860352\r\nStep 658, loss: 5.126527786254883\r\nStep 659, loss: 5.216240882873535\r\nStep 660, loss: 5.198899269104004\r\nStep 661, loss: 5.200804710388184\r\nStep 662, loss: 5.3883256912231445\r\nStep 663, loss: 5.612620830535889\r\nStep 664, loss: 5.255474090576172\r\nStep 665, loss: 5.145941734313965\r\nStep 666, loss: 5.431971549987793\r\nStep 667, loss: 5.3477983474731445\r\nStep 668, loss: 5.1120734214782715\r\nStep 669, loss: 5.181546211242676\r\nStep 670, loss: 5.1419501304626465\r\nStep 671, loss: 5.471360206604004\r\nStep 672, loss: 5.205795764923096\r\nStep 673, loss: 5.430771350860596\r\nStep 674, loss: 5.19976282119751\r\nStep 675, loss: 5.763006210327148\r\nStep 676, loss: 5.463747024536133\r\nStep 677, loss: 5.308181285858154\r\nStep 678, loss: 5.46526575088501\r\nStep 679, loss: 5.533337116241455\r\nStep 680, loss: 5.622751235961914\r\nStep 681, loss: 5.210545063018799\r\nStep 682, loss: 5.1722002029418945\r\nStep 683, loss: 5.419670581817627\r\nStep 684, loss: 5.494363307952881\r\nStep 685, loss: 5.412539005279541\r\nStep 686, loss: 5.154549598693848\r\nStep 687, loss: 5.436656475067139\r\nStep 688, loss: 5.432368278503418\r\nStep 689, loss: 5.210556507110596\r\nStep 690, loss: 5.408603668212891\r\nStep 691, loss: 5.160996913909912\r\nStep 662, loss: 5.3883256912231445\r\nStep 663, loss: 5.612620830535889\r\nStep 664, loss: 5.255474090576172\r\nStep 665, loss: 5.145941734313965\r\nStep 666, loss: 5.431971549987793\r\nStep 667, loss: 5.3477983474731445\r\nStep 668, loss: 5.1120734214782715\r\nStep 669, loss: 5.181546211242676\r\nStep 670, loss: 5.1419501304626465\r\nStep 671, loss: 5.471360206604004\r\nStep 672, loss: 5.205795764923096\r\nStep 673, loss: 5.430771350860596\r\nStep 674, loss: 5.19976282119751\r\nStep 675, loss: 5.763006210327148\r\nStep 676, loss: 5.463747024536133\r\nStep 677, loss: 5.308181285858154\r\nStep 678, loss: 5.46526575088501\r\nStep 679, loss: 5.533337116241455\r\nStep 680, loss: 5.622751235961914\r\nStep 681, loss: 5.210545063018799\r\nStep 682, loss: 5.1722002029418945\r\nStep 683, loss: 5.419670581817627\r\nStep 684, loss: 5.494363307952881\r\nStep 685, loss: 5.412539005279541\r\nStep 686, loss: 5.154549598693848\r\nStep 687, loss: 5.436656475067139\r\nStep 688, loss: 5.432368278503418\r\nStep 689, loss: 5.210556507110596\r\nStep 690, loss: 5.408603668212891\r\nStep 691, loss: 5.160996913909912\r\nStep 692, loss: 5.424027919769287\r\nStep 693, loss: 5.170679092407227\r\nStep 694, loss: 5.356256008148193\r\nStep 695, loss: 5.472571849822998\r\nStep 696, loss: 5.417373180389404\r\nStep 697, loss: 5.388498783111572\r\nStep 698, loss: 5.262965679168701\r\nStep 699, loss: 5.498363494873047\r\nStep 700, loss: 5.313055038452148\r\nStep 701, loss: 5.241844177246094\r\nStep 702, loss: 5.448908805847168\r\nStep 703, loss: 5.337677955627441\r\nStep 704, loss: 5.359001636505127\r\nStep 705, loss: 5.145549774169922\r\nStep 706, loss: 5.362729549407959\r\nStep 707, loss: 5.109035491943359\r\nStep 708, loss: 5.128541946411133\r\nStep 709, loss: 5.284796237945557\r\nStep 710, loss: 5.48291540145874\r\nStep 711, loss: 5.406717777252197\r\nStep 712, loss: 5.267970085144043\r\nStep 713, loss: 5.198863506317139\r\nStep 714, loss: 5.358706474304199\r\nStep 715, loss: 5.192930221557617\r\nStep 716, loss: 5.401691913604736\r\nStep 717, loss: 5.32148551940918\r\nStep 718, loss: 5.543673515319824\r\nStep 719, loss: 5.136351585388184\r\nStep 720, loss: 5.422923564910889\r\nStep 721, loss: 5.2997212409973145\r\nStep 692, loss: 5.424027919769287\r\nStep 693, loss: 5.170679092407227\r\nStep 694, loss: 5.356256008148193\r\nStep 695, loss: 5.472571849822998\r\nStep 696, loss: 5.417373180389404\r\nStep 697, loss: 5.388498783111572\r\nStep 698, loss: 5.262965679168701\r\nStep 699, loss: 5.498363494873047\r\nStep 700, loss: 5.313055038452148\r\nStep 701, loss: 5.241844177246094\r\nStep 702, loss: 5.448908805847168\r\nStep 703, loss: 5.337677955627441\r\nStep 704, loss: 5.359001636505127\r\nStep 705, loss: 5.145549774169922\r\nStep 706, loss: 5.362729549407959\r\nStep 707, loss: 5.109035491943359\r\nStep 708, loss: 5.128541946411133\r\nStep 709, loss: 5.284796237945557\r\nStep 710, loss: 5.48291540145874\r\nStep 711, loss: 5.406717777252197\r\nStep 712, loss: 5.267970085144043\r\nStep 713, loss: 5.198863506317139\r\nStep 714, loss: 5.358706474304199\r\nStep 715, loss: 5.192930221557617\r\nStep 716, loss: 5.401691913604736\r\nStep 717, loss: 5.32148551940918\r\nStep 718, loss: 5.543673515319824\r\nStep 719, loss: 5.136351585388184\r\nStep 720, loss: 5.422923564910889\r\nStep 721, loss: 5.2997212409973145\r\nStep 482, loss: 5.577463626861572\r\nStep 483, loss: 5.537669658660889\r\nStep 484, loss: 5.941127300262451\r\nStep 485, loss: 5.490207195281982\r\nStep 486, loss: 5.543642044067383\r\nStep 487, loss: 5.644903659820557\r\nStep 488, loss: 5.602255821228027\r\nStep 489, loss: 5.7087721824646\r\nStep 490, loss: 5.508888244628906\r\nStep 491, loss: 5.632482528686523\r\nStep 492, loss: 5.39115571975708\r\nStep 493, loss: 5.692109107971191\r\nStep 494, loss: 5.630179405212402\r\nStep 495, loss: 5.478193759918213\r\nStep 496, loss: 5.638604164123535\r\nStep 497, loss: 5.54039192199707\r\nStep 498, loss: 5.653471946716309\r\nStep 499, loss: 5.654304027557373\r\nStep 500, loss: 5.648343086242676\r\nStep 501, loss: 5.6619086265563965\r\nStep 502, loss: 5.326898574829102\r\nStep 503, loss: 5.6919264793396\r\nStep 504, loss: 5.549910068511963\r\nStep 505, loss: 5.593076229095459\r\nStep 506, loss: 5.713366508483887\r\nStep 507, loss: 5.6414794921875\r\nStep 508, loss: 5.393517017364502\r\nStep 509, loss: 5.7733964920043945\r\nStep 510, loss: 5.574671745300293\r\nStep 511, loss: 5.704086780548096\r\nStep 482, loss: 5.577463626861572\r\nStep 483, loss: 5.537669658660889\r\nStep 484, loss: 5.941127300262451\r\nStep 485, loss: 5.490207195281982\r\nStep 486, loss: 5.543642044067383\r\nStep 487, loss: 5.644903659820557\r\nStep 488, loss: 5.602255821228027\r\nStep 489, loss: 5.7087721824646\r\nStep 490, loss: 5.508888244628906\r\nStep 491, loss: 5.632482528686523\r\nStep 492, loss: 5.39115571975708\r\nStep 493, loss: 5.692109107971191\r\nStep 494, loss: 5.630179405212402\r\nStep 495, loss: 5.478193759918213\r\nStep 496, loss: 5.638604164123535\r\nStep 497, loss: 5.54039192199707\r\nStep 498, loss: 5.653471946716309\r\nStep 499, loss: 5.654304027557373\r\nStep 500, loss: 5.648343086242676\r\nStep 501, loss: 5.6619086265563965\r\nStep 502, loss: 5.326898574829102\r\nStep 503, loss: 5.6919264793396\r\nStep 504, loss: 5.549910068511963\r\nStep 505, loss: 5.593076229095459\r\nStep 506, loss: 5.713366508483887\r\nStep 507, loss: 5.6414794921875\r\nStep 508, loss: 5.393517017364502\r\nStep 509, loss: 5.7733964920043945\r\nStep 510, loss: 5.574671745300293\r\nStep 511, loss: 5.704086780548096\r\nStep 512, loss: 5.621325492858887\r\nStep 513, loss: 5.49443244934082\r\nStep 514, loss: 5.439146518707275\r\nStep 515, loss: 5.49471378326416\r\nStep 516, loss: 5.5571980476379395\r\nStep 517, loss: 5.517596244812012\r\nStep 518, loss: 5.4848527908325195\r\nStep 519, loss: 5.647706031799316\r\nStep 520, loss: 5.3533124923706055\r\nStep 521, loss: 5.372676372528076\r\nStep 522, loss: 5.403843879699707\r\nStep 523, loss: 5.5930867195129395\r\nStep 524, loss: 5.593081474304199\r\nStep 525, loss: 5.538381099700928\r\nStep 526, loss: 5.677945613861084\r\nStep 527, loss: 5.529627323150635\r\nStep 528, loss: 5.611823558807373\r\nStep 529, loss: 5.472219944000244\r\nStep 530, loss: 5.792049884796143\r\nStep 531, loss: 5.482363224029541\r\nStep 532, loss: 5.5036163330078125\r\nStep 533, loss: 5.576632499694824\r\nStep 534, loss: 5.6589226722717285\r\nStep 535, loss: 5.598089694976807\r\nStep 536, loss: 5.496463298797607\r\nStep 537, loss: 5.403911590576172\r\nStep 538, loss: 5.450916290283203\r\nStep 539, loss: 5.476700782775879\r\nStep 540, loss: 5.646368503570557\r\nStep 512, loss: 5.621325492858887\r\nStep 513, loss: 5.49443244934082\r\nStep 514, loss: 5.439146518707275\r\nStep 515, loss: 5.49471378326416\r\nStep 516, loss: 5.5571980476379395\r\nStep 517, loss: 5.517596244812012\r\nStep 518, loss: 5.4848527908325195\r\nStep 519, loss: 5.647706031799316\r\nStep 520, loss: 5.3533124923706055\r\nStep 521, loss: 5.372676372528076\r\nStep 522, loss: 5.403843879699707\r\nStep 523, loss: 5.5930867195129395\r\nStep 524, loss: 5.593081474304199\r\nStep 525, loss: 5.538381099700928\r\nStep 526, loss: 5.677945613861084\r\nStep 527, loss: 5.529627323150635\r\nStep 528, loss: 5.611823558807373\r\nStep 529, loss: 5.472219944000244\r\nStep 530, loss: 5.792049884796143\r\nStep 531, loss: 5.482363224029541\r\nStep 532, loss: 5.5036163330078125\r\nStep 533, loss: 5.576632499694824\r\nStep 534, loss: 5.6589226722717285\r\nStep 535, loss: 5.598089694976807\r\nStep 536, loss: 5.496463298797607\r\nStep 537, loss: 5.403911590576172\r\nStep 538, loss: 5.450916290283203\r\nStep 539, loss: 5.476700782775879\r\nStep 540, loss: 5.646368503570557\r\nStep 541, loss: 5.7761688232421875\r\nStep 542, loss: 5.410217761993408\r\nStep 543, loss: 5.411986827850342\r\nStep 544, loss: 5.435133457183838\r\nStep 545, loss: 5.382740020751953\r\nStep 546, loss: 5.442299842834473\r\nStep 547, loss: 5.272629261016846\r\nStep 548, loss: 5.307929039001465\r\nStep 549, loss: 5.5719523429870605\r\nStep 550, loss: 5.59519100189209\r\nStep 551, loss: 5.573168754577637\r\nStep 552, loss: 5.326801776885986\r\nStep 553, loss: 5.4029083251953125\r\nStep 554, loss: 5.825585842132568\r\nStep 555, loss: 5.606141090393066\r\nStep 556, loss: 5.3618693351745605\r\nStep 557, loss: 5.3861494064331055\r\nStep 558, loss: 5.679725646972656\r\nStep 559, loss: 5.418980121612549\r\nStep 560, loss: 5.386108875274658\r\nStep 561, loss: 5.720038414001465\r\nStep 562, loss: 5.490231990814209\r\nStep 563, loss: 5.3749494552612305\r\nStep 564, loss: 5.465063571929932\r\nStep 565, loss: 5.612521648406982\r\nStep 566, loss: 5.725881099700928\r\nStep 567, loss: 5.359390735626221\r\nStep 568, loss: 5.486212253570557\r\nStep 569, loss: 5.615264415740967\r\nStep 541, loss: 5.7761688232421875\r\nStep 542, loss: 5.410217761993408\r\nStep 543, loss: 5.411986827850342\r\nStep 544, loss: 5.435133457183838\r\nStep 545, loss: 5.382740020751953\r\nStep 546, loss: 5.442299842834473\r\nStep 547, loss: 5.272629261016846\r\nStep 548, loss: 5.307929039001465\r\nStep 549, loss: 5.5719523429870605\r\nStep 550, loss: 5.59519100189209\r\nStep 551, loss: 5.573168754577637\r\nStep 552, loss: 5.326801776885986\r\nStep 553, loss: 5.4029083251953125\r\nStep 554, loss: 5.825585842132568\r\nStep 555, loss: 5.606141090393066\r\nStep 556, loss: 5.3618693351745605\r\nStep 557, loss: 5.3861494064331055\r\nStep 558, loss: 5.679725646972656\r\nStep 559, loss: 5.418980121612549\r\nStep 560, loss: 5.386108875274658\r\nStep 561, loss: 5.720038414001465\r\nStep 562, loss: 5.490231990814209\r\nStep 563, loss: 5.3749494552612305\r\nStep 564, loss: 5.465063571929932\r\nStep 565, loss: 5.612521648406982\r\nStep 566, loss: 5.725881099700928\r\nStep 567, loss: 5.359390735626221\r\nStep 568, loss: 5.486212253570557\r\nStep 569, loss: 5.615264415740967\r\nStep 570, loss: 5.597212791442871\r\nStep 571, loss: 5.368577003479004\r\nStep 572, loss: 5.364856243133545\r\nStep 573, loss: 5.5607733726501465\r\nStep 574, loss: 5.502553462982178\r\nStep 575, loss: 5.4582037925720215\r\nStep 576, loss: 5.623786449432373\r\nStep 577, loss: 5.269536972045898\r\nStep 578, loss: 5.4617390632629395\r\nStep 579, loss: 5.435409069061279\r\nStep 580, loss: 5.52940034866333\r\nStep 581, loss: 5.613592147827148\r\nStep 582, loss: 5.687880516052246\r\nStep 583, loss: 5.469274044036865\r\nStep 584, loss: 5.551769256591797\r\nStep 585, loss: 5.4086995124816895\r\nStep 586, loss: 5.383148670196533\r\nStep 587, loss: 5.516772747039795\r\nStep 588, loss: 5.366397857666016\r\nStep 589, loss: 5.504728317260742\r\nStep 590, loss: 5.628626823425293\r\nStep 591, loss: 5.551731109619141\r\nStep 592, loss: 5.3733110427856445\r\nStep 593, loss: 5.490641117095947\r\nStep 594, loss: 5.366167068481445\r\nStep 595, loss: 5.499856472015381\r\nStep 596, loss: 5.299220085144043\r\nStep 597, loss: 5.158448219299316\r\nStep 598, loss: 5.433690547943115\r\nStep 570, loss: 5.597212791442871\r\nStep 571, loss: 5.368577003479004\r\nStep 572, loss: 5.364856243133545\r\nStep 573, loss: 5.5607733726501465\r\nStep 574, loss: 5.502553462982178\r\nStep 575, loss: 5.4582037925720215\r\nStep 576, loss: 5.623786449432373\r\nStep 577, loss: 5.269536972045898\r\nStep 578, loss: 5.4617390632629395\r\nStep 579, loss: 5.435409069061279\r\nStep 580, loss: 5.52940034866333\r\nStep 581, loss: 5.613592147827148\r\nStep 582, loss: 5.687880516052246\r\nStep 583, loss: 5.469274044036865\r\nStep 584, loss: 5.551769256591797\r\nStep 585, loss: 5.4086995124816895\r\nStep 586, loss: 5.383148670196533\r\nStep 587, loss: 5.516772747039795\r\nStep 588, loss: 5.366397857666016\r\nStep 589, loss: 5.504728317260742\r\nStep 590, loss: 5.628626823425293\r\nStep 591, loss: 5.551731109619141\r\nStep 592, loss: 5.3733110427856445\r\nStep 593, loss: 5.490641117095947\r\nStep 594, loss: 5.366167068481445\r\nStep 595, loss: 5.499856472015381\r\nStep 596, loss: 5.299220085144043\r\nStep 597, loss: 5.158448219299316\r\nStep 598, loss: 5.433690547943115\r\nStep 599, loss: 5.329139232635498\r\nStep 600, loss: 5.557008266448975\r\nStep 601, loss: 5.331687927246094\r\nStep 599, loss: 5.329139232635498\r\nStep 600, loss: 5.557008266448975\r\nStep 601, loss: 5.331687927246094\r\nStep 602, loss: 5.6626973152160645\r\nStep 603, loss: 5.384581089019775\r\nStep 604, loss: 5.374907970428467\r\nStep 605, loss: 5.254096984863281\r\nStep 606, loss: 5.541470527648926\r\nStep 607, loss: 5.296286106109619\r\nStep 608, loss: 5.154754638671875\r\nStep 609, loss: 5.358010292053223\r\nStep 610, loss: 5.432250022888184\r\nStep 611, loss: 5.315914630889893\r\nStep 612, loss: 5.464038848876953\r\nStep 613, loss: 5.403680324554443\r\nStep 614, loss: 5.4701828956604\r\nStep 615, loss: 5.374019622802734\r\nStep 616, loss: 5.368079662322998\r\nStep 617, loss: 5.493638515472412\r\nStep 618, loss: 5.4959821701049805\r\nStep 619, loss: 5.174816608428955\r\nStep 620, loss: 5.175816059112549\r\nStep 621, loss: 5.408751964569092\r\nStep 622, loss: 5.620441913604736\r\nStep 623, loss: 5.198512554168701\r\nStep 624, loss: 5.317488670349121\r\nStep 625, loss: 5.265618801116943\r\nStep 626, loss: 5.509090900421143\r\nStep 627, loss: 5.488482475280762\r\nStep 628, loss: 5.2329254150390625\r\nStep 629, loss: 5.395798206329346\r\nStep 630, loss: 5.485594749450684\r\nStep 631, loss: 5.229202747344971\r\nStep 602, loss: 5.6626973152160645\r\nStep 603, loss: 5.384581089019775\r\nStep 604, loss: 5.374907970428467\r\nStep 605, loss: 5.254096984863281\r\nStep 606, loss: 5.541470527648926\r\nStep 607, loss: 5.296286106109619\r\nStep 608, loss: 5.154754638671875\r\nStep 609, loss: 5.358010292053223\r\nStep 610, loss: 5.432250022888184\r\nStep 611, loss: 5.315914630889893\r\nStep 612, loss: 5.464038848876953\r\nStep 613, loss: 5.403680324554443\r\nStep 614, loss: 5.4701828956604\r\nStep 615, loss: 5.374019622802734\r\nStep 616, loss: 5.368079662322998\r\nStep 617, loss: 5.493638515472412\r\nStep 618, loss: 5.4959821701049805\r\nStep 619, loss: 5.174816608428955\r\nStep 620, loss: 5.175816059112549\r\nStep 621, loss: 5.408751964569092\r\nStep 622, loss: 5.620441913604736\r\nStep 623, loss: 5.198512554168701\r\nStep 624, loss: 5.317488670349121\r\nStep 625, loss: 5.265618801116943\r\nStep 626, loss: 5.509090900421143\r\nStep 627, loss: 5.488482475280762\r\nStep 628, loss: 5.2329254150390625\r\nStep 629, loss: 5.395798206329346\r\nStep 630, loss: 5.485594749450684\r\nStep 631, loss: 5.229202747344971\r\nStep 632, loss: 5.387337684631348\r\nStep 633, loss: 5.488489151000977\r\nStep 634, loss: 5.307542324066162\r\nStep 635, loss: 5.517372131347656\r\nStep 636, loss: 5.626887798309326\r\nStep 637, loss: 5.219997882843018\r\nStep 638, loss: 5.400931358337402\r\nStep 639, loss: 5.269510269165039\r\nStep 640, loss: 5.465200424194336\r\nStep 641, loss: 5.52509069442749\r\nStep 642, loss: 5.433084964752197\r\nStep 643, loss: 5.42751932144165\r\nStep 644, loss: 5.291433334350586\r\nStep 645, loss: 5.4225239753723145\r\nStep 646, loss: 5.282150745391846\r\nStep 647, loss: 5.644361972808838\r\nStep 648, loss: 5.237563610076904\r\nStep 649, loss: 5.553693771362305\r\nStep 650, loss: 5.0682220458984375\r\nStep 651, loss: 5.371950149536133\r\nStep 652, loss: 5.409918308258057\r\nStep 653, loss: 5.362102508544922\r\nStep 654, loss: 5.344768047332764\r\nStep 655, loss: 5.31875467300415\r\nStep 656, loss: 5.197376728057861\r\nStep 657, loss: 5.290582656860352\r\nStep 658, loss: 5.126527786254883\r\nStep 659, loss: 5.216240882873535\r\nStep 660, loss: 5.198899269104004\r\nStep 661, loss: 5.200804710388184\r\nStep 632, loss: 5.387337684631348\r\nStep 633, loss: 5.488489151000977\r\nStep 634, loss: 5.307542324066162\r\nStep 635, loss: 5.517372131347656\r\nStep 636, loss: 5.626887798309326\r\nStep 637, loss: 5.219997882843018\r\nStep 638, loss: 5.400931358337402\r\nStep 639, loss: 5.269510269165039\r\nStep 640, loss: 5.465200424194336\r\nStep 641, loss: 5.52509069442749\r\nStep 642, loss: 5.433084964752197\r\nStep 643, loss: 5.42751932144165\r\nStep 644, loss: 5.291433334350586\r\nStep 645, loss: 5.4225239753723145\r\nStep 646, loss: 5.282150745391846\r\nStep 647, loss: 5.644361972808838\r\nStep 648, loss: 5.237563610076904\r\nStep 649, loss: 5.553693771362305\r\nStep 650, loss: 5.0682220458984375\r\nStep 651, loss: 5.371950149536133\r\nStep 652, loss: 5.409918308258057\r\nStep 653, loss: 5.362102508544922\r\nStep 654, loss: 5.344768047332764\r\nStep 655, loss: 5.31875467300415\r\nStep 656, loss: 5.197376728057861\r\nStep 657, loss: 5.290582656860352\r\nStep 658, loss: 5.126527786254883\r\nStep 659, loss: 5.216240882873535\r\nStep 660, loss: 5.198899269104004\r\nStep 661, loss: 5.200804710388184\r\nStep 662, loss: 5.3883256912231445\r\nStep 663, loss: 5.612620830535889\r\nStep 664, loss: 5.255474090576172\r\nStep 665, loss: 5.145941734313965\r\nStep 666, loss: 5.431971549987793\r\nStep 667, loss: 5.3477983474731445\r\nStep 668, loss: 5.1120734214782715\r\nStep 669, loss: 5.181546211242676\r\nStep 670, loss: 5.1419501304626465\r\nStep 671, loss: 5.471360206604004\r\nStep 672, loss: 5.205795764923096\r\nStep 673, loss: 5.430771350860596\r\nStep 674, loss: 5.19976282119751\r\nStep 675, loss: 5.763006210327148\r\nStep 676, loss: 5.463747024536133\r\nStep 677, loss: 5.308181285858154\r\nStep 678, loss: 5.46526575088501\r\nStep 679, loss: 5.533337116241455\r\nStep 680, loss: 5.622751235961914\r\nStep 681, loss: 5.210545063018799\r\nStep 682, loss: 5.1722002029418945\r\nStep 683, loss: 5.419670581817627\r\nStep 684, loss: 5.494363307952881\r\nStep 685, loss: 5.412539005279541\r\nStep 686, loss: 5.154549598693848\r\nStep 687, loss: 5.436656475067139\r\nStep 688, loss: 5.432368278503418\r\nStep 689, loss: 5.210556507110596\r\nStep 690, loss: 5.408603668212891\r\nStep 691, loss: 5.160996913909912\r\nStep 662, loss: 5.3883256912231445\r\nStep 663, loss: 5.612620830535889\r\nStep 664, loss: 5.255474090576172\r\nStep 665, loss: 5.145941734313965\r\nStep 666, loss: 5.431971549987793\r\nStep 667, loss: 5.3477983474731445\r\nStep 668, loss: 5.1120734214782715\r\nStep 669, loss: 5.181546211242676\r\nStep 670, loss: 5.1419501304626465\r\nStep 671, loss: 5.471360206604004\r\nStep 672, loss: 5.205795764923096\r\nStep 673, loss: 5.430771350860596\r\nStep 674, loss: 5.19976282119751\r\nStep 675, loss: 5.763006210327148\r\nStep 676, loss: 5.463747024536133\r\nStep 677, loss: 5.308181285858154\r\nStep 678, loss: 5.46526575088501\r\nStep 679, loss: 5.533337116241455\r\nStep 680, loss: 5.622751235961914\r\nStep 681, loss: 5.210545063018799\r\nStep 682, loss: 5.1722002029418945\r\nStep 683, loss: 5.419670581817627\r\nStep 684, loss: 5.494363307952881\r\nStep 685, loss: 5.412539005279541\r\nStep 686, loss: 5.154549598693848\r\nStep 687, loss: 5.436656475067139\r\nStep 688, loss: 5.432368278503418\r\nStep 689, loss: 5.210556507110596\r\nStep 690, loss: 5.408603668212891\r\nStep 691, loss: 5.160996913909912\r\nStep 692, loss: 5.424027919769287\r\nStep 693, loss: 5.170679092407227\r\nStep 694, loss: 5.356256008148193\r\nStep 695, loss: 5.472571849822998\r\nStep 696, loss: 5.417373180389404\r\nStep 697, loss: 5.388498783111572\r\nStep 698, loss: 5.262965679168701\r\nStep 699, loss: 5.498363494873047\r\nStep 700, loss: 5.313055038452148\r\nStep 701, loss: 5.241844177246094\r\nStep 702, loss: 5.448908805847168\r\nStep 703, loss: 5.337677955627441\r\nStep 704, loss: 5.359001636505127\r\nStep 705, loss: 5.145549774169922\r\nStep 706, loss: 5.362729549407959\r\nStep 707, loss: 5.109035491943359\r\nStep 708, loss: 5.128541946411133\r\nStep 709, loss: 5.284796237945557\r\nStep 710, loss: 5.48291540145874\r\nStep 711, loss: 5.406717777252197\r\nStep 712, loss: 5.267970085144043\r\nStep 713, loss: 5.198863506317139\r\nStep 714, loss: 5.358706474304199\r\nStep 715, loss: 5.192930221557617\r\nStep 716, loss: 5.401691913604736\r\nStep 717, loss: 5.32148551940918\r\nStep 718, loss: 5.543673515319824\r\nStep 719, loss: 5.136351585388184\r\nStep 720, loss: 5.422923564910889\r\nStep 721, loss: 5.2997212409973145\r\nStep 692, loss: 5.424027919769287\r\nStep 693, loss: 5.170679092407227\r\nStep 694, loss: 5.356256008148193\r\nStep 695, loss: 5.472571849822998\r\nStep 696, loss: 5.417373180389404\r\nStep 697, loss: 5.388498783111572\r\nStep 698, loss: 5.262965679168701\r\nStep 699, loss: 5.498363494873047\r\nStep 700, loss: 5.313055038452148\r\nStep 701, loss: 5.241844177246094\r\nStep 702, loss: 5.448908805847168\r\nStep 703, loss: 5.337677955627441\r\nStep 704, loss: 5.359001636505127\r\nStep 705, loss: 5.145549774169922\r\nStep 706, loss: 5.362729549407959\r\nStep 707, loss: 5.109035491943359\r\nStep 708, loss: 5.128541946411133\r\nStep 709, loss: 5.284796237945557\r\nStep 710, loss: 5.48291540145874\r\nStep 711, loss: 5.406717777252197\r\nStep 712, loss: 5.267970085144043\r\nStep 713, loss: 5.198863506317139\r\nStep 714, loss: 5.358706474304199\r\nStep 715, loss: 5.192930221557617\r\nStep 716, loss: 5.401691913604736\r\nStep 717, loss: 5.32148551940918\r\nStep 718, loss: 5.543673515319824\r\nStep 719, loss: 5.136351585388184\r\nStep 720, loss: 5.422923564910889\r\nStep 721, loss: 5.2997212409973145\r\nStep 482, loss: 5.577463626861572\r\nStep 483, loss: 5.537669658660889\r\nStep 484, loss: 5.941127300262451\r\nStep 485, loss: 5.490207195281982\r\nStep 486, loss: 5.543642044067383\r\nStep 487, loss: 5.644903659820557\r\nStep 488, loss: 5.602255821228027\r\nStep 489, loss: 5.7087721824646\r\nStep 490, loss: 5.508888244628906\r\nStep 491, loss: 5.632482528686523\r\nStep 492, loss: 5.39115571975708\r\nStep 493, loss: 5.692109107971191\r\nStep 494, loss: 5.630179405212402\r\nStep 495, loss: 5.478193759918213\r\nStep 496, loss: 5.638604164123535\r\nStep 497, loss: 5.54039192199707\r\nStep 498, loss: 5.653471946716309\r\nStep 499, loss: 5.654304027557373\r\nStep 500, loss: 5.648343086242676\r\nStep 501, loss: 5.6619086265563965\r\nStep 502, loss: 5.326898574829102\r\nStep 503, loss: 5.6919264793396\r\nStep 504, loss: 5.549910068511963\r\nStep 505, loss: 5.593076229095459\r\nStep 506, loss: 5.713366508483887\r\nStep 507, loss: 5.6414794921875\r\nStep 508, loss: 5.393517017364502\r\nStep 509, loss: 5.7733964920043945\r\nStep 510, loss: 5.574671745300293\r\nStep 511, loss: 5.704086780548096\r\nStep 512, loss: 5.621325492858887\r\nStep 513, loss: 5.49443244934082\r\nStep 514, loss: 5.439146518707275\r\nStep 515, loss: 5.49471378326416\r\nStep 516, loss: 5.5571980476379395\r\nStep 517, loss: 5.517596244812012\r\nStep 518, loss: 5.4848527908325195\r\nStep 519, loss: 5.647706031799316\r\nStep 520, loss: 5.3533124923706055\r\nStep 521, loss: 5.372676372528076\r\nStep 522, loss: 5.403843879699707\r\nStep 523, loss: 5.5930867195129395\r\nStep 524, loss: 5.593081474304199\r\nStep 525, loss: 5.538381099700928\r\nStep 526, loss: 5.677945613861084\r\nStep 527, loss: 5.529627323150635\r\nStep 528, loss: 5.611823558807373\r\nStep 529, loss: 5.472219944000244\r\nStep 530, loss: 5.792049884796143\r\nStep 531, loss: 5.482363224029541\r\nStep 532, loss: 5.5036163330078125\r\nStep 533, loss: 5.576632499694824\r\nStep 534, loss: 5.6589226722717285\r\nStep 535, loss: 5.598089694976807\r\nStep 536, loss: 5.496463298797607\r\nStep 537, loss: 5.403911590576172\r\nStep 538, loss: 5.450916290283203\r\nStep 539, loss: 5.476700782775879\r\nStep 540, loss: 5.646368503570557\r\nStep 482, loss: 5.577463626861572\r\nStep 483, loss: 5.537669658660889\r\nStep 484, loss: 5.941127300262451\r\nStep 485, loss: 5.490207195281982\r\nStep 486, loss: 5.543642044067383\r\nStep 487, loss: 5.644903659820557\r\nStep 488, loss: 5.602255821228027\r\nStep 489, loss: 5.7087721824646\r\nStep 490, loss: 5.508888244628906\r\nStep 491, loss: 5.632482528686523\r\nStep 492, loss: 5.39115571975708\r\nStep 493, loss: 5.692109107971191\r\nStep 494, loss: 5.630179405212402\r\nStep 495, loss: 5.478193759918213\r\nStep 496, loss: 5.638604164123535\r\nStep 497, loss: 5.54039192199707\r\nStep 498, loss: 5.653471946716309\r\nStep 499, loss: 5.654304027557373\r\nStep 500, loss: 5.648343086242676\r\nStep 501, loss: 5.6619086265563965\r\nStep 502, loss: 5.326898574829102\r\nStep 503, loss: 5.6919264793396\r\nStep 504, loss: 5.549910068511963\r\nStep 505, loss: 5.593076229095459\r\nStep 506, loss: 5.713366508483887\r\nStep 507, loss: 5.6414794921875\r\nStep 508, loss: 5.393517017364502\r\nStep 509, loss: 5.7733964920043945\r\nStep 510, loss: 5.574671745300293\r\nStep 511, loss: 5.704086780548096\r\nStep 541, loss: 5.7761688232421875\r\nStep 542, loss: 5.410217761993408\r\nStep 543, loss: 5.411986827850342\r\nStep 544, loss: 5.435133457183838\r\nStep 545, loss: 5.382740020751953\r\nStep 546, loss: 5.442299842834473\r\nStep 547, loss: 5.272629261016846\r\nStep 548, loss: 5.307929039001465\r\nStep 549, loss: 5.5719523429870605\r\nStep 550, loss: 5.59519100189209\r\nStep 551, loss: 5.573168754577637\r\nStep 552, loss: 5.326801776885986\r\nStep 553, loss: 5.4029083251953125\r\nStep 554, loss: 5.825585842132568\r\nStep 555, loss: 5.606141090393066\r\nStep 556, loss: 5.3618693351745605\r\nStep 557, loss: 5.3861494064331055\r\nStep 558, loss: 5.679725646972656\r\nStep 559, loss: 5.418980121612549\r\nStep 560, loss: 5.386108875274658\r\nStep 561, loss: 5.720038414001465\r\nStep 562, loss: 5.490231990814209\r\nStep 563, loss: 5.3749494552612305\r\nStep 564, loss: 5.465063571929932\r\nStep 565, loss: 5.612521648406982\r\nStep 566, loss: 5.725881099700928\r\nStep 567, loss: 5.359390735626221\r\nStep 568, loss: 5.486212253570557\r\nStep 569, loss: 5.615264415740967\r\nStep 512, loss: 5.621325492858887\r\nStep 513, loss: 5.49443244934082\r\nStep 514, loss: 5.439146518707275\r\nStep 515, loss: 5.49471378326416\r\nStep 516, loss: 5.5571980476379395\r\nStep 517, loss: 5.517596244812012\r\nStep 518, loss: 5.4848527908325195\r\nStep 519, loss: 5.647706031799316\r\nStep 520, loss: 5.3533124923706055\r\nStep 521, loss: 5.372676372528076\r\nStep 522, loss: 5.403843879699707\r\nStep 523, loss: 5.5930867195129395\r\nStep 524, loss: 5.593081474304199\r\nStep 525, loss: 5.538381099700928\r\nStep 526, loss: 5.677945613861084\r\nStep 527, loss: 5.529627323150635\r\nStep 528, loss: 5.611823558807373\r\nStep 529, loss: 5.472219944000244\r\nStep 530, loss: 5.792049884796143\r\nStep 531, loss: 5.482363224029541\r\nStep 532, loss: 5.5036163330078125\r\nStep 533, loss: 5.576632499694824\r\nStep 534, loss: 5.6589226722717285\r\nStep 535, loss: 5.598089694976807\r\nStep 536, loss: 5.496463298797607\r\nStep 537, loss: 5.403911590576172\r\nStep 538, loss: 5.450916290283203\r\nStep 539, loss: 5.476700782775879\r\nStep 540, loss: 5.646368503570557\r\nStep 570, loss: 5.597212791442871\r\nStep 571, loss: 5.368577003479004\r\nStep 572, loss: 5.364856243133545\r\nStep 573, loss: 5.5607733726501465\r\nStep 574, loss: 5.502553462982178\r\nStep 575, loss: 5.4582037925720215\r\nStep 576, loss: 5.623786449432373\r\nStep 577, loss: 5.269536972045898\r\nStep 578, loss: 5.4617390632629395\r\nStep 579, loss: 5.435409069061279\r\nStep 580, loss: 5.52940034866333\r\nStep 581, loss: 5.613592147827148\r\nStep 582, loss: 5.687880516052246\r\nStep 583, loss: 5.469274044036865\r\nStep 584, loss: 5.551769256591797\r\nStep 585, loss: 5.4086995124816895\r\nStep 586, loss: 5.383148670196533\r\nStep 587, loss: 5.516772747039795\r\nStep 588, loss: 5.366397857666016\r\nStep 589, loss: 5.504728317260742\r\nStep 590, loss: 5.628626823425293\r\nStep 591, loss: 5.551731109619141\r\nStep 592, loss: 5.3733110427856445\r\nStep 593, loss: 5.490641117095947\r\nStep 594, loss: 5.366167068481445\r\nStep 595, loss: 5.499856472015381\r\nStep 596, loss: 5.299220085144043\r\nStep 597, loss: 5.158448219299316\r\nStep 598, loss: 5.433690547943115\r\nStep 541, loss: 5.7761688232421875\r\nStep 542, loss: 5.410217761993408\r\nStep 543, loss: 5.411986827850342\r\nStep 544, loss: 5.435133457183838\r\nStep 545, loss: 5.382740020751953\r\nStep 546, loss: 5.442299842834473\r\nStep 547, loss: 5.272629261016846\r\nStep 548, loss: 5.307929039001465\r\nStep 549, loss: 5.5719523429870605\r\nStep 550, loss: 5.59519100189209\r\nStep 551, loss: 5.573168754577637\r\nStep 552, loss: 5.326801776885986\r\nStep 553, loss: 5.4029083251953125\r\nStep 554, loss: 5.825585842132568\r\nStep 555, loss: 5.606141090393066\r\nStep 556, loss: 5.3618693351745605\r\nStep 557, loss: 5.3861494064331055\r\nStep 558, loss: 5.679725646972656\r\nStep 559, loss: 5.418980121612549\r\nStep 560, loss: 5.386108875274658\r\nStep 561, loss: 5.720038414001465\r\nStep 562, loss: 5.490231990814209\r\nStep 563, loss: 5.3749494552612305\r\nStep 564, loss: 5.465063571929932\r\nStep 565, loss: 5.612521648406982\r\nStep 566, loss: 5.725881099700928\r\nStep 567, loss: 5.359390735626221\r\nStep 568, loss: 5.486212253570557\r\nStep 569, loss: 5.615264415740967\r\nStep 599, loss: 5.329139232635498\r\nStep 600, loss: 5.557008266448975\r\nStep 601, loss: 5.331687927246094\r\nStep 570, loss: 5.597212791442871\r\nStep 571, loss: 5.368577003479004\r\nStep 572, loss: 5.364856243133545\r\nStep 573, loss: 5.5607733726501465\r\nStep 574, loss: 5.502553462982178\r\nStep 575, loss: 5.4582037925720215\r\nStep 576, loss: 5.623786449432373\r\nStep 577, loss: 5.269536972045898\r\nStep 578, loss: 5.4617390632629395\r\nStep 579, loss: 5.435409069061279\r\nStep 580, loss: 5.52940034866333\r\nStep 581, loss: 5.613592147827148\r\nStep 582, loss: 5.687880516052246\r\nStep 583, loss: 5.469274044036865\r\nStep 584, loss: 5.551769256591797\r\nStep 585, loss: 5.4086995124816895\r\nStep 586, loss: 5.383148670196533\r\nStep 587, loss: 5.516772747039795\r\nStep 588, loss: 5.366397857666016\r\nStep 589, loss: 5.504728317260742\r\nStep 590, loss: 5.628626823425293\r\nStep 591, loss: 5.551731109619141\r\nStep 592, loss: 5.3733110427856445\r\nStep 593, loss: 5.490641117095947\r\nStep 594, loss: 5.366167068481445\r\nStep 595, loss: 5.499856472015381\r\nStep 596, loss: 5.299220085144043\r\nStep 597, loss: 5.158448219299316\r\nStep 598, loss: 5.433690547943115\r\nStep 602, loss: 5.6626973152160645\r\nStep 603, loss: 5.384581089019775\r\nStep 604, loss: 5.374907970428467\r\nStep 605, loss: 5.254096984863281\r\nStep 606, loss: 5.541470527648926\r\nStep 607, loss: 5.296286106109619\r\nStep 608, loss: 5.154754638671875\r\nStep 609, loss: 5.358010292053223\r\nStep 610, loss: 5.432250022888184\r\nStep 611, loss: 5.315914630889893\r\nStep 612, loss: 5.464038848876953\r\nStep 613, loss: 5.403680324554443\r\nStep 614, loss: 5.4701828956604\r\nStep 615, loss: 5.374019622802734\r\nStep 616, loss: 5.368079662322998\r\nStep 617, loss: 5.493638515472412\r\nStep 618, loss: 5.4959821701049805\r\nStep 619, loss: 5.174816608428955\r\nStep 620, loss: 5.175816059112549\r\nStep 621, loss: 5.408751964569092\r\nStep 622, loss: 5.620441913604736\r\nStep 623, loss: 5.198512554168701\r\nStep 624, loss: 5.317488670349121\r\nStep 625, loss: 5.265618801116943\r\nStep 626, loss: 5.509090900421143\r\nStep 627, loss: 5.488482475280762\r\nStep 628, loss: 5.2329254150390625\r\nStep 629, loss: 5.395798206329346\r\nStep 630, loss: 5.485594749450684\r\nStep 631, loss: 5.229202747344971\r\nStep 599, loss: 5.329139232635498\r\nStep 600, loss: 5.557008266448975\r\nStep 601, loss: 5.331687927246094\r\nStep 632, loss: 5.387337684631348\r\nStep 633, loss: 5.488489151000977\r\nStep 634, loss: 5.307542324066162\r\nStep 635, loss: 5.517372131347656\r\nStep 636, loss: 5.626887798309326\r\nStep 637, loss: 5.219997882843018\r\nStep 638, loss: 5.400931358337402\r\nStep 639, loss: 5.269510269165039\r\nStep 640, loss: 5.465200424194336\r\nStep 641, loss: 5.52509069442749\r\nStep 642, loss: 5.433084964752197\r\nStep 643, loss: 5.42751932144165\r\nStep 644, loss: 5.291433334350586\r\nStep 645, loss: 5.4225239753723145\r\nStep 646, loss: 5.282150745391846\r\nStep 647, loss: 5.644361972808838\r\nStep 648, loss: 5.237563610076904\r\nStep 649, loss: 5.553693771362305\r\nStep 650, loss: 5.0682220458984375\r\nStep 651, loss: 5.371950149536133\r\nStep 652, loss: 5.409918308258057\r\nStep 653, loss: 5.362102508544922\r\nStep 654, loss: 5.344768047332764\r\nStep 655, loss: 5.31875467300415\r\nStep 656, loss: 5.197376728057861\r\nStep 657, loss: 5.290582656860352\r\nStep 658, loss: 5.126527786254883\r\nStep 659, loss: 5.216240882873535\r\nStep 660, loss: 5.198899269104004\r\nStep 661, loss: 5.200804710388184\r\nStep 602, loss: 5.6626973152160645\r\nStep 603, loss: 5.384581089019775\r\nStep 604, loss: 5.374907970428467\r\nStep 605, loss: 5.254096984863281\r\nStep 606, loss: 5.541470527648926\r\nStep 607, loss: 5.296286106109619\r\nStep 608, loss: 5.154754638671875\r\nStep 609, loss: 5.358010292053223\r\nStep 610, loss: 5.432250022888184\r\nStep 611, loss: 5.315914630889893\r\nStep 612, loss: 5.464038848876953\r\nStep 613, loss: 5.403680324554443\r\nStep 614, loss: 5.4701828956604\r\nStep 615, loss: 5.374019622802734\r\nStep 616, loss: 5.368079662322998\r\nStep 617, loss: 5.493638515472412\r\nStep 618, loss: 5.4959821701049805\r\nStep 619, loss: 5.174816608428955\r\nStep 620, loss: 5.175816059112549\r\nStep 621, loss: 5.408751964569092\r\nStep 622, loss: 5.620441913604736\r\nStep 623, loss: 5.198512554168701\r\nStep 624, loss: 5.317488670349121\r\nStep 625, loss: 5.265618801116943\r\nStep 626, loss: 5.509090900421143\r\nStep 627, loss: 5.488482475280762\r\nStep 628, loss: 5.2329254150390625\r\nStep 629, loss: 5.395798206329346\r\nStep 630, loss: 5.485594749450684\r\nStep 631, loss: 5.229202747344971\r\nStep 662, loss: 5.3883256912231445\r\nStep 663, loss: 5.612620830535889\r\nStep 664, loss: 5.255474090576172\r\nStep 665, loss: 5.145941734313965\r\nStep 666, loss: 5.431971549987793\r\nStep 667, loss: 5.3477983474731445\r\nStep 668, loss: 5.1120734214782715\r\nStep 669, loss: 5.181546211242676\r\nStep 670, loss: 5.1419501304626465\r\nStep 671, loss: 5.471360206604004\r\nStep 672, loss: 5.205795764923096\r\nStep 673, loss: 5.430771350860596\r\nStep 674, loss: 5.19976282119751\r\nStep 675, loss: 5.763006210327148\r\nStep 676, loss: 5.463747024536133\r\nStep 677, loss: 5.308181285858154\r\nStep 678, loss: 5.46526575088501\r\nStep 679, loss: 5.533337116241455\r\nStep 680, loss: 5.622751235961914\r\nStep 681, loss: 5.210545063018799\r\nStep 682, loss: 5.1722002029418945\r\nStep 683, loss: 5.419670581817627\r\nStep 684, loss: 5.494363307952881\r\nStep 685, loss: 5.412539005279541\r\nStep 686, loss: 5.154549598693848\r\nStep 687, loss: 5.436656475067139\r\nStep 688, loss: 5.432368278503418\r\nStep 689, loss: 5.210556507110596\r\nStep 690, loss: 5.408603668212891\r\nStep 691, loss: 5.160996913909912\r\nStep 632, loss: 5.387337684631348\r\nStep 633, loss: 5.488489151000977\r\nStep 634, loss: 5.307542324066162\r\nStep 635, loss: 5.517372131347656\r\nStep 636, loss: 5.626887798309326\r\nStep 637, loss: 5.219997882843018\r\nStep 638, loss: 5.400931358337402\r\nStep 639, loss: 5.269510269165039\r\nStep 640, loss: 5.465200424194336\r\nStep 641, loss: 5.52509069442749\r\nStep 642, loss: 5.433084964752197\r\nStep 643, loss: 5.42751932144165\r\nStep 644, loss: 5.291433334350586\r\nStep 645, loss: 5.4225239753723145\r\nStep 646, loss: 5.282150745391846\r\nStep 647, loss: 5.644361972808838\r\nStep 648, loss: 5.237563610076904\r\nStep 649, loss: 5.553693771362305\r\nStep 650, loss: 5.0682220458984375\r\nStep 651, loss: 5.371950149536133\r\nStep 652, loss: 5.409918308258057\r\nStep 653, loss: 5.362102508544922\r\nStep 654, loss: 5.344768047332764\r\nStep 655, loss: 5.31875467300415\r\nStep 656, loss: 5.197376728057861\r\nStep 657, loss: 5.290582656860352\r\nStep 658, loss: 5.126527786254883\r\nStep 659, loss: 5.216240882873535\r\nStep 660, loss: 5.198899269104004\r\nStep 661, loss: 5.200804710388184\r\nStep 692, loss: 5.424027919769287\r\nStep 693, loss: 5.170679092407227\r\nStep 694, loss: 5.356256008148193\r\nStep 695, loss: 5.472571849822998\r\nStep 696, loss: 5.417373180389404\r\nStep 697, loss: 5.388498783111572\r\nStep 698, loss: 5.262965679168701\r\nStep 699, loss: 5.498363494873047\r\nStep 700, loss: 5.313055038452148\r\nStep 701, loss: 5.241844177246094\r\nStep 702, loss: 5.448908805847168\r\nStep 703, loss: 5.337677955627441\r\nStep 704, loss: 5.359001636505127\r\nStep 705, loss: 5.145549774169922\r\nStep 706, loss: 5.362729549407959\r\nStep 707, loss: 5.109035491943359\r\nStep 708, loss: 5.128541946411133\r\nStep 709, loss: 5.284796237945557\r\nStep 710, loss: 5.48291540145874\r\nStep 711, loss: 5.406717777252197\r\nStep 712, loss: 5.267970085144043\r\nStep 713, loss: 5.198863506317139\r\nStep 714, loss: 5.358706474304199\r\nStep 715, loss: 5.192930221557617\r\nStep 716, loss: 5.401691913604736\r\nStep 717, loss: 5.32148551940918\r\nStep 718, loss: 5.543673515319824\r\nStep 719, loss: 5.136351585388184\r\nStep 720, loss: 5.422923564910889\r\nStep 721, loss: 5.2997212409973145\r\nStep 662, loss: 5.3883256912231445\r\nStep 663, loss: 5.612620830535889\r\nStep 664, loss: 5.255474090576172\r\nStep 665, loss: 5.145941734313965\r\nStep 666, loss: 5.431971549987793\r\nStep 667, loss: 5.3477983474731445\r\nStep 668, loss: 5.1120734214782715\r\nStep 669, loss: 5.181546211242676\r\nStep 670, loss: 5.1419501304626465\r\nStep 671, loss: 5.471360206604004\r\nStep 672, loss: 5.205795764923096\r\nStep 673, loss: 5.430771350860596\r\nStep 674, loss: 5.19976282119751\r\nStep 675, loss: 5.763006210327148\r\nStep 676, loss: 5.463747024536133\r\nStep 677, loss: 5.308181285858154\r\nStep 678, loss: 5.46526575088501\r\nStep 679, loss: 5.533337116241455\r\nStep 680, loss: 5.622751235961914\r\nStep 681, loss: 5.210545063018799\r\nStep 682, loss: 5.1722002029418945\r\nStep 683, loss: 5.419670581817627\r\nStep 684, loss: 5.494363307952881\r\nStep 685, loss: 5.412539005279541\r\nStep 686, loss: 5.154549598693848\r\nStep 687, loss: 5.436656475067139\r\nStep 688, loss: 5.432368278503418\r\nStep 689, loss: 5.210556507110596\r\nStep 690, loss: 5.408603668212891\r\nStep 691, loss: 5.160996913909912\r\nStep 692, loss: 5.424027919769287\r\nStep 693, loss: 5.170679092407227\r\nStep 694, loss: 5.356256008148193\r\nStep 695, loss: 5.472571849822998\r\nStep 696, loss: 5.417373180389404\r\nStep 697, loss: 5.388498783111572\r\nStep 698, loss: 5.262965679168701\r\nStep 699, loss: 5.498363494873047\r\nStep 700, loss: 5.313055038452148\r\nStep 701, loss: 5.241844177246094\r\nStep 702, loss: 5.448908805847168\r\nStep 703, loss: 5.337677955627441\r\nStep 704, loss: 5.359001636505127\r\nStep 705, loss: 5.145549774169922\r\nStep 706, loss: 5.362729549407959\r\nStep 707, loss: 5.109035491943359\r\nStep 708, loss: 5.128541946411133\r\nStep 709, loss: 5.284796237945557\r\nStep 710, loss: 5.48291540145874\r\nStep 711, loss: 5.406717777252197\r\nStep 712, loss: 5.267970085144043\r\nStep 713, loss: 5.198863506317139\r\nStep 714, loss: 5.358706474304199\r\nStep 715, loss: 5.192930221557617\r\nStep 716, loss: 5.401691913604736\r\nStep 717, loss: 5.32148551940918\r\nStep 718, loss: 5.543673515319824\r\nStep 719, loss: 5.136351585388184\r\nStep 720, loss: 5.422923564910889\r\nStep 721, loss: 5.2997212409973145\r\n",,terminal_output +2974,4340175,"TERMINAL",0,0,"srun",,terminal_focus +2975,4344028,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2976,4344883,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2977,4349875,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"",shellscript,tab +2978,4351751,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +2979,4365352,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +2980,4366789,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1152,0,"",shellscript,selection_mouse +2981,4366954,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1152,1,"=",shellscript,selection_mouse +2982,4366955,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1152,2,"=1",shellscript,selection_mouse +2983,4366965,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1152,3,"=1.",shellscript,selection_mouse +2984,4367028,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1152,4,"=1.5",shellscript,selection_mouse +2985,4367029,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1152,5,"=1.5e",shellscript,selection_mouse +2986,4367030,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1152,6,"=1.5e-",shellscript,selection_mouse +2987,4367344,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1152,7,"=1.5e-5",shellscript,selection_mouse +2988,4367448,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1152,8,"=1.5e-5 ",shellscript,selection_mouse +2989,4367892,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1160,0,"",shellscript,selection_mouse +2990,4368672,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1159,0,"",shellscript,selection_mouse +2991,4368814,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1158,1,"5",shellscript,selection_mouse +2992,4368815,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1157,2,"-5",shellscript,selection_mouse +2993,4368824,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1156,3,"e-5",shellscript,selection_mouse +2994,4368841,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1155,4,"5e-5",shellscript,selection_mouse +2995,4368900,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1154,5,".5e-5",shellscript,selection_mouse +2996,4368906,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1153,6,"1.5e-5",shellscript,selection_mouse +2997,4368907,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1152,7,"=1.5e-5",shellscript,selection_mouse +2998,4368985,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1151,8,"r=1.5e-5",shellscript,selection_mouse +2999,4369271,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1152,7,"=1.5e-5",shellscript,selection_mouse +3000,4369731,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1153,6,"1.5e-5",shellscript,selection_mouse +3001,4375452,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",0,0,"",shellscript,tab +3002,4377079,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1176,0,"",shellscript,selection_mouse +3003,4378298,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1175,1,"",shellscript,content +3004,4378446,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1174,1,"",shellscript,content +3005,4378676,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1172,2,"",shellscript,content +3006,4379008,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1171,1,"",shellscript,content +3007,4379359,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1170,1,"",shellscript,content +3008,4379665,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1170,0,"1.5e-5",shellscript,content +3009,4381269,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",0,0,"",shellscript,tab +3010,4382495,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1176,0,"",shellscript,selection_mouse +3011,4382626,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1175,1,"4",shellscript,selection_mouse +3012,4382658,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1174,2,"-4",shellscript,selection_mouse +3013,4382676,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1173,3,"e-4",shellscript,selection_mouse +3014,4382734,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1172,4,"4e-4",shellscript,selection_mouse +3015,4382807,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1171,5,".4e-4",shellscript,selection_mouse +3016,4382810,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1176,17," \\n --log_imag",shellscript,selection_mouse +3017,4382829,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1176,16," \\n --log_ima",shellscript,selection_mouse +3018,4382908,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1176,15," \\n --log_im",shellscript,selection_mouse +3019,4383259,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1169,7,"=1.4e-4",shellscript,selection_mouse +3020,4383621,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1170,6,"1.4e-4",shellscript,selection_mouse +3021,4384359,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1170,6,"",shellscript,content +3022,4384872,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1170,0,"1.5e-5",shellscript,content +3023,4386900,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",0,0,"",shellscript,tab +3024,4388055,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1176,0,"",shellscript,selection_mouse +3025,4388212,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1175,1,"4",shellscript,selection_mouse +3026,4388231,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1174,2,"-4",shellscript,selection_mouse +3027,4388268,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1173,3,"e-4",shellscript,selection_mouse +3028,4388332,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1172,4,"4e-4",shellscript,selection_mouse +3029,4388394,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1171,5,".4e-4",shellscript,selection_mouse +3030,4388461,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1170,6,"1.4e-4",shellscript,selection_mouse +3031,4389593,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1170,6,"",shellscript,content +3032,4390903,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1170,0,"1.5e-5",shellscript,content +3033,4392274,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",0,0,"",shellscript,tab +3034,4393655,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1176,0,"",shellscript,selection_mouse +3035,4393821,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1175,1,"4",shellscript,selection_mouse +3036,4393835,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1174,2,"-4",shellscript,selection_mouse +3037,4393883,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1173,3,"e-4",shellscript,selection_mouse +3038,4394017,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1172,4,"4e-4",shellscript,selection_mouse +3039,4394074,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1171,5,".4e-4",shellscript,selection_mouse +3040,4394082,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1170,6,"1.4e-4",shellscript,selection_mouse +3041,4394995,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1170,6,"",shellscript,content +3042,4395361,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1170,0,"1.5e-5",shellscript,content +3043,4396794,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",0,0,"",shellscript,tab +3044,4398427,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1176,0,"",shellscript,selection_mouse +3045,4398607,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1175,1,"4",shellscript,selection_mouse +3046,4398636,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1174,2,"-4",shellscript,selection_mouse +3047,4398702,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1173,3,"e-4",shellscript,selection_mouse +3048,4398785,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1172,4,"4e-4",shellscript,selection_mouse +3049,4398869,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1171,5,".4e-4",shellscript,selection_mouse +3050,4399021,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1170,6,"1.4e-4",shellscript,selection_mouse +3051,4399791,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1170,6,"",shellscript,content +3052,4400200,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1170,0,"1.5e-5",shellscript,content +3053,4660846,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",0,0,"",shellscript,tab +3054,4661756,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",0,0,"",shellscript,tab +3055,4662231,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",0,0,"",shellscript,tab +3056,4662797,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",0,0,"",shellscript,tab +3057,4664191,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",0,0,"",shellscript,tab +3058,4700094,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4",0,0,"",plaintext,tab +3059,4734772,"slurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/train_tokenizer_37M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_model_size_scaling_37M\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/open_ai_minecraft_tfrecord\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-model-size-scaling-38M-$slurm_job_id \\n --tags tokenizer model-size-scaling 38M \\n --entity instant-uv \\n --project jafar \\n --data_dir $tf_records_dir\n",shellscript,tab +3060,4736968,"slurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/train_tokenizer_37M.sbatch",1249,0,"",shellscript,selection_mouse +3061,4737146,"slurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/train_tokenizer_37M.sbatch",1091,158,"size-scaling-38M-$slurm_job_id \\n --tags tokenizer model-size-scaling 38M \\n --entity instant-uv \\n --project jafar \\n --data_dir $tf_records_dir\n",shellscript,selection_mouse +3062,4737167,"slurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/train_tokenizer_37M.sbatch",825,424,"HECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-model-size-scaling-38M-$slurm_job_id \\n --tags tokenizer model-size-scaling 38M \\n --entity instant-uv \\n --project jafar \\n --data_dir $tf_records_dir\n",shellscript,selection_mouse +3063,4737181,"slurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/train_tokenizer_37M.sbatch",544,705,"\ntf_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/open_ai_minecraft_tfrecord\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-model-size-scaling-38M-$slurm_job_id \\n --tags tokenizer model-size-scaling 38M \\n --entity instant-uv \\n --project jafar \\n --data_dir $tf_records_dir\n",shellscript,selection_mouse +3064,4737203,"slurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/train_tokenizer_37M.sbatch",149,1100,"#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_model_size_scaling_37M\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/open_ai_minecraft_tfrecord\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-model-size-scaling-38M-$slurm_job_id \\n --tags tokenizer model-size-scaling 38M \\n --entity instant-uv \\n --project jafar \\n --data_dir $tf_records_dir\n",shellscript,selection_mouse +3065,4737231,"slurm/jobs/mihir/horeka/modelsize_scaling/tokenizer/train_tokenizer_37M.sbatch",0,1249,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_model_size_scaling_37M\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/open_ai_minecraft_tfrecord\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-model-size-scaling-38M-$slurm_job_id \\n --tags tokenizer model-size-scaling 38M \\n --entity instant-uv \\n --project jafar \\n --data_dir $tf_records_dir\n",shellscript,selection_mouse +3066,4740822,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4",0,0,"",plaintext,tab +3067,4749556,"slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_2_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\n# array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked_subset\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +3068,4750301,"slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_2_nodes.sbatch",1442,0,"",shellscript,selection_mouse +3069,4750553,"slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_2_nodes.sbatch",1435,7,"ds_dir\n",shellscript,selection_mouse +3070,4750564,"slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_2_nodes.sbatch",1008,434," SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3071,4750580,"slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_2_nodes.sbatch",810,632,"_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3072,4750638,"slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_2_nodes.sbatch",548,894,"# array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked_subset\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3073,4750639,"slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_2_nodes.sbatch",367,1075,"#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\n# array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked_subset\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3074,4750639,"slurm/jobs/mihir/horeka/batchsize_scaling/tokenizer/sqrt_lr/train_tokenizer_2_nodes.sbatch",0,1442,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\n# array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked_subset\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3075,4756727,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4",0,0,"",plaintext,tab +3076,4758420,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\n# array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked_subset\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",plaintext,content +3077,4768099,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\n# array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords/10fps_160x90\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked_subset\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +3078,4771812,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"",shellscript,tab +3079,4775168,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",665,0,"",shellscript,selection_mouse +3080,4775330,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",631,38,"open_ai_minecraft_arrayrecords_chunked",shellscript,selection_mouse +3081,4775464,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",548,122,"array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n",shellscript,selection_mouse +3082,4777770,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3083,4778771,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",597,0,"",shellscript,selection_mouse +3084,4779976,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",732,0,"",shellscript,selection_mouse +3085,4780817,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",805,0,"\n",shellscript,content +3086,4781090,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",806,0,"array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n",shellscript,content +3087,4782206,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",806,0,"",shellscript,selection_command +3088,4782390,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",677,0,"",shellscript,selection_command +3089,4784457,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",548,258,"",shellscript,content +3090,4789364,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",696,0,"",shellscript,selection_mouse +3091,4789370,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",695,0,"",shellscript,selection_command +3092,4790102,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",671,0,"",shellscript,selection_mouse +3093,4790908,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",671,1,"",shellscript,content +3094,4792345,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1306,0,"",shellscript,selection_mouse +3095,4792885,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1305,0,"",shellscript,selection_mouse +3096,4792886,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1304,0,"",shellscript,selection_command +3097,4793042,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1304,1,"r",shellscript,selection_mouse +3098,4793042,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1305,0,"",shellscript,selection_command +3099,4793062,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1271,34,"\n --data_dir $array_records_dir",shellscript,selection_mouse +3100,4793093,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1249,56,"\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3101,4793109,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1121,184,"caling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3102,4793124,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1087,218,"\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3103,4793139,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1059,246,"_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3104,4793154,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1057,248,"nt_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3105,4793171,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1054,251,"point_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3106,4793185,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1021,284,"e_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3107,4793200,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1020,285,"ge_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3108,4793216,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",996,309,"r=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3109,4793231,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",978,327,"lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3110,4793293,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",955,350,"ch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3111,4793294,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",954,351,"tch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3112,4793295,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",920,385,"kpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3113,4793299,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",919,386,"ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3114,4793317,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",918,387,"-ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3115,4793374,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",884,421," python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3116,4793374,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",883,422,"n python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3117,4793396,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",882,423,"un python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3118,4793417,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",879,426,"\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3119,4794240,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",881,0,"",shellscript,selection_mouse +3120,4794241,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",880,4,"srun",shellscript,selection_mouse +3121,4794399,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",880,38,"srun python train_tokenizer.py \\n -",shellscript,selection_mouse +3122,4794400,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",880,100,"srun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr",shellscript,selection_mouse +3123,4794425,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",880,151,"srun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval",shellscript,selection_mouse +3124,4794449,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",880,207,"srun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \",shellscript,selection_mouse +3125,4794465,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",880,239,"srun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size",shellscript,selection_mouse +3126,4794479,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",880,318,"srun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling",shellscript,selection_mouse +3127,4794538,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",880,369,"srun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \",shellscript,selection_mouse +3128,4794539,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",880,391,"srun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \",shellscript,selection_mouse +3129,4794539,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",880,425,"srun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir",shellscript,selection_mouse +3130,4794764,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1305,0,"",shellscript,selection_mouse +3131,4794765,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1304,0,"",shellscript,selection_command +3132,4796737,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1114,0,"",shellscript,selection_mouse +3133,4797221,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1110,0,"",shellscript,selection_mouse +3134,4798151,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1306,0,"",shellscript,selection_mouse +3135,4799120,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1305,1,"\n",shellscript,selection_mouse +3136,4799130,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1189,117,"e-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3137,4799146,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",933,373,"KPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3138,4799204,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",457,849,"$0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3139,4799204,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,1306,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.4e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-batch-size-scaling-2-node-sqrt-lr-$slurm_job_id \\n --tags tokenizer batch-size-scaling 2-node sqrt-lr-scaling \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3140,4800671,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,selection_command +3141,4822965,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",367,0,"#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n",shellscript,content +3142,4822967,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",416,61,"",shellscript,content +3143,4823882,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",973,0," --max_lr=1e-4 \\n",shellscript,content +3144,4823939,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",993,22,"",shellscript,content +3145,4824599,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1074,0," --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n",shellscript,content +3146,4824808,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1125,0," --tags tokenizer lr-sweep 1e-4 \\n",shellscript,content +3147,4825079,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1162,136,"",shellscript,content +3148,4832338,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4 copy.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +3149,4839854,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +3150,4840528,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1244,0,"",shellscript,selection_mouse +3151,4840687,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1209,35,"\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3152,4840707,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",969,275,"0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3153,4840797,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",719,525,"INT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3154,4840798,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",417,827,"# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3155,4840798,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",149,1095,"#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3156,4840798,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",67,1177,"#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3157,4840816,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",21,1223,"#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3158,4840816,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",20,1224,"\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3159,4842124,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1244,0,"",shellscript,selection_mouse +3160,4842624,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1187,57,"\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3161,4842683,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",972,272,"\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3162,4842683,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",711,533,"\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3163,4842683,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",517,727,"venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3164,4842690,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",426,818," sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3165,4842706,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",416,828,"\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3166,4842762,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",375,869,"--job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3167,4842769,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",277,967,"--error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3168,4842772,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",178,1066,"--output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3169,4842791,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",157,1087,"--gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3170,4842849,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",130,1114," --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3171,4842920,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",124,1120,"SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3172,4842939,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",92,1152,"SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3173,4842995,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",68,1176,"SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3174,4842996,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",40,1204,"SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3175,4842996,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",22,1222,"SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3176,4843056,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",20,1224,"\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3177,4843073,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",2,1242,"/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3178,4843140,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1,1243,"!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3179,4843154,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",20,1224,"\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3180,4843516,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,1244,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,selection_mouse +3181,4844480,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,selection_command +3182,4860898,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",367,0,"#SBATCH --job-name=train_tokenizer_lr_sweep_5e-5\n",shellscript,content +3183,4860898,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",416,49,"",shellscript,content +3184,4861841,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",973,0," --max_lr=5e-5 \\n",shellscript,content +3185,4861864,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",993,20,"",shellscript,content +3186,4862099,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1074,0," --name=tokenizer-lr-sweep-5e-5-$slurm_job_id \\n",shellscript,content +3187,4862130,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1125,0," --tags tokenizer lr-sweep 5e-5 \\n",shellscript,content +3188,4862185,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1162,88,"",shellscript,content +3189,4868510,"TERMINAL",0,0,"Step 722, loss: 5.520607948303223\r\nStep 723, loss: 5.279177188873291\r\nStep 724, loss: 5.26333475112915\r\nStep 725, loss: 5.184632778167725\r\nStep 726, loss: 5.329742431640625\r\nStep 727, loss: 5.339062213897705\r\nStep 728, loss: 5.266631603240967\r\nStep 729, loss: 5.203587055206299\r\nStep 730, loss: 5.071401119232178\r\nStep 731, loss: 5.185279846191406\r\nStep 732, loss: 5.403427600860596\r\nStep 733, loss: 5.377686500549316\r\nStep 734, loss: 5.246264934539795\r\nStep 735, loss: 5.169823169708252\r\nStep 736, loss: 5.027594566345215\r\nStep 737, loss: 5.25385856628418\r\nStep 738, loss: 5.292280197143555\r\nStep 739, loss: 5.387245178222656\r\nStep 740, loss: 5.163019180297852\r\nStep 741, loss: 5.0033721923828125\r\nStep 742, loss: 5.450179576873779\r\nStep 743, loss: 5.4586567878723145\r\nStep 744, loss: 5.1958513259887695\r\nStep 745, loss: 5.50988245010376\r\nStep 746, loss: 5.329402446746826\r\nStep 747, loss: 5.505457878112793\r\nStep 748, loss: 5.3953537940979\r\nStep 749, loss: 5.421640396118164\r\nStep 750, loss: 4.844815731048584\r\nStep 751, loss: 5.168392658233643\r\nStep 752, loss: 5.35214900970459\r\nStep 753, loss: 5.326869964599609\r\nStep 754, loss: 5.134158134460449\r\nStep 755, loss: 4.939837455749512\r\nStep 756, loss: 5.239572525024414\r\nStep 757, loss: 5.154784202575684\r\nStep 758, loss: 5.339284420013428\r\nStep 759, loss: 5.508555889129639\r\nStep 760, loss: 5.005982398986816\r\nStep 761, loss: 5.133974075317383\r\nStep 762, loss: 5.095065593719482\r\nStep 763, loss: 5.198653697967529\r\nStep 764, loss: 5.1999616622924805\r\nStep 765, loss: 5.3742570877075195\r\nStep 766, loss: 5.18704891204834\r\nStep 767, loss: 5.39532995223999\r\nStep 768, loss: 5.305274963378906\r\nStep 769, loss: 5.349462509155273\r\nStep 770, loss: 5.104123592376709\r\nStep 771, loss: 5.199271202087402\r\nStep 772, loss: 5.092228889465332\r\nStep 773, loss: 5.3029913902282715\r\nStep 774, loss: 5.354883193969727\r\nStep 775, loss: 5.052666187286377\r\nStep 776, loss: 5.35020637512207\r\nStep 777, loss: 5.49293851852417\r\nStep 778, loss: 5.183391094207764\r\nStep 779, loss: 5.375885009765625\r\nStep 780, loss: 5.115752220153809\r\nStep 781, loss: 5.134352684020996\r\nStep 782, loss: 4.961171627044678\r\nStep 783, loss: 5.193238735198975\r\nStep 784, loss: 5.077812194824219\r\nStep 785, loss: 5.231456756591797\r\nStep 786, loss: 5.293372631072998\r\nStep 787, loss: 5.198167324066162\r\nStep 788, loss: 5.177676200866699\r\nStep 789, loss: 5.366830825805664\r\nStep 790, loss: 5.297807216644287\r\nStep 791, loss: 5.202820777893066\r\nStep 792, loss: 5.1975321769714355\r\nStep 793, loss: 5.511959075927734\r\nStep 794, loss: 5.435663223266602\r\nStep 795, loss: 5.210607528686523\r\nStep 796, loss: 5.22182035446167\r\nStep 797, loss: 5.280124187469482\r\nStep 798, loss: 5.483152389526367\r\nStep 799, loss: 5.372382640838623\r\nStep 800, loss: 5.3446364402771\r\nStep 801, loss: 5.2136759757995605\r\nStep 802, loss: 5.432462215423584\r\nStep 803, loss: 5.094836235046387\r\nStep 804, loss: 5.033899307250977\r\nStep 805, loss: 5.34099006652832\r\nStep 806, loss: 4.993969917297363\r\nStep 807, loss: 5.139680862426758\r\nStep 808, loss: 5.169064998626709\r\nStep 809, loss: 5.114250183105469\r\nStep 810, loss: 5.067643165588379\r\nStep 811, loss: 5.102647304534912\r\nStep 812, loss: 4.900321006774902\r\nStep 813, loss: 5.187741756439209\r\nStep 814, loss: 5.1775970458984375\r\nStep 815, loss: 5.301947593688965\r\nStep 816, loss: 5.487864971160889\r\nStep 817, loss: 5.195375442504883\r\nStep 818, loss: 5.214444160461426\r\nStep 819, loss: 5.186010837554932\r\nStep 820, loss: 5.233945846557617\r\nStep 821, loss: 5.107511520385742\r\nStep 822, loss: 5.087286472320557\r\nStep 823, loss: 5.018490314483643\r\nStep 824, loss: 5.2226762771606445\r\nStep 825, loss: 5.2764387130737305\r\nStep 826, loss: 5.0772833824157715\r\nStep 827, loss: 5.010648250579834\r\nStep 828, loss: 5.3085856437683105\r\nStep 829, loss: 5.063631057739258\r\nStep 830, loss: 5.158314228057861\r\nStep 831, loss: 5.056264400482178\r\nStep 832, loss: 5.233462810516357\r\nStep 833, loss: 5.12919282913208\r\nStep 834, loss: 5.364454746246338\r\nStep 835, loss: 5.389515399932861\r\nStep 836, loss: 5.048567771911621\r\nStep 837, loss: 5.149641990661621\r\nStep 838, loss: 5.039944171905518\r\nStep 839, loss: 5.245511531829834\r\nStep 840, loss: 5.165370464324951\r\nStep 841, loss: 5.112173557281494\r\nStep 842, loss: 4.9662394523620605\r\nStep 843, loss: 5.026354789733887\r\nStep 844, loss: 5.0263519287109375\r\nStep 845, loss: 5.264150619506836\r\nStep 846, loss: 5.310177803039551\r\nStep 847, loss: 5.203787803649902\r\nStep 848, loss: 5.234409332275391\r\nStep 849, loss: 5.48568868637085\r\nStep 850, loss: 4.754512310028076\r\nStep 851, loss: 5.181948661804199\r\nStep 852, loss: 5.177431106567383\r\nStep 853, loss: 5.1184773445129395\r\nStep 854, loss: 5.161408424377441\r\nStep 855, loss: 5.284380912780762\r\nStep 856, loss: 5.365041255950928\r\nStep 857, loss: 4.994603633880615\r\nStep 858, loss: 5.119741916656494\r\nStep 859, loss: 5.4332475662231445\r\nStep 860, loss: 4.946900844573975\r\nStep 861, loss: 5.184002876281738\r\nStep 862, loss: 5.185951232910156\r\nStep 863, loss: 5.291336536407471\r\nStep 864, loss: 4.999337196350098\r\nStep 865, loss: 5.166353702545166\r\nStep 866, loss: 5.531929016113281\r\nStep 867, loss: 5.285906791687012\r\nStep 868, loss: 5.155843734741211\r\nStep 869, loss: 5.214123725891113\r\nStep 870, loss: 5.001017093658447\r\nStep 871, loss: 5.293190956115723\r\nStep 872, loss: 5.077353477478027\r\nStep 873, loss: 5.047084331512451\r\nStep 874, loss: 4.81112003326416\r\nStep 875, loss: 5.0530524253845215\r\nStep 876, loss: 5.3802690505981445\r\nStep 877, loss: 5.140942573547363\r\nStep 878, loss: 5.007487773895264\r\nStep 879, loss: 4.9970245361328125\r\nStep 880, loss: 5.022059440612793\r\nStep 881, loss: 4.9126386642456055\r\nStep 882, loss: 5.345889568328857\r\nStep 883, loss: 5.15145206451416\r\nStep 884, loss: 5.181049346923828\r\nStep 885, loss: 4.934937000274658\r\nStep 886, loss: 4.96279239654541\r\nStep 887, loss: 4.8947906494140625\r\nStep 888, loss: 5.221316337585449\r\nStep 889, loss: 4.922238349914551\r\nStep 890, loss: 5.096713542938232\r\nStep 891, loss: 5.419372081756592\r\nStep 892, loss: 5.152041435241699\r\nStep 893, loss: 4.9967474937438965\r\nStep 894, loss: 5.041221618652344\r\nStep 895, loss: 5.1730804443359375\r\nStep 896, loss: 5.100033283233643\r\nStep 897, loss: 5.236809253692627\r\nStep 898, loss: 5.174713611602783\r\nStep 899, loss: 4.951124668121338\r\nStep 900, loss: 4.98465633392334\r\nStep 901, loss: 5.190182209014893\r\nStep 902, loss: 5.194697380065918\r\nStep 903, loss: 5.08566951751709\r\nStep 904, loss: 5.178715229034424\r\nStep 905, loss: 5.028305530548096\r\nStep 906, loss: 5.365531921386719\r\nStep 907, loss: 5.162856578826904\r\nStep 908, loss: 5.212813854217529\r\nStep 909, loss: 4.955593109130859\r\nStep 910, loss: 4.896980285644531\r\nStep 911, loss: 5.0863165855407715\r\nStep 912, loss: 5.099323749542236\r\nStep 913, loss: 5.119900703430176\r\nStep 914, loss: 5.37208890914917\r\nStep 915, loss: 5.117595195770264\r\nStep 916, loss: 5.058721542358398\r\nStep 917, loss: 5.050705909729004\r\nStep 918, loss: 5.170966625213623\r\nStep 919, loss: 5.2613205909729\r\nStep 920, loss: 5.0097880363464355\r\nStep 921, loss: 5.356467247009277\r\nStep 922, loss: 5.3333563804626465\r\nStep 923, loss: 4.88981294631958\r\nStep 924, loss: 5.183024883270264\r\nStep 925, loss: 5.0197858810424805\r\nStep 926, loss: 5.204542636871338\r\nStep 927, loss: 4.9824676513671875\r\nStep 928, loss: 5.0433549880981445\r\nStep 929, loss: 5.069318771362305\r\nStep 930, loss: 5.1959099769592285\r\nStep 931, loss: 5.107265472412109\r\nStep 932, loss: 5.068860054016113\r\nStep 933, loss: 5.199520587921143\r\nStep 934, loss: 4.9961018562316895\r\nStep 935, loss: 5.081637859344482\r\nStep 936, loss: 5.0613484382629395\r\nStep 937, loss: 5.130383014678955\r\nStep 938, loss: 5.093724727630615\r\nStep 939, loss: 4.736340522766113\r\nStep 940, loss: 5.0766215324401855\r\nStep 941, loss: 4.86067533493042\r\nStep 942, loss: 4.996115684509277\r\nStep 943, loss: 5.155579090118408\r\nStep 944, loss: 5.194488048553467\r\nStep 945, loss: 5.429854869842529\r\nStep 946, loss: 4.887845993041992\r\nStep 947, loss: 5.2252912521362305\r\nStep 948, loss: 5.225298881530762\r\nStep 949, loss: 5.137124061584473\r\nStep 950, loss: 5.239169120788574\r\nStep 951, loss: 5.131504058837891\r\nStep 952, loss: 5.070860385894775\r\nStep 953, loss: 5.186777591705322\r\nStep 954, loss: 4.983127593994141\r\nStep 955, loss: 4.94017219543457\r\nStep 956, loss: 5.011066913604736\r\nStep 957, loss: 5.030396461486816\r\nStep 958, loss: 5.042568206787109\r\nStep 959, loss: 4.927941799163818\r\nStep 960, loss: 5.052581310272217\r\nStep 961, loss: 4.9004340171813965\r\nStep 722, loss: 5.520607948303223\r\nStep 723, loss: 5.279177188873291\r\nStep 724, loss: 5.26333475112915\r\nStep 725, loss: 5.184632778167725\r\nStep 726, loss: 5.329742431640625\r\nStep 727, loss: 5.339062213897705\r\nStep 728, loss: 5.266631603240967\r\nStep 729, loss: 5.203587055206299\r\nStep 730, loss: 5.071401119232178\r\nStep 731, loss: 5.185279846191406\r\nStep 732, loss: 5.403427600860596\r\nStep 733, loss: 5.377686500549316\r\nStep 734, loss: 5.246264934539795\r\nStep 735, loss: 5.169823169708252\r\nStep 736, loss: 5.027594566345215\r\nStep 737, loss: 5.25385856628418\r\nStep 738, loss: 5.292280197143555\r\nStep 739, loss: 5.387245178222656\r\nStep 740, loss: 5.163019180297852\r\nStep 741, loss: 5.0033721923828125\r\nStep 742, loss: 5.450179576873779\r\nStep 743, loss: 5.4586567878723145\r\nStep 744, loss: 5.1958513259887695\r\nStep 745, loss: 5.50988245010376\r\nStep 746, loss: 5.329402446746826\r\nStep 747, loss: 5.505457878112793\r\nStep 748, loss: 5.3953537940979\r\nStep 749, loss: 5.421640396118164\r\nStep 750, loss: 4.844815731048584\r\nStep 751, loss: 5.168392658233643\r\nStep 722, loss: 5.520607948303223\r\nStep 723, loss: 5.279177188873291\r\nStep 724, loss: 5.26333475112915\r\nStep 725, loss: 5.184632778167725\r\nStep 726, loss: 5.329742431640625\r\nStep 727, loss: 5.339062213897705\r\nStep 728, loss: 5.266631603240967\r\nStep 729, loss: 5.203587055206299\r\nStep 730, loss: 5.071401119232178\r\nStep 731, loss: 5.185279846191406\r\nStep 732, loss: 5.403427600860596\r\nStep 733, loss: 5.377686500549316\r\nStep 734, loss: 5.246264934539795\r\nStep 735, loss: 5.169823169708252\r\nStep 736, loss: 5.027594566345215\r\nStep 737, loss: 5.25385856628418\r\nStep 738, loss: 5.292280197143555\r\nStep 739, loss: 5.387245178222656\r\nStep 740, loss: 5.163019180297852\r\nStep 741, loss: 5.0033721923828125\r\nStep 742, loss: 5.450179576873779\r\nStep 743, loss: 5.4586567878723145\r\nStep 744, loss: 5.1958513259887695\r\nStep 745, loss: 5.50988245010376\r\nStep 746, loss: 5.329402446746826\r\nStep 747, loss: 5.505457878112793\r\nStep 748, loss: 5.3953537940979\r\nStep 749, loss: 5.421640396118164\r\nStep 750, loss: 4.844815731048584\r\nStep 751, loss: 5.168392658233643\r\nStep 752, loss: 5.35214900970459\r\nStep 753, loss: 5.326869964599609\r\nStep 754, loss: 5.134158134460449\r\nStep 755, loss: 4.939837455749512\r\nStep 756, loss: 5.239572525024414\r\nStep 757, loss: 5.154784202575684\r\nStep 758, loss: 5.339284420013428\r\nStep 759, loss: 5.508555889129639\r\nStep 760, loss: 5.005982398986816\r\nStep 761, loss: 5.133974075317383\r\nStep 762, loss: 5.095065593719482\r\nStep 763, loss: 5.198653697967529\r\nStep 764, loss: 5.1999616622924805\r\nStep 765, loss: 5.3742570877075195\r\nStep 766, loss: 5.18704891204834\r\nStep 767, loss: 5.39532995223999\r\nStep 768, loss: 5.305274963378906\r\nStep 769, loss: 5.349462509155273\r\nStep 770, loss: 5.104123592376709\r\nStep 771, loss: 5.199271202087402\r\nStep 772, loss: 5.092228889465332\r\nStep 773, loss: 5.3029913902282715\r\nStep 774, loss: 5.354883193969727\r\nStep 775, loss: 5.052666187286377\r\nStep 776, loss: 5.35020637512207\r\nStep 777, loss: 5.49293851852417\r\nStep 778, loss: 5.183391094207764\r\nStep 779, loss: 5.375885009765625\r\nStep 780, loss: 5.115752220153809\r\nStep 781, loss: 5.134352684020996\r\nStep 752, loss: 5.35214900970459\r\nStep 753, loss: 5.326869964599609\r\nStep 754, loss: 5.134158134460449\r\nStep 755, loss: 4.939837455749512\r\nStep 756, loss: 5.239572525024414\r\nStep 757, loss: 5.154784202575684\r\nStep 758, loss: 5.339284420013428\r\nStep 759, loss: 5.508555889129639\r\nStep 760, loss: 5.005982398986816\r\nStep 761, loss: 5.133974075317383\r\nStep 762, loss: 5.095065593719482\r\nStep 763, loss: 5.198653697967529\r\nStep 764, loss: 5.1999616622924805\r\nStep 765, loss: 5.3742570877075195\r\nStep 766, loss: 5.18704891204834\r\nStep 767, loss: 5.39532995223999\r\nStep 768, loss: 5.305274963378906\r\nStep 769, loss: 5.349462509155273\r\nStep 770, loss: 5.104123592376709\r\nStep 771, loss: 5.199271202087402\r\nStep 772, loss: 5.092228889465332\r\nStep 773, loss: 5.3029913902282715\r\nStep 774, loss: 5.354883193969727\r\nStep 775, loss: 5.052666187286377\r\nStep 776, loss: 5.35020637512207\r\nStep 777, loss: 5.49293851852417\r\nStep 778, loss: 5.183391094207764\r\nStep 779, loss: 5.375885009765625\r\nStep 780, loss: 5.115752220153809\r\nStep 781, loss: 5.134352684020996\r\nStep 782, loss: 4.961171627044678\r\nStep 783, loss: 5.193238735198975\r\nStep 784, loss: 5.077812194824219\r\nStep 785, loss: 5.231456756591797\r\nStep 786, loss: 5.293372631072998\r\nStep 787, loss: 5.198167324066162\r\nStep 788, loss: 5.177676200866699\r\nStep 789, loss: 5.366830825805664\r\nStep 790, loss: 5.297807216644287\r\nStep 791, loss: 5.202820777893066\r\nStep 792, loss: 5.1975321769714355\r\nStep 793, loss: 5.511959075927734\r\nStep 794, loss: 5.435663223266602\r\nStep 795, loss: 5.210607528686523\r\nStep 796, loss: 5.22182035446167\r\nStep 797, loss: 5.280124187469482\r\nStep 798, loss: 5.483152389526367\r\nStep 799, loss: 5.372382640838623\r\nStep 800, loss: 5.3446364402771\r\nStep 801, loss: 5.2136759757995605\r\nStep 802, loss: 5.432462215423584\r\nStep 803, loss: 5.094836235046387\r\nStep 804, loss: 5.033899307250977\r\nStep 805, loss: 5.34099006652832\r\nStep 806, loss: 4.993969917297363\r\nStep 807, loss: 5.139680862426758\r\nStep 808, loss: 5.169064998626709\r\nStep 809, loss: 5.114250183105469\r\nStep 810, loss: 5.067643165588379\r\nStep 811, loss: 5.102647304534912\r\nStep 782, loss: 4.961171627044678\r\nStep 783, loss: 5.193238735198975\r\nStep 784, loss: 5.077812194824219\r\nStep 785, loss: 5.231456756591797\r\nStep 786, loss: 5.293372631072998\r\nStep 787, loss: 5.198167324066162\r\nStep 788, loss: 5.177676200866699\r\nStep 789, loss: 5.366830825805664\r\nStep 790, loss: 5.297807216644287\r\nStep 791, loss: 5.202820777893066\r\nStep 792, loss: 5.1975321769714355\r\nStep 793, loss: 5.511959075927734\r\nStep 794, loss: 5.435663223266602\r\nStep 795, loss: 5.210607528686523\r\nStep 796, loss: 5.22182035446167\r\nStep 797, loss: 5.280124187469482\r\nStep 798, loss: 5.483152389526367\r\nStep 799, loss: 5.372382640838623\r\nStep 800, loss: 5.3446364402771\r\nStep 801, loss: 5.2136759757995605\r\nStep 802, loss: 5.432462215423584\r\nStep 803, loss: 5.094836235046387\r\nStep 804, loss: 5.033899307250977\r\nStep 805, loss: 5.34099006652832\r\nStep 806, loss: 4.993969917297363\r\nStep 807, loss: 5.139680862426758\r\nStep 808, loss: 5.169064998626709\r\nStep 809, loss: 5.114250183105469\r\nStep 810, loss: 5.067643165588379\r\nStep 811, loss: 5.102647304534912\r\nStep 812, loss: 4.900321006774902\r\nStep 813, loss: 5.187741756439209\r\nStep 814, loss: 5.1775970458984375\r\nStep 815, loss: 5.301947593688965\r\nStep 816, loss: 5.487864971160889\r\nStep 817, loss: 5.195375442504883\r\nStep 818, loss: 5.214444160461426\r\nStep 819, loss: 5.186010837554932\r\nStep 820, loss: 5.233945846557617\r\nStep 821, loss: 5.107511520385742\r\nStep 822, loss: 5.087286472320557\r\nStep 823, loss: 5.018490314483643\r\nStep 824, loss: 5.2226762771606445\r\nStep 825, loss: 5.2764387130737305\r\nStep 826, loss: 5.0772833824157715\r\nStep 827, loss: 5.010648250579834\r\nStep 828, loss: 5.3085856437683105\r\nStep 829, loss: 5.063631057739258\r\nStep 830, loss: 5.158314228057861\r\nStep 831, loss: 5.056264400482178\r\nStep 832, loss: 5.233462810516357\r\nStep 833, loss: 5.12919282913208\r\nStep 834, loss: 5.364454746246338\r\nStep 835, loss: 5.389515399932861\r\nStep 836, loss: 5.048567771911621\r\nStep 837, loss: 5.149641990661621\r\nStep 838, loss: 5.039944171905518\r\nStep 839, loss: 5.245511531829834\r\nStep 840, loss: 5.165370464324951\r\nStep 812, loss: 4.900321006774902\r\nStep 813, loss: 5.187741756439209\r\nStep 814, loss: 5.1775970458984375\r\nStep 815, loss: 5.301947593688965\r\nStep 816, loss: 5.487864971160889\r\nStep 817, loss: 5.195375442504883\r\nStep 818, loss: 5.214444160461426\r\nStep 819, loss: 5.186010837554932\r\nStep 820, loss: 5.233945846557617\r\nStep 821, loss: 5.107511520385742\r\nStep 822, loss: 5.087286472320557\r\nStep 823, loss: 5.018490314483643\r\nStep 824, loss: 5.2226762771606445\r\nStep 825, loss: 5.2764387130737305\r\nStep 826, loss: 5.0772833824157715\r\nStep 827, loss: 5.010648250579834\r\nStep 828, loss: 5.3085856437683105\r\nStep 829, loss: 5.063631057739258\r\nStep 830, loss: 5.158314228057861\r\nStep 831, loss: 5.056264400482178\r\nStep 832, loss: 5.233462810516357\r\nStep 833, loss: 5.12919282913208\r\nStep 834, loss: 5.364454746246338\r\nStep 835, loss: 5.389515399932861\r\nStep 836, loss: 5.048567771911621\r\nStep 837, loss: 5.149641990661621\r\nStep 838, loss: 5.039944171905518\r\nStep 839, loss: 5.245511531829834\r\nStep 840, loss: 5.165370464324951\r\nStep 841, loss: 5.112173557281494\r\nStep 841, loss: 5.112173557281494\r\nStep 722, loss: 5.520607948303223\r\nStep 723, loss: 5.279177188873291\r\nStep 724, loss: 5.26333475112915\r\nStep 725, loss: 5.184632778167725\r\nStep 726, loss: 5.329742431640625\r\nStep 727, loss: 5.339062213897705\r\nStep 728, loss: 5.266631603240967\r\nStep 729, loss: 5.203587055206299\r\nStep 730, loss: 5.071401119232178\r\nStep 731, loss: 5.185279846191406\r\nStep 732, loss: 5.403427600860596\r\nStep 733, loss: 5.377686500549316\r\nStep 734, loss: 5.246264934539795\r\nStep 735, loss: 5.169823169708252\r\nStep 736, loss: 5.027594566345215\r\nStep 737, loss: 5.25385856628418\r\nStep 738, loss: 5.292280197143555\r\nStep 739, loss: 5.387245178222656\r\nStep 740, loss: 5.163019180297852\r\nStep 741, loss: 5.0033721923828125\r\nStep 742, loss: 5.450179576873779\r\nStep 743, loss: 5.4586567878723145\r\nStep 744, loss: 5.1958513259887695\r\nStep 745, loss: 5.50988245010376\r\nStep 746, loss: 5.329402446746826\r\nStep 747, loss: 5.505457878112793\r\nStep 748, loss: 5.3953537940979\r\nStep 749, loss: 5.421640396118164\r\nStep 750, loss: 4.844815731048584\r\nStep 751, loss: 5.168392658233643\r\nStep 842, loss: 4.9662394523620605\r\nStep 843, loss: 5.026354789733887\r\nStep 844, loss: 5.0263519287109375\r\nStep 845, loss: 5.264150619506836\r\nStep 846, loss: 5.310177803039551\r\nStep 847, loss: 5.203787803649902\r\nStep 848, loss: 5.234409332275391\r\nStep 849, loss: 5.48568868637085\r\nStep 850, loss: 4.754512310028076\r\nStep 851, loss: 5.181948661804199\r\nStep 852, loss: 5.177431106567383\r\nStep 853, loss: 5.1184773445129395\r\nStep 854, loss: 5.161408424377441\r\nStep 855, loss: 5.284380912780762\r\nStep 856, loss: 5.365041255950928\r\nStep 857, loss: 4.994603633880615\r\nStep 858, loss: 5.119741916656494\r\nStep 859, loss: 5.4332475662231445\r\nStep 860, loss: 4.946900844573975\r\nStep 861, loss: 5.184002876281738\r\nStep 862, loss: 5.185951232910156\r\nStep 863, loss: 5.291336536407471\r\nStep 864, loss: 4.999337196350098\r\nStep 865, loss: 5.166353702545166\r\nStep 866, loss: 5.531929016113281\r\nStep 867, loss: 5.285906791687012\r\nStep 868, loss: 5.155843734741211\r\nStep 869, loss: 5.214123725891113\r\nStep 870, loss: 5.001017093658447\r\nStep 871, loss: 5.293190956115723\r\nStep 752, loss: 5.35214900970459\r\nStep 753, loss: 5.326869964599609\r\nStep 754, loss: 5.134158134460449\r\nStep 755, loss: 4.939837455749512\r\nStep 756, loss: 5.239572525024414\r\nStep 757, loss: 5.154784202575684\r\nStep 758, loss: 5.339284420013428\r\nStep 759, loss: 5.508555889129639\r\nStep 760, loss: 5.005982398986816\r\nStep 761, loss: 5.133974075317383\r\nStep 762, loss: 5.095065593719482\r\nStep 763, loss: 5.198653697967529\r\nStep 764, loss: 5.1999616622924805\r\nStep 765, loss: 5.3742570877075195\r\nStep 766, loss: 5.18704891204834\r\nStep 767, loss: 5.39532995223999\r\nStep 768, loss: 5.305274963378906\r\nStep 769, loss: 5.349462509155273\r\nStep 770, loss: 5.104123592376709\r\nStep 771, loss: 5.199271202087402\r\nStep 772, loss: 5.092228889465332\r\nStep 773, loss: 5.3029913902282715\r\nStep 774, loss: 5.354883193969727\r\nStep 775, loss: 5.052666187286377\r\nStep 776, loss: 5.35020637512207\r\nStep 777, loss: 5.49293851852417\r\nStep 778, loss: 5.183391094207764\r\nStep 779, loss: 5.375885009765625\r\nStep 780, loss: 5.115752220153809\r\nStep 781, loss: 5.134352684020996\r\nStep 872, loss: 5.077353477478027\r\nStep 873, loss: 5.047084331512451\r\nStep 874, loss: 4.81112003326416\r\nStep 875, loss: 5.0530524253845215\r\nStep 876, loss: 5.3802690505981445\r\nStep 877, loss: 5.140942573547363\r\nStep 878, loss: 5.007487773895264\r\nStep 879, loss: 4.9970245361328125\r\nStep 880, loss: 5.022059440612793\r\nStep 881, loss: 4.9126386642456055\r\nStep 882, loss: 5.345889568328857\r\nStep 883, loss: 5.15145206451416\r\nStep 884, loss: 5.181049346923828\r\nStep 885, loss: 4.934937000274658\r\nStep 886, loss: 4.96279239654541\r\nStep 887, loss: 4.8947906494140625\r\nStep 888, loss: 5.221316337585449\r\nStep 889, loss: 4.922238349914551\r\nStep 890, loss: 5.096713542938232\r\nStep 891, loss: 5.419372081756592\r\nStep 892, loss: 5.152041435241699\r\nStep 893, loss: 4.9967474937438965\r\nStep 894, loss: 5.041221618652344\r\nStep 895, loss: 5.1730804443359375\r\nStep 896, loss: 5.100033283233643\r\nStep 897, loss: 5.236809253692627\r\nStep 898, loss: 5.174713611602783\r\nStep 899, loss: 4.951124668121338\r\nStep 900, loss: 4.98465633392334\r\nStep 901, loss: 5.190182209014893\r\nStep 782, loss: 4.961171627044678\r\nStep 783, loss: 5.193238735198975\r\nStep 784, loss: 5.077812194824219\r\nStep 785, loss: 5.231456756591797\r\nStep 786, loss: 5.293372631072998\r\nStep 787, loss: 5.198167324066162\r\nStep 788, loss: 5.177676200866699\r\nStep 789, loss: 5.366830825805664\r\nStep 790, loss: 5.297807216644287\r\nStep 791, loss: 5.202820777893066\r\nStep 792, loss: 5.1975321769714355\r\nStep 793, loss: 5.511959075927734\r\nStep 794, loss: 5.435663223266602\r\nStep 795, loss: 5.210607528686523\r\nStep 796, loss: 5.22182035446167\r\nStep 797, loss: 5.280124187469482\r\nStep 798, loss: 5.483152389526367\r\nStep 799, loss: 5.372382640838623\r\nStep 800, loss: 5.3446364402771\r\nStep 801, loss: 5.2136759757995605\r\nStep 802, loss: 5.432462215423584\r\nStep 803, loss: 5.094836235046387\r\nStep 804, loss: 5.033899307250977\r\nStep 805, loss: 5.34099006652832\r\nStep 806, loss: 4.993969917297363\r\nStep 807, loss: 5.139680862426758\r\nStep 808, loss: 5.169064998626709\r\nStep 809, loss: 5.114250183105469\r\nStep 810, loss: 5.067643165588379\r\nStep 811, loss: 5.102647304534912\r\nStep 902, loss: 5.194697380065918\r\nStep 903, loss: 5.08566951751709\r\nStep 904, loss: 5.178715229034424\r\nStep 905, loss: 5.028305530548096\r\nStep 906, loss: 5.365531921386719\r\nStep 907, loss: 5.162856578826904\r\nStep 908, loss: 5.212813854217529\r\nStep 909, loss: 4.955593109130859\r\nStep 910, loss: 4.896980285644531\r\nStep 911, loss: 5.0863165855407715\r\nStep 912, loss: 5.099323749542236\r\nStep 913, loss: 5.119900703430176\r\nStep 914, loss: 5.37208890914917\r\nStep 915, loss: 5.117595195770264\r\nStep 916, loss: 5.058721542358398\r\nStep 917, loss: 5.050705909729004\r\nStep 918, loss: 5.170966625213623\r\nStep 919, loss: 5.2613205909729\r\nStep 920, loss: 5.0097880363464355\r\nStep 921, loss: 5.356467247009277\r\nStep 922, loss: 5.3333563804626465\r\nStep 923, loss: 4.88981294631958\r\nStep 924, loss: 5.183024883270264\r\nStep 925, loss: 5.0197858810424805\r\nStep 926, loss: 5.204542636871338\r\nStep 927, loss: 4.9824676513671875\r\nStep 928, loss: 5.0433549880981445\r\nStep 929, loss: 5.069318771362305\r\nStep 930, loss: 5.1959099769592285\r\nStep 931, loss: 5.107265472412109\r\nStep 812, loss: 4.900321006774902\r\nStep 813, loss: 5.187741756439209\r\nStep 814, loss: 5.1775970458984375\r\nStep 815, loss: 5.301947593688965\r\nStep 816, loss: 5.487864971160889\r\nStep 817, loss: 5.195375442504883\r\nStep 818, loss: 5.214444160461426\r\nStep 819, loss: 5.186010837554932\r\nStep 820, loss: 5.233945846557617\r\nStep 821, loss: 5.107511520385742\r\nStep 822, loss: 5.087286472320557\r\nStep 823, loss: 5.018490314483643\r\nStep 824, loss: 5.2226762771606445\r\nStep 825, loss: 5.2764387130737305\r\nStep 826, loss: 5.0772833824157715\r\nStep 827, loss: 5.010648250579834\r\nStep 828, loss: 5.3085856437683105\r\nStep 829, loss: 5.063631057739258\r\nStep 830, loss: 5.158314228057861\r\nStep 831, loss: 5.056264400482178\r\nStep 832, loss: 5.233462810516357\r\nStep 833, loss: 5.12919282913208\r\nStep 834, loss: 5.364454746246338\r\nStep 835, loss: 5.389515399932861\r\nStep 836, loss: 5.048567771911621\r\nStep 837, loss: 5.149641990661621\r\nStep 838, loss: 5.039944171905518\r\nStep 839, loss: 5.245511531829834\r\nStep 840, loss: 5.165370464324951\r\nStep 932, loss: 5.068860054016113\r\nStep 933, loss: 5.199520587921143\r\nStep 934, loss: 4.9961018562316895\r\nStep 935, loss: 5.081637859344482\r\nStep 936, loss: 5.0613484382629395\r\nStep 937, loss: 5.130383014678955\r\nStep 938, loss: 5.093724727630615\r\nStep 939, loss: 4.736340522766113\r\nStep 940, loss: 5.0766215324401855\r\nStep 941, loss: 4.86067533493042\r\nStep 942, loss: 4.996115684509277\r\nStep 943, loss: 5.155579090118408\r\nStep 944, loss: 5.194488048553467\r\nStep 945, loss: 5.429854869842529\r\nStep 946, loss: 4.887845993041992\r\nStep 947, loss: 5.2252912521362305\r\nStep 948, loss: 5.225298881530762\r\nStep 949, loss: 5.137124061584473\r\nStep 950, loss: 5.239169120788574\r\nStep 951, loss: 5.131504058837891\r\nStep 952, loss: 5.070860385894775\r\nStep 953, loss: 5.186777591705322\r\nStep 954, loss: 4.983127593994141\r\nStep 955, loss: 4.94017219543457\r\nStep 956, loss: 5.011066913604736\r\nStep 957, loss: 5.030396461486816\r\nStep 958, loss: 5.042568206787109\r\nStep 959, loss: 4.927941799163818\r\nStep 960, loss: 5.052581310272217\r\nStep 961, loss: 4.9004340171813965\r\nStep 841, loss: 5.112173557281494\r\nStep 842, loss: 4.9662394523620605\r\nStep 843, loss: 5.026354789733887\r\nStep 844, loss: 5.0263519287109375\r\nStep 845, loss: 5.264150619506836\r\nStep 846, loss: 5.310177803039551\r\nStep 847, loss: 5.203787803649902\r\nStep 848, loss: 5.234409332275391\r\nStep 849, loss: 5.48568868637085\r\nStep 850, loss: 4.754512310028076\r\nStep 851, loss: 5.181948661804199\r\nStep 852, loss: 5.177431106567383\r\nStep 853, loss: 5.1184773445129395\r\nStep 854, loss: 5.161408424377441\r\nStep 855, loss: 5.284380912780762\r\nStep 856, loss: 5.365041255950928\r\nStep 857, loss: 4.994603633880615\r\nStep 858, loss: 5.119741916656494\r\nStep 859, loss: 5.4332475662231445\r\nStep 860, loss: 4.946900844573975\r\nStep 861, loss: 5.184002876281738\r\nStep 862, loss: 5.185951232910156\r\nStep 863, loss: 5.291336536407471\r\nStep 864, loss: 4.999337196350098\r\nStep 865, loss: 5.166353702545166\r\nStep 866, loss: 5.531929016113281\r\nStep 867, loss: 5.285906791687012\r\nStep 868, loss: 5.155843734741211\r\nStep 869, loss: 5.214123725891113\r\nStep 870, loss: 5.001017093658447\r\nStep 871, loss: 5.293190956115723\r\nStep 872, loss: 5.077353477478027\r\nStep 873, loss: 5.047084331512451\r\nStep 874, loss: 4.81112003326416\r\nStep 875, loss: 5.0530524253845215\r\nStep 876, loss: 5.3802690505981445\r\nStep 877, loss: 5.140942573547363\r\nStep 878, loss: 5.007487773895264\r\nStep 879, loss: 4.9970245361328125\r\nStep 880, loss: 5.022059440612793\r\nStep 881, loss: 4.9126386642456055\r\nStep 882, loss: 5.345889568328857\r\nStep 883, loss: 5.15145206451416\r\nStep 884, loss: 5.181049346923828\r\nStep 885, loss: 4.934937000274658\r\nStep 886, loss: 4.96279239654541\r\nStep 887, loss: 4.8947906494140625\r\nStep 888, loss: 5.221316337585449\r\nStep 889, loss: 4.922238349914551\r\nStep 890, loss: 5.096713542938232\r\nStep 891, loss: 5.419372081756592\r\nStep 892, loss: 5.152041435241699\r\nStep 893, loss: 4.9967474937438965\r\nStep 894, loss: 5.041221618652344\r\nStep 895, loss: 5.1730804443359375\r\nStep 896, loss: 5.100033283233643\r\nStep 897, loss: 5.236809253692627\r\nStep 898, loss: 5.174713611602783\r\nStep 899, loss: 4.951124668121338\r\nStep 900, loss: 4.98465633392334\r\nStep 901, loss: 5.190182209014893\r\nStep 902, loss: 5.194697380065918\r\nStep 903, loss: 5.08566951751709\r\nStep 904, loss: 5.178715229034424\r\nStep 905, loss: 5.028305530548096\r\nStep 906, loss: 5.365531921386719\r\nStep 907, loss: 5.162856578826904\r\nStep 908, loss: 5.212813854217529\r\nStep 909, loss: 4.955593109130859\r\nStep 910, loss: 4.896980285644531\r\nStep 911, loss: 5.0863165855407715\r\nStep 912, loss: 5.099323749542236\r\nStep 913, loss: 5.119900703430176\r\nStep 914, loss: 5.37208890914917\r\nStep 915, loss: 5.117595195770264\r\nStep 916, loss: 5.058721542358398\r\nStep 917, loss: 5.050705909729004\r\nStep 918, loss: 5.170966625213623\r\nStep 919, loss: 5.2613205909729\r\nStep 920, loss: 5.0097880363464355\r\nStep 921, loss: 5.356467247009277\r\nStep 922, loss: 5.3333563804626465\r\nStep 923, loss: 4.88981294631958\r\nStep 924, loss: 5.183024883270264\r\nStep 925, loss: 5.0197858810424805\r\nStep 926, loss: 5.204542636871338\r\nStep 927, loss: 4.9824676513671875\r\nStep 928, loss: 5.0433549880981445\r\nStep 929, loss: 5.069318771362305\r\nStep 930, loss: 5.1959099769592285\r\nStep 931, loss: 5.107265472412109\r\nStep 932, loss: 5.068860054016113\r\nStep 933, loss: 5.199520587921143\r\nStep 934, loss: 4.9961018562316895\r\nStep 935, loss: 5.081637859344482\r\nStep 936, loss: 5.0613484382629395\r\nStep 937, loss: 5.130383014678955\r\nStep 938, loss: 5.093724727630615\r\nStep 939, loss: 4.736340522766113\r\nStep 940, loss: 5.0766215324401855\r\nStep 941, loss: 4.86067533493042\r\nStep 942, loss: 4.996115684509277\r\nStep 943, loss: 5.155579090118408\r\nStep 944, loss: 5.194488048553467\r\nStep 945, loss: 5.429854869842529\r\nStep 946, loss: 4.887845993041992\r\nStep 947, loss: 5.2252912521362305\r\nStep 948, loss: 5.225298881530762\r\nStep 949, loss: 5.137124061584473\r\nStep 950, loss: 5.239169120788574\r\nStep 951, loss: 5.131504058837891\r\nStep 952, loss: 5.070860385894775\r\nStep 953, loss: 5.186777591705322\r\nStep 954, loss: 4.983127593994141\r\nStep 955, loss: 4.94017219543457\r\nStep 956, loss: 5.011066913604736\r\nStep 957, loss: 5.030396461486816\r\nStep 958, loss: 5.042568206787109\r\nStep 959, loss: 4.927941799163818\r\nStep 960, loss: 5.052581310272217\r\nStep 961, loss: 4.9004340171813965\r\nStep 842, loss: 4.9662394523620605\r\nStep 843, loss: 5.026354789733887\r\nStep 844, loss: 5.0263519287109375\r\nStep 845, loss: 5.264150619506836\r\nStep 846, loss: 5.310177803039551\r\nStep 847, loss: 5.203787803649902\r\nStep 848, loss: 5.234409332275391\r\nStep 849, loss: 5.48568868637085\r\nStep 850, loss: 4.754512310028076\r\nStep 851, loss: 5.181948661804199\r\nStep 852, loss: 5.177431106567383\r\nStep 853, loss: 5.1184773445129395\r\nStep 854, loss: 5.161408424377441\r\nStep 855, loss: 5.284380912780762\r\nStep 856, loss: 5.365041255950928\r\nStep 857, loss: 4.994603633880615\r\nStep 858, loss: 5.119741916656494\r\nStep 859, loss: 5.4332475662231445\r\nStep 860, loss: 4.946900844573975\r\nStep 861, loss: 5.184002876281738\r\nStep 862, loss: 5.185951232910156\r\nStep 863, loss: 5.291336536407471\r\nStep 864, loss: 4.999337196350098\r\nStep 865, loss: 5.166353702545166\r\nStep 866, loss: 5.531929016113281\r\nStep 867, loss: 5.285906791687012\r\nStep 868, loss: 5.155843734741211\r\nStep 869, loss: 5.214123725891113\r\nStep 870, loss: 5.001017093658447\r\nStep 871, loss: 5.293190956115723\r\nStep 872, loss: 5.077353477478027\r\nStep 873, loss: 5.047084331512451\r\nStep 874, loss: 4.81112003326416\r\nStep 875, loss: 5.0530524253845215\r\nStep 876, loss: 5.3802690505981445\r\nStep 877, loss: 5.140942573547363\r\nStep 878, loss: 5.007487773895264\r\nStep 879, loss: 4.9970245361328125\r\nStep 880, loss: 5.022059440612793\r\nStep 881, loss: 4.9126386642456055\r\nStep 882, loss: 5.345889568328857\r\nStep 883, loss: 5.15145206451416\r\nStep 884, loss: 5.181049346923828\r\nStep 885, loss: 4.934937000274658\r\nStep 886, loss: 4.96279239654541\r\nStep 887, loss: 4.8947906494140625\r\nStep 888, loss: 5.221316337585449\r\nStep 889, loss: 4.922238349914551\r\nStep 890, loss: 5.096713542938232\r\nStep 891, loss: 5.419372081756592\r\nStep 892, loss: 5.152041435241699\r\nStep 893, loss: 4.9967474937438965\r\nStep 894, loss: 5.041221618652344\r\nStep 895, loss: 5.1730804443359375\r\nStep 896, loss: 5.100033283233643\r\nStep 897, loss: 5.236809253692627\r\nStep 898, loss: 5.174713611602783\r\nStep 899, loss: 4.951124668121338\r\nStep 900, loss: 4.98465633392334\r\nStep 901, loss: 5.190182209014893\r\nStep 902, loss: 5.194697380065918\r\nStep 903, loss: 5.08566951751709\r\nStep 904, loss: 5.178715229034424\r\nStep 905, loss: 5.028305530548096\r\nStep 906, loss: 5.365531921386719\r\nStep 907, loss: 5.162856578826904\r\nStep 908, loss: 5.212813854217529\r\nStep 909, loss: 4.955593109130859\r\nStep 910, loss: 4.896980285644531\r\nStep 911, loss: 5.0863165855407715\r\nStep 912, loss: 5.099323749542236\r\nStep 913, loss: 5.119900703430176\r\nStep 914, loss: 5.37208890914917\r\nStep 915, loss: 5.117595195770264\r\nStep 916, loss: 5.058721542358398\r\nStep 917, loss: 5.050705909729004\r\nStep 918, loss: 5.170966625213623\r\nStep 919, loss: 5.2613205909729\r\nStep 920, loss: 5.0097880363464355\r\nStep 921, loss: 5.356467247009277\r\nStep 922, loss: 5.3333563804626465\r\nStep 923, loss: 4.88981294631958\r\nStep 924, loss: 5.183024883270264\r\nStep 925, loss: 5.0197858810424805\r\nStep 926, loss: 5.204542636871338\r\nStep 927, loss: 4.9824676513671875\r\nStep 928, loss: 5.0433549880981445\r\nStep 929, loss: 5.069318771362305\r\nStep 930, loss: 5.1959099769592285\r\nStep 931, loss: 5.107265472412109\r\nStep 932, loss: 5.068860054016113\r\nStep 933, loss: 5.199520587921143\r\nStep 934, loss: 4.9961018562316895\r\nStep 935, loss: 5.081637859344482\r\nStep 936, loss: 5.0613484382629395\r\nStep 937, loss: 5.130383014678955\r\nStep 938, loss: 5.093724727630615\r\nStep 939, loss: 4.736340522766113\r\nStep 940, loss: 5.0766215324401855\r\nStep 941, loss: 4.86067533493042\r\nStep 942, loss: 4.996115684509277\r\nStep 943, loss: 5.155579090118408\r\nStep 944, loss: 5.194488048553467\r\nStep 945, loss: 5.429854869842529\r\nStep 946, loss: 4.887845993041992\r\nStep 947, loss: 5.2252912521362305\r\nStep 948, loss: 5.225298881530762\r\nStep 949, loss: 5.137124061584473\r\nStep 950, loss: 5.239169120788574\r\nStep 951, loss: 5.131504058837891\r\nStep 952, loss: 5.070860385894775\r\nStep 953, loss: 5.186777591705322\r\nStep 954, loss: 4.983127593994141\r\nStep 955, loss: 4.94017219543457\r\nStep 956, loss: 5.011066913604736\r\nStep 957, loss: 5.030396461486816\r\nStep 958, loss: 5.042568206787109\r\nStep 959, loss: 4.927941799163818\r\nStep 960, loss: 5.052581310272217\r\nStep 961, loss: 4.9004340171813965\r\nStep 722, loss: 5.520607948303223\r\nStep 723, loss: 5.279177188873291\r\nStep 724, loss: 5.26333475112915\r\nStep 725, loss: 5.184632778167725\r\nStep 726, loss: 5.329742431640625\r\nStep 727, loss: 5.339062213897705\r\nStep 728, loss: 5.266631603240967\r\nStep 729, loss: 5.203587055206299\r\nStep 730, loss: 5.071401119232178\r\nStep 731, loss: 5.185279846191406\r\nStep 732, loss: 5.403427600860596\r\nStep 733, loss: 5.377686500549316\r\nStep 734, loss: 5.246264934539795\r\nStep 735, loss: 5.169823169708252\r\nStep 736, loss: 5.027594566345215\r\nStep 737, loss: 5.25385856628418\r\nStep 738, loss: 5.292280197143555\r\nStep 739, loss: 5.387245178222656\r\nStep 740, loss: 5.163019180297852\r\nStep 741, loss: 5.0033721923828125\r\nStep 742, loss: 5.450179576873779\r\nStep 743, loss: 5.4586567878723145\r\nStep 744, loss: 5.1958513259887695\r\nStep 745, loss: 5.50988245010376\r\nStep 746, loss: 5.329402446746826\r\nStep 747, loss: 5.505457878112793\r\nStep 748, loss: 5.3953537940979\r\nStep 749, loss: 5.421640396118164\r\nStep 750, loss: 4.844815731048584\r\nStep 751, loss: 5.168392658233643\r\nStep 752, loss: 5.35214900970459\r\nStep 753, loss: 5.326869964599609\r\nStep 754, loss: 5.134158134460449\r\nStep 755, loss: 4.939837455749512\r\nStep 756, loss: 5.239572525024414\r\nStep 757, loss: 5.154784202575684\r\nStep 758, loss: 5.339284420013428\r\nStep 759, loss: 5.508555889129639\r\nStep 760, loss: 5.005982398986816\r\nStep 761, loss: 5.133974075317383\r\nStep 762, loss: 5.095065593719482\r\nStep 763, loss: 5.198653697967529\r\nStep 764, loss: 5.1999616622924805\r\nStep 765, loss: 5.3742570877075195\r\nStep 766, loss: 5.18704891204834\r\nStep 767, loss: 5.39532995223999\r\nStep 768, loss: 5.305274963378906\r\nStep 769, loss: 5.349462509155273\r\nStep 770, loss: 5.104123592376709\r\nStep 771, loss: 5.199271202087402\r\nStep 772, loss: 5.092228889465332\r\nStep 773, loss: 5.3029913902282715\r\nStep 774, loss: 5.354883193969727\r\nStep 775, loss: 5.052666187286377\r\nStep 776, loss: 5.35020637512207\r\nStep 777, loss: 5.49293851852417\r\nStep 778, loss: 5.183391094207764\r\nStep 779, loss: 5.375885009765625\r\nStep 780, loss: 5.115752220153809\r\nStep 781, loss: 5.134352684020996\r\nStep 782, loss: 4.961171627044678\r\nStep 783, loss: 5.193238735198975\r\nStep 784, loss: 5.077812194824219\r\nStep 785, loss: 5.231456756591797\r\nStep 786, loss: 5.293372631072998\r\nStep 787, loss: 5.198167324066162\r\nStep 788, loss: 5.177676200866699\r\nStep 789, loss: 5.366830825805664\r\nStep 790, loss: 5.297807216644287\r\nStep 791, loss: 5.202820777893066\r\nStep 792, loss: 5.1975321769714355\r\nStep 793, loss: 5.511959075927734\r\nStep 794, loss: 5.435663223266602\r\nStep 795, loss: 5.210607528686523\r\nStep 796, loss: 5.22182035446167\r\nStep 797, loss: 5.280124187469482\r\nStep 798, loss: 5.483152389526367\r\nStep 799, loss: 5.372382640838623\r\nStep 800, loss: 5.3446364402771\r\nStep 801, loss: 5.2136759757995605\r\nStep 802, loss: 5.432462215423584\r\nStep 803, loss: 5.094836235046387\r\nStep 804, loss: 5.033899307250977\r\nStep 805, loss: 5.34099006652832\r\nStep 806, loss: 4.993969917297363\r\nStep 807, loss: 5.139680862426758\r\nStep 808, loss: 5.169064998626709\r\nStep 809, loss: 5.114250183105469\r\nStep 810, loss: 5.067643165588379\r\nStep 811, loss: 5.102647304534912\r\nStep 812, loss: 4.900321006774902\r\nStep 813, loss: 5.187741756439209\r\nStep 814, loss: 5.1775970458984375\r\nStep 815, loss: 5.301947593688965\r\nStep 816, loss: 5.487864971160889\r\nStep 817, loss: 5.195375442504883\r\nStep 818, loss: 5.214444160461426\r\nStep 819, loss: 5.186010837554932\r\nStep 820, loss: 5.233945846557617\r\nStep 821, loss: 5.107511520385742\r\nStep 822, loss: 5.087286472320557\r\nStep 823, loss: 5.018490314483643\r\nStep 824, loss: 5.2226762771606445\r\nStep 825, loss: 5.2764387130737305\r\nStep 826, loss: 5.0772833824157715\r\nStep 827, loss: 5.010648250579834\r\nStep 828, loss: 5.3085856437683105\r\nStep 829, loss: 5.063631057739258\r\nStep 830, loss: 5.158314228057861\r\nStep 831, loss: 5.056264400482178\r\nStep 832, loss: 5.233462810516357\r\nStep 833, loss: 5.12919282913208\r\nStep 834, loss: 5.364454746246338\r\nStep 835, loss: 5.389515399932861\r\nStep 836, loss: 5.048567771911621\r\nStep 837, loss: 5.149641990661621\r\nStep 838, loss: 5.039944171905518\r\nStep 839, loss: 5.245511531829834\r\nStep 840, loss: 5.165370464324951\r\nStep 841, loss: 5.112173557281494\r\nStep 842, loss: 4.9662394523620605\r\nStep 843, loss: 5.026354789733887\r\nStep 844, loss: 5.0263519287109375\r\nStep 845, loss: 5.264150619506836\r\nStep 846, loss: 5.310177803039551\r\nStep 847, loss: 5.203787803649902\r\nStep 848, loss: 5.234409332275391\r\nStep 849, loss: 5.48568868637085\r\nStep 850, loss: 4.754512310028076\r\nStep 851, loss: 5.181948661804199\r\nStep 852, loss: 5.177431106567383\r\nStep 853, loss: 5.1184773445129395\r\nStep 854, loss: 5.161408424377441\r\nStep 855, loss: 5.284380912780762\r\nStep 856, loss: 5.365041255950928\r\nStep 857, loss: 4.994603633880615\r\nStep 858, loss: 5.119741916656494\r\nStep 859, loss: 5.4332475662231445\r\nStep 860, loss: 4.946900844573975\r\nStep 861, loss: 5.184002876281738\r\nStep 862, loss: 5.185951232910156\r\nStep 863, loss: 5.291336536407471\r\nStep 864, loss: 4.999337196350098\r\nStep 865, loss: 5.166353702545166\r\nStep 866, loss: 5.531929016113281\r\nStep 867, loss: 5.285906791687012\r\nStep 868, loss: 5.155843734741211\r\nStep 869, loss: 5.214123725891113\r\nStep 870, loss: 5.001017093658447\r\nStep 871, loss: 5.293190956115723\r\nStep 872, loss: 5.077353477478027\r\nStep 873, loss: 5.047084331512451\r\nStep 874, loss: 4.81112003326416\r\nStep 875, loss: 5.0530524253845215\r\nStep 876, loss: 5.3802690505981445\r\nStep 877, loss: 5.140942573547363\r\nStep 878, loss: 5.007487773895264\r\nStep 879, loss: 4.9970245361328125\r\nStep 880, loss: 5.022059440612793\r\nStep 881, loss: 4.9126386642456055\r\nStep 882, loss: 5.345889568328857\r\nStep 883, loss: 5.15145206451416\r\nStep 884, loss: 5.181049346923828\r\nStep 885, loss: 4.934937000274658\r\nStep 886, loss: 4.96279239654541\r\nStep 887, loss: 4.8947906494140625\r\nStep 888, loss: 5.221316337585449\r\nStep 889, loss: 4.922238349914551\r\nStep 890, loss: 5.096713542938232\r\nStep 891, loss: 5.419372081756592\r\nStep 892, loss: 5.152041435241699\r\nStep 893, loss: 4.9967474937438965\r\nStep 894, loss: 5.041221618652344\r\nStep 895, loss: 5.1730804443359375\r\nStep 896, loss: 5.100033283233643\r\nStep 897, loss: 5.236809253692627\r\nStep 898, loss: 5.174713611602783\r\nStep 899, loss: 4.951124668121338\r\nStep 900, loss: 4.98465633392334\r\nStep 901, loss: 5.190182209014893\r\nStep 902, loss: 5.194697380065918\r\nStep 903, loss: 5.08566951751709\r\nStep 904, loss: 5.178715229034424\r\nStep 905, loss: 5.028305530548096\r\nStep 906, loss: 5.365531921386719\r\nStep 907, loss: 5.162856578826904\r\nStep 908, loss: 5.212813854217529\r\nStep 909, loss: 4.955593109130859\r\nStep 910, loss: 4.896980285644531\r\nStep 911, loss: 5.0863165855407715\r\nStep 912, loss: 5.099323749542236\r\nStep 913, loss: 5.119900703430176\r\nStep 914, loss: 5.37208890914917\r\nStep 915, loss: 5.117595195770264\r\nStep 916, loss: 5.058721542358398\r\nStep 917, loss: 5.050705909729004\r\nStep 918, loss: 5.170966625213623\r\nStep 919, loss: 5.2613205909729\r\nStep 920, loss: 5.0097880363464355\r\nStep 921, loss: 5.356467247009277\r\nStep 922, loss: 5.3333563804626465\r\nStep 923, loss: 4.88981294631958\r\nStep 924, loss: 5.183024883270264\r\nStep 925, loss: 5.0197858810424805\r\nStep 926, loss: 5.204542636871338\r\nStep 927, loss: 4.9824676513671875\r\nStep 928, loss: 5.0433549880981445\r\nStep 929, loss: 5.069318771362305\r\nStep 930, loss: 5.1959099769592285\r\nStep 931, loss: 5.107265472412109\r\nStep 932, loss: 5.068860054016113\r\nStep 933, loss: 5.199520587921143\r\nStep 934, loss: 4.9961018562316895\r\nStep 935, loss: 5.081637859344482\r\nStep 936, loss: 5.0613484382629395\r\nStep 937, loss: 5.130383014678955\r\nStep 938, loss: 5.093724727630615\r\nStep 939, loss: 4.736340522766113\r\nStep 940, loss: 5.0766215324401855\r\nStep 941, loss: 4.86067533493042\r\nStep 942, loss: 4.996115684509277\r\nStep 943, loss: 5.155579090118408\r\nStep 944, loss: 5.194488048553467\r\nStep 945, loss: 5.429854869842529\r\nStep 946, loss: 4.887845993041992\r\nStep 947, loss: 5.2252912521362305\r\nStep 948, loss: 5.225298881530762\r\nStep 949, loss: 5.137124061584473\r\nStep 950, loss: 5.239169120788574\r\nStep 951, loss: 5.131504058837891\r\nStep 952, loss: 5.070860385894775\r\nStep 953, loss: 5.186777591705322\r\nStep 954, loss: 4.983127593994141\r\nStep 955, loss: 4.94017219543457\r\nStep 956, loss: 5.011066913604736\r\nStep 957, loss: 5.030396461486816\r\nStep 958, loss: 5.042568206787109\r\nStep 959, loss: 4.927941799163818\r\nStep 960, loss: 5.052581310272217\r\nStep 961, loss: 4.9004340171813965\r\nStep 722, loss: 5.520607948303223\r\nStep 723, loss: 5.279177188873291\r\nStep 724, loss: 5.26333475112915\r\nStep 725, loss: 5.184632778167725\r\nStep 726, loss: 5.329742431640625\r\nStep 727, loss: 5.339062213897705\r\nStep 728, loss: 5.266631603240967\r\nStep 729, loss: 5.203587055206299\r\nStep 730, loss: 5.071401119232178\r\nStep 731, loss: 5.185279846191406\r\nStep 732, loss: 5.403427600860596\r\nStep 733, loss: 5.377686500549316\r\nStep 734, loss: 5.246264934539795\r\nStep 735, loss: 5.169823169708252\r\nStep 736, loss: 5.027594566345215\r\nStep 737, loss: 5.25385856628418\r\nStep 738, loss: 5.292280197143555\r\nStep 739, loss: 5.387245178222656\r\nStep 740, loss: 5.163019180297852\r\nStep 741, loss: 5.0033721923828125\r\nStep 742, loss: 5.450179576873779\r\nStep 743, loss: 5.4586567878723145\r\nStep 744, loss: 5.1958513259887695\r\nStep 745, loss: 5.50988245010376\r\nStep 746, loss: 5.329402446746826\r\nStep 747, loss: 5.505457878112793\r\nStep 748, loss: 5.3953537940979\r\nStep 749, loss: 5.421640396118164\r\nStep 750, loss: 4.844815731048584\r\nStep 751, loss: 5.168392658233643\r\nStep 752, loss: 5.35214900970459\r\nStep 753, loss: 5.326869964599609\r\nStep 754, loss: 5.134158134460449\r\nStep 755, loss: 4.939837455749512\r\nStep 756, loss: 5.239572525024414\r\nStep 757, loss: 5.154784202575684\r\nStep 758, loss: 5.339284420013428\r\nStep 759, loss: 5.508555889129639\r\nStep 760, loss: 5.005982398986816\r\nStep 761, loss: 5.133974075317383\r\nStep 762, loss: 5.095065593719482\r\nStep 763, loss: 5.198653697967529\r\nStep 764, loss: 5.1999616622924805\r\nStep 765, loss: 5.3742570877075195\r\nStep 766, loss: 5.18704891204834\r\nStep 767, loss: 5.39532995223999\r\nStep 768, loss: 5.305274963378906\r\nStep 769, loss: 5.349462509155273\r\nStep 770, loss: 5.104123592376709\r\nStep 771, loss: 5.199271202087402\r\nStep 772, loss: 5.092228889465332\r\nStep 773, loss: 5.3029913902282715\r\nStep 774, loss: 5.354883193969727\r\nStep 775, loss: 5.052666187286377\r\nStep 776, loss: 5.35020637512207\r\nStep 777, loss: 5.49293851852417\r\nStep 778, loss: 5.183391094207764\r\nStep 779, loss: 5.375885009765625\r\nStep 780, loss: 5.115752220153809\r\nStep 781, loss: 5.134352684020996\r\nStep 782, loss: 4.961171627044678\r\nStep 783, loss: 5.193238735198975\r\nStep 784, loss: 5.077812194824219\r\nStep 785, loss: 5.231456756591797\r\nStep 786, loss: 5.293372631072998\r\nStep 787, loss: 5.198167324066162\r\nStep 788, loss: 5.177676200866699\r\nStep 789, loss: 5.366830825805664\r\nStep 790, loss: 5.297807216644287\r\nStep 791, loss: 5.202820777893066\r\nStep 792, loss: 5.1975321769714355\r\nStep 793, loss: 5.511959075927734\r\nStep 794, loss: 5.435663223266602\r\nStep 795, loss: 5.210607528686523\r\nStep 796, loss: 5.22182035446167\r\nStep 797, loss: 5.280124187469482\r\nStep 798, loss: 5.483152389526367\r\nStep 799, loss: 5.372382640838623\r\nStep 800, loss: 5.3446364402771\r\nStep 801, loss: 5.2136759757995605\r\nStep 802, loss: 5.432462215423584\r\nStep 803, loss: 5.094836235046387\r\nStep 804, loss: 5.033899307250977\r\nStep 805, loss: 5.34099006652832\r\nStep 806, loss: 4.993969917297363\r\nStep 807, loss: 5.139680862426758\r\nStep 808, loss: 5.169064998626709\r\nStep 809, loss: 5.114250183105469\r\nStep 810, loss: 5.067643165588379\r\nStep 811, loss: 5.102647304534912\r\nStep 812, loss: 4.900321006774902\r\nStep 813, loss: 5.187741756439209\r\nStep 814, loss: 5.1775970458984375\r\nStep 815, loss: 5.301947593688965\r\nStep 816, loss: 5.487864971160889\r\nStep 817, loss: 5.195375442504883\r\nStep 818, loss: 5.214444160461426\r\nStep 819, loss: 5.186010837554932\r\nStep 820, loss: 5.233945846557617\r\nStep 821, loss: 5.107511520385742\r\nStep 822, loss: 5.087286472320557\r\nStep 823, loss: 5.018490314483643\r\nStep 824, loss: 5.2226762771606445\r\nStep 825, loss: 5.2764387130737305\r\nStep 826, loss: 5.0772833824157715\r\nStep 827, loss: 5.010648250579834\r\nStep 828, loss: 5.3085856437683105\r\nStep 829, loss: 5.063631057739258\r\nStep 830, loss: 5.158314228057861\r\nStep 831, loss: 5.056264400482178\r\nStep 832, loss: 5.233462810516357\r\nStep 833, loss: 5.12919282913208\r\nStep 834, loss: 5.364454746246338\r\nStep 835, loss: 5.389515399932861\r\nStep 836, loss: 5.048567771911621\r\nStep 837, loss: 5.149641990661621\r\nStep 838, loss: 5.039944171905518\r\nStep 839, loss: 5.245511531829834\r\nStep 840, loss: 5.165370464324951\r\nStep 841, loss: 5.112173557281494\r\nStep 842, loss: 4.9662394523620605\r\nStep 843, loss: 5.026354789733887\r\nStep 844, loss: 5.0263519287109375\r\nStep 845, loss: 5.264150619506836\r\nStep 846, loss: 5.310177803039551\r\nStep 847, loss: 5.203787803649902\r\nStep 848, loss: 5.234409332275391\r\nStep 849, loss: 5.48568868637085\r\nStep 850, loss: 4.754512310028076\r\nStep 851, loss: 5.181948661804199\r\nStep 852, loss: 5.177431106567383\r\nStep 853, loss: 5.1184773445129395\r\nStep 854, loss: 5.161408424377441\r\nStep 855, loss: 5.284380912780762\r\nStep 856, loss: 5.365041255950928\r\nStep 857, loss: 4.994603633880615\r\nStep 858, loss: 5.119741916656494\r\nStep 859, loss: 5.4332475662231445\r\nStep 860, loss: 4.946900844573975\r\nStep 861, loss: 5.184002876281738\r\nStep 862, loss: 5.185951232910156\r\nStep 863, loss: 5.291336536407471\r\nStep 864, loss: 4.999337196350098\r\nStep 865, loss: 5.166353702545166\r\nStep 866, loss: 5.531929016113281\r\nStep 867, loss: 5.285906791687012\r\nStep 868, loss: 5.155843734741211\r\nStep 869, loss: 5.214123725891113\r\nStep 870, loss: 5.001017093658447\r\nStep 871, loss: 5.293190956115723\r\nStep 872, loss: 5.077353477478027\r\nStep 873, loss: 5.047084331512451\r\nStep 874, loss: 4.81112003326416\r\nStep 875, loss: 5.0530524253845215\r\nStep 876, loss: 5.3802690505981445\r\nStep 877, loss: 5.140942573547363\r\nStep 878, loss: 5.007487773895264\r\nStep 879, loss: 4.9970245361328125\r\nStep 880, loss: 5.022059440612793\r\nStep 881, loss: 4.9126386642456055\r\nStep 882, loss: 5.345889568328857\r\nStep 883, loss: 5.15145206451416\r\nStep 884, loss: 5.181049346923828\r\nStep 885, loss: 4.934937000274658\r\nStep 886, loss: 4.96279239654541\r\nStep 887, loss: 4.8947906494140625\r\nStep 888, loss: 5.221316337585449\r\nStep 889, loss: 4.922238349914551\r\nStep 890, loss: 5.096713542938232\r\nStep 891, loss: 5.419372081756592\r\nStep 892, loss: 5.152041435241699\r\nStep 893, loss: 4.9967474937438965\r\nStep 894, loss: 5.041221618652344\r\nStep 895, loss: 5.1730804443359375\r\nStep 896, loss: 5.100033283233643\r\nStep 897, loss: 5.236809253692627\r\nStep 898, loss: 5.174713611602783\r\nStep 899, loss: 4.951124668121338\r\nStep 900, loss: 4.98465633392334\r\nStep 901, loss: 5.190182209014893\r\nStep 902, loss: 5.194697380065918\r\nStep 903, loss: 5.08566951751709\r\nStep 904, loss: 5.178715229034424\r\nStep 905, loss: 5.028305530548096\r\nStep 906, loss: 5.365531921386719\r\nStep 907, loss: 5.162856578826904\r\nStep 908, loss: 5.212813854217529\r\nStep 909, loss: 4.955593109130859\r\nStep 910, loss: 4.896980285644531\r\nStep 911, loss: 5.0863165855407715\r\nStep 912, loss: 5.099323749542236\r\nStep 913, loss: 5.119900703430176\r\nStep 914, loss: 5.37208890914917\r\nStep 915, loss: 5.117595195770264\r\nStep 916, loss: 5.058721542358398\r\nStep 917, loss: 5.050705909729004\r\nStep 918, loss: 5.170966625213623\r\nStep 919, loss: 5.2613205909729\r\nStep 920, loss: 5.0097880363464355\r\nStep 921, loss: 5.356467247009277\r\nStep 922, loss: 5.3333563804626465\r\nStep 923, loss: 4.88981294631958\r\nStep 924, loss: 5.183024883270264\r\nStep 925, loss: 5.0197858810424805\r\nStep 926, loss: 5.204542636871338\r\nStep 927, loss: 4.9824676513671875\r\nStep 928, loss: 5.0433549880981445\r\nStep 929, loss: 5.069318771362305\r\nStep 930, loss: 5.1959099769592285\r\nStep 931, loss: 5.107265472412109\r\nStep 932, loss: 5.068860054016113\r\nStep 933, loss: 5.199520587921143\r\nStep 934, loss: 4.9961018562316895\r\nStep 935, loss: 5.081637859344482\r\nStep 936, loss: 5.0613484382629395\r\nStep 937, loss: 5.130383014678955\r\nStep 938, loss: 5.093724727630615\r\nStep 939, loss: 4.736340522766113\r\nStep 940, loss: 5.0766215324401855\r\nStep 941, loss: 4.86067533493042\r\nStep 942, loss: 4.996115684509277\r\nStep 943, loss: 5.155579090118408\r\nStep 944, loss: 5.194488048553467\r\nStep 945, loss: 5.429854869842529\r\nStep 946, loss: 4.887845993041992\r\nStep 947, loss: 5.2252912521362305\r\nStep 948, loss: 5.225298881530762\r\nStep 949, loss: 5.137124061584473\r\nStep 950, loss: 5.239169120788574\r\nStep 951, loss: 5.131504058837891\r\nStep 952, loss: 5.070860385894775\r\nStep 953, loss: 5.186777591705322\r\nStep 954, loss: 4.983127593994141\r\nStep 955, loss: 4.94017219543457\r\nStep 956, loss: 5.011066913604736\r\nStep 957, loss: 5.030396461486816\r\nStep 958, loss: 5.042568206787109\r\nStep 959, loss: 4.927941799163818\r\nStep 960, loss: 5.052581310272217\r\nStep 961, loss: 4.9004340171813965\r\nStep 722, loss: 5.520607948303223\r\nStep 723, loss: 5.279177188873291\r\nStep 724, loss: 5.26333475112915\r\nStep 725, loss: 5.184632778167725\r\nStep 726, loss: 5.329742431640625\r\nStep 727, loss: 5.339062213897705\r\nStep 728, loss: 5.266631603240967\r\nStep 729, loss: 5.203587055206299\r\nStep 730, loss: 5.071401119232178\r\nStep 731, loss: 5.185279846191406\r\nStep 732, loss: 5.403427600860596\r\nStep 733, loss: 5.377686500549316\r\nStep 734, loss: 5.246264934539795\r\nStep 735, loss: 5.169823169708252\r\nStep 736, loss: 5.027594566345215\r\nStep 737, loss: 5.25385856628418\r\nStep 738, loss: 5.292280197143555\r\nStep 739, loss: 5.387245178222656\r\nStep 740, loss: 5.163019180297852\r\nStep 741, loss: 5.0033721923828125\r\nStep 742, loss: 5.450179576873779\r\nStep 743, loss: 5.4586567878723145\r\nStep 744, loss: 5.1958513259887695\r\nStep 745, loss: 5.50988245010376\r\nStep 746, loss: 5.329402446746826\r\nStep 747, loss: 5.505457878112793\r\nStep 748, loss: 5.3953537940979\r\nStep 749, loss: 5.421640396118164\r\nStep 750, loss: 4.844815731048584\r\nStep 751, loss: 5.168392658233643\r\nStep 752, loss: 5.35214900970459\r\nStep 753, loss: 5.326869964599609\r\nStep 754, loss: 5.134158134460449\r\nStep 755, loss: 4.939837455749512\r\nStep 756, loss: 5.239572525024414\r\nStep 757, loss: 5.154784202575684\r\nStep 758, loss: 5.339284420013428\r\nStep 759, loss: 5.508555889129639\r\nStep 760, loss: 5.005982398986816\r\nStep 761, loss: 5.133974075317383\r\nStep 762, loss: 5.095065593719482\r\nStep 763, loss: 5.198653697967529\r\nStep 764, loss: 5.1999616622924805\r\nStep 765, loss: 5.3742570877075195\r\nStep 766, loss: 5.18704891204834\r\nStep 767, loss: 5.39532995223999\r\nStep 768, loss: 5.305274963378906\r\nStep 769, loss: 5.349462509155273\r\nStep 770, loss: 5.104123592376709\r\nStep 771, loss: 5.199271202087402\r\nStep 772, loss: 5.092228889465332\r\nStep 773, loss: 5.3029913902282715\r\nStep 774, loss: 5.354883193969727\r\nStep 775, loss: 5.052666187286377\r\nStep 776, loss: 5.35020637512207\r\nStep 777, loss: 5.49293851852417\r\nStep 778, loss: 5.183391094207764\r\nStep 779, loss: 5.375885009765625\r\nStep 780, loss: 5.115752220153809\r\nStep 781, loss: 5.134352684020996\r\nStep 782, loss: 4.961171627044678\r\nStep 783, loss: 5.193238735198975\r\nStep 784, loss: 5.077812194824219\r\nStep 785, loss: 5.231456756591797\r\nStep 786, loss: 5.293372631072998\r\nStep 787, loss: 5.198167324066162\r\nStep 788, loss: 5.177676200866699\r\nStep 789, loss: 5.366830825805664\r\nStep 790, loss: 5.297807216644287\r\nStep 791, loss: 5.202820777893066\r\nStep 792, loss: 5.1975321769714355\r\nStep 793, loss: 5.511959075927734\r\nStep 794, loss: 5.435663223266602\r\nStep 795, loss: 5.210607528686523\r\nStep 796, loss: 5.22182035446167\r\nStep 797, loss: 5.280124187469482\r\nStep 798, loss: 5.483152389526367\r\nStep 799, loss: 5.372382640838623\r\nStep 800, loss: 5.3446364402771\r\nStep 801, loss: 5.2136759757995605\r\nStep 802, loss: 5.432462215423584\r\nStep 803, loss: 5.094836235046387\r\nStep 804, loss: 5.033899307250977\r\nStep 805, loss: 5.34099006652832\r\nStep 806, loss: 4.993969917297363\r\nStep 807, loss: 5.139680862426758\r\nStep 808, loss: 5.169064998626709\r\nStep 809, loss: 5.114250183105469\r\nStep 810, loss: 5.067643165588379\r\nStep 811, loss: 5.102647304534912\r\nStep 812, loss: 4.900321006774902\r\nStep 813, loss: 5.187741756439209\r\nStep 814, loss: 5.1775970458984375\r\nStep 815, loss: 5.301947593688965\r\nStep 816, loss: 5.487864971160889\r\nStep 817, loss: 5.195375442504883\r\nStep 818, loss: 5.214444160461426\r\nStep 819, loss: 5.186010837554932\r\nStep 820, loss: 5.233945846557617\r\nStep 821, loss: 5.107511520385742\r\nStep 822, loss: 5.087286472320557\r\nStep 823, loss: 5.018490314483643\r\nStep 824, loss: 5.2226762771606445\r\nStep 825, loss: 5.2764387130737305\r\nStep 826, loss: 5.0772833824157715\r\nStep 827, loss: 5.010648250579834\r\nStep 828, loss: 5.3085856437683105\r\nStep 829, loss: 5.063631057739258\r\nStep 830, loss: 5.158314228057861\r\nStep 831, loss: 5.056264400482178\r\nStep 832, loss: 5.233462810516357\r\nStep 833, loss: 5.12919282913208\r\nStep 834, loss: 5.364454746246338\r\nStep 835, loss: 5.389515399932861\r\nStep 836, loss: 5.048567771911621\r\nStep 837, loss: 5.149641990661621\r\nStep 838, loss: 5.039944171905518\r\nStep 839, loss: 5.245511531829834\r\nStep 840, loss: 5.165370464324951\r\nStep 841, loss: 5.112173557281494\r\nStep 842, loss: 4.9662394523620605\r\nStep 843, loss: 5.026354789733887\r\nStep 844, loss: 5.0263519287109375\r\nStep 845, loss: 5.264150619506836\r\nStep 846, loss: 5.310177803039551\r\nStep 847, loss: 5.203787803649902\r\nStep 848, loss: 5.234409332275391\r\nStep 849, loss: 5.48568868637085\r\nStep 850, loss: 4.754512310028076\r\nStep 851, loss: 5.181948661804199\r\nStep 852, loss: 5.177431106567383\r\nStep 853, loss: 5.1184773445129395\r\nStep 854, loss: 5.161408424377441\r\nStep 855, loss: 5.284380912780762\r\nStep 856, loss: 5.365041255950928\r\nStep 857, loss: 4.994603633880615\r\nStep 858, loss: 5.119741916656494\r\nStep 859, loss: 5.4332475662231445\r\nStep 860, loss: 4.946900844573975\r\nStep 861, loss: 5.184002876281738\r\nStep 862, loss: 5.185951232910156\r\nStep 863, loss: 5.291336536407471\r\nStep 864, loss: 4.999337196350098\r\nStep 865, loss: 5.166353702545166\r\nStep 866, loss: 5.531929016113281\r\nStep 867, loss: 5.285906791687012\r\nStep 868, loss: 5.155843734741211\r\nStep 869, loss: 5.214123725891113\r\nStep 870, loss: 5.001017093658447\r\nStep 871, loss: 5.293190956115723\r\nStep 872, loss: 5.077353477478027\r\nStep 873, loss: 5.047084331512451\r\nStep 874, loss: 4.81112003326416\r\nStep 875, loss: 5.0530524253845215\r\nStep 876, loss: 5.3802690505981445\r\nStep 877, loss: 5.140942573547363\r\nStep 878, loss: 5.007487773895264\r\nStep 879, loss: 4.9970245361328125\r\nStep 880, loss: 5.022059440612793\r\nStep 881, loss: 4.9126386642456055\r\nStep 882, loss: 5.345889568328857\r\nStep 883, loss: 5.15145206451416\r\nStep 884, loss: 5.181049346923828\r\nStep 885, loss: 4.934937000274658\r\nStep 886, loss: 4.96279239654541\r\nStep 887, loss: 4.8947906494140625\r\nStep 888, loss: 5.221316337585449\r\nStep 889, loss: 4.922238349914551\r\nStep 890, loss: 5.096713542938232\r\nStep 891, loss: 5.419372081756592\r\nStep 892, loss: 5.152041435241699\r\nStep 893, loss: 4.9967474937438965\r\nStep 894, loss: 5.041221618652344\r\nStep 895, loss: 5.1730804443359375\r\nStep 896, loss: 5.100033283233643\r\nStep 897, loss: 5.236809253692627\r\nStep 898, loss: 5.174713611602783\r\nStep 899, loss: 4.951124668121338\r\nStep 900, loss: 4.98465633392334\r\nStep 901, loss: 5.190182209014893\r\nStep 902, loss: 5.194697380065918\r\nStep 903, loss: 5.08566951751709\r\nStep 904, loss: 5.178715229034424\r\nStep 905, loss: 5.028305530548096\r\nStep 906, loss: 5.365531921386719\r\nStep 907, loss: 5.162856578826904\r\nStep 908, loss: 5.212813854217529\r\nStep 909, loss: 4.955593109130859\r\nStep 910, loss: 4.896980285644531\r\nStep 911, loss: 5.0863165855407715\r\nStep 912, loss: 5.099323749542236\r\nStep 913, loss: 5.119900703430176\r\nStep 914, loss: 5.37208890914917\r\nStep 915, loss: 5.117595195770264\r\nStep 916, loss: 5.058721542358398\r\nStep 917, loss: 5.050705909729004\r\nStep 918, loss: 5.170966625213623\r\nStep 919, loss: 5.2613205909729\r\nStep 920, loss: 5.0097880363464355\r\nStep 921, loss: 5.356467247009277\r\nStep 922, loss: 5.3333563804626465\r\nStep 923, loss: 4.88981294631958\r\nStep 924, loss: 5.183024883270264\r\nStep 925, loss: 5.0197858810424805\r\nStep 926, loss: 5.204542636871338\r\nStep 927, loss: 4.9824676513671875\r\nStep 928, loss: 5.0433549880981445\r\nStep 929, loss: 5.069318771362305\r\nStep 930, loss: 5.1959099769592285\r\nStep 931, loss: 5.107265472412109\r\nStep 932, loss: 5.068860054016113\r\nStep 933, loss: 5.199520587921143\r\nStep 934, loss: 4.9961018562316895\r\nStep 935, loss: 5.081637859344482\r\nStep 936, loss: 5.0613484382629395\r\nStep 937, loss: 5.130383014678955\r\nStep 938, loss: 5.093724727630615\r\nStep 939, loss: 4.736340522766113\r\nStep 940, loss: 5.0766215324401855\r\nStep 941, loss: 4.86067533493042\r\nStep 942, loss: 4.996115684509277\r\nStep 943, loss: 5.155579090118408\r\nStep 944, loss: 5.194488048553467\r\nStep 945, loss: 5.429854869842529\r\nStep 946, loss: 4.887845993041992\r\nStep 947, loss: 5.2252912521362305\r\nStep 948, loss: 5.225298881530762\r\nStep 949, loss: 5.137124061584473\r\nStep 950, loss: 5.239169120788574\r\nStep 951, loss: 5.131504058837891\r\nStep 952, loss: 5.070860385894775\r\nStep 953, loss: 5.186777591705322\r\nStep 954, loss: 4.983127593994141\r\nStep 955, loss: 4.94017219543457\r\nStep 956, loss: 5.011066913604736\r\nStep 957, loss: 5.030396461486816\r\nStep 958, loss: 5.042568206787109\r\nStep 959, loss: 4.927941799163818\r\nStep 960, loss: 5.052581310272217\r\nStep 961, loss: 4.9004340171813965\r\nStep 722, loss: 5.520607948303223\r\nStep 723, loss: 5.279177188873291\r\nStep 724, loss: 5.26333475112915\r\nStep 725, loss: 5.184632778167725\r\nStep 726, loss: 5.329742431640625\r\nStep 727, loss: 5.339062213897705\r\nStep 728, loss: 5.266631603240967\r\nStep 729, loss: 5.203587055206299\r\nStep 730, loss: 5.071401119232178\r\nStep 731, loss: 5.185279846191406\r\nStep 732, loss: 5.403427600860596\r\nStep 733, loss: 5.377686500549316\r\nStep 734, loss: 5.246264934539795\r\nStep 735, loss: 5.169823169708252\r\nStep 736, loss: 5.027594566345215\r\nStep 737, loss: 5.25385856628418\r\nStep 738, loss: 5.292280197143555\r\nStep 739, loss: 5.387245178222656\r\nStep 740, loss: 5.163019180297852\r\nStep 741, loss: 5.0033721923828125\r\nStep 742, loss: 5.450179576873779\r\nStep 743, loss: 5.4586567878723145\r\nStep 744, loss: 5.1958513259887695\r\nStep 745, loss: 5.50988245010376\r\nStep 746, loss: 5.329402446746826\r\nStep 747, loss: 5.505457878112793\r\nStep 748, loss: 5.3953537940979\r\nStep 749, loss: 5.421640396118164\r\nStep 750, loss: 4.844815731048584\r\nStep 751, loss: 5.168392658233643\r\nStep 752, loss: 5.35214900970459\r\nStep 753, loss: 5.326869964599609\r\nStep 754, loss: 5.134158134460449\r\nStep 755, loss: 4.939837455749512\r\nStep 756, loss: 5.239572525024414\r\nStep 757, loss: 5.154784202575684\r\nStep 758, loss: 5.339284420013428\r\nStep 759, loss: 5.508555889129639\r\nStep 760, loss: 5.005982398986816\r\nStep 761, loss: 5.133974075317383\r\nStep 762, loss: 5.095065593719482\r\nStep 763, loss: 5.198653697967529\r\nStep 764, loss: 5.1999616622924805\r\nStep 765, loss: 5.3742570877075195\r\nStep 766, loss: 5.18704891204834\r\nStep 767, loss: 5.39532995223999\r\nStep 768, loss: 5.305274963378906\r\nStep 769, loss: 5.349462509155273\r\nStep 770, loss: 5.104123592376709\r\nStep 771, loss: 5.199271202087402\r\nStep 772, loss: 5.092228889465332\r\nStep 773, loss: 5.3029913902282715\r\nStep 774, loss: 5.354883193969727\r\nStep 775, loss: 5.052666187286377\r\nStep 776, loss: 5.35020637512207\r\nStep 777, loss: 5.49293851852417\r\nStep 778, loss: 5.183391094207764\r\nStep 779, loss: 5.375885009765625\r\nStep 780, loss: 5.115752220153809\r\nStep 781, loss: 5.134352684020996\r\nStep 782, loss: 4.961171627044678\r\nStep 783, loss: 5.193238735198975\r\nStep 784, loss: 5.077812194824219\r\nStep 785, loss: 5.231456756591797\r\nStep 786, loss: 5.293372631072998\r\nStep 787, loss: 5.198167324066162\r\nStep 788, loss: 5.177676200866699\r\nStep 789, loss: 5.366830825805664\r\nStep 790, loss: 5.297807216644287\r\nStep 791, loss: 5.202820777893066\r\nStep 792, loss: 5.1975321769714355\r\nStep 793, loss: 5.511959075927734\r\nStep 794, loss: 5.435663223266602\r\nStep 795, loss: 5.210607528686523\r\nStep 796, loss: 5.22182035446167\r\nStep 797, loss: 5.280124187469482\r\nStep 798, loss: 5.483152389526367\r\nStep 799, loss: 5.372382640838623\r\nStep 800, loss: 5.3446364402771\r\nStep 801, loss: 5.2136759757995605\r\nStep 802, loss: 5.432462215423584\r\nStep 803, loss: 5.094836235046387\r\nStep 804, loss: 5.033899307250977\r\nStep 805, loss: 5.34099006652832\r\nStep 806, loss: 4.993969917297363\r\nStep 807, loss: 5.139680862426758\r\nStep 808, loss: 5.169064998626709\r\nStep 809, loss: 5.114250183105469\r\nStep 810, loss: 5.067643165588379\r\nStep 811, loss: 5.102647304534912\r\nStep 812, loss: 4.900321006774902\r\nStep 813, loss: 5.187741756439209\r\nStep 814, loss: 5.1775970458984375\r\nStep 815, loss: 5.301947593688965\r\nStep 816, loss: 5.487864971160889\r\nStep 817, loss: 5.195375442504883\r\nStep 818, loss: 5.214444160461426\r\nStep 819, loss: 5.186010837554932\r\nStep 820, loss: 5.233945846557617\r\nStep 821, loss: 5.107511520385742\r\nStep 822, loss: 5.087286472320557\r\nStep 823, loss: 5.018490314483643\r\nStep 824, loss: 5.2226762771606445\r\nStep 825, loss: 5.2764387130737305\r\nStep 826, loss: 5.0772833824157715\r\nStep 827, loss: 5.010648250579834\r\nStep 828, loss: 5.3085856437683105\r\nStep 829, loss: 5.063631057739258\r\nStep 830, loss: 5.158314228057861\r\nStep 831, loss: 5.056264400482178\r\nStep 832, loss: 5.233462810516357\r\nStep 833, loss: 5.12919282913208\r\nStep 834, loss: 5.364454746246338\r\nStep 835, loss: 5.389515399932861\r\nStep 836, loss: 5.048567771911621\r\nStep 837, loss: 5.149641990661621\r\nStep 838, loss: 5.039944171905518\r\nStep 839, loss: 5.245511531829834\r\nStep 840, loss: 5.165370464324951\r\nStep 841, loss: 5.112173557281494\r\nStep 842, loss: 4.9662394523620605\r\nStep 843, loss: 5.026354789733887\r\nStep 844, loss: 5.0263519287109375\r\nStep 845, loss: 5.264150619506836\r\nStep 846, loss: 5.310177803039551\r\nStep 847, loss: 5.203787803649902\r\nStep 848, loss: 5.234409332275391\r\nStep 849, loss: 5.48568868637085\r\nStep 850, loss: 4.754512310028076\r\nStep 851, loss: 5.181948661804199\r\nStep 852, loss: 5.177431106567383\r\nStep 853, loss: 5.1184773445129395\r\nStep 854, loss: 5.161408424377441\r\nStep 855, loss: 5.284380912780762\r\nStep 856, loss: 5.365041255950928\r\nStep 857, loss: 4.994603633880615\r\nStep 858, loss: 5.119741916656494\r\nStep 859, loss: 5.4332475662231445\r\nStep 860, loss: 4.946900844573975\r\nStep 861, loss: 5.184002876281738\r\nStep 862, loss: 5.185951232910156\r\nStep 863, loss: 5.291336536407471\r\nStep 864, loss: 4.999337196350098\r\nStep 865, loss: 5.166353702545166\r\nStep 866, loss: 5.531929016113281\r\nStep 867, loss: 5.285906791687012\r\nStep 868, loss: 5.155843734741211\r\nStep 869, loss: 5.214123725891113\r\nStep 870, loss: 5.001017093658447\r\nStep 871, loss: 5.293190956115723\r\nStep 872, loss: 5.077353477478027\r\nStep 873, loss: 5.047084331512451\r\nStep 874, loss: 4.81112003326416\r\nStep 875, loss: 5.0530524253845215\r\nStep 876, loss: 5.3802690505981445\r\nStep 877, loss: 5.140942573547363\r\nStep 878, loss: 5.007487773895264\r\nStep 879, loss: 4.9970245361328125\r\nStep 880, loss: 5.022059440612793\r\nStep 881, loss: 4.9126386642456055\r\nStep 882, loss: 5.345889568328857\r\nStep 883, loss: 5.15145206451416\r\nStep 884, loss: 5.181049346923828\r\nStep 885, loss: 4.934937000274658\r\nStep 886, loss: 4.96279239654541\r\nStep 887, loss: 4.8947906494140625\r\nStep 888, loss: 5.221316337585449\r\nStep 889, loss: 4.922238349914551\r\nStep 890, loss: 5.096713542938232\r\nStep 891, loss: 5.419372081756592\r\nStep 892, loss: 5.152041435241699\r\nStep 893, loss: 4.9967474937438965\r\nStep 894, loss: 5.041221618652344\r\nStep 895, loss: 5.1730804443359375\r\nStep 896, loss: 5.100033283233643\r\nStep 897, loss: 5.236809253692627\r\nStep 898, loss: 5.174713611602783\r\nStep 899, loss: 4.951124668121338\r\nStep 900, loss: 4.98465633392334\r\nStep 901, loss: 5.190182209014893\r\nStep 902, loss: 5.194697380065918\r\nStep 903, loss: 5.08566951751709\r\nStep 904, loss: 5.178715229034424\r\nStep 905, loss: 5.028305530548096\r\nStep 906, loss: 5.365531921386719\r\nStep 907, loss: 5.162856578826904\r\nStep 908, loss: 5.212813854217529\r\nStep 909, loss: 4.955593109130859\r\nStep 910, loss: 4.896980285644531\r\nStep 911, loss: 5.0863165855407715\r\nStep 912, loss: 5.099323749542236\r\nStep 913, loss: 5.119900703430176\r\nStep 914, loss: 5.37208890914917\r\nStep 915, loss: 5.117595195770264\r\nStep 916, loss: 5.058721542358398\r\nStep 917, loss: 5.050705909729004\r\nStep 918, loss: 5.170966625213623\r\nStep 919, loss: 5.2613205909729\r\nStep 920, loss: 5.0097880363464355\r\nStep 921, loss: 5.356467247009277\r\nStep 922, loss: 5.3333563804626465\r\nStep 923, loss: 4.88981294631958\r\nStep 924, loss: 5.183024883270264\r\nStep 925, loss: 5.0197858810424805\r\nStep 926, loss: 5.204542636871338\r\nStep 927, loss: 4.9824676513671875\r\nStep 928, loss: 5.0433549880981445\r\nStep 929, loss: 5.069318771362305\r\nStep 930, loss: 5.1959099769592285\r\nStep 931, loss: 5.107265472412109\r\nStep 932, loss: 5.068860054016113\r\nStep 933, loss: 5.199520587921143\r\nStep 934, loss: 4.9961018562316895\r\nStep 935, loss: 5.081637859344482\r\nStep 936, loss: 5.0613484382629395\r\nStep 937, loss: 5.130383014678955\r\nStep 938, loss: 5.093724727630615\r\nStep 939, loss: 4.736340522766113\r\nStep 940, loss: 5.0766215324401855\r\nStep 941, loss: 4.86067533493042\r\nStep 942, loss: 4.996115684509277\r\nStep 943, loss: 5.155579090118408\r\nStep 944, loss: 5.194488048553467\r\nStep 945, loss: 5.429854869842529\r\nStep 946, loss: 4.887845993041992\r\nStep 947, loss: 5.2252912521362305\r\nStep 948, loss: 5.225298881530762\r\nStep 949, loss: 5.137124061584473\r\nStep 950, loss: 5.239169120788574\r\nStep 951, loss: 5.131504058837891\r\nStep 952, loss: 5.070860385894775\r\nStep 953, loss: 5.186777591705322\r\nStep 954, loss: 4.983127593994141\r\nStep 955, loss: 4.94017219543457\r\nStep 956, loss: 5.011066913604736\r\nStep 957, loss: 5.030396461486816\r\nStep 958, loss: 5.042568206787109\r\nStep 959, loss: 4.927941799163818\r\nStep 960, loss: 5.052581310272217\r\nStep 961, loss: 4.9004340171813965\r\n",,terminal_output +3190,4873223,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1244,0,"",shellscript,selection_mouse +3191,4941362,"TERMINAL",0,0,"bash",,terminal_focus +3192,5182026,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,tab +3193,5182027,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1187,0,"",shellscript,selection_mouse +3194,5182028,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1186,0,"",shellscript,selection_command +3195,5246678,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,tab +3196,5248331,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1112,0,"",shellscript,selection_mouse +3197,5248932,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1061,0,"",shellscript,selection_mouse +3198,5248934,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1060,0,"",shellscript,selection_command +3199,5270528,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,tab +3200,5270529,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",933,0,"",shellscript,selection_mouse +3201,5270545,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",932,0,"",shellscript,selection_command +3202,5278368,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",580,0,"",shellscript,selection_mouse +3203,5278531,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",575,7,"scratch",shellscript,selection_mouse +3204,5279122,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",580,0,"",shellscript,selection_mouse +3205,5279381,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",575,7,"scratch",shellscript,selection_mouse +3206,5279562,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",536,122,"array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n",shellscript,selection_mouse +3207,5282665,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3208,5284977,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,tab +3209,5286116,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3210,5287230,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,tab +3211,5305174,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",972,0,"",shellscript,selection_mouse +3212,5305193,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",971,0,"",shellscript,selection_command +3213,5314379,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",955,0,"",shellscript,selection_mouse +3214,5314399,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",954,0,"",shellscript,selection_command +3215,5324206,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"",shellscript,tab +3216,5325098,"TERMINAL",0,0,"srun",,terminal_focus +3217,5338115,"TERMINAL",0,0,"bash",,terminal_focus +3218,5338116,"TERMINAL",0,0,"srun",,terminal_focus +3219,5345448,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3220,5348389,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3221,5349971,"TERMINAL",0,0,"bash",,terminal_focus +3222,5351009,"TERMINAL",0,0,"bash",,terminal_focus +3223,5351749,"TERMINAL",0,0,"bash",,terminal_focus +3224,5356995,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3225,5356996,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",900,0,"",shellscript,selection_mouse +3226,5357011,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",899,0,"",shellscript,selection_command +3227,5358773,"TERMINAL",0,0,"srun",,terminal_focus +3228,5360142,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3229,5360143,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",972,0,"",shellscript,selection_mouse +3230,5360145,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",971,0,"",shellscript,selection_command +3231,5360716,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",992,0,"",shellscript,selection_mouse +3232,5360718,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",991,0,"",shellscript,selection_command +3233,5361219,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",897,0,"",shellscript,selection_mouse +3234,5362940,"TERMINAL",0,0,"bash",,terminal_focus +3235,5366013,"TERMINAL",0,0,"git checkout main",,terminal_command +3236,5366063,"TERMINAL",0,0,"]633;E;2025-07-10 17:47:40 git checkout main;16c08be6-3885-420f-ac53-f5272aed6e54]633;C",,terminal_output +3237,5366856,"TERMINAL",0,0,"Switched to branch 'main'\r\n",,terminal_output +3238,5367030,"TERMINAL",0,0,"Your branch is behind 'origin/main' by 28 commits, and can be fast-forwarded.\r\n (use ""git pull"" to update your local branch)\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +3239,5368577,"TERMINAL",0,0,"git pull",,terminal_command +3240,5368628,"TERMINAL",0,0,"]633;E;2025-07-10 17:47:42 git pull;16c08be6-3885-420f-ac53-f5272aed6e54]633;C",,terminal_output +3241,5370271,"TERMINAL",0,0,"remote: Enumerating objects: 10, done.\r\nremote: Counting objects: 10% (1/10)\rremote: Counting objects: 20% (2/10)\rremote: Counting objects: 30% (3/10)\rremote: Counting objects: 40% (4/10)\rremote: Counting objects: 50% (5/10)\rremote: Counting objects: 60% (6/10)\rremote: Counting objects: 70% (7/10)\rremote: Counting objects: 80% (8/10)\rremote: Counting objects: 90% (9/10)\rremote: Counting objects: 100% (10/10)\rremote: Counting objects: 100% (10/10), done.\r\nremote: Compressing objects: 50% (1/2)\rremote: Compressing objects: 100% (2/2)\rremote: Compressing objects: 100% (2/2), done.\r\nremote: Total 10 (delta 8), reused 9 (delta 8), pack-reused 0 (from 0)\r\nUnpacking objects: 10% (1/10)\rUnpacking objects: 20% (2/10)\rUnpacking objects: 30% (3/10)\r",,terminal_output +3242,5370378,"TERMINAL",0,0,"Unpacking objects: 40% (4/10)\rUnpacking objects: 50% (5/10)\rUnpacking objects: 60% (6/10)\rUnpacking objects: 70% (7/10)\rUnpacking objects: 80% (8/10)\r",,terminal_output +3243,5370443,"TERMINAL",0,0,"Unpacking objects: 90% (9/10)\rUnpacking objects: 100% (10/10)\rUnpacking objects: 100% (10/10), 2.38 KiB | 14.00 KiB/s, done.\r\n",,terminal_output +3244,5370552,"TERMINAL",0,0,"From github.com:p-doom/jafar\r\n 55a787f..a7c5361 correct-step-time-logging -> origin/correct-step-time-logging\r\n",,terminal_output +3245,5370623,"",0,0,"Switched from branch 'runner-grain' to 'main'",,git_branch_checkout +3246,5370675,"TERMINAL",0,0,"Updating 01461aa..eb5163c\r\n",,terminal_output +3247,5370728,"TERMINAL",0,0,"Fast-forward\r\n",,terminal_output +3248,5370763,"TERMINAL",0,0," genie.py | 181 +++++++++++++++++++++++++++++++------------------\r\n requirements.txt | 3 +-\r\n sample.py | 36 +++++-----\r\n tests/data/generate_dummy_data.py | 93 ++++++++++++--------------\r\n tests/test_dataloader.py | 33 ++++++---\r\n train_dynamics.py | 97 ++++++++++++++++++++-------\r\n train_lam.py | 102 ++++++++++++++++++----------\r\n train_tokenizer.py | 101 ++++++++++++++++++----------\r\n utils/dataloader.py | 214 +++++++++++++++++++++++++++++++++-------------------------\r\n utils/dataset_utils.py | 255 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\r\n utils/preprocess_dataset.py | 59 ++++++++--------\r\n 11 files changed, 821 insertions(+), 353 deletions(-)\r\n create mode 100644 utils/dataset_utils.py\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0",,terminal_output +3249,5388122,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +3250,5388123,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1244,0,"",shellscript,selection_mouse +3251,5392651,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",972,0,"",shellscript,selection_mouse +3252,5392666,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",971,0,"",shellscript,selection_command +3253,5393128,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",933,0,"",shellscript,selection_mouse +3254,5393129,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",932,0,"",shellscript,selection_command +3255,5393705,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",900,0,"",shellscript,selection_mouse +3256,5393706,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",899,0,"",shellscript,selection_command +3257,5394329,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",933,0,"",shellscript,selection_mouse +3258,5394331,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",932,0,"",shellscript,selection_command +3259,5394920,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",900,0,"",shellscript,selection_mouse +3260,5394925,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",899,0,"",shellscript,selection_command +3261,5397836,"train_tokenizer.py",0,0,"from dataclasses import dataclass, field\nimport os\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n min_lr: float = 0.0\n max_lr: float = 3e-4\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 8\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n\n\nargs = tyro.cli(Args)\n\n\ndef tokenizer_loss_fn(params, state, inputs):\n # --- Compute loss ---\n outputs = state.apply_fn(\n params,\n inputs,\n training=True,\n rngs={""params"": inputs[""rng""], ""dropout"": inputs[""dropout_rng""]},\n )\n mse = jnp.square(inputs[""videos""] - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = inputs[""videos""].clip(0, 1).reshape(-1, *inputs[""videos""].shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = pix.psnr(gt, recon).mean() # type: ignore\n ssim = pix.ssim(gt, recon).mean() # type: ignore\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n grad_fn = jax.value_and_grad(tokenizer_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n\n # --- Initialize model ---\n tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_height, args.image_width, args.image_channels)\n inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=jnp.float32,\n ),\n )\n init_params = tokenizer.init(_rng, inputs)\n\n param_counts = count_parameters_by_component(init_params)\n\n if args.log and jax.process_index() == 0:\n wandb.init(\n entity=args.entity,\n project=args.project,\n name=args.name,\n tags=args.tags,\n group=""debug"",\n config=args,\n )\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=tokenizer.apply, params=init_params, tx=tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(\n mesh, PartitionSpec(""data"", None, None, None, None)\n )\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add('model_state', ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler)\n handler_registry.add('model_state', ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler)\n handler_registry.add('dataloader_state', grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add('dataloader_state', grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler) # type: ignore\n \n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n \n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n \n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_train_state = jax.tree_util.tree_map(ocp.utils.to_shape_dtype_struct, train_state)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_train_state),\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator),\n )\n )\n train_state = restored[""model_state""]\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader = (jax.make_array_from_process_local_data(videos_sharding, elem) for elem in grain_iterator) # type: ignore\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _rng_dropout = jax.random.split(rng, 3)\n\n inputs = dict(videos=videos, rng=_rng, dropout_rng=_rng_dropout)\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.StandardSave(train_state),\n dataloader_state=grain.checkpoint.CheckpointSave(grain_iterator),\n )\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()",python,tab +3262,5401767,"train_tokenizer.py",807,0,"",python,selection_mouse +3263,5401882,"train_tokenizer.py",802,9,"save_ckpt",python,selection_mouse +3264,5403700,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3265,5403701,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",900,0,"",shellscript,selection_mouse +3266,5403717,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",899,0,"",shellscript,selection_command +3267,5404722,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",900,0,"\n",shellscript,content +3268,5405304,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",901,0," ",shellscript,content +3269,5406055,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",905,0,"-",shellscript,content +3270,5406056,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",906,0,"",shellscript,selection_keyboard +3271,5406179,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",906,0,"-",shellscript,content +3272,5406180,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",907,0,"",shellscript,selection_keyboard +3273,5406442,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",907,0," ",shellscript,content +3274,5406442,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",908,0,"",shellscript,selection_keyboard +3275,5406767,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",907,1,"",shellscript,content +3276,5407035,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",907,0,"save_ckpt",shellscript,content +3277,5409691,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",916,0,"?",shellscript,content +3278,5409693,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",917,0,"",shellscript,selection_keyboard +3279,5410268,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",916,1,"",shellscript,content +3280,5410913,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",916,0,"=",shellscript,content +3281,5410914,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",917,0,"",shellscript,selection_keyboard +3282,5411251,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",917,0,"t",shellscript,content +3283,5411252,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",918,0,"",shellscript,selection_keyboard +3284,5411326,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",918,0,"r",shellscript,content +3285,5411327,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",919,0,"",shellscript,selection_keyboard +3286,5411416,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",919,0,"u",shellscript,content +3287,5411417,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",920,0,"",shellscript,selection_keyboard +3288,5411505,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",920,0,"e",shellscript,content +3289,5411506,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",921,0,"",shellscript,selection_keyboard +3290,5412679,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",921,0," ",shellscript,content +3291,5412680,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",922,0,"",shellscript,selection_keyboard +3292,5412776,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",922,0,"\",shellscript,content +3293,5412776,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",923,0,"",shellscript,selection_keyboard +3294,5413585,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",922,0,"",shellscript,selection_command +3295,5416201,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",921,0,"",shellscript,selection_command +3296,5416374,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",920,0,"",shellscript,selection_command +3297,5416516,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",919,0,"",shellscript,selection_command +3298,5416654,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",918,0,"",shellscript,selection_command +3299,5416793,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",917,0,"",shellscript,selection_command +3300,5416945,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",916,0,"",shellscript,selection_command +3301,5417199,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",916,1,"",shellscript,content +3302,5417449,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",916,1,"",shellscript,content +3303,5417640,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",916,1,"",shellscript,content +3304,5417815,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",916,1,"",shellscript,content +3305,5418000,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",916,1,"",shellscript,content +3306,5423368,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",940,0,"",shellscript,selection_mouse +3307,5423987,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",929,0,"",shellscript,selection_mouse +3308,5424502,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",915,0,"",shellscript,selection_mouse +3309,5425091,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",938,0,"",shellscript,selection_mouse +3310,5425631,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",915,0,"",shellscript,selection_mouse +3311,5445022,"TERMINAL",0,0,"Step 962, loss: 5.1568522453308105\r\nStep 963, loss: 4.816960334777832\r\nStep 964, loss: 5.064596652984619\r\nStep 965, loss: 4.939937114715576\r\nStep 966, loss: 5.27336311340332\r\nStep 967, loss: 5.090263366699219\r\nStep 968, loss: 4.881126880645752\r\nStep 969, loss: 5.1100921630859375\r\nStep 970, loss: 5.195235252380371\r\nStep 971, loss: 5.071698188781738\r\nStep 972, loss: 4.915706157684326\r\nStep 973, loss: 5.233161926269531\r\nStep 974, loss: 5.001443386077881\r\nStep 975, loss: 5.120009899139404\r\nStep 976, loss: 5.074319362640381\r\nStep 977, loss: 4.791117191314697\r\nStep 978, loss: 5.0973615646362305\r\nStep 979, loss: 5.003950119018555\r\nStep 980, loss: 5.08306884765625\r\nStep 981, loss: 4.865263938903809\r\nStep 982, loss: 5.179849624633789\r\nStep 983, loss: 5.135115623474121\r\nStep 984, loss: 4.828519821166992\r\nStep 985, loss: 5.138967037200928\r\nStep 986, loss: 4.932805061340332\r\nStep 987, loss: 5.0530877113342285\r\nStep 988, loss: 5.030508041381836\r\nStep 989, loss: 5.121220588684082\r\nStep 990, loss: 5.019931316375732\r\nStep 991, loss: 4.990509986877441\r\nStep 992, loss: 4.882480621337891\r\nStep 993, loss: 5.091967582702637\r\nStep 994, loss: 4.786419868469238\r\nStep 995, loss: 4.806129455566406\r\nStep 996, loss: 5.057449817657471\r\nStep 997, loss: 4.943149566650391\r\nStep 998, loss: 4.96529483795166\r\nStep 999, loss: 5.127987384796143\r\nStep 1000, loss: 5.115264892578125\r\nStep 1001, loss: 5.172677993774414\r\nStep 1002, loss: 5.104112148284912\r\nStep 1003, loss: 5.088464736938477\r\nStep 1004, loss: 4.918788433074951\r\nStep 1005, loss: 5.007358074188232\r\nStep 1006, loss: 5.114655494689941\r\nStep 1007, loss: 4.874712944030762\r\nStep 1008, loss: 5.253972053527832\r\nStep 1009, loss: 4.990752696990967\r\nStep 1010, loss: 5.004293918609619\r\nStep 1011, loss: 5.132078647613525\r\nStep 1012, loss: 4.958162784576416\r\nStep 1013, loss: 4.8234992027282715\r\nStep 1014, loss: 5.221305847167969\r\nStep 1015, loss: 4.886592388153076\r\nStep 1016, loss: 4.896177768707275\r\nStep 1017, loss: 5.032711029052734\r\nStep 1018, loss: 4.961817741394043\r\nStep 1019, loss: 5.0470805168151855\r\nStep 1020, loss: 5.076366424560547\r\nStep 1021, loss: 4.905526161193848\r\nStep 1022, loss: 5.042372703552246\r\nStep 1023, loss: 5.161309242248535\r\nStep 1024, loss: 4.904819488525391\r\nStep 1025, loss: 4.9067277908325195\r\nStep 1026, loss: 5.095753192901611\r\nStep 1027, loss: 5.177361488342285\r\nStep 1028, loss: 4.811687469482422\r\nStep 1029, loss: 4.922520637512207\r\nStep 1030, loss: 5.0681610107421875\r\nStep 1031, loss: 4.868681907653809\r\nStep 1032, loss: 5.1157732009887695\r\nStep 1033, loss: 5.085040092468262\r\nStep 1034, loss: 5.087116241455078\r\nStep 1035, loss: 5.127306938171387\r\nStep 1036, loss: 4.932193756103516\r\nStep 1037, loss: 4.980825424194336\r\nStep 1038, loss: 4.920042991638184\r\nStep 1039, loss: 4.951450347900391\r\nStep 1040, loss: 4.9700727462768555\r\nStep 1041, loss: 4.698306083679199\r\nStep 1042, loss: 4.857691287994385\r\nStep 1043, loss: 4.852866172790527\r\nStep 1044, loss: 4.981034278869629\r\nStep 1045, loss: 4.956365585327148\r\nStep 1046, loss: 4.646729469299316\r\nStep 1047, loss: 4.902242183685303\r\nStep 1048, loss: 4.928100109100342\r\nStep 1049, loss: 5.10159969329834\r\nStep 962, loss: 5.1568522453308105\r\nStep 963, loss: 4.816960334777832\r\nStep 964, loss: 5.064596652984619\r\nStep 965, loss: 4.939937114715576\r\nStep 966, loss: 5.27336311340332\r\nStep 967, loss: 5.090263366699219\r\nStep 968, loss: 4.881126880645752\r\nStep 969, loss: 5.1100921630859375\r\nStep 970, loss: 5.195235252380371\r\nStep 971, loss: 5.071698188781738\r\nStep 972, loss: 4.915706157684326\r\nStep 973, loss: 5.233161926269531\r\nStep 974, loss: 5.001443386077881\r\nStep 975, loss: 5.120009899139404\r\nStep 976, loss: 5.074319362640381\r\nStep 977, loss: 4.791117191314697\r\nStep 978, loss: 5.0973615646362305\r\nStep 979, loss: 5.003950119018555\r\nStep 980, loss: 5.08306884765625\r\nStep 981, loss: 4.865263938903809\r\nStep 982, loss: 5.179849624633789\r\nStep 983, loss: 5.135115623474121\r\nStep 984, loss: 4.828519821166992\r\nStep 985, loss: 5.138967037200928\r\nStep 986, loss: 4.932805061340332\r\nStep 987, loss: 5.0530877113342285\r\nStep 988, loss: 5.030508041381836\r\nStep 989, loss: 5.121220588684082\r\nStep 990, loss: 5.019931316375732\r\nStep 991, loss: 4.990509986877441\r\nStep 1050, loss: 4.694330215454102\r\nStep 1051, loss: 4.912756443023682\r\nStep 1052, loss: 4.8436279296875\r\nStep 1053, loss: 4.9195427894592285\r\nStep 1054, loss: 4.9399309158325195\r\nStep 1055, loss: 5.0726518630981445\r\nStep 1056, loss: 5.081210136413574\r\nStep 1057, loss: 5.1472086906433105\r\nStep 1058, loss: 5.104522228240967\r\nStep 1059, loss: 4.850593090057373\r\nStep 1060, loss: 4.889288425445557\r\nStep 1061, loss: 5.073827743530273\r\nStep 1062, loss: 5.254583835601807\r\nStep 1063, loss: 4.925671100616455\r\nStep 1064, loss: 5.084733963012695\r\nStep 1065, loss: 4.838853359222412\r\nStep 1066, loss: 5.06833028793335\r\nStep 1067, loss: 4.889566421508789\r\nStep 1068, loss: 5.001903057098389\r\nStep 1069, loss: 5.257298946380615\r\nStep 1070, loss: 5.1954803466796875\r\nStep 1071, loss: 4.670302867889404\r\nStep 1072, loss: 4.818952560424805\r\nStep 1073, loss: 5.172548294067383\r\nStep 1074, loss: 4.8732733726501465\r\nStep 1075, loss: 5.093983173370361\r\nStep 1076, loss: 4.797558784484863\r\nStep 1077, loss: 5.1093010902404785\r\nStep 1078, loss: 4.9869561195373535\r\nStep 992, loss: 4.882480621337891\r\nStep 993, loss: 5.091967582702637\r\nStep 994, loss: 4.786419868469238\r\nStep 995, loss: 4.806129455566406\r\nStep 996, loss: 5.057449817657471\r\nStep 997, loss: 4.943149566650391\r\nStep 998, loss: 4.96529483795166\r\nStep 999, loss: 5.127987384796143\r\nStep 1000, loss: 5.115264892578125\r\nStep 1001, loss: 5.172677993774414\r\nStep 1002, loss: 5.104112148284912\r\nStep 1003, loss: 5.088464736938477\r\nStep 1004, loss: 4.918788433074951\r\nStep 1005, loss: 5.007358074188232\r\nStep 1006, loss: 5.114655494689941\r\nStep 1007, loss: 4.874712944030762\r\nStep 1008, loss: 5.253972053527832\r\nStep 1009, loss: 4.990752696990967\r\nStep 1010, loss: 5.004293918609619\r\nStep 1011, loss: 5.132078647613525\r\nStep 1012, loss: 4.958162784576416\r\nStep 1013, loss: 4.8234992027282715\r\nStep 1014, loss: 5.221305847167969\r\nStep 1015, loss: 4.886592388153076\r\nStep 1016, loss: 4.896177768707275\r\nStep 1017, loss: 5.032711029052734\r\nStep 1018, loss: 4.961817741394043\r\nStep 1019, loss: 5.0470805168151855\r\nStep 1020, loss: 5.076366424560547\r\nStep 1079, loss: 4.784158706665039\r\nStep 1080, loss: 4.877741813659668\r\nStep 1081, loss: 5.154816150665283\r\nStep 1082, loss: 4.751720428466797\r\nStep 1083, loss: 5.125522613525391\r\nStep 1084, loss: 4.889678955078125\r\nStep 1085, loss: 5.0008063316345215\r\nStep 1086, loss: 4.946294784545898\r\nStep 1087, loss: 5.081248760223389\r\nStep 1088, loss: 5.148041248321533\r\nStep 1089, loss: 4.9699811935424805\r\nStep 1090, loss: 5.008603572845459\r\nStep 1091, loss: 5.2874908447265625\r\nStep 1092, loss: 4.957520961761475\r\nStep 1093, loss: 4.772770404815674\r\nStep 1094, loss: 4.826335906982422\r\nStep 1095, loss: 4.773747444152832\r\nStep 1096, loss: 4.727792739868164\r\nStep 1097, loss: 4.867959976196289\r\nStep 1098, loss: 4.990972995758057\r\nStep 1099, loss: 5.051648139953613\r\nStep 1100, loss: 4.934340953826904\r\nStep 1101, loss: 4.967705726623535\r\nStep 1102, loss: 4.928976058959961\r\nStep 1103, loss: 5.019003868103027\r\nStep 1104, loss: 4.9384446144104\r\nStep 1105, loss: 5.105501174926758\r\nStep 1106, loss: 4.737863540649414\r\nStep 1107, loss: 5.02548885345459\r\nStep 1021, loss: 4.905526161193848\r\nStep 1022, loss: 5.042372703552246\r\nStep 1023, loss: 5.161309242248535\r\nStep 1024, loss: 4.904819488525391\r\nStep 1025, loss: 4.9067277908325195\r\nStep 1026, loss: 5.095753192901611\r\nStep 1027, loss: 5.177361488342285\r\nStep 1028, loss: 4.811687469482422\r\nStep 1029, loss: 4.922520637512207\r\nStep 1030, loss: 5.0681610107421875\r\nStep 1031, loss: 4.868681907653809\r\nStep 1032, loss: 5.1157732009887695\r\nStep 1033, loss: 5.085040092468262\r\nStep 1034, loss: 5.087116241455078\r\nStep 1035, loss: 5.127306938171387\r\nStep 1036, loss: 4.932193756103516\r\nStep 1037, loss: 4.980825424194336\r\nStep 1038, loss: 4.920042991638184\r\nStep 1039, loss: 4.951450347900391\r\nStep 1040, loss: 4.9700727462768555\r\nStep 1041, loss: 4.698306083679199\r\nStep 1042, loss: 4.857691287994385\r\nStep 1043, loss: 4.852866172790527\r\nStep 1044, loss: 4.981034278869629\r\nStep 1045, loss: 4.956365585327148\r\nStep 1046, loss: 4.646729469299316\r\nStep 1047, loss: 4.902242183685303\r\nStep 1048, loss: 4.928100109100342\r\nStep 1049, loss: 5.10159969329834\r\nStep 1108, loss: 4.888282775878906\r\nStep 1109, loss: 5.192723274230957\r\nStep 1110, loss: 4.966083526611328\r\nStep 1111, loss: 4.8921966552734375\r\nStep 1112, loss: 4.865470886230469\r\nStep 1113, loss: 4.701240062713623\r\nStep 1114, loss: 5.101388454437256\r\nStep 1115, loss: 4.849363803863525\r\nStep 1116, loss: 5.013510704040527\r\nStep 1117, loss: 5.012414932250977\r\nStep 1118, loss: 4.933382987976074\r\nStep 1119, loss: 4.641906261444092\r\nStep 1120, loss: 4.92882776260376\r\nStep 1121, loss: 4.703517913818359\r\nStep 1122, loss: 4.972881317138672\r\nStep 1123, loss: 5.049308776855469\r\nStep 1124, loss: 4.712826728820801\r\nStep 1125, loss: 4.8883585929870605\r\nStep 1126, loss: 5.122026443481445\r\nStep 1127, loss: 5.010403156280518\r\nStep 1128, loss: 4.868868827819824\r\nStep 1129, loss: 5.078428268432617\r\nStep 1130, loss: 4.8772053718566895\r\nStep 1131, loss: 4.8967671394348145\r\nStep 1132, loss: 4.88825798034668\r\nStep 1133, loss: 4.8844780921936035\r\nStep 1134, loss: 4.8840436935424805\r\nStep 1135, loss: 4.7452850341796875\r\nStep 1136, loss: 5.050529479980469\r\nStep 1050, loss: 4.694330215454102\r\nStep 1051, loss: 4.912756443023682\r\nStep 1052, loss: 4.8436279296875\r\nStep 1053, loss: 4.9195427894592285\r\nStep 1054, loss: 4.9399309158325195\r\nStep 1055, loss: 5.0726518630981445\r\nStep 1056, loss: 5.081210136413574\r\nStep 1057, loss: 5.1472086906433105\r\nStep 1058, loss: 5.104522228240967\r\nStep 1059, loss: 4.850593090057373\r\nStep 1060, loss: 4.889288425445557\r\nStep 1061, loss: 5.073827743530273\r\nStep 1062, loss: 5.254583835601807\r\nStep 1063, loss: 4.925671100616455\r\nStep 1064, loss: 5.084733963012695\r\nStep 1065, loss: 4.838853359222412\r\nStep 1066, loss: 5.06833028793335\r\nStep 1067, loss: 4.889566421508789\r\nStep 1068, loss: 5.001903057098389\r\nStep 1069, loss: 5.257298946380615\r\nStep 1070, loss: 5.1954803466796875\r\nStep 1071, loss: 4.670302867889404\r\nStep 1072, loss: 4.818952560424805\r\nStep 1073, loss: 5.172548294067383\r\nStep 1074, loss: 4.8732733726501465\r\nStep 1075, loss: 5.093983173370361\r\nStep 1076, loss: 4.797558784484863\r\nStep 1077, loss: 5.1093010902404785\r\nStep 1078, loss: 4.9869561195373535\r\nStep 1137, loss: 4.917625904083252\r\nStep 1138, loss: 4.829206943511963\r\nStep 1139, loss: 4.8869781494140625\r\nStep 1140, loss: 4.968005657196045\r\nStep 1141, loss: 4.84523868560791\r\nStep 1142, loss: 4.98886251449585\r\nStep 1143, loss: 4.957108497619629\r\nStep 1144, loss: 5.097373008728027\r\nStep 1145, loss: 4.651041507720947\r\nStep 1146, loss: 4.8319783210754395\r\nStep 1147, loss: 4.675942897796631\r\nStep 1148, loss: 5.058229446411133\r\nStep 1149, loss: 4.955995082855225\r\nStep 1150, loss: 4.861543655395508\r\nStep 1151, loss: 4.811530590057373\r\nStep 1152, loss: 4.847794532775879\r\nStep 1153, loss: 5.00905704498291\r\nStep 1154, loss: 4.871018886566162\r\nStep 1155, loss: 4.829107284545898\r\nStep 1156, loss: 4.926144599914551\r\nStep 1157, loss: 4.929976940155029\r\nStep 1158, loss: 4.581959247589111\r\nStep 1159, loss: 4.8127546310424805\r\nStep 1160, loss: 4.829219341278076\r\nStep 1161, loss: 5.095797061920166\r\nStep 1162, loss: 5.034276008605957\r\nStep 1163, loss: 4.616420269012451\r\nStep 1164, loss: 5.0754852294921875\r\nStep 1165, loss: 4.784201622009277\r\nStep 1079, loss: 4.784158706665039\r\nStep 1080, loss: 4.877741813659668\r\nStep 1081, loss: 5.154816150665283\r\nStep 1082, loss: 4.751720428466797\r\nStep 1083, loss: 5.125522613525391\r\nStep 1084, loss: 4.889678955078125\r\nStep 1085, loss: 5.0008063316345215\r\nStep 1086, loss: 4.946294784545898\r\nStep 1087, loss: 5.081248760223389\r\nStep 1088, loss: 5.148041248321533\r\nStep 1089, loss: 4.9699811935424805\r\nStep 1090, loss: 5.008603572845459\r\nStep 1091, loss: 5.2874908447265625\r\nStep 1092, loss: 4.957520961761475\r\nStep 1093, loss: 4.772770404815674\r\nStep 1094, loss: 4.826335906982422\r\nStep 1095, loss: 4.773747444152832\r\nStep 1096, loss: 4.727792739868164\r\nStep 1097, loss: 4.867959976196289\r\nStep 1098, loss: 4.990972995758057\r\nStep 1099, loss: 5.051648139953613\r\nStep 1100, loss: 4.934340953826904\r\nStep 1101, loss: 4.967705726623535\r\nStep 1102, loss: 4.928976058959961\r\nStep 1103, loss: 5.019003868103027\r\nStep 1104, loss: 4.9384446144104\r\nStep 1105, loss: 5.105501174926758\r\nStep 1106, loss: 4.737863540649414\r\nStep 1107, loss: 5.02548885345459\r\nStep 1166, loss: 4.93991231918335\r\nStep 1167, loss: 4.745754241943359\r\nStep 1168, loss: 4.946484088897705\r\nStep 1169, loss: 4.959706783294678\r\nStep 1170, loss: 5.136929512023926\r\nStep 1171, loss: 4.963235855102539\r\nStep 1172, loss: 4.689530372619629\r\nStep 1173, loss: 4.873706340789795\r\nStep 1174, loss: 5.097964286804199\r\nStep 1175, loss: 4.9454545974731445\r\nStep 1176, loss: 4.84202766418457\r\nStep 1177, loss: 5.002065181732178\r\nStep 1178, loss: 5.085189342498779\r\nStep 1179, loss: 5.059033393859863\r\nStep 1180, loss: 5.012739181518555\r\nStep 1181, loss: 4.655002117156982\r\nStep 1182, loss: 5.119167804718018\r\nStep 1183, loss: 4.922578811645508\r\nStep 1184, loss: 4.863205909729004\r\nStep 1185, loss: 4.81952428817749\r\nStep 1186, loss: 4.806148052215576\r\nStep 1187, loss: 4.85841703414917\r\nStep 1188, loss: 5.1563920974731445\r\nStep 1189, loss: 4.81588077545166\r\nStep 1190, loss: 4.9441118240356445\r\nStep 1191, loss: 4.791698932647705\r\nStep 1192, loss: 5.128923416137695\r\nStep 1193, loss: 4.806921005249023\r\nStep 1194, loss: 5.237899303436279\r\nStep 1108, loss: 4.888282775878906\r\nStep 1109, loss: 5.192723274230957\r\nStep 1110, loss: 4.966083526611328\r\nStep 1111, loss: 4.8921966552734375\r\nStep 1112, loss: 4.865470886230469\r\nStep 1113, loss: 4.701240062713623\r\nStep 1114, loss: 5.101388454437256\r\nStep 1115, loss: 4.849363803863525\r\nStep 1116, loss: 5.013510704040527\r\nStep 1117, loss: 5.012414932250977\r\nStep 1118, loss: 4.933382987976074\r\nStep 1119, loss: 4.641906261444092\r\nStep 1120, loss: 4.92882776260376\r\nStep 1121, loss: 4.703517913818359\r\nStep 1122, loss: 4.972881317138672\r\nStep 1123, loss: 5.049308776855469\r\nStep 1124, loss: 4.712826728820801\r\nStep 1125, loss: 4.8883585929870605\r\nStep 1126, loss: 5.122026443481445\r\nStep 1127, loss: 5.010403156280518\r\nStep 1128, loss: 4.868868827819824\r\nStep 1129, loss: 5.078428268432617\r\nStep 1130, loss: 4.8772053718566895\r\nStep 1131, loss: 4.8967671394348145\r\nStep 1132, loss: 4.88825798034668\r\nStep 1133, loss: 4.8844780921936035\r\nStep 1134, loss: 4.8840436935424805\r\nStep 1135, loss: 4.7452850341796875\r\nStep 1136, loss: 5.050529479980469\r\nStep 1195, loss: 4.842819690704346\r\nStep 1137, loss: 4.917625904083252\r\nStep 1138, loss: 4.829206943511963\r\nStep 1139, loss: 4.8869781494140625\r\nStep 1140, loss: 4.968005657196045\r\nStep 1141, loss: 4.84523868560791\r\nStep 1142, loss: 4.98886251449585\r\nStep 1143, loss: 4.957108497619629\r\nStep 1144, loss: 5.097373008728027\r\nStep 1145, loss: 4.651041507720947\r\nStep 1146, loss: 4.8319783210754395\r\nStep 1147, loss: 4.675942897796631\r\nStep 1148, loss: 5.058229446411133\r\nStep 1149, loss: 4.955995082855225\r\nStep 1150, loss: 4.861543655395508\r\nStep 1151, loss: 4.811530590057373\r\nStep 1152, loss: 4.847794532775879\r\nStep 1153, loss: 5.00905704498291\r\nStep 1154, loss: 4.871018886566162\r\nStep 1155, loss: 4.829107284545898\r\nStep 1156, loss: 4.926144599914551\r\nStep 1157, loss: 4.929976940155029\r\nStep 1158, loss: 4.581959247589111\r\nStep 1159, loss: 4.8127546310424805\r\nStep 1160, loss: 4.829219341278076\r\nStep 1161, loss: 5.095797061920166\r\nStep 1162, loss: 5.034276008605957\r\nStep 1163, loss: 4.616420269012451\r\nStep 1164, loss: 5.0754852294921875\r\nStep 1165, loss: 4.784201622009277\r\nStep 1166, loss: 4.93991231918335\r\nStep 1167, loss: 4.745754241943359\r\nStep 1168, loss: 4.946484088897705\r\nStep 1169, loss: 4.959706783294678\r\nStep 1170, loss: 5.136929512023926\r\nStep 1171, loss: 4.963235855102539\r\nStep 1172, loss: 4.689530372619629\r\nStep 1173, loss: 4.873706340789795\r\nStep 1174, loss: 5.097964286804199\r\nStep 1175, loss: 4.9454545974731445\r\nStep 1176, loss: 4.84202766418457\r\nStep 1177, loss: 5.002065181732178\r\nStep 1178, loss: 5.085189342498779\r\nStep 1179, loss: 5.059033393859863\r\nStep 1180, loss: 5.012739181518555\r\nStep 1181, loss: 4.655002117156982\r\nStep 1182, loss: 5.119167804718018\r\nStep 1183, loss: 4.922578811645508\r\nStep 1184, loss: 4.863205909729004\r\nStep 1185, loss: 4.81952428817749\r\nStep 1186, loss: 4.806148052215576\r\nStep 1187, loss: 4.85841703414917\r\nStep 1188, loss: 5.1563920974731445\r\nStep 1189, loss: 4.81588077545166\r\nStep 1190, loss: 4.9441118240356445\r\nStep 1191, loss: 4.791698932647705\r\nStep 1192, loss: 5.128923416137695\r\nStep 1193, loss: 4.806921005249023\r\nStep 1194, loss: 5.237899303436279\r\nStep 1195, loss: 4.842819690704346\r\nStep 962, loss: 5.1568522453308105\r\nStep 963, loss: 4.816960334777832\r\nStep 964, loss: 5.064596652984619\r\nStep 965, loss: 4.939937114715576\r\nStep 966, loss: 5.27336311340332\r\nStep 967, loss: 5.090263366699219\r\nStep 968, loss: 4.881126880645752\r\nStep 969, loss: 5.1100921630859375\r\nStep 970, loss: 5.195235252380371\r\nStep 971, loss: 5.071698188781738\r\nStep 972, loss: 4.915706157684326\r\nStep 973, loss: 5.233161926269531\r\nStep 974, loss: 5.001443386077881\r\nStep 975, loss: 5.120009899139404\r\nStep 976, loss: 5.074319362640381\r\nStep 977, loss: 4.791117191314697\r\nStep 978, loss: 5.0973615646362305\r\nStep 979, loss: 5.003950119018555\r\nStep 980, loss: 5.08306884765625\r\nStep 981, loss: 4.865263938903809\r\nStep 982, loss: 5.179849624633789\r\nStep 983, loss: 5.135115623474121\r\nStep 984, loss: 4.828519821166992\r\nStep 985, loss: 5.138967037200928\r\nStep 986, loss: 4.932805061340332\r\nStep 987, loss: 5.0530877113342285\r\nStep 988, loss: 5.030508041381836\r\nStep 989, loss: 5.121220588684082\r\nStep 990, loss: 5.019931316375732\r\nStep 991, loss: 4.990509986877441\r\nStep 992, loss: 4.882480621337891\r\nStep 993, loss: 5.091967582702637\r\nStep 994, loss: 4.786419868469238\r\nStep 995, loss: 4.806129455566406\r\nStep 996, loss: 5.057449817657471\r\nStep 997, loss: 4.943149566650391\r\nStep 998, loss: 4.96529483795166\r\nStep 999, loss: 5.127987384796143\r\nStep 1000, loss: 5.115264892578125\r\nStep 1001, loss: 5.172677993774414\r\nStep 1002, loss: 5.104112148284912\r\nStep 1003, loss: 5.088464736938477\r\nStep 1004, loss: 4.918788433074951\r\nStep 1005, loss: 5.007358074188232\r\nStep 1006, loss: 5.114655494689941\r\nStep 1007, loss: 4.874712944030762\r\nStep 1008, loss: 5.253972053527832\r\nStep 1009, loss: 4.990752696990967\r\nStep 1010, loss: 5.004293918609619\r\nStep 1011, loss: 5.132078647613525\r\nStep 1012, loss: 4.958162784576416\r\nStep 1013, loss: 4.8234992027282715\r\nStep 1014, loss: 5.221305847167969\r\nStep 1015, loss: 4.886592388153076\r\nStep 1016, loss: 4.896177768707275\r\nStep 1017, loss: 5.032711029052734\r\nStep 1018, loss: 4.961817741394043\r\nStep 1019, loss: 5.0470805168151855\r\nStep 1020, loss: 5.076366424560547\r\nStep 1021, loss: 4.905526161193848\r\nStep 1022, loss: 5.042372703552246\r\nStep 1023, loss: 5.161309242248535\r\nStep 1024, loss: 4.904819488525391\r\nStep 1025, loss: 4.9067277908325195\r\nStep 1026, loss: 5.095753192901611\r\nStep 1027, loss: 5.177361488342285\r\nStep 1028, loss: 4.811687469482422\r\nStep 1029, loss: 4.922520637512207\r\nStep 1030, loss: 5.0681610107421875\r\nStep 1031, loss: 4.868681907653809\r\nStep 1032, loss: 5.1157732009887695\r\nStep 1033, loss: 5.085040092468262\r\nStep 1034, loss: 5.087116241455078\r\nStep 1035, loss: 5.127306938171387\r\nStep 1036, loss: 4.932193756103516\r\nStep 1037, loss: 4.980825424194336\r\nStep 1038, loss: 4.920042991638184\r\nStep 1039, loss: 4.951450347900391\r\nStep 1040, loss: 4.9700727462768555\r\nStep 1041, loss: 4.698306083679199\r\nStep 1042, loss: 4.857691287994385\r\nStep 1043, loss: 4.852866172790527\r\nStep 1044, loss: 4.981034278869629\r\nStep 1045, loss: 4.956365585327148\r\nStep 1046, loss: 4.646729469299316\r\nStep 1047, loss: 4.902242183685303\r\nStep 1048, loss: 4.928100109100342\r\nStep 1049, loss: 5.10159969329834\r\nStep 1050, loss: 4.694330215454102\r\nStep 1051, loss: 4.912756443023682\r\nStep 1052, loss: 4.8436279296875\r\nStep 1053, loss: 4.9195427894592285\r\nStep 1054, loss: 4.9399309158325195\r\nStep 1055, loss: 5.0726518630981445\r\nStep 1056, loss: 5.081210136413574\r\nStep 1057, loss: 5.1472086906433105\r\nStep 1058, loss: 5.104522228240967\r\nStep 1059, loss: 4.850593090057373\r\nStep 1060, loss: 4.889288425445557\r\nStep 1061, loss: 5.073827743530273\r\nStep 1062, loss: 5.254583835601807\r\nStep 1063, loss: 4.925671100616455\r\nStep 1064, loss: 5.084733963012695\r\nStep 1065, loss: 4.838853359222412\r\nStep 1066, loss: 5.06833028793335\r\nStep 1067, loss: 4.889566421508789\r\nStep 1068, loss: 5.001903057098389\r\nStep 1069, loss: 5.257298946380615\r\nStep 1070, loss: 5.1954803466796875\r\nStep 1071, loss: 4.670302867889404\r\nStep 1072, loss: 4.818952560424805\r\nStep 1073, loss: 5.172548294067383\r\nStep 1074, loss: 4.8732733726501465\r\nStep 1075, loss: 5.093983173370361\r\nStep 1076, loss: 4.797558784484863\r\nStep 1077, loss: 5.1093010902404785\r\nStep 1078, loss: 4.9869561195373535\r\nStep 1079, loss: 4.784158706665039\r\nStep 1080, loss: 4.877741813659668\r\nStep 1081, loss: 5.154816150665283\r\nStep 1082, loss: 4.751720428466797\r\nStep 1083, loss: 5.125522613525391\r\nStep 1084, loss: 4.889678955078125\r\nStep 1085, loss: 5.0008063316345215\r\nStep 1086, loss: 4.946294784545898\r\nStep 1087, loss: 5.081248760223389\r\nStep 1088, loss: 5.148041248321533\r\nStep 1089, loss: 4.9699811935424805\r\nStep 1090, loss: 5.008603572845459\r\nStep 1091, loss: 5.2874908447265625\r\nStep 1092, loss: 4.957520961761475\r\nStep 1093, loss: 4.772770404815674\r\nStep 1094, loss: 4.826335906982422\r\nStep 1095, loss: 4.773747444152832\r\nStep 1096, loss: 4.727792739868164\r\nStep 1097, loss: 4.867959976196289\r\nStep 1098, loss: 4.990972995758057\r\nStep 1099, loss: 5.051648139953613\r\nStep 1100, loss: 4.934340953826904\r\nStep 1101, loss: 4.967705726623535\r\nStep 1102, loss: 4.928976058959961\r\nStep 1103, loss: 5.019003868103027\r\nStep 1104, loss: 4.9384446144104\r\nStep 1105, loss: 5.105501174926758\r\nStep 1106, loss: 4.737863540649414\r\nStep 1107, loss: 5.02548885345459\r\nStep 1108, loss: 4.888282775878906\r\nStep 1109, loss: 5.192723274230957\r\nStep 1110, loss: 4.966083526611328\r\nStep 1111, loss: 4.8921966552734375\r\nStep 1112, loss: 4.865470886230469\r\nStep 1113, loss: 4.701240062713623\r\nStep 1114, loss: 5.101388454437256\r\nStep 1115, loss: 4.849363803863525\r\nStep 1116, loss: 5.013510704040527\r\nStep 1117, loss: 5.012414932250977\r\nStep 1118, loss: 4.933382987976074\r\nStep 1119, loss: 4.641906261444092\r\nStep 1120, loss: 4.92882776260376\r\nStep 1121, loss: 4.703517913818359\r\nStep 1122, loss: 4.972881317138672\r\nStep 1123, loss: 5.049308776855469\r\nStep 1124, loss: 4.712826728820801\r\nStep 1125, loss: 4.8883585929870605\r\nStep 1126, loss: 5.122026443481445\r\nStep 1127, loss: 5.010403156280518\r\nStep 1128, loss: 4.868868827819824\r\nStep 1129, loss: 5.078428268432617\r\nStep 1130, loss: 4.8772053718566895\r\nStep 1131, loss: 4.8967671394348145\r\nStep 1132, loss: 4.88825798034668\r\nStep 1133, loss: 4.8844780921936035\r\nStep 1134, loss: 4.8840436935424805\r\nStep 1135, loss: 4.7452850341796875\r\nStep 1136, loss: 5.050529479980469\r\nStep 1137, loss: 4.917625904083252\r\nStep 1138, loss: 4.829206943511963\r\nStep 1139, loss: 4.8869781494140625\r\nStep 1140, loss: 4.968005657196045\r\nStep 1141, loss: 4.84523868560791\r\nStep 1142, loss: 4.98886251449585\r\nStep 1143, loss: 4.957108497619629\r\nStep 1144, loss: 5.097373008728027\r\nStep 1145, loss: 4.651041507720947\r\nStep 1146, loss: 4.8319783210754395\r\nStep 1147, loss: 4.675942897796631\r\nStep 1148, loss: 5.058229446411133\r\nStep 1149, loss: 4.955995082855225\r\nStep 1150, loss: 4.861543655395508\r\nStep 1151, loss: 4.811530590057373\r\nStep 1152, loss: 4.847794532775879\r\nStep 1153, loss: 5.00905704498291\r\nStep 1154, loss: 4.871018886566162\r\nStep 1155, loss: 4.829107284545898\r\nStep 1156, loss: 4.926144599914551\r\nStep 1157, loss: 4.929976940155029\r\nStep 1158, loss: 4.581959247589111\r\nStep 1159, loss: 4.8127546310424805\r\nStep 1160, loss: 4.829219341278076\r\nStep 1161, loss: 5.095797061920166\r\nStep 1162, loss: 5.034276008605957\r\nStep 1163, loss: 4.616420269012451\r\nStep 1164, loss: 5.0754852294921875\r\nStep 1165, loss: 4.784201622009277\r\nStep 1166, loss: 4.93991231918335\r\nStep 1167, loss: 4.745754241943359\r\nStep 1168, loss: 4.946484088897705\r\nStep 1169, loss: 4.959706783294678\r\nStep 1170, loss: 5.136929512023926\r\nStep 1171, loss: 4.963235855102539\r\nStep 1172, loss: 4.689530372619629\r\nStep 1173, loss: 4.873706340789795\r\nStep 1174, loss: 5.097964286804199\r\nStep 1175, loss: 4.9454545974731445\r\nStep 1176, loss: 4.84202766418457\r\nStep 1177, loss: 5.002065181732178\r\nStep 1178, loss: 5.085189342498779\r\nStep 1179, loss: 5.059033393859863\r\nStep 1180, loss: 5.012739181518555\r\nStep 1181, loss: 4.655002117156982\r\nStep 1182, loss: 5.119167804718018\r\nStep 1183, loss: 4.922578811645508\r\nStep 1184, loss: 4.863205909729004\r\nStep 1185, loss: 4.81952428817749\r\nStep 1186, loss: 4.806148052215576\r\nStep 1187, loss: 4.85841703414917\r\nStep 1188, loss: 5.1563920974731445\r\nStep 1189, loss: 4.81588077545166\r\nStep 1190, loss: 4.9441118240356445\r\nStep 1191, loss: 4.791698932647705\r\nStep 1192, loss: 5.128923416137695\r\nStep 1193, loss: 4.806921005249023\r\nStep 1194, loss: 5.237899303436279\r\nStep 1195, loss: 4.842819690704346\r\nStep 962, loss: 5.1568522453308105\r\nStep 963, loss: 4.816960334777832\r\nStep 964, loss: 5.064596652984619\r\nStep 965, loss: 4.939937114715576\r\nStep 966, loss: 5.27336311340332\r\nStep 967, loss: 5.090263366699219\r\nStep 968, loss: 4.881126880645752\r\nStep 969, loss: 5.1100921630859375\r\nStep 970, loss: 5.195235252380371\r\nStep 971, loss: 5.071698188781738\r\nStep 972, loss: 4.915706157684326\r\nStep 973, loss: 5.233161926269531\r\nStep 974, loss: 5.001443386077881\r\nStep 975, loss: 5.120009899139404\r\nStep 976, loss: 5.074319362640381\r\nStep 977, loss: 4.791117191314697\r\nStep 978, loss: 5.0973615646362305\r\nStep 979, loss: 5.003950119018555\r\nStep 980, loss: 5.08306884765625\r\nStep 981, loss: 4.865263938903809\r\nStep 982, loss: 5.179849624633789\r\nStep 983, loss: 5.135115623474121\r\nStep 984, loss: 4.828519821166992\r\nStep 985, loss: 5.138967037200928\r\nStep 986, loss: 4.932805061340332\r\nStep 987, loss: 5.0530877113342285\r\nStep 988, loss: 5.030508041381836\r\nStep 989, loss: 5.121220588684082\r\nStep 990, loss: 5.019931316375732\r\nStep 991, loss: 4.990509986877441\r\nStep 992, loss: 4.882480621337891\r\nStep 993, loss: 5.091967582702637\r\nStep 994, loss: 4.786419868469238\r\nStep 995, loss: 4.806129455566406\r\nStep 996, loss: 5.057449817657471\r\nStep 997, loss: 4.943149566650391\r\nStep 998, loss: 4.96529483795166\r\nStep 999, loss: 5.127987384796143\r\nStep 1000, loss: 5.115264892578125\r\nStep 1001, loss: 5.172677993774414\r\nStep 1002, loss: 5.104112148284912\r\nStep 1003, loss: 5.088464736938477\r\nStep 1004, loss: 4.918788433074951\r\nStep 1005, loss: 5.007358074188232\r\nStep 1006, loss: 5.114655494689941\r\nStep 1007, loss: 4.874712944030762\r\nStep 1008, loss: 5.253972053527832\r\nStep 1009, loss: 4.990752696990967\r\nStep 1010, loss: 5.004293918609619\r\nStep 1011, loss: 5.132078647613525\r\nStep 1012, loss: 4.958162784576416\r\nStep 1013, loss: 4.8234992027282715\r\nStep 1014, loss: 5.221305847167969\r\nStep 1015, loss: 4.886592388153076\r\nStep 1016, loss: 4.896177768707275\r\nStep 1017, loss: 5.032711029052734\r\nStep 1018, loss: 4.961817741394043\r\nStep 1019, loss: 5.0470805168151855\r\nStep 1020, loss: 5.076366424560547\r\nStep 1021, loss: 4.905526161193848\r\nStep 1022, loss: 5.042372703552246\r\nStep 1023, loss: 5.161309242248535\r\nStep 1024, loss: 4.904819488525391\r\nStep 1025, loss: 4.9067277908325195\r\nStep 1026, loss: 5.095753192901611\r\nStep 1027, loss: 5.177361488342285\r\nStep 1028, loss: 4.811687469482422\r\nStep 1029, loss: 4.922520637512207\r\nStep 1030, loss: 5.0681610107421875\r\nStep 1031, loss: 4.868681907653809\r\nStep 1032, loss: 5.1157732009887695\r\nStep 1033, loss: 5.085040092468262\r\nStep 1034, loss: 5.087116241455078\r\nStep 1035, loss: 5.127306938171387\r\nStep 1036, loss: 4.932193756103516\r\nStep 1037, loss: 4.980825424194336\r\nStep 1038, loss: 4.920042991638184\r\nStep 1039, loss: 4.951450347900391\r\nStep 1040, loss: 4.9700727462768555\r\nStep 1041, loss: 4.698306083679199\r\nStep 1042, loss: 4.857691287994385\r\nStep 1043, loss: 4.852866172790527\r\nStep 1044, loss: 4.981034278869629\r\nStep 1045, loss: 4.956365585327148\r\nStep 1046, loss: 4.646729469299316\r\nStep 1047, loss: 4.902242183685303\r\nStep 1048, loss: 4.928100109100342\r\nStep 1049, loss: 5.10159969329834\r\nStep 962, loss: 5.1568522453308105\r\nStep 963, loss: 4.816960334777832\r\nStep 964, loss: 5.064596652984619\r\nStep 965, loss: 4.939937114715576\r\nStep 966, loss: 5.27336311340332\r\nStep 967, loss: 5.090263366699219\r\nStep 968, loss: 4.881126880645752\r\nStep 969, loss: 5.1100921630859375\r\nStep 970, loss: 5.195235252380371\r\nStep 971, loss: 5.071698188781738\r\nStep 972, loss: 4.915706157684326\r\nStep 973, loss: 5.233161926269531\r\nStep 974, loss: 5.001443386077881\r\nStep 975, loss: 5.120009899139404\r\nStep 976, loss: 5.074319362640381\r\nStep 977, loss: 4.791117191314697\r\nStep 978, loss: 5.0973615646362305\r\nStep 979, loss: 5.003950119018555\r\nStep 980, loss: 5.08306884765625\r\nStep 981, loss: 4.865263938903809\r\nStep 982, loss: 5.179849624633789\r\nStep 983, loss: 5.135115623474121\r\nStep 984, loss: 4.828519821166992\r\nStep 985, loss: 5.138967037200928\r\nStep 986, loss: 4.932805061340332\r\nStep 987, loss: 5.0530877113342285\r\nStep 988, loss: 5.030508041381836\r\nStep 989, loss: 5.121220588684082\r\nStep 990, loss: 5.019931316375732\r\nStep 991, loss: 4.990509986877441\r\nStep 1050, loss: 4.694330215454102\r\nStep 1051, loss: 4.912756443023682\r\nStep 1052, loss: 4.8436279296875\r\nStep 1053, loss: 4.9195427894592285\r\nStep 1054, loss: 4.9399309158325195\r\nStep 1055, loss: 5.0726518630981445\r\nStep 1056, loss: 5.081210136413574\r\nStep 1057, loss: 5.1472086906433105\r\nStep 1058, loss: 5.104522228240967\r\nStep 1059, loss: 4.850593090057373\r\nStep 1060, loss: 4.889288425445557\r\nStep 1061, loss: 5.073827743530273\r\nStep 1062, loss: 5.254583835601807\r\nStep 1063, loss: 4.925671100616455\r\nStep 1064, loss: 5.084733963012695\r\nStep 1065, loss: 4.838853359222412\r\nStep 1066, loss: 5.06833028793335\r\nStep 1067, loss: 4.889566421508789\r\nStep 1068, loss: 5.001903057098389\r\nStep 1069, loss: 5.257298946380615\r\nStep 1070, loss: 5.1954803466796875\r\nStep 1071, loss: 4.670302867889404\r\nStep 1072, loss: 4.818952560424805\r\nStep 1073, loss: 5.172548294067383\r\nStep 1074, loss: 4.8732733726501465\r\nStep 1075, loss: 5.093983173370361\r\nStep 1076, loss: 4.797558784484863\r\nStep 1077, loss: 5.1093010902404785\r\nStep 1078, loss: 4.9869561195373535\r\nStep 992, loss: 4.882480621337891\r\nStep 993, loss: 5.091967582702637\r\nStep 994, loss: 4.786419868469238\r\nStep 995, loss: 4.806129455566406\r\nStep 996, loss: 5.057449817657471\r\nStep 997, loss: 4.943149566650391\r\nStep 998, loss: 4.96529483795166\r\nStep 999, loss: 5.127987384796143\r\nStep 1000, loss: 5.115264892578125\r\nStep 1001, loss: 5.172677993774414\r\nStep 1002, loss: 5.104112148284912\r\nStep 1003, loss: 5.088464736938477\r\nStep 1004, loss: 4.918788433074951\r\nStep 1005, loss: 5.007358074188232\r\nStep 1006, loss: 5.114655494689941\r\nStep 1007, loss: 4.874712944030762\r\nStep 1008, loss: 5.253972053527832\r\nStep 1009, loss: 4.990752696990967\r\nStep 1010, loss: 5.004293918609619\r\nStep 1011, loss: 5.132078647613525\r\nStep 1012, loss: 4.958162784576416\r\nStep 1013, loss: 4.8234992027282715\r\nStep 1014, loss: 5.221305847167969\r\nStep 1015, loss: 4.886592388153076\r\nStep 1016, loss: 4.896177768707275\r\nStep 1017, loss: 5.032711029052734\r\nStep 1018, loss: 4.961817741394043\r\nStep 1019, loss: 5.0470805168151855\r\nStep 1020, loss: 5.076366424560547\r\nStep 962, loss: 5.1568522453308105\r\nStep 963, loss: 4.816960334777832\r\nStep 964, loss: 5.064596652984619\r\nStep 965, loss: 4.939937114715576\r\nStep 966, loss: 5.27336311340332\r\nStep 967, loss: 5.090263366699219\r\nStep 968, loss: 4.881126880645752\r\nStep 969, loss: 5.1100921630859375\r\nStep 970, loss: 5.195235252380371\r\nStep 971, loss: 5.071698188781738\r\nStep 972, loss: 4.915706157684326\r\nStep 973, loss: 5.233161926269531\r\nStep 974, loss: 5.001443386077881\r\nStep 975, loss: 5.120009899139404\r\nStep 976, loss: 5.074319362640381\r\nStep 977, loss: 4.791117191314697\r\nStep 978, loss: 5.0973615646362305\r\nStep 979, loss: 5.003950119018555\r\nStep 980, loss: 5.08306884765625\r\nStep 981, loss: 4.865263938903809\r\nStep 982, loss: 5.179849624633789\r\nStep 983, loss: 5.135115623474121\r\nStep 984, loss: 4.828519821166992\r\nStep 985, loss: 5.138967037200928\r\nStep 986, loss: 4.932805061340332\r\nStep 987, loss: 5.0530877113342285\r\nStep 988, loss: 5.030508041381836\r\nStep 989, loss: 5.121220588684082\r\nStep 990, loss: 5.019931316375732\r\nStep 991, loss: 4.990509986877441\r\nStep 1021, loss: 4.905526161193848\r\nStep 1022, loss: 5.042372703552246\r\nStep 1023, loss: 5.161309242248535\r\nStep 1024, loss: 4.904819488525391\r\nStep 1025, loss: 4.9067277908325195\r\nStep 1026, loss: 5.095753192901611\r\nStep 1027, loss: 5.177361488342285\r\nStep 1028, loss: 4.811687469482422\r\nStep 1029, loss: 4.922520637512207\r\nStep 1030, loss: 5.0681610107421875\r\nStep 1031, loss: 4.868681907653809\r\nStep 1032, loss: 5.1157732009887695\r\nStep 1033, loss: 5.085040092468262\r\nStep 1034, loss: 5.087116241455078\r\nStep 1035, loss: 5.127306938171387\r\nStep 1036, loss: 4.932193756103516\r\nStep 1037, loss: 4.980825424194336\r\nStep 1038, loss: 4.920042991638184\r\nStep 1039, loss: 4.951450347900391\r\nStep 1040, loss: 4.9700727462768555\r\nStep 1041, loss: 4.698306083679199\r\nStep 1042, loss: 4.857691287994385\r\nStep 1043, loss: 4.852866172790527\r\nStep 1044, loss: 4.981034278869629\r\nStep 1045, loss: 4.956365585327148\r\nStep 1046, loss: 4.646729469299316\r\nStep 1047, loss: 4.902242183685303\r\nStep 1048, loss: 4.928100109100342\r\nStep 1049, loss: 5.10159969329834\r\nStep 992, loss: 4.882480621337891\r\nStep 993, loss: 5.091967582702637\r\nStep 994, loss: 4.786419868469238\r\nStep 995, loss: 4.806129455566406\r\nStep 996, loss: 5.057449817657471\r\nStep 997, loss: 4.943149566650391\r\nStep 998, loss: 4.96529483795166\r\nStep 999, loss: 5.127987384796143\r\nStep 1000, loss: 5.115264892578125\r\nStep 1001, loss: 5.172677993774414\r\nStep 1002, loss: 5.104112148284912\r\nStep 1003, loss: 5.088464736938477\r\nStep 1004, loss: 4.918788433074951\r\nStep 1005, loss: 5.007358074188232\r\nStep 1006, loss: 5.114655494689941\r\nStep 1007, loss: 4.874712944030762\r\nStep 1008, loss: 5.253972053527832\r\nStep 1009, loss: 4.990752696990967\r\nStep 1010, loss: 5.004293918609619\r\nStep 1011, loss: 5.132078647613525\r\nStep 1012, loss: 4.958162784576416\r\nStep 1013, loss: 4.8234992027282715\r\nStep 1014, loss: 5.221305847167969\r\nStep 1015, loss: 4.886592388153076\r\nStep 1016, loss: 4.896177768707275\r\nStep 1017, loss: 5.032711029052734\r\nStep 1018, loss: 4.961817741394043\r\nStep 1019, loss: 5.0470805168151855\r\nStep 1020, loss: 5.076366424560547\r\nStep 1050, loss: 4.694330215454102\r\nStep 1051, loss: 4.912756443023682\r\nStep 1052, loss: 4.8436279296875\r\nStep 1053, loss: 4.9195427894592285\r\nStep 1054, loss: 4.9399309158325195\r\nStep 1055, loss: 5.0726518630981445\r\nStep 1056, loss: 5.081210136413574\r\nStep 1057, loss: 5.1472086906433105\r\nStep 1058, loss: 5.104522228240967\r\nStep 1059, loss: 4.850593090057373\r\nStep 1060, loss: 4.889288425445557\r\nStep 1061, loss: 5.073827743530273\r\nStep 1062, loss: 5.254583835601807\r\nStep 1063, loss: 4.925671100616455\r\nStep 1064, loss: 5.084733963012695\r\nStep 1065, loss: 4.838853359222412\r\nStep 1066, loss: 5.06833028793335\r\nStep 1067, loss: 4.889566421508789\r\nStep 1068, loss: 5.001903057098389\r\nStep 1069, loss: 5.257298946380615\r\nStep 1070, loss: 5.1954803466796875\r\nStep 1071, loss: 4.670302867889404\r\nStep 1072, loss: 4.818952560424805\r\nStep 1073, loss: 5.172548294067383\r\nStep 1074, loss: 4.8732733726501465\r\nStep 1075, loss: 5.093983173370361\r\nStep 1076, loss: 4.797558784484863\r\nStep 1077, loss: 5.1093010902404785\r\nStep 1078, loss: 4.9869561195373535\r\nStep 1021, loss: 4.905526161193848\r\nStep 1022, loss: 5.042372703552246\r\nStep 1023, loss: 5.161309242248535\r\nStep 1024, loss: 4.904819488525391\r\nStep 1025, loss: 4.9067277908325195\r\nStep 1026, loss: 5.095753192901611\r\nStep 1027, loss: 5.177361488342285\r\nStep 1028, loss: 4.811687469482422\r\nStep 1029, loss: 4.922520637512207\r\nStep 1030, loss: 5.0681610107421875\r\nStep 1031, loss: 4.868681907653809\r\nStep 1032, loss: 5.1157732009887695\r\nStep 1033, loss: 5.085040092468262\r\nStep 1034, loss: 5.087116241455078\r\nStep 1035, loss: 5.127306938171387\r\nStep 1036, loss: 4.932193756103516\r\nStep 1037, loss: 4.980825424194336\r\nStep 1038, loss: 4.920042991638184\r\nStep 1039, loss: 4.951450347900391\r\nStep 1040, loss: 4.9700727462768555\r\nStep 1041, loss: 4.698306083679199\r\nStep 1042, loss: 4.857691287994385\r\nStep 1043, loss: 4.852866172790527\r\nStep 1044, loss: 4.981034278869629\r\nStep 1045, loss: 4.956365585327148\r\nStep 1046, loss: 4.646729469299316\r\nStep 1047, loss: 4.902242183685303\r\nStep 1048, loss: 4.928100109100342\r\nStep 1049, loss: 5.10159969329834\r\nStep 962, loss: 5.1568522453308105\r\nStep 963, loss: 4.816960334777832\r\nStep 964, loss: 5.064596652984619\r\nStep 965, loss: 4.939937114715576\r\nStep 966, loss: 5.27336311340332\r\nStep 967, loss: 5.090263366699219\r\nStep 968, loss: 4.881126880645752\r\nStep 969, loss: 5.1100921630859375\r\nStep 970, loss: 5.195235252380371\r\nStep 971, loss: 5.071698188781738\r\nStep 972, loss: 4.915706157684326\r\nStep 973, loss: 5.233161926269531\r\nStep 974, loss: 5.001443386077881\r\nStep 975, loss: 5.120009899139404\r\nStep 976, loss: 5.074319362640381\r\nStep 977, loss: 4.791117191314697\r\nStep 978, loss: 5.0973615646362305\r\nStep 979, loss: 5.003950119018555\r\nStep 980, loss: 5.08306884765625\r\nStep 981, loss: 4.865263938903809\r\nStep 982, loss: 5.179849624633789\r\nStep 983, loss: 5.135115623474121\r\nStep 984, loss: 4.828519821166992\r\nStep 985, loss: 5.138967037200928\r\nStep 986, loss: 4.932805061340332\r\nStep 987, loss: 5.0530877113342285\r\nStep 988, loss: 5.030508041381836\r\nStep 989, loss: 5.121220588684082\r\nStep 990, loss: 5.019931316375732\r\nStep 991, loss: 4.990509986877441\r\nStep 1050, loss: 4.694330215454102\r\nStep 1051, loss: 4.912756443023682\r\nStep 1052, loss: 4.8436279296875\r\nStep 1053, loss: 4.9195427894592285\r\nStep 1054, loss: 4.9399309158325195\r\nStep 1055, loss: 5.0726518630981445\r\nStep 1056, loss: 5.081210136413574\r\nStep 1057, loss: 5.1472086906433105\r\nStep 1058, loss: 5.104522228240967\r\nStep 1059, loss: 4.850593090057373\r\nStep 1060, loss: 4.889288425445557\r\nStep 1061, loss: 5.073827743530273\r\nStep 1062, loss: 5.254583835601807\r\nStep 1063, loss: 4.925671100616455\r\nStep 1064, loss: 5.084733963012695\r\nStep 1065, loss: 4.838853359222412\r\nStep 1066, loss: 5.06833028793335\r\nStep 1067, loss: 4.889566421508789\r\nStep 1068, loss: 5.001903057098389\r\nStep 1069, loss: 5.257298946380615\r\nStep 1070, loss: 5.1954803466796875\r\nStep 1071, loss: 4.670302867889404\r\nStep 1072, loss: 4.818952560424805\r\nStep 1073, loss: 5.172548294067383\r\nStep 1074, loss: 4.8732733726501465\r\nStep 1075, loss: 5.093983173370361\r\nStep 1076, loss: 4.797558784484863\r\nStep 1077, loss: 5.1093010902404785\r\nStep 1078, loss: 4.9869561195373535\r\nStep 992, loss: 4.882480621337891\r\nStep 993, loss: 5.091967582702637\r\nStep 994, loss: 4.786419868469238\r\nStep 995, loss: 4.806129455566406\r\nStep 996, loss: 5.057449817657471\r\nStep 997, loss: 4.943149566650391\r\nStep 998, loss: 4.96529483795166\r\nStep 999, loss: 5.127987384796143\r\nStep 1000, loss: 5.115264892578125\r\nStep 1001, loss: 5.172677993774414\r\nStep 1002, loss: 5.104112148284912\r\nStep 1003, loss: 5.088464736938477\r\nStep 1004, loss: 4.918788433074951\r\nStep 1005, loss: 5.007358074188232\r\nStep 1006, loss: 5.114655494689941\r\nStep 1007, loss: 4.874712944030762\r\nStep 1008, loss: 5.253972053527832\r\nStep 1009, loss: 4.990752696990967\r\nStep 1010, loss: 5.004293918609619\r\nStep 1011, loss: 5.132078647613525\r\nStep 1012, loss: 4.958162784576416\r\nStep 1013, loss: 4.8234992027282715\r\nStep 1014, loss: 5.221305847167969\r\nStep 1015, loss: 4.886592388153076\r\nStep 1016, loss: 4.896177768707275\r\nStep 1017, loss: 5.032711029052734\r\nStep 1018, loss: 4.961817741394043\r\nStep 1019, loss: 5.0470805168151855\r\nStep 1020, loss: 5.076366424560547\r\nStep 962, loss: 5.1568522453308105\r\nStep 963, loss: 4.816960334777832\r\nStep 964, loss: 5.064596652984619\r\nStep 965, loss: 4.939937114715576\r\nStep 966, loss: 5.27336311340332\r\nStep 967, loss: 5.090263366699219\r\nStep 968, loss: 4.881126880645752\r\nStep 969, loss: 5.1100921630859375\r\nStep 970, loss: 5.195235252380371\r\nStep 971, loss: 5.071698188781738\r\nStep 972, loss: 4.915706157684326\r\nStep 973, loss: 5.233161926269531\r\nStep 974, loss: 5.001443386077881\r\nStep 975, loss: 5.120009899139404\r\nStep 976, loss: 5.074319362640381\r\nStep 977, loss: 4.791117191314697\r\nStep 978, loss: 5.0973615646362305\r\nStep 979, loss: 5.003950119018555\r\nStep 980, loss: 5.08306884765625\r\nStep 981, loss: 4.865263938903809\r\nStep 982, loss: 5.179849624633789\r\nStep 983, loss: 5.135115623474121\r\nStep 984, loss: 4.828519821166992\r\nStep 985, loss: 5.138967037200928\r\nStep 986, loss: 4.932805061340332\r\nStep 987, loss: 5.0530877113342285\r\nStep 988, loss: 5.030508041381836\r\nStep 989, loss: 5.121220588684082\r\nStep 990, loss: 5.019931316375732\r\nStep 991, loss: 4.990509986877441\r\nStep 1021, loss: 4.905526161193848\r\nStep 1022, loss: 5.042372703552246\r\nStep 1023, loss: 5.161309242248535\r\nStep 1024, loss: 4.904819488525391\r\nStep 1025, loss: 4.9067277908325195\r\nStep 1026, loss: 5.095753192901611\r\nStep 1027, loss: 5.177361488342285\r\nStep 1028, loss: 4.811687469482422\r\nStep 1029, loss: 4.922520637512207\r\nStep 1030, loss: 5.0681610107421875\r\nStep 1031, loss: 4.868681907653809\r\nStep 1032, loss: 5.1157732009887695\r\nStep 1033, loss: 5.085040092468262\r\nStep 1034, loss: 5.087116241455078\r\nStep 1035, loss: 5.127306938171387\r\nStep 1036, loss: 4.932193756103516\r\nStep 1037, loss: 4.980825424194336\r\nStep 1038, loss: 4.920042991638184\r\nStep 1039, loss: 4.951450347900391\r\nStep 1040, loss: 4.9700727462768555\r\nStep 1041, loss: 4.698306083679199\r\nStep 1042, loss: 4.857691287994385\r\nStep 1043, loss: 4.852866172790527\r\nStep 1044, loss: 4.981034278869629\r\nStep 1045, loss: 4.956365585327148\r\nStep 1046, loss: 4.646729469299316\r\nStep 1047, loss: 4.902242183685303\r\nStep 1048, loss: 4.928100109100342\r\nStep 1049, loss: 5.10159969329834\r\nStep 992, loss: 4.882480621337891\r\nStep 993, loss: 5.091967582702637\r\nStep 994, loss: 4.786419868469238\r\nStep 995, loss: 4.806129455566406\r\nStep 996, loss: 5.057449817657471\r\nStep 997, loss: 4.943149566650391\r\nStep 998, loss: 4.96529483795166\r\nStep 999, loss: 5.127987384796143\r\nStep 1000, loss: 5.115264892578125\r\nStep 1001, loss: 5.172677993774414\r\nStep 1002, loss: 5.104112148284912\r\nStep 1003, loss: 5.088464736938477\r\nStep 1004, loss: 4.918788433074951\r\nStep 1005, loss: 5.007358074188232\r\nStep 1006, loss: 5.114655494689941\r\nStep 1007, loss: 4.874712944030762\r\nStep 1008, loss: 5.253972053527832\r\nStep 1009, loss: 4.990752696990967\r\nStep 1010, loss: 5.004293918609619\r\nStep 1011, loss: 5.132078647613525\r\nStep 1012, loss: 4.958162784576416\r\nStep 1013, loss: 4.8234992027282715\r\nStep 1014, loss: 5.221305847167969\r\nStep 1015, loss: 4.886592388153076\r\nStep 1016, loss: 4.896177768707275\r\nStep 1017, loss: 5.032711029052734\r\nStep 1018, loss: 4.961817741394043\r\nStep 1019, loss: 5.0470805168151855\r\nStep 1020, loss: 5.076366424560547\r\nStep 1050, loss: 4.694330215454102\r\nStep 1051, loss: 4.912756443023682\r\nStep 1052, loss: 4.8436279296875\r\nStep 1053, loss: 4.9195427894592285\r\nStep 1054, loss: 4.9399309158325195\r\nStep 1055, loss: 5.0726518630981445\r\nStep 1056, loss: 5.081210136413574\r\nStep 1057, loss: 5.1472086906433105\r\nStep 1058, loss: 5.104522228240967\r\nStep 1059, loss: 4.850593090057373\r\nStep 1060, loss: 4.889288425445557\r\nStep 1061, loss: 5.073827743530273\r\nStep 1062, loss: 5.254583835601807\r\nStep 1063, loss: 4.925671100616455\r\nStep 1064, loss: 5.084733963012695\r\nStep 1065, loss: 4.838853359222412\r\nStep 1066, loss: 5.06833028793335\r\nStep 1067, loss: 4.889566421508789\r\nStep 1068, loss: 5.001903057098389\r\nStep 1069, loss: 5.257298946380615\r\nStep 1070, loss: 5.1954803466796875\r\nStep 1071, loss: 4.670302867889404\r\nStep 1072, loss: 4.818952560424805\r\nStep 1073, loss: 5.172548294067383\r\nStep 1074, loss: 4.8732733726501465\r\nStep 1075, loss: 5.093983173370361\r\nStep 1076, loss: 4.797558784484863\r\nStep 1077, loss: 5.1093010902404785\r\nStep 1078, loss: 4.9869561195373535\r\nStep 1021, loss: 4.905526161193848\r\nStep 1022, loss: 5.042372703552246\r\nStep 1023, loss: 5.161309242248535\r\nStep 1024, loss: 4.904819488525391\r\nStep 1025, loss: 4.9067277908325195\r\nStep 1026, loss: 5.095753192901611\r\nStep 1027, loss: 5.177361488342285\r\nStep 1028, loss: 4.811687469482422\r\nStep 1029, loss: 4.922520637512207\r\nStep 1030, loss: 5.0681610107421875\r\nStep 1031, loss: 4.868681907653809\r\nStep 1032, loss: 5.1157732009887695\r\nStep 1033, loss: 5.085040092468262\r\nStep 1034, loss: 5.087116241455078\r\nStep 1035, loss: 5.127306938171387\r\nStep 1036, loss: 4.932193756103516\r\nStep 1037, loss: 4.980825424194336\r\nStep 1038, loss: 4.920042991638184\r\nStep 1039, loss: 4.951450347900391\r\nStep 1040, loss: 4.9700727462768555\r\nStep 1041, loss: 4.698306083679199\r\nStep 1042, loss: 4.857691287994385\r\nStep 1043, loss: 4.852866172790527\r\nStep 1044, loss: 4.981034278869629\r\nStep 1045, loss: 4.956365585327148\r\nStep 1046, loss: 4.646729469299316\r\nStep 1047, loss: 4.902242183685303\r\nStep 1048, loss: 4.928100109100342\r\nStep 1049, loss: 5.10159969329834\r\nStep 1079, loss: 4.784158706665039\r\nStep 1080, loss: 4.877741813659668\r\nStep 1081, loss: 5.154816150665283\r\nStep 1082, loss: 4.751720428466797\r\nStep 1083, loss: 5.125522613525391\r\nStep 1084, loss: 4.889678955078125\r\nStep 1085, loss: 5.0008063316345215\r\nStep 1086, loss: 4.946294784545898\r\nStep 1087, loss: 5.081248760223389\r\nStep 1088, loss: 5.148041248321533\r\nStep 1089, loss: 4.9699811935424805\r\nStep 1090, loss: 5.008603572845459\r\nStep 1091, loss: 5.2874908447265625\r\nStep 1092, loss: 4.957520961761475\r\nStep 1093, loss: 4.772770404815674\r\nStep 1094, loss: 4.826335906982422\r\nStep 1095, loss: 4.773747444152832\r\nStep 1096, loss: 4.727792739868164\r\nStep 1097, loss: 4.867959976196289\r\nStep 1098, loss: 4.990972995758057\r\nStep 1099, loss: 5.051648139953613\r\nStep 1100, loss: 4.934340953826904\r\nStep 1101, loss: 4.967705726623535\r\nStep 1102, loss: 4.928976058959961\r\nStep 1103, loss: 5.019003868103027\r\nStep 1104, loss: 4.9384446144104\r\nStep 1105, loss: 5.105501174926758\r\nStep 1106, loss: 4.737863540649414\r\nStep 1107, loss: 5.02548885345459\r\nStep 1050, loss: 4.694330215454102\r\nStep 1051, loss: 4.912756443023682\r\nStep 1052, loss: 4.8436279296875\r\nStep 1053, loss: 4.9195427894592285\r\nStep 1054, loss: 4.9399309158325195\r\nStep 1055, loss: 5.0726518630981445\r\nStep 1056, loss: 5.081210136413574\r\nStep 1057, loss: 5.1472086906433105\r\nStep 1058, loss: 5.104522228240967\r\nStep 1059, loss: 4.850593090057373\r\nStep 1060, loss: 4.889288425445557\r\nStep 1061, loss: 5.073827743530273\r\nStep 1062, loss: 5.254583835601807\r\nStep 1063, loss: 4.925671100616455\r\nStep 1064, loss: 5.084733963012695\r\nStep 1065, loss: 4.838853359222412\r\nStep 1066, loss: 5.06833028793335\r\nStep 1067, loss: 4.889566421508789\r\nStep 1068, loss: 5.001903057098389\r\nStep 1069, loss: 5.257298946380615\r\nStep 1070, loss: 5.1954803466796875\r\nStep 1071, loss: 4.670302867889404\r\nStep 1072, loss: 4.818952560424805\r\nStep 1073, loss: 5.172548294067383\r\nStep 1074, loss: 4.8732733726501465\r\nStep 1075, loss: 5.093983173370361\r\nStep 1076, loss: 4.797558784484863\r\nStep 1077, loss: 5.1093010902404785\r\nStep 1078, loss: 4.9869561195373535\r\nStep 1108, loss: 4.888282775878906\r\nStep 1109, loss: 5.192723274230957\r\nStep 1110, loss: 4.966083526611328\r\nStep 1111, loss: 4.8921966552734375\r\nStep 1112, loss: 4.865470886230469\r\nStep 1113, loss: 4.701240062713623\r\nStep 1114, loss: 5.101388454437256\r\nStep 1115, loss: 4.849363803863525\r\nStep 1116, loss: 5.013510704040527\r\nStep 1117, loss: 5.012414932250977\r\nStep 1118, loss: 4.933382987976074\r\nStep 1119, loss: 4.641906261444092\r\nStep 1120, loss: 4.92882776260376\r\nStep 1121, loss: 4.703517913818359\r\nStep 1122, loss: 4.972881317138672\r\nStep 1123, loss: 5.049308776855469\r\nStep 1124, loss: 4.712826728820801\r\nStep 1125, loss: 4.8883585929870605\r\nStep 1126, loss: 5.122026443481445\r\nStep 1127, loss: 5.010403156280518\r\nStep 1128, loss: 4.868868827819824\r\nStep 1129, loss: 5.078428268432617\r\nStep 1130, loss: 4.8772053718566895\r\nStep 1131, loss: 4.8967671394348145\r\nStep 1132, loss: 4.88825798034668\r\nStep 1133, loss: 4.8844780921936035\r\nStep 1134, loss: 4.8840436935424805\r\nStep 1135, loss: 4.7452850341796875\r\nStep 1136, loss: 5.050529479980469\r\nStep 1079, loss: 4.784158706665039\r\nStep 1080, loss: 4.877741813659668\r\nStep 1081, loss: 5.154816150665283\r\nStep 1082, loss: 4.751720428466797\r\nStep 1083, loss: 5.125522613525391\r\nStep 1084, loss: 4.889678955078125\r\nStep 1085, loss: 5.0008063316345215\r\nStep 1086, loss: 4.946294784545898\r\nStep 1087, loss: 5.081248760223389\r\nStep 1088, loss: 5.148041248321533\r\nStep 1089, loss: 4.9699811935424805\r\nStep 1090, loss: 5.008603572845459\r\nStep 1091, loss: 5.2874908447265625\r\nStep 1092, loss: 4.957520961761475\r\nStep 1093, loss: 4.772770404815674\r\nStep 1094, loss: 4.826335906982422\r\nStep 1095, loss: 4.773747444152832\r\nStep 1096, loss: 4.727792739868164\r\nStep 1097, loss: 4.867959976196289\r\nStep 1098, loss: 4.990972995758057\r\nStep 1099, loss: 5.051648139953613\r\nStep 1100, loss: 4.934340953826904\r\nStep 1101, loss: 4.967705726623535\r\nStep 1102, loss: 4.928976058959961\r\nStep 1103, loss: 5.019003868103027\r\nStep 1104, loss: 4.9384446144104\r\nStep 1105, loss: 5.105501174926758\r\nStep 1106, loss: 4.737863540649414\r\nStep 1107, loss: 5.02548885345459\r\nStep 1137, loss: 4.917625904083252\r\nStep 1138, loss: 4.829206943511963\r\nStep 1139, loss: 4.8869781494140625\r\nStep 1140, loss: 4.968005657196045\r\nStep 1141, loss: 4.84523868560791\r\nStep 1142, loss: 4.98886251449585\r\nStep 1143, loss: 4.957108497619629\r\nStep 1144, loss: 5.097373008728027\r\nStep 1145, loss: 4.651041507720947\r\nStep 1146, loss: 4.8319783210754395\r\nStep 1147, loss: 4.675942897796631\r\nStep 1148, loss: 5.058229446411133\r\nStep 1149, loss: 4.955995082855225\r\nStep 1150, loss: 4.861543655395508\r\nStep 1151, loss: 4.811530590057373\r\nStep 1152, loss: 4.847794532775879\r\nStep 1153, loss: 5.00905704498291\r\nStep 1154, loss: 4.871018886566162\r\nStep 1155, loss: 4.829107284545898\r\nStep 1156, loss: 4.926144599914551\r\nStep 1157, loss: 4.929976940155029\r\nStep 1158, loss: 4.581959247589111\r\nStep 1159, loss: 4.8127546310424805\r\nStep 1160, loss: 4.829219341278076\r\nStep 1161, loss: 5.095797061920166\r\nStep 1162, loss: 5.034276008605957\r\nStep 1163, loss: 4.616420269012451\r\nStep 1164, loss: 5.0754852294921875\r\nStep 1165, loss: 4.784201622009277\r\nStep 1108, loss: 4.888282775878906\r\nStep 1109, loss: 5.192723274230957\r\nStep 1110, loss: 4.966083526611328\r\nStep 1111, loss: 4.8921966552734375\r\nStep 1112, loss: 4.865470886230469\r\nStep 1113, loss: 4.701240062713623\r\nStep 1114, loss: 5.101388454437256\r\nStep 1115, loss: 4.849363803863525\r\nStep 1116, loss: 5.013510704040527\r\nStep 1117, loss: 5.012414932250977\r\nStep 1118, loss: 4.933382987976074\r\nStep 1119, loss: 4.641906261444092\r\nStep 1120, loss: 4.92882776260376\r\nStep 1121, loss: 4.703517913818359\r\nStep 1122, loss: 4.972881317138672\r\nStep 1123, loss: 5.049308776855469\r\nStep 1124, loss: 4.712826728820801\r\nStep 1125, loss: 4.8883585929870605\r\nStep 1126, loss: 5.122026443481445\r\nStep 1127, loss: 5.010403156280518\r\nStep 1128, loss: 4.868868827819824\r\nStep 1129, loss: 5.078428268432617\r\nStep 1130, loss: 4.8772053718566895\r\nStep 1131, loss: 4.8967671394348145\r\nStep 1132, loss: 4.88825798034668\r\nStep 1133, loss: 4.8844780921936035\r\nStep 1134, loss: 4.8840436935424805\r\nStep 1135, loss: 4.7452850341796875\r\nStep 1136, loss: 5.050529479980469\r\nStep 1166, loss: 4.93991231918335\r\nStep 1167, loss: 4.745754241943359\r\nStep 1168, loss: 4.946484088897705\r\nStep 1169, loss: 4.959706783294678\r\nStep 1170, loss: 5.136929512023926\r\nStep 1171, loss: 4.963235855102539\r\nStep 1172, loss: 4.689530372619629\r\nStep 1173, loss: 4.873706340789795\r\nStep 1174, loss: 5.097964286804199\r\nStep 1175, loss: 4.9454545974731445\r\nStep 1176, loss: 4.84202766418457\r\nStep 1177, loss: 5.002065181732178\r\nStep 1178, loss: 5.085189342498779\r\nStep 1179, loss: 5.059033393859863\r\nStep 1180, loss: 5.012739181518555\r\nStep 1181, loss: 4.655002117156982\r\nStep 1182, loss: 5.119167804718018\r\nStep 1183, loss: 4.922578811645508\r\nStep 1184, loss: 4.863205909729004\r\nStep 1185, loss: 4.81952428817749\r\nStep 1186, loss: 4.806148052215576\r\nStep 1187, loss: 4.85841703414917\r\nStep 1188, loss: 5.1563920974731445\r\nStep 1189, loss: 4.81588077545166\r\nStep 1190, loss: 4.9441118240356445\r\nStep 1191, loss: 4.791698932647705\r\nStep 1192, loss: 5.128923416137695\r\nStep 1193, loss: 4.806921005249023\r\nStep 1194, loss: 5.237899303436279\r\nStep 1137, loss: 4.917625904083252\r\nStep 1138, loss: 4.829206943511963\r\nStep 1139, loss: 4.8869781494140625\r\nStep 1140, loss: 4.968005657196045\r\nStep 1141, loss: 4.84523868560791\r\nStep 1142, loss: 4.98886251449585\r\nStep 1143, loss: 4.957108497619629\r\nStep 1144, loss: 5.097373008728027\r\nStep 1145, loss: 4.651041507720947\r\nStep 1146, loss: 4.8319783210754395\r\nStep 1147, loss: 4.675942897796631\r\nStep 1148, loss: 5.058229446411133\r\nStep 1149, loss: 4.955995082855225\r\nStep 1150, loss: 4.861543655395508\r\nStep 1151, loss: 4.811530590057373\r\nStep 1152, loss: 4.847794532775879\r\nStep 1153, loss: 5.00905704498291\r\nStep 1154, loss: 4.871018886566162\r\nStep 1155, loss: 4.829107284545898\r\nStep 1156, loss: 4.926144599914551\r\nStep 1157, loss: 4.929976940155029\r\nStep 1158, loss: 4.581959247589111\r\nStep 1159, loss: 4.8127546310424805\r\nStep 1160, loss: 4.829219341278076\r\nStep 1161, loss: 5.095797061920166\r\nStep 1162, loss: 5.034276008605957\r\nStep 1163, loss: 4.616420269012451\r\nStep 1164, loss: 5.0754852294921875\r\nStep 1165, loss: 4.784201622009277\r\nStep 1079, loss: 4.784158706665039\r\nStep 1080, loss: 4.877741813659668\r\nStep 1081, loss: 5.154816150665283\r\nStep 1082, loss: 4.751720428466797\r\nStep 1083, loss: 5.125522613525391\r\nStep 1084, loss: 4.889678955078125\r\nStep 1085, loss: 5.0008063316345215\r\nStep 1086, loss: 4.946294784545898\r\nStep 1087, loss: 5.081248760223389\r\nStep 1088, loss: 5.148041248321533\r\nStep 1089, loss: 4.9699811935424805\r\nStep 1090, loss: 5.008603572845459\r\nStep 1091, loss: 5.2874908447265625\r\nStep 1092, loss: 4.957520961761475\r\nStep 1093, loss: 4.772770404815674\r\nStep 1094, loss: 4.826335906982422\r\nStep 1095, loss: 4.773747444152832\r\nStep 1096, loss: 4.727792739868164\r\nStep 1097, loss: 4.867959976196289\r\nStep 1098, loss: 4.990972995758057\r\nStep 1099, loss: 5.051648139953613\r\nStep 1100, loss: 4.934340953826904\r\nStep 1101, loss: 4.967705726623535\r\nStep 1102, loss: 4.928976058959961\r\nStep 1103, loss: 5.019003868103027\r\nStep 1104, loss: 4.9384446144104\r\nStep 1105, loss: 5.105501174926758\r\nStep 1106, loss: 4.737863540649414\r\nStep 1107, loss: 5.02548885345459\r\nStep 1166, loss: 4.93991231918335\r\nStep 1167, loss: 4.745754241943359\r\nStep 1168, loss: 4.946484088897705\r\nStep 1169, loss: 4.959706783294678\r\nStep 1170, loss: 5.136929512023926\r\nStep 1171, loss: 4.963235855102539\r\nStep 1172, loss: 4.689530372619629\r\nStep 1173, loss: 4.873706340789795\r\nStep 1174, loss: 5.097964286804199\r\nStep 1175, loss: 4.9454545974731445\r\nStep 1176, loss: 4.84202766418457\r\nStep 1177, loss: 5.002065181732178\r\nStep 1178, loss: 5.085189342498779\r\nStep 1179, loss: 5.059033393859863\r\nStep 1180, loss: 5.012739181518555\r\nStep 1181, loss: 4.655002117156982\r\nStep 1182, loss: 5.119167804718018\r\nStep 1183, loss: 4.922578811645508\r\nStep 1184, loss: 4.863205909729004\r\nStep 1185, loss: 4.81952428817749\r\nStep 1186, loss: 4.806148052215576\r\nStep 1187, loss: 4.85841703414917\r\nStep 1188, loss: 5.1563920974731445\r\nStep 1189, loss: 4.81588077545166\r\nStep 1190, loss: 4.9441118240356445\r\nStep 1191, loss: 4.791698932647705\r\nStep 1192, loss: 5.128923416137695\r\nStep 1193, loss: 4.806921005249023\r\nStep 1194, loss: 5.237899303436279\r\nStep 1108, loss: 4.888282775878906\r\nStep 1109, loss: 5.192723274230957\r\nStep 1110, loss: 4.966083526611328\r\nStep 1111, loss: 4.8921966552734375\r\nStep 1112, loss: 4.865470886230469\r\nStep 1113, loss: 4.701240062713623\r\nStep 1114, loss: 5.101388454437256\r\nStep 1115, loss: 4.849363803863525\r\nStep 1116, loss: 5.013510704040527\r\nStep 1117, loss: 5.012414932250977\r\nStep 1118, loss: 4.933382987976074\r\nStep 1119, loss: 4.641906261444092\r\nStep 1120, loss: 4.92882776260376\r\nStep 1121, loss: 4.703517913818359\r\nStep 1122, loss: 4.972881317138672\r\nStep 1123, loss: 5.049308776855469\r\nStep 1124, loss: 4.712826728820801\r\nStep 1125, loss: 4.8883585929870605\r\nStep 1126, loss: 5.122026443481445\r\nStep 1127, loss: 5.010403156280518\r\nStep 1128, loss: 4.868868827819824\r\nStep 1129, loss: 5.078428268432617\r\nStep 1130, loss: 4.8772053718566895\r\nStep 1131, loss: 4.8967671394348145\r\nStep 1132, loss: 4.88825798034668\r\nStep 1133, loss: 4.8844780921936035\r\nStep 1134, loss: 4.8840436935424805\r\nStep 1135, loss: 4.7452850341796875\r\nStep 1136, loss: 5.050529479980469\r\nStep 1079, loss: 4.784158706665039\r\nStep 1080, loss: 4.877741813659668\r\nStep 1081, loss: 5.154816150665283\r\nStep 1082, loss: 4.751720428466797\r\nStep 1083, loss: 5.125522613525391\r\nStep 1084, loss: 4.889678955078125\r\nStep 1085, loss: 5.0008063316345215\r\nStep 1086, loss: 4.946294784545898\r\nStep 1087, loss: 5.081248760223389\r\nStep 1088, loss: 5.148041248321533\r\nStep 1089, loss: 4.9699811935424805\r\nStep 1090, loss: 5.008603572845459\r\nStep 1091, loss: 5.2874908447265625\r\nStep 1092, loss: 4.957520961761475\r\nStep 1093, loss: 4.772770404815674\r\nStep 1094, loss: 4.826335906982422\r\nStep 1095, loss: 4.773747444152832\r\nStep 1096, loss: 4.727792739868164\r\nStep 1097, loss: 4.867959976196289\r\nStep 1098, loss: 4.990972995758057\r\nStep 1099, loss: 5.051648139953613\r\nStep 1100, loss: 4.934340953826904\r\nStep 1101, loss: 4.967705726623535\r\nStep 1102, loss: 4.928976058959961\r\nStep 1103, loss: 5.019003868103027\r\nStep 1104, loss: 4.9384446144104\r\nStep 1105, loss: 5.105501174926758\r\nStep 1106, loss: 4.737863540649414\r\nStep 1107, loss: 5.02548885345459\r\nStep 1137, loss: 4.917625904083252\r\nStep 1138, loss: 4.829206943511963\r\nStep 1139, loss: 4.8869781494140625\r\nStep 1140, loss: 4.968005657196045\r\nStep 1141, loss: 4.84523868560791\r\nStep 1142, loss: 4.98886251449585\r\nStep 1143, loss: 4.957108497619629\r\nStep 1144, loss: 5.097373008728027\r\nStep 1145, loss: 4.651041507720947\r\nStep 1146, loss: 4.8319783210754395\r\nStep 1147, loss: 4.675942897796631\r\nStep 1148, loss: 5.058229446411133\r\nStep 1149, loss: 4.955995082855225\r\nStep 1150, loss: 4.861543655395508\r\nStep 1151, loss: 4.811530590057373\r\nStep 1152, loss: 4.847794532775879\r\nStep 1153, loss: 5.00905704498291\r\nStep 1154, loss: 4.871018886566162\r\nStep 1155, loss: 4.829107284545898\r\nStep 1156, loss: 4.926144599914551\r\nStep 1157, loss: 4.929976940155029\r\nStep 1158, loss: 4.581959247589111\r\nStep 1159, loss: 4.8127546310424805\r\nStep 1160, loss: 4.829219341278076\r\nStep 1161, loss: 5.095797061920166\r\nStep 1162, loss: 5.034276008605957\r\nStep 1163, loss: 4.616420269012451\r\nStep 1164, loss: 5.0754852294921875\r\nStep 1165, loss: 4.784201622009277\r\nStep 1108, loss: 4.888282775878906\r\nStep 1109, loss: 5.192723274230957\r\nStep 1110, loss: 4.966083526611328\r\nStep 1111, loss: 4.8921966552734375\r\nStep 1112, loss: 4.865470886230469\r\nStep 1113, loss: 4.701240062713623\r\nStep 1114, loss: 5.101388454437256\r\nStep 1115, loss: 4.849363803863525\r\nStep 1116, loss: 5.013510704040527\r\nStep 1117, loss: 5.012414932250977\r\nStep 1118, loss: 4.933382987976074\r\nStep 1119, loss: 4.641906261444092\r\nStep 1120, loss: 4.92882776260376\r\nStep 1121, loss: 4.703517913818359\r\nStep 1122, loss: 4.972881317138672\r\nStep 1123, loss: 5.049308776855469\r\nStep 1124, loss: 4.712826728820801\r\nStep 1125, loss: 4.8883585929870605\r\nStep 1126, loss: 5.122026443481445\r\nStep 1127, loss: 5.010403156280518\r\nStep 1128, loss: 4.868868827819824\r\nStep 1129, loss: 5.078428268432617\r\nStep 1130, loss: 4.8772053718566895\r\nStep 1131, loss: 4.8967671394348145\r\nStep 1132, loss: 4.88825798034668\r\nStep 1133, loss: 4.8844780921936035\r\nStep 1134, loss: 4.8840436935424805\r\nStep 1135, loss: 4.7452850341796875\r\nStep 1136, loss: 5.050529479980469\r\nStep 1166, loss: 4.93991231918335\r\nStep 1167, loss: 4.745754241943359\r\nStep 1168, loss: 4.946484088897705\r\nStep 1169, loss: 4.959706783294678\r\nStep 1170, loss: 5.136929512023926\r\nStep 1171, loss: 4.963235855102539\r\nStep 1172, loss: 4.689530372619629\r\nStep 1173, loss: 4.873706340789795\r\nStep 1174, loss: 5.097964286804199\r\nStep 1175, loss: 4.9454545974731445\r\nStep 1176, loss: 4.84202766418457\r\nStep 1177, loss: 5.002065181732178\r\nStep 1178, loss: 5.085189342498779\r\nStep 1179, loss: 5.059033393859863\r\nStep 1180, loss: 5.012739181518555\r\nStep 1181, loss: 4.655002117156982\r\nStep 1182, loss: 5.119167804718018\r\nStep 1183, loss: 4.922578811645508\r\nStep 1184, loss: 4.863205909729004\r\nStep 1185, loss: 4.81952428817749\r\nStep 1186, loss: 4.806148052215576\r\nStep 1187, loss: 4.85841703414917\r\nStep 1188, loss: 5.1563920974731445\r\nStep 1189, loss: 4.81588077545166\r\nStep 1190, loss: 4.9441118240356445\r\nStep 1191, loss: 4.791698932647705\r\nStep 1192, loss: 5.128923416137695\r\nStep 1193, loss: 4.806921005249023\r\nStep 1194, loss: 5.237899303436279\r\nStep 1137, loss: 4.917625904083252\r\nStep 1138, loss: 4.829206943511963\r\nStep 1139, loss: 4.8869781494140625\r\nStep 1140, loss: 4.968005657196045\r\nStep 1141, loss: 4.84523868560791\r\nStep 1142, loss: 4.98886251449585\r\nStep 1143, loss: 4.957108497619629\r\nStep 1144, loss: 5.097373008728027\r\nStep 1145, loss: 4.651041507720947\r\nStep 1146, loss: 4.8319783210754395\r\nStep 1147, loss: 4.675942897796631\r\nStep 1148, loss: 5.058229446411133\r\nStep 1149, loss: 4.955995082855225\r\nStep 1150, loss: 4.861543655395508\r\nStep 1151, loss: 4.811530590057373\r\nStep 1152, loss: 4.847794532775879\r\nStep 1153, loss: 5.00905704498291\r\nStep 1154, loss: 4.871018886566162\r\nStep 1155, loss: 4.829107284545898\r\nStep 1156, loss: 4.926144599914551\r\nStep 1157, loss: 4.929976940155029\r\nStep 1158, loss: 4.581959247589111\r\nStep 1159, loss: 4.8127546310424805\r\nStep 1160, loss: 4.829219341278076\r\nStep 1161, loss: 5.095797061920166\r\nStep 1162, loss: 5.034276008605957\r\nStep 1163, loss: 4.616420269012451\r\nStep 1164, loss: 5.0754852294921875\r\nStep 1165, loss: 4.784201622009277\r\nStep 1195, loss: 4.842819690704346\r\nStep 1166, loss: 4.93991231918335\r\nStep 1167, loss: 4.745754241943359\r\nStep 1168, loss: 4.946484088897705\r\nStep 1169, loss: 4.959706783294678\r\nStep 1170, loss: 5.136929512023926\r\nStep 1171, loss: 4.963235855102539\r\nStep 1172, loss: 4.689530372619629\r\nStep 1173, loss: 4.873706340789795\r\nStep 1174, loss: 5.097964286804199\r\nStep 1175, loss: 4.9454545974731445\r\nStep 1176, loss: 4.84202766418457\r\nStep 1177, loss: 5.002065181732178\r\nStep 1178, loss: 5.085189342498779\r\nStep 1179, loss: 5.059033393859863\r\nStep 1180, loss: 5.012739181518555\r\nStep 1181, loss: 4.655002117156982\r\nStep 1182, loss: 5.119167804718018\r\nStep 1183, loss: 4.922578811645508\r\nStep 1184, loss: 4.863205909729004\r\nStep 1185, loss: 4.81952428817749\r\nStep 1186, loss: 4.806148052215576\r\nStep 1187, loss: 4.85841703414917\r\nStep 1188, loss: 5.1563920974731445\r\nStep 1189, loss: 4.81588077545166\r\nStep 1190, loss: 4.9441118240356445\r\nStep 1191, loss: 4.791698932647705\r\nStep 1192, loss: 5.128923416137695\r\nStep 1193, loss: 4.806921005249023\r\nStep 1194, loss: 5.237899303436279\r\nStep 1195, loss: 4.842819690704346\r\nStep 1079, loss: 4.784158706665039\r\nStep 1080, loss: 4.877741813659668\r\nStep 1081, loss: 5.154816150665283\r\nStep 1082, loss: 4.751720428466797\r\nStep 1083, loss: 5.125522613525391\r\nStep 1084, loss: 4.889678955078125\r\nStep 1085, loss: 5.0008063316345215\r\nStep 1086, loss: 4.946294784545898\r\nStep 1087, loss: 5.081248760223389\r\nStep 1088, loss: 5.148041248321533\r\nStep 1089, loss: 4.9699811935424805\r\nStep 1090, loss: 5.008603572845459\r\nStep 1091, loss: 5.2874908447265625\r\nStep 1092, loss: 4.957520961761475\r\nStep 1093, loss: 4.772770404815674\r\nStep 1094, loss: 4.826335906982422\r\nStep 1095, loss: 4.773747444152832\r\nStep 1096, loss: 4.727792739868164\r\nStep 1097, loss: 4.867959976196289\r\nStep 1098, loss: 4.990972995758057\r\nStep 1099, loss: 5.051648139953613\r\nStep 1100, loss: 4.934340953826904\r\nStep 1101, loss: 4.967705726623535\r\nStep 1102, loss: 4.928976058959961\r\nStep 1103, loss: 5.019003868103027\r\nStep 1104, loss: 4.9384446144104\r\nStep 1105, loss: 5.105501174926758\r\nStep 1106, loss: 4.737863540649414\r\nStep 1107, loss: 5.02548885345459\r\nStep 1108, loss: 4.888282775878906\r\nStep 1109, loss: 5.192723274230957\r\nStep 1110, loss: 4.966083526611328\r\nStep 1111, loss: 4.8921966552734375\r\nStep 1112, loss: 4.865470886230469\r\nStep 1113, loss: 4.701240062713623\r\nStep 1114, loss: 5.101388454437256\r\nStep 1115, loss: 4.849363803863525\r\nStep 1116, loss: 5.013510704040527\r\nStep 1117, loss: 5.012414932250977\r\nStep 1118, loss: 4.933382987976074\r\nStep 1119, loss: 4.641906261444092\r\nStep 1120, loss: 4.92882776260376\r\nStep 1121, loss: 4.703517913818359\r\nStep 1122, loss: 4.972881317138672\r\nStep 1123, loss: 5.049308776855469\r\nStep 1124, loss: 4.712826728820801\r\nStep 1125, loss: 4.8883585929870605\r\nStep 1126, loss: 5.122026443481445\r\nStep 1127, loss: 5.010403156280518\r\nStep 1128, loss: 4.868868827819824\r\nStep 1129, loss: 5.078428268432617\r\nStep 1130, loss: 4.8772053718566895\r\nStep 1131, loss: 4.8967671394348145\r\nStep 1132, loss: 4.88825798034668\r\nStep 1133, loss: 4.8844780921936035\r\nStep 1134, loss: 4.8840436935424805\r\nStep 1135, loss: 4.7452850341796875\r\nStep 1136, loss: 5.050529479980469\r\nStep 1137, loss: 4.917625904083252\r\nStep 1138, loss: 4.829206943511963\r\nStep 1139, loss: 4.8869781494140625\r\nStep 1140, loss: 4.968005657196045\r\nStep 1141, loss: 4.84523868560791\r\nStep 1142, loss: 4.98886251449585\r\nStep 1143, loss: 4.957108497619629\r\nStep 1144, loss: 5.097373008728027\r\nStep 1145, loss: 4.651041507720947\r\nStep 1146, loss: 4.8319783210754395\r\nStep 1147, loss: 4.675942897796631\r\nStep 1148, loss: 5.058229446411133\r\nStep 1149, loss: 4.955995082855225\r\nStep 1150, loss: 4.861543655395508\r\nStep 1151, loss: 4.811530590057373\r\nStep 1152, loss: 4.847794532775879\r\nStep 1153, loss: 5.00905704498291\r\nStep 1154, loss: 4.871018886566162\r\nStep 1155, loss: 4.829107284545898\r\nStep 1156, loss: 4.926144599914551\r\nStep 1157, loss: 4.929976940155029\r\nStep 1158, loss: 4.581959247589111\r\nStep 1159, loss: 4.8127546310424805\r\nStep 1160, loss: 4.829219341278076\r\nStep 1161, loss: 5.095797061920166\r\nStep 1162, loss: 5.034276008605957\r\nStep 1163, loss: 4.616420269012451\r\nStep 1164, loss: 5.0754852294921875\r\nStep 1165, loss: 4.784201622009277\r\nStep 1166, loss: 4.93991231918335\r\nStep 1167, loss: 4.745754241943359\r\nStep 1168, loss: 4.946484088897705\r\nStep 1169, loss: 4.959706783294678\r\nStep 1170, loss: 5.136929512023926\r\nStep 1171, loss: 4.963235855102539\r\nStep 1172, loss: 4.689530372619629\r\nStep 1173, loss: 4.873706340789795\r\nStep 1174, loss: 5.097964286804199\r\nStep 1175, loss: 4.9454545974731445\r\nStep 1176, loss: 4.84202766418457\r\nStep 1177, loss: 5.002065181732178\r\nStep 1178, loss: 5.085189342498779\r\nStep 1179, loss: 5.059033393859863\r\nStep 1180, loss: 5.012739181518555\r\nStep 1181, loss: 4.655002117156982\r\nStep 1182, loss: 5.119167804718018\r\nStep 1183, loss: 4.922578811645508\r\nStep 1184, loss: 4.863205909729004\r\nStep 1185, loss: 4.81952428817749\r\nStep 1186, loss: 4.806148052215576\r\nStep 1187, loss: 4.85841703414917\r\nStep 1188, loss: 5.1563920974731445\r\nStep 1189, loss: 4.81588077545166\r\nStep 1190, loss: 4.9441118240356445\r\nStep 1191, loss: 4.791698932647705\r\nStep 1192, loss: 5.128923416137695\r\nStep 1193, loss: 4.806921005249023\r\nStep 1194, loss: 5.237899303436279\r\nStep 1195, loss: 4.842819690704346\r\nStep 1195, loss: 4.842819690704346\r\nStep 1195, loss: 4.842819690704346\r\n",,terminal_output +3312,5470472,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_5e-5\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=5e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-5e-5-$slurm_job_id \\n --tags tokenizer lr-sweep 5e-5 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +3313,5471223,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3314,5471224,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",912,0,"",shellscript,selection_mouse +3315,5473965,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,tab +3316,5473966,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",945,0,"",shellscript,selection_mouse +3317,5474460,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",914,0,"",shellscript,selection_mouse +3318,5474977,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",884,0,"",shellscript,selection_mouse +3319,5475685,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",900,0,"\n --save_ckpt \",shellscript,content +3320,5475698,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",905,0,"",shellscript,selection_command +3321,5477115,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3322,5479031,"TERMINAL",0,0,"srun",,terminal_focus +3323,5482245,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3335335.0 tasks 0-7: running\r\n",,terminal_output +3324,5482434,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.0\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-5:\r\nsrun: forcing job termination\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-6:\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3335335.0 ON hkn0425 CANCELLED AT 2025-07-10T17:49:36 ***\r\n",,terminal_output +3325,5482598,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.0\r\nsrun: job abort in progress\r\n",,terminal_output +3326,5482769,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.0\r\n",,terminal_output +3327,5483625,"TERMINAL",0,0,"^C",,terminal_output +3328,5483747,"TERMINAL",0,0,"]0;tum_cte0515@hkn0425:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +3329,5483868,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0425:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +3330,5484035,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0425:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +3331,5484832,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/yolo-runs/tester.sh",,terminal_output +3332,5485883,"TERMINAL",0,0,"\r",,terminal_output +3333,5487122,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3334,5490815,"TERMINAL",0,0,"[?25lsh[?25h",,terminal_output +3335,5490911,"TERMINAL",0,0,"[?25lh[?25h[?25l [?25h",,terminal_output +3336,5491271,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",,terminal_output +3337,5494362,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=2\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=36:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:4\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\r\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nenv | grep SLURM\r\n\r\nsrun python train_tokenizer.py \\r\n --save_ckpt \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --min_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=1000 \\r\n --log_checkpoint_interval=1000 \\r\n --log \\r\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\r\n --tags tokenizer lr-sweep 1e-4 \\r\n --entity instant-uv \\r\n --project jafar \\r\n --data_dir $array_records_dir\r\n",,terminal_output +3338,5495062,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1196654\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\r\nSLURMD_NODENAME=hkn0425\r\nSLURM_JOB_START_TIME=1752159082\r\nSLURM_STEP_NODELIST=hkn0425\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752195082\r\nSLURM_PMI2_SRUN_PORT=33587\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3335335\r\nSLURM_PTY_PORT=45071\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=48\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e13.hkn0425\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=228\r\nSLURM_NODELIST=hkn[0425,0428]\r\nSLURM_SRUN_COMM_PORT=37803\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3335335\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0425\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_STEP_LAUNCHER_PORT=37803\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0425,0428]\r\n",,terminal_output +3339,5495218,"TERMINAL",0,0,"GpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +3340,5547011,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +3341,5547728,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.19.11\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250710_175041-le1aybcw\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run tokenizer-lr-sweep-1e-4-3335335\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/le1aybcw\r\n",,terminal_output +3342,5636717,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +3343,5637018,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\nRunning on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\nRunning on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\nRunning on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\n",,terminal_output +3344,5654393,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +3345,5654874,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\nRunning on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\n",,terminal_output +3346,5658991,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +3347,5659339,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\n",,terminal_output +3348,5663639,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +3349,5664026,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\n",,terminal_output +3350,5895307,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3335335.1 tasks 0-7: running\r\n",,terminal_output +3351,5895520,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.1\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-1:\r\nsrun: forcing job termination\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-8:\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3335335.1 ON hkn0425 CANCELLED AT 2025-07-10T17:56:29 ***\r\n",,terminal_output +3352,5895675,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.1\r\nsrun: job abort in progress\r\n",,terminal_output +3353,5895846,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.1\r\n",,terminal_output +3354,5896120,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.1\r\nsrun: job abort in progress\r\n",,terminal_output +3355,5896323,"TERMINAL",0,0,"^C",,terminal_output +3356,5896578,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.1\r\n",,terminal_output +3357,5896820,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.1\r\n]0;tum_cte0515@hkn0425:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +3358,5898713,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3359,5898714,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1075,0,"",shellscript,selection_mouse +3360,5899876,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1074,1,"",shellscript,content +3361,5900694,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1074,1,"",shellscript,content +3362,5901085,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1042,0,"",shellscript,selection_command +3363,5901747,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1040,0,"",shellscript,selection_command +3364,5902006,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1039,1,"",shellscript,content +3365,5902143,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1038,1,"",shellscript,content +3366,5903544,"TERMINAL",0,0,"sh slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",,terminal_output +3367,5904347,"TERMINAL",0,0,"\r\n[?2004l\r#!/usr/bin/env bash\r\n\r\n#SBATCH --nodes=2\r\n#SBATCH --ntasks-per-node=4\r\n#SBATCH --time=36:00:00\r\n#SBATCH --partition=accelerated\r\n#SBATCH --cpus-per-task=5\r\n#SBATCH --gres=gpu:4\r\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\r\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\r\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\r\n\r\n# Log the sbatch script\r\ncat $0\r\n\r\nmodule unload mpi/openmpi/5.0\r\nmodule unload devel/cuda/12.4\r\nsource .venv/bin/activate\r\n\r\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\n\r\njob_name=$SLURM_JOB_NAME\r\nslurm_job_id=$SLURM_JOB_ID\r\n\r\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\r\nmkdir -p $CHECKPOINT_DIR\r\n\r\nenv | grep SLURM\r\n\r\nsrun python train_tokenizer.py \\r\n --save_ckpt \\r\n --ckpt_dir $CHECKPOINT_DIR \\r\n --batch_size=96 \\r\n --min_lr=0 \\r\n --max_lr=1e-4 \\r\n --log_image_interval=10 \\r\n --log_checkpoint_interval=10 \\r\n --log \\r\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\r\n --tags tokenizer lr-sweep 1e-4 \\r\n --entity instant-uv \\r\n --project jafar \\r\n --data_dir $array_records_dir\r\n",,terminal_output +3368,5904516,"TERMINAL",0,0,"SLURM_STEP_NUM_TASKS=1\r\nSLURM_JOB_USER=tum_cte0515\r\nSLURM_TASKS_PER_NODE=4(x2)\r\nSLURM_JOB_UID=999226\r\nSLURM_TASK_PID=1196654\r\nSLURM_JOB_GPUS=0,1,2,3\r\nSLURM_LOCALID=0\r\nSLURM_SUBMIT_DIR=/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar\r\nSLURMD_NODENAME=hkn0425\r\nSLURM_JOB_START_TIME=1752159082\r\nSLURM_STEP_NODELIST=hkn0425\r\nSLURM_CLUSTER_NAME=hk\r\nSLURM_JOB_END_TIME=1752195082\r\nSLURM_PMI2_SRUN_PORT=33587\r\nSLURM_CPUS_ON_NODE=24\r\nSLURM_JOB_CPUS_PER_NODE=24(x2)\r\nSLURM_GPUS_ON_NODE=4\r\nSLURM_GTIDS=0\r\nSLURM_JOB_PARTITION=accelerated\r\nSLURM_TRES_PER_TASK=cpu=5\r\nSLURM_OOM_KILL_STEP=0\r\nSLURM_JOB_NUM_NODES=2\r\nSLURM_STEPID=4294967290\r\nSLURM_JOBID=3335335\r\nSLURM_PTY_PORT=45071\r\nSLURM_JOB_QOS=normal\r\nSLURM_LAUNCH_NODE_IPADDR=10.0.7.201\r\nSLURM_PTY_WIN_ROW=48\r\nSLURM_PMI2_PROC_MAPPING=(vector,(0,1,1))\r\nSLURMD_DEBUG=2\r\nSLURM_PROCID=0\r\nSLURM_CPUS_PER_TASK=5\r\nSLURM_NTASKS=8\r\nSLURM_TOPOLOGY_ADDR=hkibb.hkibbi1.hkibbi1e13.hkn0425\r\nSLURM_TOPOLOGY_ADDR_PATTERN=switch.switch.switch.node\r\nSLURM_SRUN_COMM_HOST=10.0.7.201\r\nSLURM_SCRIPT_CONTEXT=prolog_task\r\nSLURM_PTY_WIN_COL=228\r\nSLURM_NODELIST=hkn[0425,0428]\r\nSLURM_SRUN_COMM_PORT=37803\r\nSLURM_STEP_ID=4294967290\r\nSLURM_JOB_ACCOUNT=hk-project-p0023960\r\nSLURM_PRIO_PROCESS=0\r\nSLURM_NPROCS=8\r\nSLURM_NNODES=2\r\nSLURM_SUBMIT_HOST=hkn1993.localdomain\r\nSLURM_JOB_ID=3335335\r\nSLURM_NODEID=0\r\nSLURM_STEP_NUM_NODES=1\r\nSLURM_STEP_TASKS_PER_NODE=1\r\nSLURM_MPI_TYPE=pmi2\r\nSLURM_PMI2_STEP_NODES=hkn0425\r\nSLURM_CONF=/etc/slurm/slurm.conf\r\nSLURM_JOB_NAME=interactive\r\nSLURM_NTASKS_PER_NODE=4\r\nSLURM_STEP_LAUNCHER_PORT=37803\r\nSLURM_JOB_GID=502226\r\nSLURM_JOB_NODELIST=hkn[0425,0428]\r\n",,terminal_output +3369,5904651,"TERMINAL",0,0,"GpuFreq=control_disabled\r\nGpuFreq=control_disabled\r\n",,terminal_output +3370,5909460,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3371,5909460,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1072,0,"",shellscript,selection_mouse +3372,5917973,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1138,0,"",shellscript,selection_mouse +3373,5942582,"TERMINAL",0,0,"wandb: Currently logged in as: mihir-mahajan2002 (instant-uv) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin\r\n",,terminal_output +3374,5943276,"TERMINAL",0,0,"wandb: Tracking run with wandb version 0.19.11\r\nwandb: Run data is saved locally in /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/wandb/run-20250710_175716-a488tl5y\r\nwandb: Run `wandb offline` to turn off syncing.\r\nwandb: Syncing run tokenizer-lr-sweep-1e-4-3335335\r\nwandb: ⭐️ View project at https://wandb.ai/instant-uv/jafar\r\nwandb: 🚀 View run at https://wandb.ai/instant-uv/jafar/runs/a488tl5y\r\n",,terminal_output +3375,6010380,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3376,6010380,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",878,0,"",shellscript,selection_mouse +3377,6010920,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",911,0,"",shellscript,selection_mouse +3378,6011295,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",910,0,"",shellscript,selection_command +3379,6043413,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\nWARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +3380,6043473,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +3381,6043614,"TERMINAL",0,0,"WARNING:absl:Dropping 2 examples of 89394 examples (shard 8).\r\n",,terminal_output +3382,6043751,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\nRunning on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\nRunning on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\nRunning on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\nRunning on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\n",,terminal_output +3383,6043859,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\nRunning on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\n",,terminal_output +3384,6044165,"TERMINAL",0,0,"Running on 8 devices.\r\nCounting all components: ['encoder', 'vq', 'decoder']\r\nParameter counts:\r\n{'encoder': 18978432, 'vq': 32768, 'decoder': 18978416, 'total': 37989616}\r\nStarting training from step 0...\r\n",,terminal_output +3385,6081246,"TERMINAL",0,0,"bash",,terminal_focus +3386,6082724,"TERMINAL",0,0,"python",,terminal_command +3387,6082775,"TERMINAL",0,0,"]633;E;2025-07-10 17:59:36 python;d0af5434-9a5b-4b8f-9412-c82c37ee36e4]633;C",,terminal_output +3388,6082912,"TERMINAL",0,0,"Python 3.10.18 (main, Jun 4 2025, 17:36:27) [Clang 20.1.4 ] on linux\r\nType ""help"", ""copyright"", ""credits"" or ""license"" for more information.\r\n",,terminal_output +3389,6082979,"TERMINAL",0,0,">>> ",,terminal_output +3390,6084120,"TERMINAL",0,0,"8",,terminal_output +3391,6084401,"TERMINAL",0,0,"[?25l*[?25h",,terminal_output +3392,6084777,"TERMINAL",0,0,"[?25l4[?25h",,terminal_output +3393,6085044,"TERMINAL",0,0,"\r\n32\r\n>>> ",,terminal_output +3394,6085786,"TERMINAL",0,0,"^D\r\n]0;tum_cte0515@hkn1993:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +3395,6192480,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1175,0,"",shellscript,selection_mouse +3396,6192482,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1174,0,"",shellscript,selection_command +3397,6193609,"TERMINAL",0,0,"srun",,terminal_focus +3398,6223557,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,tab +3399,6226968,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,tab +3400,6229034,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3401,6230135,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,tab +3402,6231085,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",912,0,"",shellscript,selection_mouse +3403,6231221,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",907,9,"save_ckpt",shellscript,selection_mouse +3404,6231355,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",901,18," --save_ckpt \\n",shellscript,selection_mouse +3405,6233608,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3406,6251242,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=48 \\n --min_lr=0 \\n --max_lr=1e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-batch-size-scaling-1-node-$slurm_job_id \\n --tags dynamics debug \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +3407,6252584,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1085,0,"",shellscript,selection_mouse +3408,6253437,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1062,0,"",shellscript,selection_mouse +3409,6254353,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1067,0,"\n",shellscript,content +3410,6254654,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1068,0," --save_ckpt \\n",shellscript,content +3411,6255251,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1085,1,"",shellscript,content +3412,6258676,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,tab +3413,6259493,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1062,0,"",shellscript,selection_mouse +3414,6260226,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1067,0,"\n",shellscript,content +3415,6260505,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1068,0," --save_ckpt \\n",shellscript,content +3416,6260934,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1085,1,"",shellscript,content +3417,6265098,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=4\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_4_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=192 \\n --min_lr=0 \\n --max_lr=2.1e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-batch-size-scaling-4-node-$slurm_job_id \\n --tags dynamics debug \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +3418,6266165,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1084,0,"",shellscript,selection_mouse +3419,6266864,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1063,0,"",shellscript,selection_mouse +3420,6267569,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1062,0,"",shellscript,selection_command +3421,6267773,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1067,0,"\n",shellscript,content +3422,6268500,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1068,0," --save_ckpt \\n",shellscript,content +3423,6269080,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1085,1,"",shellscript,content +3424,6273175,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-batch-size-scaling-8-node-$slurm_job_id \\n --tags dynamics debug \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +3425,6275154,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1063,0,"",shellscript,selection_mouse +3426,6275467,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1062,0,"",shellscript,selection_command +3427,6275911,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1067,0,"\n",shellscript,content +3428,6276285,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1068,0," --save_ckpt \\n",shellscript,content +3429,6276680,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1085,1,"",shellscript,content +3430,6279238,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=16\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_16_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=768 \\n --min_lr=0 \\n --max_lr=4.2e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-batch-size-scaling-16-node-$slurm_job_id \\n --tags dynamics debug \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +3431,6280377,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1058,0,"",shellscript,selection_mouse +3432,6280572,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1057,0,"",shellscript,selection_command +3433,6281241,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1069,0,"\n",shellscript,content +3434,6281511,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1070,0," --save_ckpt \\n",shellscript,content +3435,6281905,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1087,1,"",shellscript,content +3436,6296483,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=6 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-6-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 6 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +3437,6299390,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1113,0,"",shellscript,selection_mouse +3438,6301354,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1084,0,"\n",shellscript,content +3439,6301364,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1085,0," ",shellscript,content +3440,6302081,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1089,0," --save_ckpt \\n",shellscript,content +3441,6302496,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1106,1,"",shellscript,content +3442,6304055,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1094,0,"",shellscript,selection_command +3443,6304268,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1093,0,"",shellscript,selection_command +3444,6304494,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1089,4,"",shellscript,content +3445,6304686,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1085,4,"",shellscript,content +3446,6305237,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1085,0," ",shellscript,content +3447,6305983,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1088,0,"",shellscript,selection_command +3448,6308356,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=8 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-8-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 8 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +3449,6309196,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1066,0,"",shellscript,selection_mouse +3450,6309907,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1084,0,"\n --save_ckpt \",shellscript,content +3451,6309909,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1089,0,"",shellscript,selection_command +3452,6312223,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=12 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-12-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 12 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +3453,6313287,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1095,0,"",shellscript,selection_mouse +3454,6314005,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1064,0,"",shellscript,selection_mouse +3455,6314802,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1084,0,"\n --save_ckpt \",shellscript,content +3456,6314818,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1089,0,"",shellscript,selection_command +3457,6316391,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-20-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 20 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +3458,6317284,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1069,0,"",shellscript,selection_mouse +3459,6318005,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1084,0,"\n --save_ckpt \",shellscript,content +3460,6318006,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1089,0,"",shellscript,selection_command +3461,6319872,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=50 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-50-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 50 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +3462,6320776,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1066,0,"",shellscript,selection_mouse +3463,6321496,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1084,0,"\n --save_ckpt \",shellscript,content +3464,6321516,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1089,0,"",shellscript,selection_command +3465,6394258,"TERMINAL",0,0,"bash",,terminal_focus +3466,6395085,"TERMINAL",0,0,"srun",,terminal_focus +3467,6405389,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_36M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,tab +3468,6409868,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1214,0,"",shellscript,selection_mouse +3469,6409869,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1213,0,"",shellscript,selection_command +3470,6410397,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1234,0,"",shellscript,selection_mouse +3471,6410399,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1233,0,"",shellscript,selection_command +3472,6411095,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1607,0,"",shellscript,selection_mouse +3473,6411111,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1606,0,"",shellscript,selection_command +3474,6660791,"TERMINAL",0,0,"Step 0, loss: 0.33292484283447266\r\nStep 1, loss: 0.33973613381385803\r\nStep 2, loss: 0.3459185063838959\r\nStep 3, loss: 0.33234596252441406\r\nStep 4, loss: 0.3411635756492615\r\nStep 5, loss: 0.3499658405780792\r\nStep 6, loss: 0.3499256670475006\r\nStep 7, loss: 0.34394073486328125\r\nStep 8, loss: 0.3490101397037506\r\nStep 9, loss: 0.3501622974872589\r\nSaved checkpoint at step 10\r\nStep 10, loss: 0.3454197645187378\r\nStep 11, loss: 0.3307967782020569\r\nStep 12, loss: 0.3328101336956024\r\nStep 13, loss: 0.34072747826576233\r\nStep 14, loss: 0.3415687382221222\r\nStep 15, loss: 0.3421204090118408\r\nStep 16, loss: 0.34998252987861633\r\nStep 17, loss: 0.3370188772678375\r\nStep 18, loss: 0.3368116021156311\r\nStep 19, loss: 0.3397890627384186\r\nSaved checkpoint at step 20\r\nStep 20, loss: 0.33846819400787354\r\nStep 21, loss: 0.328185111284256\r\nStep 22, loss: 0.33603933453559875\r\nStep 23, loss: 0.3364322781562805\r\nStep 24, loss: 0.3406607210636139\r\nStep 25, loss: 0.3376489579677582\r\nStep 26, loss: 0.3372756540775299\r\nStep 27, loss: 0.3242914080619812\r\nStep 28, loss: 0.32896149158477783\r\nStep 29, loss: 0.32680416107177734\r\nSaved checkpoint at step 30\r\nStep 30, loss: 0.31527817249298096\r\nStep 31, loss: 0.32453596591949463\r\nStep 32, loss: 0.32291722297668457\r\nStep 33, loss: 0.3086308240890503\r\nStep 34, loss: 0.3201759457588196\r\nStep 35, loss: 0.32361075282096863\r\nStep 36, loss: 0.3231964409351349\r\nStep 37, loss: 0.32022643089294434\r\nStep 38, loss: 0.31821075081825256\r\nStep 39, loss: 0.3172142505645752\r\nSaved checkpoint at step 40\r\nStep 40, loss: 0.32197457551956177\r\nStep 41, loss: 0.3097772002220154\r\nStep 42, loss: 0.3095897436141968\r\nStep 43, loss: 0.30802375078201294\r\nStep 44, loss: 0.3085505962371826\r\nStep 45, loss: 0.30728694796562195\r\nStep 46, loss: 0.30426204204559326\r\nStep 47, loss: 0.31653067469596863\r\nStep 48, loss: 0.29350337386131287\r\nStep 49, loss: 0.2900897264480591\r\nSaved checkpoint at step 50\r\nStep 50, loss: 0.3072112500667572\r\nStep 51, loss: 0.29866600036621094\r\nStep 52, loss: 0.2933558225631714\r\nStep 53, loss: 0.28990650177001953\r\nStep 54, loss: 0.2917138338088989\r\nStep 55, loss: 0.29527750611305237\r\nStep 56, loss: 0.2855745553970337\r\nStep 57, loss: 0.29226770997047424\r\nStep 58, loss: 0.28528568148612976\r\nStep 59, loss: 0.2834138870239258\r\nSaved checkpoint at step 60\r\nStep 60, loss: 0.2725016474723816\r\nStep 61, loss: 0.28022363781929016\r\nStep 62, loss: 0.27179959416389465\r\nStep 63, loss: 0.2702350914478302\r\nStep 64, loss: 0.2684958875179291\r\nStep 65, loss: 0.26719269156455994\r\nStep 66, loss: 0.25736168026924133\r\nStep 67, loss: 0.26021021604537964\r\nStep 68, loss: 0.2600043714046478\r\nStep 69, loss: 0.2554631233215332\r\nSaved checkpoint at step 70\r\nStep 70, loss: 0.25915834307670593\r\nStep 71, loss: 0.2538088262081146\r\nStep 72, loss: 0.2527981102466583\r\nStep 73, loss: 0.24763624370098114\r\nStep 74, loss: 0.23982031643390656\r\nStep 75, loss: 0.2431134581565857\r\nStep 76, loss: 0.24418558180332184\r\nStep 77, loss: 0.23978117108345032\r\nStep 78, loss: 0.23628409206867218\r\nStep 79, loss: 0.22659100592136383\r\nSaved checkpoint at step 80\r\nStep 80, loss: 0.23002658784389496\r\nStep 81, loss: 0.2237771600484848\r\nStep 82, loss: 0.21777285635471344\r\nStep 83, loss: 0.21955353021621704\r\nStep 84, loss: 0.2197633534669876\r\nStep 85, loss: 0.21693949401378632\r\nStep 86, loss: 0.2156858891248703\r\nStep 87, loss: 0.20907285809516907\r\nStep 88, loss: 0.20304414629936218\r\nStep 89, loss: 0.20950746536254883\r\nSaved checkpoint at step 90\r\nStep 90, loss: 0.20157180726528168\r\nStep 91, loss: 0.19864089787006378\r\nStep 92, loss: 0.1961870640516281\r\nStep 93, loss: 0.1970020979642868\r\nStep 94, loss: 0.19751623272895813\r\nStep 95, loss: 0.1926620602607727\r\nStep 96, loss: 0.1907958984375\r\nStep 97, loss: 0.18797270953655243\r\nStep 98, loss: 0.18371835350990295\r\nStep 99, loss: 0.19347549974918365\r\nSaved checkpoint at step 100\r\nStep 100, loss: 0.17737938463687897\r\nStep 101, loss: 0.18215568363666534\r\nStep 102, loss: 0.17679424583911896\r\nStep 103, loss: 0.17569106817245483\r\nStep 104, loss: 0.17612037062644958\r\nStep 105, loss: 0.1708865910768509\r\nStep 106, loss: 0.17582806944847107\r\nStep 107, loss: 0.16020779311656952\r\nStep 108, loss: 0.16374938189983368\r\nStep 0, loss: 0.33292484283447266\r\nStep 1, loss: 0.33973613381385803\r\nStep 2, loss: 0.3459185063838959\r\nStep 3, loss: 0.33234596252441406\r\nStep 4, loss: 0.3411635756492615\r\nStep 5, loss: 0.3499658405780792\r\nStep 6, loss: 0.3499256670475006\r\nStep 7, loss: 0.34394073486328125\r\nStep 8, loss: 0.3490101397037506\r\nStep 9, loss: 0.3501622974872589\r\nSaved checkpoint at step 10\r\nStep 10, loss: 0.3454197645187378\r\nStep 11, loss: 0.3307967782020569\r\nStep 12, loss: 0.3328101336956024\r\nStep 13, loss: 0.34072747826576233\r\nStep 14, loss: 0.3415687382221222\r\nStep 15, loss: 0.3421204090118408\r\nStep 16, loss: 0.34998252987861633\r\nStep 17, loss: 0.3370188772678375\r\nStep 18, loss: 0.3368116021156311\r\nStep 19, loss: 0.3397890627384186\r\nSaved checkpoint at step 20\r\nStep 20, loss: 0.33846819400787354\r\nStep 21, loss: 0.328185111284256\r\nStep 22, loss: 0.33603933453559875\r\nStep 23, loss: 0.3364322781562805\r\nStep 24, loss: 0.3406607210636139\r\nStep 25, loss: 0.3376489579677582\r\nStep 26, loss: 0.3372756540775299\r\nStep 27, loss: 0.3242914080619812\r\nStep 109, loss: 0.16316014528274536\r\nSaved checkpoint at step 110\r\nStep 110, loss: 0.15965776145458221\r\nStep 111, loss: 0.16084066033363342\r\nStep 112, loss: 0.15732374787330627\r\nStep 113, loss: 0.1501660794019699\r\nStep 114, loss: 0.15066221356391907\r\nStep 115, loss: 0.15051837265491486\r\nStep 116, loss: 0.1540840119123459\r\nStep 117, loss: 0.15616631507873535\r\nStep 118, loss: 0.14677441120147705\r\nStep 119, loss: 0.15072251856327057\r\nSaved checkpoint at step 120\r\nStep 120, loss: 0.14081387221813202\r\nStep 121, loss: 0.1404900848865509\r\nStep 122, loss: 0.13735362887382507\r\nStep 123, loss: 0.14047737419605255\r\nStep 124, loss: 0.14097315073013306\r\nStep 125, loss: 0.13012643158435822\r\nStep 126, loss: 0.13319577276706696\r\nStep 127, loss: 0.12502308189868927\r\nStep 128, loss: 0.13217462599277496\r\nStep 129, loss: 0.12322020530700684\r\nSaved checkpoint at step 130\r\nStep 130, loss: 0.1309303492307663\r\nStep 131, loss: 0.12471090257167816\r\nStep 132, loss: 0.1259678304195404\r\nStep 133, loss: 0.11994366347789764\r\nStep 134, loss: 0.12396806478500366\r\nStep 135, loss: 0.12951606512069702\r\nStep 28, loss: 0.32896149158477783\r\nStep 29, loss: 0.32680416107177734\r\nSaved checkpoint at step 30\r\nStep 30, loss: 0.31527817249298096\r\nStep 31, loss: 0.32453596591949463\r\nStep 32, loss: 0.32291722297668457\r\nStep 33, loss: 0.3086308240890503\r\nStep 34, loss: 0.3201759457588196\r\nStep 35, loss: 0.32361075282096863\r\nStep 36, loss: 0.3231964409351349\r\nStep 37, loss: 0.32022643089294434\r\nStep 38, loss: 0.31821075081825256\r\nStep 39, loss: 0.3172142505645752\r\nSaved checkpoint at step 40\r\nStep 40, loss: 0.32197457551956177\r\nStep 41, loss: 0.3097772002220154\r\nStep 42, loss: 0.3095897436141968\r\nStep 43, loss: 0.30802375078201294\r\nStep 44, loss: 0.3085505962371826\r\nStep 45, loss: 0.30728694796562195\r\nStep 46, loss: 0.30426204204559326\r\nStep 47, loss: 0.31653067469596863\r\nStep 48, loss: 0.29350337386131287\r\nStep 49, loss: 0.2900897264480591\r\nSaved checkpoint at step 50\r\nStep 50, loss: 0.3072112500667572\r\nStep 51, loss: 0.29866600036621094\r\nStep 52, loss: 0.2933558225631714\r\nStep 53, loss: 0.28990650177001953\r\nStep 54, loss: 0.2917138338088989\r\nStep 136, loss: 0.12501047551631927\r\nStep 137, loss: 0.12460850179195404\r\nStep 138, loss: 0.12312853336334229\r\nStep 139, loss: 0.11870359629392624\r\nSaved checkpoint at step 140\r\nStep 140, loss: 0.10972652584314346\r\nStep 141, loss: 0.11810508370399475\r\nStep 142, loss: 0.11488064378499985\r\nStep 143, loss: 0.10100933909416199\r\nStep 144, loss: 0.11190460622310638\r\nStep 145, loss: 0.10655320435762405\r\nStep 146, loss: 0.11590567231178284\r\nStep 147, loss: 0.11223439872264862\r\nStep 148, loss: 0.11101178079843521\r\nStep 149, loss: 0.11008752137422562\r\nSaved checkpoint at step 150\r\nStep 150, loss: 0.1229330375790596\r\nStep 151, loss: 0.12016700953245163\r\nStep 152, loss: 0.11195394396781921\r\nStep 153, loss: 0.11655457317829132\r\nStep 154, loss: 0.10912562161684036\r\nStep 155, loss: 0.12283866107463837\r\nStep 156, loss: 0.10622447729110718\r\nStep 157, loss: 0.10579507797956467\r\nStep 158, loss: 0.09965544939041138\r\nStep 159, loss: 0.10587912797927856\r\nSaved checkpoint at step 160\r\nStep 160, loss: 0.11324921250343323\r\nStep 161, loss: 0.11093202978372574\r\nStep 162, loss: 0.10980116575956345\r\nStep 163, loss: 0.10331201553344727\r\nStep 164, loss: 0.09606180340051651\r\nStep 165, loss: 0.10288108885288239\r\nStep 166, loss: 0.10788475722074509\r\nStep 167, loss: 0.11626137793064117\r\nStep 168, loss: 0.11271797865629196\r\nStep 169, loss: 0.10126489400863647\r\nSaved checkpoint at step 170\r\nStep 170, loss: 0.10251966863870621\r\nStep 171, loss: 0.09720180183649063\r\nStep 172, loss: 0.10119358450174332\r\nStep 173, loss: 0.10164882242679596\r\nStep 174, loss: 0.09646020829677582\r\nStep 175, loss: 0.09435637295246124\r\nStep 176, loss: 0.09778910130262375\r\nStep 177, loss: 0.09556873142719269\r\nStep 178, loss: 0.10166315734386444\r\nStep 179, loss: 0.09307191520929337\r\nSaved checkpoint at step 180\r\nStep 180, loss: 0.09001213312149048\r\nStep 181, loss: 0.09384429454803467\r\nStep 182, loss: 0.09797736257314682\r\nStep 183, loss: 0.09913866221904755\r\nStep 184, loss: 0.09100943058729172\r\nStep 185, loss: 0.09598495066165924\r\nStep 186, loss: 0.09197870641946793\r\nStep 187, loss: 0.09248428791761398\r\nStep 188, loss: 0.0881965383887291\r\nStep 189, loss: 0.08620244264602661\r\nSaved checkpoint at step 190\r\nStep 190, loss: 0.0936160534620285\r\nStep 191, loss: 0.08297702670097351\r\nStep 192, loss: 0.08549270033836365\r\nStep 193, loss: 0.09282968193292618\r\nStep 194, loss: 0.09096062183380127\r\nStep 195, loss: 0.08280432224273682\r\nStep 196, loss: 0.08810421824455261\r\nStep 197, loss: 0.08211309462785721\r\nStep 198, loss: 0.07822512835264206\r\nStep 199, loss: 0.08037614822387695\r\nSaved checkpoint at step 200\r\nStep 200, loss: 0.08436369150876999\r\nStep 201, loss: 0.0888621136546135\r\nStep 202, loss: 0.08680012822151184\r\nStep 203, loss: 0.08875889331102371\r\nStep 204, loss: 0.08150099217891693\r\nStep 205, loss: 0.08258917927742004\r\nStep 206, loss: 0.07716099172830582\r\nStep 207, loss: 0.07317893952131271\r\nStep 208, loss: 0.07029818743467331\r\nStep 209, loss: 0.07618950307369232\r\nSaved checkpoint at step 210\r\nStep 210, loss: 0.07858575135469437\r\nStep 211, loss: 0.07911297678947449\r\nStep 212, loss: 0.09101250022649765\r\nStep 213, loss: 0.07573471963405609\r\nStep 55, loss: 0.29527750611305237\r\nStep 56, loss: 0.2855745553970337\r\nStep 57, loss: 0.29226770997047424\r\nStep 58, loss: 0.28528568148612976\r\nStep 59, loss: 0.2834138870239258\r\nSaved checkpoint at step 60\r\nStep 60, loss: 0.2725016474723816\r\nStep 61, loss: 0.28022363781929016\r\nStep 62, loss: 0.27179959416389465\r\nStep 63, loss: 0.2702350914478302\r\nStep 64, loss: 0.2684958875179291\r\nStep 65, loss: 0.26719269156455994\r\nStep 66, loss: 0.25736168026924133\r\nStep 67, loss: 0.26021021604537964\r\nStep 68, loss: 0.2600043714046478\r\nStep 69, loss: 0.2554631233215332\r\nSaved checkpoint at step 70\r\nStep 70, loss: 0.25915834307670593\r\nStep 71, loss: 0.2538088262081146\r\nStep 72, loss: 0.2527981102466583\r\nStep 73, loss: 0.24763624370098114\r\nStep 74, loss: 0.23982031643390656\r\nStep 75, loss: 0.2431134581565857\r\nStep 76, loss: 0.24418558180332184\r\nStep 77, loss: 0.23978117108345032\r\nStep 78, loss: 0.23628409206867218\r\nStep 79, loss: 0.22659100592136383\r\nSaved checkpoint at step 80\r\nStep 80, loss: 0.23002658784389496\r\nStep 81, loss: 0.2237771600484848\r\nStep 214, loss: 0.07784590125083923\r\nStep 82, loss: 0.21777285635471344\r\nStep 83, loss: 0.21955353021621704\r\nStep 84, loss: 0.2197633534669876\r\nStep 85, loss: 0.21693949401378632\r\nStep 86, loss: 0.2156858891248703\r\nStep 87, loss: 0.20907285809516907\r\nStep 88, loss: 0.20304414629936218\r\nStep 89, loss: 0.20950746536254883\r\nSaved checkpoint at step 90\r\nStep 90, loss: 0.20157180726528168\r\nStep 91, loss: 0.19864089787006378\r\nStep 92, loss: 0.1961870640516281\r\nStep 93, loss: 0.1970020979642868\r\nStep 94, loss: 0.19751623272895813\r\nStep 95, loss: 0.1926620602607727\r\nStep 96, loss: 0.1907958984375\r\nStep 97, loss: 0.18797270953655243\r\nStep 98, loss: 0.18371835350990295\r\nStep 99, loss: 0.19347549974918365\r\nSaved checkpoint at step 100\r\nStep 100, loss: 0.17737938463687897\r\nStep 101, loss: 0.18215568363666534\r\nStep 102, loss: 0.17679424583911896\r\nStep 103, loss: 0.17569106817245483\r\nStep 104, loss: 0.17612037062644958\r\nStep 105, loss: 0.1708865910768509\r\nStep 106, loss: 0.17582806944847107\r\nStep 107, loss: 0.16020779311656952\r\nStep 108, loss: 0.16374938189983368\r\nStep 0, loss: 0.33292484283447266\r\nStep 1, loss: 0.33973613381385803\r\nStep 2, loss: 0.3459185063838959\r\nStep 3, loss: 0.33234596252441406\r\nStep 4, loss: 0.3411635756492615\r\nStep 5, loss: 0.3499658405780792\r\nStep 6, loss: 0.3499256670475006\r\nStep 7, loss: 0.34394073486328125\r\nStep 8, loss: 0.3490101397037506\r\nStep 9, loss: 0.3501622974872589\r\nSaved checkpoint at step 10\r\nStep 10, loss: 0.3454197645187378\r\nStep 11, loss: 0.3307967782020569\r\nStep 12, loss: 0.3328101336956024\r\nStep 13, loss: 0.34072747826576233\r\nStep 14, loss: 0.3415687382221222\r\nStep 15, loss: 0.3421204090118408\r\nStep 16, loss: 0.34998252987861633\r\nStep 17, loss: 0.3370188772678375\r\nStep 18, loss: 0.3368116021156311\r\nStep 19, loss: 0.3397890627384186\r\nSaved checkpoint at step 20\r\nStep 20, loss: 0.33846819400787354\r\nStep 21, loss: 0.328185111284256\r\nStep 22, loss: 0.33603933453559875\r\nStep 23, loss: 0.3364322781562805\r\nStep 24, loss: 0.3406607210636139\r\nStep 25, loss: 0.3376489579677582\r\nStep 26, loss: 0.3372756540775299\r\nStep 27, loss: 0.3242914080619812\r\nStep 109, loss: 0.16316014528274536\r\nStep 28, loss: 0.32896149158477783\r\nStep 29, loss: 0.32680416107177734\r\nSaved checkpoint at step 30\r\nStep 30, loss: 0.31527817249298096\r\nStep 31, loss: 0.32453596591949463\r\nStep 32, loss: 0.32291722297668457\r\nStep 33, loss: 0.3086308240890503\r\nStep 34, loss: 0.3201759457588196\r\nStep 35, loss: 0.32361075282096863\r\nStep 36, loss: 0.3231964409351349\r\nStep 37, loss: 0.32022643089294434\r\nStep 38, loss: 0.31821075081825256\r\nStep 39, loss: 0.3172142505645752\r\nSaved checkpoint at step 40\r\nStep 40, loss: 0.32197457551956177\r\nStep 41, loss: 0.3097772002220154\r\nStep 42, loss: 0.3095897436141968\r\nStep 43, loss: 0.30802375078201294\r\nStep 44, loss: 0.3085505962371826\r\nStep 45, loss: 0.30728694796562195\r\nStep 46, loss: 0.30426204204559326\r\nStep 47, loss: 0.31653067469596863\r\nStep 48, loss: 0.29350337386131287\r\nStep 49, loss: 0.2900897264480591\r\nSaved checkpoint at step 50\r\nStep 50, loss: 0.3072112500667572\r\nStep 51, loss: 0.29866600036621094\r\nStep 52, loss: 0.2933558225631714\r\nStep 53, loss: 0.28990650177001953\r\nStep 54, loss: 0.2917138338088989\r\nSaved checkpoint at step 110\r\nStep 110, loss: 0.15965776145458221\r\nStep 111, loss: 0.16084066033363342\r\nStep 112, loss: 0.15732374787330627\r\nStep 113, loss: 0.1501660794019699\r\nStep 114, loss: 0.15066221356391907\r\nStep 115, loss: 0.15051837265491486\r\nStep 116, loss: 0.1540840119123459\r\nStep 117, loss: 0.15616631507873535\r\nStep 118, loss: 0.14677441120147705\r\nStep 119, loss: 0.15072251856327057\r\nSaved checkpoint at step 120\r\nStep 120, loss: 0.14081387221813202\r\nStep 121, loss: 0.1404900848865509\r\nStep 122, loss: 0.13735362887382507\r\nStep 123, loss: 0.14047737419605255\r\nStep 124, loss: 0.14097315073013306\r\nStep 125, loss: 0.13012643158435822\r\nStep 126, loss: 0.13319577276706696\r\nStep 127, loss: 0.12502308189868927\r\nStep 128, loss: 0.13217462599277496\r\nStep 129, loss: 0.12322020530700684\r\nSaved checkpoint at step 130\r\nStep 130, loss: 0.1309303492307663\r\nStep 131, loss: 0.12471090257167816\r\nStep 132, loss: 0.1259678304195404\r\nStep 133, loss: 0.11994366347789764\r\nStep 134, loss: 0.12396806478500366\r\nStep 135, loss: 0.12951606512069702\r\nStep 55, loss: 0.29527750611305237\r\nStep 56, loss: 0.2855745553970337\r\nStep 57, loss: 0.29226770997047424\r\nStep 58, loss: 0.28528568148612976\r\nStep 59, loss: 0.2834138870239258\r\nSaved checkpoint at step 60\r\nStep 60, loss: 0.2725016474723816\r\nStep 61, loss: 0.28022363781929016\r\nStep 62, loss: 0.27179959416389465\r\nStep 63, loss: 0.2702350914478302\r\nStep 64, loss: 0.2684958875179291\r\nStep 65, loss: 0.26719269156455994\r\nStep 66, loss: 0.25736168026924133\r\nStep 67, loss: 0.26021021604537964\r\nStep 68, loss: 0.2600043714046478\r\nStep 69, loss: 0.2554631233215332\r\nSaved checkpoint at step 70\r\nStep 70, loss: 0.25915834307670593\r\nStep 71, loss: 0.2538088262081146\r\nStep 72, loss: 0.2527981102466583\r\nStep 73, loss: 0.24763624370098114\r\nStep 74, loss: 0.23982031643390656\r\nStep 75, loss: 0.2431134581565857\r\nStep 76, loss: 0.24418558180332184\r\nStep 77, loss: 0.23978117108345032\r\nStep 78, loss: 0.23628409206867218\r\nStep 79, loss: 0.22659100592136383\r\nSaved checkpoint at step 80\r\nStep 80, loss: 0.23002658784389496\r\nStep 81, loss: 0.2237771600484848\r\nStep 136, loss: 0.12501047551631927\r\nStep 137, loss: 0.12460850179195404\r\nStep 138, loss: 0.12312853336334229\r\nStep 139, loss: 0.11870359629392624\r\nSaved checkpoint at step 140\r\nStep 140, loss: 0.10972652584314346\r\nStep 141, loss: 0.11810508370399475\r\nStep 142, loss: 0.11488064378499985\r\nStep 143, loss: 0.10100933909416199\r\nStep 144, loss: 0.11190460622310638\r\nStep 145, loss: 0.10655320435762405\r\nStep 146, loss: 0.11590567231178284\r\nStep 147, loss: 0.11223439872264862\r\nStep 148, loss: 0.11101178079843521\r\nStep 149, loss: 0.11008752137422562\r\nSaved checkpoint at step 150\r\nStep 150, loss: 0.1229330375790596\r\nStep 151, loss: 0.12016700953245163\r\nStep 152, loss: 0.11195394396781921\r\nStep 153, loss: 0.11655457317829132\r\nStep 154, loss: 0.10912562161684036\r\nStep 155, loss: 0.12283866107463837\r\nStep 156, loss: 0.10622447729110718\r\nStep 157, loss: 0.10579507797956467\r\nStep 158, loss: 0.09965544939041138\r\nStep 159, loss: 0.10587912797927856\r\nSaved checkpoint at step 160\r\nStep 160, loss: 0.11324921250343323\r\nStep 161, loss: 0.11093202978372574\r\nStep 82, loss: 0.21777285635471344\r\nStep 83, loss: 0.21955353021621704\r\nStep 84, loss: 0.2197633534669876\r\nStep 85, loss: 0.21693949401378632\r\nStep 86, loss: 0.2156858891248703\r\nStep 87, loss: 0.20907285809516907\r\nStep 88, loss: 0.20304414629936218\r\nStep 89, loss: 0.20950746536254883\r\nSaved checkpoint at step 90\r\nStep 90, loss: 0.20157180726528168\r\nStep 91, loss: 0.19864089787006378\r\nStep 92, loss: 0.1961870640516281\r\nStep 93, loss: 0.1970020979642868\r\nStep 94, loss: 0.19751623272895813\r\nStep 95, loss: 0.1926620602607727\r\nStep 96, loss: 0.1907958984375\r\nStep 97, loss: 0.18797270953655243\r\nStep 98, loss: 0.18371835350990295\r\nStep 99, loss: 0.19347549974918365\r\nSaved checkpoint at step 100\r\nStep 100, loss: 0.17737938463687897\r\nStep 101, loss: 0.18215568363666534\r\nStep 102, loss: 0.17679424583911896\r\nStep 103, loss: 0.17569106817245483\r\nStep 104, loss: 0.17612037062644958\r\nStep 105, loss: 0.1708865910768509\r\nStep 106, loss: 0.17582806944847107\r\nStep 107, loss: 0.16020779311656952\r\nStep 108, loss: 0.16374938189983368\r\nStep 162, loss: 0.10980116575956345\r\nStep 163, loss: 0.10331201553344727\r\nStep 164, loss: 0.09606180340051651\r\nStep 165, loss: 0.10288108885288239\r\nStep 166, loss: 0.10788475722074509\r\nStep 167, loss: 0.11626137793064117\r\nStep 168, loss: 0.11271797865629196\r\nStep 169, loss: 0.10126489400863647\r\nSaved checkpoint at step 170\r\nStep 170, loss: 0.10251966863870621\r\nStep 171, loss: 0.09720180183649063\r\nStep 172, loss: 0.10119358450174332\r\nStep 173, loss: 0.10164882242679596\r\nStep 174, loss: 0.09646020829677582\r\nStep 175, loss: 0.09435637295246124\r\nStep 176, loss: 0.09778910130262375\r\nStep 177, loss: 0.09556873142719269\r\nStep 178, loss: 0.10166315734386444\r\nStep 179, loss: 0.09307191520929337\r\nSaved checkpoint at step 180\r\nStep 180, loss: 0.09001213312149048\r\nStep 181, loss: 0.09384429454803467\r\nStep 182, loss: 0.09797736257314682\r\nStep 183, loss: 0.09913866221904755\r\nStep 184, loss: 0.09100943058729172\r\nStep 185, loss: 0.09598495066165924\r\nStep 186, loss: 0.09197870641946793\r\nStep 187, loss: 0.09248428791761398\r\nStep 109, loss: 0.16316014528274536\r\nStep 188, loss: 0.0881965383887291\r\nStep 189, loss: 0.08620244264602661\r\nSaved checkpoint at step 190\r\nStep 190, loss: 0.0936160534620285\r\nStep 191, loss: 0.08297702670097351\r\nStep 192, loss: 0.08549270033836365\r\nStep 193, loss: 0.09282968193292618\r\nStep 194, loss: 0.09096062183380127\r\nStep 195, loss: 0.08280432224273682\r\nStep 196, loss: 0.08810421824455261\r\nStep 197, loss: 0.08211309462785721\r\nStep 198, loss: 0.07822512835264206\r\nStep 199, loss: 0.08037614822387695\r\nSaved checkpoint at step 200\r\nStep 200, loss: 0.08436369150876999\r\nStep 201, loss: 0.0888621136546135\r\nStep 202, loss: 0.08680012822151184\r\nStep 203, loss: 0.08875889331102371\r\nStep 204, loss: 0.08150099217891693\r\nStep 205, loss: 0.08258917927742004\r\nStep 206, loss: 0.07716099172830582\r\nStep 207, loss: 0.07317893952131271\r\nStep 208, loss: 0.07029818743467331\r\nStep 209, loss: 0.07618950307369232\r\nSaved checkpoint at step 210\r\nStep 210, loss: 0.07858575135469437\r\nStep 211, loss: 0.07911297678947449\r\nStep 212, loss: 0.09101250022649765\r\nStep 213, loss: 0.07573471963405609\r\nSaved checkpoint at step 110\r\nStep 110, loss: 0.15965776145458221\r\nStep 111, loss: 0.16084066033363342\r\nStep 112, loss: 0.15732374787330627\r\nStep 113, loss: 0.1501660794019699\r\nStep 114, loss: 0.15066221356391907\r\nStep 115, loss: 0.15051837265491486\r\nStep 116, loss: 0.1540840119123459\r\nStep 117, loss: 0.15616631507873535\r\nStep 118, loss: 0.14677441120147705\r\nStep 119, loss: 0.15072251856327057\r\nSaved checkpoint at step 120\r\nStep 120, loss: 0.14081387221813202\r\nStep 121, loss: 0.1404900848865509\r\nStep 122, loss: 0.13735362887382507\r\nStep 123, loss: 0.14047737419605255\r\nStep 124, loss: 0.14097315073013306\r\nStep 125, loss: 0.13012643158435822\r\nStep 126, loss: 0.13319577276706696\r\nStep 127, loss: 0.12502308189868927\r\nStep 128, loss: 0.13217462599277496\r\nStep 129, loss: 0.12322020530700684\r\nSaved checkpoint at step 130\r\nStep 130, loss: 0.1309303492307663\r\nStep 131, loss: 0.12471090257167816\r\nStep 132, loss: 0.1259678304195404\r\nStep 133, loss: 0.11994366347789764\r\nStep 134, loss: 0.12396806478500366\r\nStep 135, loss: 0.12951606512069702\r\nStep 214, loss: 0.07784590125083923\r\nStep 136, loss: 0.12501047551631927\r\nStep 137, loss: 0.12460850179195404\r\nStep 138, loss: 0.12312853336334229\r\nStep 139, loss: 0.11870359629392624\r\nSaved checkpoint at step 140\r\nStep 140, loss: 0.10972652584314346\r\nStep 141, loss: 0.11810508370399475\r\nStep 142, loss: 0.11488064378499985\r\nStep 143, loss: 0.10100933909416199\r\nStep 144, loss: 0.11190460622310638\r\nStep 145, loss: 0.10655320435762405\r\nStep 146, loss: 0.11590567231178284\r\nStep 147, loss: 0.11223439872264862\r\nStep 148, loss: 0.11101178079843521\r\nStep 149, loss: 0.11008752137422562\r\nSaved checkpoint at step 150\r\nStep 150, loss: 0.1229330375790596\r\nStep 151, loss: 0.12016700953245163\r\nStep 152, loss: 0.11195394396781921\r\nStep 153, loss: 0.11655457317829132\r\nStep 154, loss: 0.10912562161684036\r\nStep 155, loss: 0.12283866107463837\r\nStep 156, loss: 0.10622447729110718\r\nStep 157, loss: 0.10579507797956467\r\nStep 158, loss: 0.09965544939041138\r\nStep 159, loss: 0.10587912797927856\r\nSaved checkpoint at step 160\r\nStep 160, loss: 0.11324921250343323\r\nStep 161, loss: 0.11093202978372574\r\nStep 0, loss: 0.33292484283447266\r\nStep 1, loss: 0.33973613381385803\r\nStep 2, loss: 0.3459185063838959\r\nStep 3, loss: 0.33234596252441406\r\nStep 4, loss: 0.3411635756492615\r\nStep 5, loss: 0.3499658405780792\r\nStep 6, loss: 0.3499256670475006\r\nStep 7, loss: 0.34394073486328125\r\nStep 8, loss: 0.3490101397037506\r\nStep 9, loss: 0.3501622974872589\r\nSaved checkpoint at step 10\r\nStep 10, loss: 0.3454197645187378\r\nStep 11, loss: 0.3307967782020569\r\nStep 12, loss: 0.3328101336956024\r\nStep 13, loss: 0.34072747826576233\r\nStep 14, loss: 0.3415687382221222\r\nStep 15, loss: 0.3421204090118408\r\nStep 16, loss: 0.34998252987861633\r\nStep 17, loss: 0.3370188772678375\r\nStep 18, loss: 0.3368116021156311\r\nStep 19, loss: 0.3397890627384186\r\nSaved checkpoint at step 20\r\nStep 20, loss: 0.33846819400787354\r\nStep 21, loss: 0.328185111284256\r\nStep 22, loss: 0.33603933453559875\r\nStep 23, loss: 0.3364322781562805\r\nStep 24, loss: 0.3406607210636139\r\nStep 25, loss: 0.3376489579677582\r\nStep 26, loss: 0.3372756540775299\r\nStep 27, loss: 0.3242914080619812\r\nStep 162, loss: 0.10980116575956345\r\nStep 163, loss: 0.10331201553344727\r\nStep 164, loss: 0.09606180340051651\r\nStep 165, loss: 0.10288108885288239\r\nStep 166, loss: 0.10788475722074509\r\nStep 167, loss: 0.11626137793064117\r\nStep 168, loss: 0.11271797865629196\r\nStep 169, loss: 0.10126489400863647\r\nSaved checkpoint at step 170\r\nStep 170, loss: 0.10251966863870621\r\nStep 171, loss: 0.09720180183649063\r\nStep 172, loss: 0.10119358450174332\r\nStep 173, loss: 0.10164882242679596\r\nStep 174, loss: 0.09646020829677582\r\nStep 175, loss: 0.09435637295246124\r\nStep 176, loss: 0.09778910130262375\r\nStep 177, loss: 0.09556873142719269\r\nStep 178, loss: 0.10166315734386444\r\nStep 179, loss: 0.09307191520929337\r\nSaved checkpoint at step 180\r\nStep 180, loss: 0.09001213312149048\r\nStep 181, loss: 0.09384429454803467\r\nStep 182, loss: 0.09797736257314682\r\nStep 183, loss: 0.09913866221904755\r\nStep 184, loss: 0.09100943058729172\r\nStep 185, loss: 0.09598495066165924\r\nStep 186, loss: 0.09197870641946793\r\nStep 187, loss: 0.09248428791761398\r\nStep 28, loss: 0.32896149158477783\r\nStep 29, loss: 0.32680416107177734\r\nSaved checkpoint at step 30\r\nStep 30, loss: 0.31527817249298096\r\nStep 31, loss: 0.32453596591949463\r\nStep 32, loss: 0.32291722297668457\r\nStep 33, loss: 0.3086308240890503\r\nStep 34, loss: 0.3201759457588196\r\nStep 35, loss: 0.32361075282096863\r\nStep 36, loss: 0.3231964409351349\r\nStep 37, loss: 0.32022643089294434\r\nStep 38, loss: 0.31821075081825256\r\nStep 39, loss: 0.3172142505645752\r\nSaved checkpoint at step 40\r\nStep 40, loss: 0.32197457551956177\r\nStep 41, loss: 0.3097772002220154\r\nStep 42, loss: 0.3095897436141968\r\nStep 43, loss: 0.30802375078201294\r\nStep 44, loss: 0.3085505962371826\r\nStep 45, loss: 0.30728694796562195\r\nStep 46, loss: 0.30426204204559326\r\nStep 47, loss: 0.31653067469596863\r\nStep 48, loss: 0.29350337386131287\r\nStep 49, loss: 0.2900897264480591\r\nSaved checkpoint at step 50\r\nStep 50, loss: 0.3072112500667572\r\nStep 51, loss: 0.29866600036621094\r\nStep 52, loss: 0.2933558225631714\r\nStep 53, loss: 0.28990650177001953\r\nStep 54, loss: 0.2917138338088989\r\nStep 188, loss: 0.0881965383887291\r\nStep 189, loss: 0.08620244264602661\r\nSaved checkpoint at step 190\r\nStep 190, loss: 0.0936160534620285\r\nStep 191, loss: 0.08297702670097351\r\nStep 192, loss: 0.08549270033836365\r\nStep 193, loss: 0.09282968193292618\r\nStep 194, loss: 0.09096062183380127\r\nStep 195, loss: 0.08280432224273682\r\nStep 196, loss: 0.08810421824455261\r\nStep 197, loss: 0.08211309462785721\r\nStep 198, loss: 0.07822512835264206\r\nStep 199, loss: 0.08037614822387695\r\nSaved checkpoint at step 200\r\nStep 200, loss: 0.08436369150876999\r\nStep 201, loss: 0.0888621136546135\r\nStep 202, loss: 0.08680012822151184\r\nStep 203, loss: 0.08875889331102371\r\nStep 204, loss: 0.08150099217891693\r\nStep 205, loss: 0.08258917927742004\r\nStep 206, loss: 0.07716099172830582\r\nStep 207, loss: 0.07317893952131271\r\nStep 208, loss: 0.07029818743467331\r\nStep 209, loss: 0.07618950307369232\r\nSaved checkpoint at step 210\r\nStep 210, loss: 0.07858575135469437\r\nStep 211, loss: 0.07911297678947449\r\nStep 212, loss: 0.09101250022649765\r\nStep 213, loss: 0.07573471963405609\r\nStep 55, loss: 0.29527750611305237\r\nStep 56, loss: 0.2855745553970337\r\nStep 57, loss: 0.29226770997047424\r\nStep 58, loss: 0.28528568148612976\r\nStep 59, loss: 0.2834138870239258\r\nSaved checkpoint at step 60\r\nStep 60, loss: 0.2725016474723816\r\nStep 61, loss: 0.28022363781929016\r\nStep 62, loss: 0.27179959416389465\r\nStep 63, loss: 0.2702350914478302\r\nStep 64, loss: 0.2684958875179291\r\nStep 65, loss: 0.26719269156455994\r\nStep 66, loss: 0.25736168026924133\r\nStep 67, loss: 0.26021021604537964\r\nStep 68, loss: 0.2600043714046478\r\nStep 69, loss: 0.2554631233215332\r\nSaved checkpoint at step 70\r\nStep 70, loss: 0.25915834307670593\r\nStep 71, loss: 0.2538088262081146\r\nStep 72, loss: 0.2527981102466583\r\nStep 73, loss: 0.24763624370098114\r\nStep 74, loss: 0.23982031643390656\r\nStep 75, loss: 0.2431134581565857\r\nStep 76, loss: 0.24418558180332184\r\nStep 77, loss: 0.23978117108345032\r\nStep 78, loss: 0.23628409206867218\r\nStep 79, loss: 0.22659100592136383\r\nSaved checkpoint at step 80\r\nStep 80, loss: 0.23002658784389496\r\nStep 81, loss: 0.2237771600484848\r\nStep 214, loss: 0.07784590125083923\r\nStep 82, loss: 0.21777285635471344\r\nStep 83, loss: 0.21955353021621704\r\nStep 84, loss: 0.2197633534669876\r\nStep 85, loss: 0.21693949401378632\r\nStep 86, loss: 0.2156858891248703\r\nStep 87, loss: 0.20907285809516907\r\nStep 88, loss: 0.20304414629936218\r\nStep 89, loss: 0.20950746536254883\r\nSaved checkpoint at step 90\r\nStep 90, loss: 0.20157180726528168\r\nStep 91, loss: 0.19864089787006378\r\nStep 92, loss: 0.1961870640516281\r\nStep 93, loss: 0.1970020979642868\r\nStep 94, loss: 0.19751623272895813\r\nStep 95, loss: 0.1926620602607727\r\nStep 96, loss: 0.1907958984375\r\nStep 97, loss: 0.18797270953655243\r\nStep 98, loss: 0.18371835350990295\r\nStep 99, loss: 0.19347549974918365\r\nSaved checkpoint at step 100\r\nStep 100, loss: 0.17737938463687897\r\nStep 101, loss: 0.18215568363666534\r\nStep 102, loss: 0.17679424583911896\r\nStep 103, loss: 0.17569106817245483\r\nStep 104, loss: 0.17612037062644958\r\nStep 105, loss: 0.1708865910768509\r\nStep 106, loss: 0.17582806944847107\r\nStep 107, loss: 0.16020779311656952\r\nStep 108, loss: 0.16374938189983368\r\nStep 0, loss: 0.33292484283447266\r\nStep 1, loss: 0.33973613381385803\r\nStep 2, loss: 0.3459185063838959\r\nStep 3, loss: 0.33234596252441406\r\nStep 4, loss: 0.3411635756492615\r\nStep 5, loss: 0.3499658405780792\r\nStep 6, loss: 0.3499256670475006\r\nStep 7, loss: 0.34394073486328125\r\nStep 8, loss: 0.3490101397037506\r\nStep 9, loss: 0.3501622974872589\r\nSaved checkpoint at step 10\r\nStep 10, loss: 0.3454197645187378\r\nStep 11, loss: 0.3307967782020569\r\nStep 12, loss: 0.3328101336956024\r\nStep 13, loss: 0.34072747826576233\r\nStep 14, loss: 0.3415687382221222\r\nStep 15, loss: 0.3421204090118408\r\nStep 16, loss: 0.34998252987861633\r\nStep 17, loss: 0.3370188772678375\r\nStep 18, loss: 0.3368116021156311\r\nStep 19, loss: 0.3397890627384186\r\nSaved checkpoint at step 20\r\nStep 20, loss: 0.33846819400787354\r\nStep 21, loss: 0.328185111284256\r\nStep 22, loss: 0.33603933453559875\r\nStep 23, loss: 0.3364322781562805\r\nStep 24, loss: 0.3406607210636139\r\nStep 25, loss: 0.3376489579677582\r\nStep 26, loss: 0.3372756540775299\r\nStep 27, loss: 0.3242914080619812\r\nStep 109, loss: 0.16316014528274536\r\nStep 28, loss: 0.32896149158477783\r\nStep 29, loss: 0.32680416107177734\r\nSaved checkpoint at step 30\r\nStep 30, loss: 0.31527817249298096\r\nStep 31, loss: 0.32453596591949463\r\nStep 32, loss: 0.32291722297668457\r\nStep 33, loss: 0.3086308240890503\r\nStep 34, loss: 0.3201759457588196\r\nStep 35, loss: 0.32361075282096863\r\nStep 36, loss: 0.3231964409351349\r\nStep 37, loss: 0.32022643089294434\r\nStep 38, loss: 0.31821075081825256\r\nStep 39, loss: 0.3172142505645752\r\nSaved checkpoint at step 40\r\nStep 40, loss: 0.32197457551956177\r\nStep 41, loss: 0.3097772002220154\r\nStep 42, loss: 0.3095897436141968\r\nStep 43, loss: 0.30802375078201294\r\nStep 44, loss: 0.3085505962371826\r\nStep 45, loss: 0.30728694796562195\r\nStep 46, loss: 0.30426204204559326\r\nStep 47, loss: 0.31653067469596863\r\nStep 48, loss: 0.29350337386131287\r\nStep 49, loss: 0.2900897264480591\r\nSaved checkpoint at step 50\r\nStep 50, loss: 0.3072112500667572\r\nStep 51, loss: 0.29866600036621094\r\nStep 52, loss: 0.2933558225631714\r\nStep 53, loss: 0.28990650177001953\r\nStep 54, loss: 0.2917138338088989\r\nStep 0, loss: 0.33292484283447266\r\nStep 1, loss: 0.33973613381385803\r\nStep 2, loss: 0.3459185063838959\r\nStep 3, loss: 0.33234596252441406\r\nStep 4, loss: 0.3411635756492615\r\nStep 5, loss: 0.3499658405780792\r\nStep 6, loss: 0.3499256670475006\r\nStep 7, loss: 0.34394073486328125\r\nStep 8, loss: 0.3490101397037506\r\nStep 9, loss: 0.3501622974872589\r\nSaved checkpoint at step 10\r\nStep 10, loss: 0.3454197645187378\r\nStep 11, loss: 0.3307967782020569\r\nStep 12, loss: 0.3328101336956024\r\nStep 13, loss: 0.34072747826576233\r\nStep 14, loss: 0.3415687382221222\r\nStep 15, loss: 0.3421204090118408\r\nStep 16, loss: 0.34998252987861633\r\nStep 17, loss: 0.3370188772678375\r\nStep 18, loss: 0.3368116021156311\r\nStep 19, loss: 0.3397890627384186\r\nSaved checkpoint at step 20\r\nStep 20, loss: 0.33846819400787354\r\nStep 21, loss: 0.328185111284256\r\nStep 22, loss: 0.33603933453559875\r\nStep 23, loss: 0.3364322781562805\r\nStep 24, loss: 0.3406607210636139\r\nStep 25, loss: 0.3376489579677582\r\nStep 26, loss: 0.3372756540775299\r\nStep 27, loss: 0.3242914080619812\r\nStep 55, loss: 0.29527750611305237\r\nStep 56, loss: 0.2855745553970337\r\nStep 57, loss: 0.29226770997047424\r\nStep 58, loss: 0.28528568148612976\r\nStep 59, loss: 0.2834138870239258\r\nSaved checkpoint at step 60\r\nStep 60, loss: 0.2725016474723816\r\nStep 61, loss: 0.28022363781929016\r\nStep 62, loss: 0.27179959416389465\r\nStep 63, loss: 0.2702350914478302\r\nStep 64, loss: 0.2684958875179291\r\nStep 65, loss: 0.26719269156455994\r\nStep 66, loss: 0.25736168026924133\r\nStep 67, loss: 0.26021021604537964\r\nStep 68, loss: 0.2600043714046478\r\nStep 69, loss: 0.2554631233215332\r\nSaved checkpoint at step 70\r\nStep 70, loss: 0.25915834307670593\r\nStep 71, loss: 0.2538088262081146\r\nStep 72, loss: 0.2527981102466583\r\nStep 73, loss: 0.24763624370098114\r\nStep 74, loss: 0.23982031643390656\r\nStep 75, loss: 0.2431134581565857\r\nStep 76, loss: 0.24418558180332184\r\nStep 77, loss: 0.23978117108345032\r\nStep 78, loss: 0.23628409206867218\r\nStep 79, loss: 0.22659100592136383\r\nSaved checkpoint at step 80\r\nStep 80, loss: 0.23002658784389496\r\nStep 81, loss: 0.2237771600484848\r\nStep 28, loss: 0.32896149158477783\r\nStep 29, loss: 0.32680416107177734\r\nSaved checkpoint at step 30\r\nStep 30, loss: 0.31527817249298096\r\nStep 31, loss: 0.32453596591949463\r\nStep 32, loss: 0.32291722297668457\r\nStep 33, loss: 0.3086308240890503\r\nStep 34, loss: 0.3201759457588196\r\nStep 35, loss: 0.32361075282096863\r\nStep 36, loss: 0.3231964409351349\r\nStep 37, loss: 0.32022643089294434\r\nStep 38, loss: 0.31821075081825256\r\nStep 39, loss: 0.3172142505645752\r\nSaved checkpoint at step 40\r\nStep 40, loss: 0.32197457551956177\r\nStep 41, loss: 0.3097772002220154\r\nStep 42, loss: 0.3095897436141968\r\nStep 43, loss: 0.30802375078201294\r\nStep 44, loss: 0.3085505962371826\r\nStep 45, loss: 0.30728694796562195\r\nStep 46, loss: 0.30426204204559326\r\nStep 47, loss: 0.31653067469596863\r\nStep 48, loss: 0.29350337386131287\r\nStep 49, loss: 0.2900897264480591\r\nSaved checkpoint at step 50\r\nStep 50, loss: 0.3072112500667572\r\nStep 51, loss: 0.29866600036621094\r\nStep 52, loss: 0.2933558225631714\r\nStep 53, loss: 0.28990650177001953\r\nStep 54, loss: 0.2917138338088989\r\nStep 82, loss: 0.21777285635471344\r\nStep 83, loss: 0.21955353021621704\r\nStep 84, loss: 0.2197633534669876\r\nStep 85, loss: 0.21693949401378632\r\nStep 86, loss: 0.2156858891248703\r\nStep 87, loss: 0.20907285809516907\r\nStep 88, loss: 0.20304414629936218\r\nStep 89, loss: 0.20950746536254883\r\nSaved checkpoint at step 90\r\nStep 90, loss: 0.20157180726528168\r\nStep 91, loss: 0.19864089787006378\r\nStep 92, loss: 0.1961870640516281\r\nStep 93, loss: 0.1970020979642868\r\nStep 94, loss: 0.19751623272895813\r\nStep 95, loss: 0.1926620602607727\r\nStep 96, loss: 0.1907958984375\r\nStep 97, loss: 0.18797270953655243\r\nStep 98, loss: 0.18371835350990295\r\nStep 99, loss: 0.19347549974918365\r\nSaved checkpoint at step 100\r\nStep 100, loss: 0.17737938463687897\r\nStep 101, loss: 0.18215568363666534\r\nStep 102, loss: 0.17679424583911896\r\nStep 103, loss: 0.17569106817245483\r\nStep 104, loss: 0.17612037062644958\r\nStep 105, loss: 0.1708865910768509\r\nStep 106, loss: 0.17582806944847107\r\nStep 107, loss: 0.16020779311656952\r\nStep 108, loss: 0.16374938189983368\r\nStep 55, loss: 0.29527750611305237\r\nStep 56, loss: 0.2855745553970337\r\nStep 57, loss: 0.29226770997047424\r\nStep 58, loss: 0.28528568148612976\r\nStep 59, loss: 0.2834138870239258\r\nSaved checkpoint at step 60\r\nStep 60, loss: 0.2725016474723816\r\nStep 61, loss: 0.28022363781929016\r\nStep 62, loss: 0.27179959416389465\r\nStep 63, loss: 0.2702350914478302\r\nStep 64, loss: 0.2684958875179291\r\nStep 65, loss: 0.26719269156455994\r\nStep 66, loss: 0.25736168026924133\r\nStep 67, loss: 0.26021021604537964\r\nStep 68, loss: 0.2600043714046478\r\nStep 69, loss: 0.2554631233215332\r\nSaved checkpoint at step 70\r\nStep 70, loss: 0.25915834307670593\r\nStep 71, loss: 0.2538088262081146\r\nStep 72, loss: 0.2527981102466583\r\nStep 73, loss: 0.24763624370098114\r\nStep 74, loss: 0.23982031643390656\r\nStep 75, loss: 0.2431134581565857\r\nStep 76, loss: 0.24418558180332184\r\nStep 77, loss: 0.23978117108345032\r\nStep 78, loss: 0.23628409206867218\r\nStep 79, loss: 0.22659100592136383\r\nSaved checkpoint at step 80\r\nStep 80, loss: 0.23002658784389496\r\nStep 81, loss: 0.2237771600484848\r\nStep 109, loss: 0.16316014528274536\r\nStep 82, loss: 0.21777285635471344\r\nStep 83, loss: 0.21955353021621704\r\nStep 84, loss: 0.2197633534669876\r\nStep 85, loss: 0.21693949401378632\r\nStep 86, loss: 0.2156858891248703\r\nStep 87, loss: 0.20907285809516907\r\nStep 88, loss: 0.20304414629936218\r\nStep 89, loss: 0.20950746536254883\r\nSaved checkpoint at step 90\r\nStep 90, loss: 0.20157180726528168\r\nStep 91, loss: 0.19864089787006378\r\nStep 92, loss: 0.1961870640516281\r\nStep 93, loss: 0.1970020979642868\r\nStep 94, loss: 0.19751623272895813\r\nStep 95, loss: 0.1926620602607727\r\nStep 96, loss: 0.1907958984375\r\nStep 97, loss: 0.18797270953655243\r\nStep 98, loss: 0.18371835350990295\r\nStep 99, loss: 0.19347549974918365\r\nSaved checkpoint at step 100\r\nStep 100, loss: 0.17737938463687897\r\nStep 101, loss: 0.18215568363666534\r\nStep 102, loss: 0.17679424583911896\r\nStep 103, loss: 0.17569106817245483\r\nStep 104, loss: 0.17612037062644958\r\nStep 105, loss: 0.1708865910768509\r\nStep 106, loss: 0.17582806944847107\r\nStep 107, loss: 0.16020779311656952\r\nStep 108, loss: 0.16374938189983368\r\nSaved checkpoint at step 110\r\nStep 110, loss: 0.15965776145458221\r\nStep 111, loss: 0.16084066033363342\r\nStep 112, loss: 0.15732374787330627\r\nStep 113, loss: 0.1501660794019699\r\nStep 114, loss: 0.15066221356391907\r\nStep 115, loss: 0.15051837265491486\r\nStep 116, loss: 0.1540840119123459\r\nStep 117, loss: 0.15616631507873535\r\nStep 118, loss: 0.14677441120147705\r\nStep 119, loss: 0.15072251856327057\r\nSaved checkpoint at step 120\r\nStep 120, loss: 0.14081387221813202\r\nStep 121, loss: 0.1404900848865509\r\nStep 122, loss: 0.13735362887382507\r\nStep 123, loss: 0.14047737419605255\r\nStep 124, loss: 0.14097315073013306\r\nStep 125, loss: 0.13012643158435822\r\nStep 126, loss: 0.13319577276706696\r\nStep 127, loss: 0.12502308189868927\r\nStep 128, loss: 0.13217462599277496\r\nStep 129, loss: 0.12322020530700684\r\nSaved checkpoint at step 130\r\nStep 130, loss: 0.1309303492307663\r\nStep 131, loss: 0.12471090257167816\r\nStep 132, loss: 0.1259678304195404\r\nStep 133, loss: 0.11994366347789764\r\nStep 134, loss: 0.12396806478500366\r\nStep 135, loss: 0.12951606512069702\r\nStep 109, loss: 0.16316014528274536\r\nStep 136, loss: 0.12501047551631927\r\nStep 137, loss: 0.12460850179195404\r\nStep 138, loss: 0.12312853336334229\r\nStep 139, loss: 0.11870359629392624\r\nSaved checkpoint at step 140\r\nStep 140, loss: 0.10972652584314346\r\nStep 141, loss: 0.11810508370399475\r\nStep 142, loss: 0.11488064378499985\r\nStep 143, loss: 0.10100933909416199\r\nStep 144, loss: 0.11190460622310638\r\nStep 145, loss: 0.10655320435762405\r\nStep 146, loss: 0.11590567231178284\r\nStep 147, loss: 0.11223439872264862\r\nStep 148, loss: 0.11101178079843521\r\nStep 149, loss: 0.11008752137422562\r\nSaved checkpoint at step 150\r\nStep 150, loss: 0.1229330375790596\r\nStep 151, loss: 0.12016700953245163\r\nStep 152, loss: 0.11195394396781921\r\nStep 153, loss: 0.11655457317829132\r\nStep 154, loss: 0.10912562161684036\r\nStep 155, loss: 0.12283866107463837\r\nStep 156, loss: 0.10622447729110718\r\nStep 157, loss: 0.10579507797956467\r\nStep 158, loss: 0.09965544939041138\r\nStep 159, loss: 0.10587912797927856\r\nSaved checkpoint at step 160\r\nStep 160, loss: 0.11324921250343323\r\nStep 161, loss: 0.11093202978372574\r\nSaved checkpoint at step 110\r\nStep 110, loss: 0.15965776145458221\r\nStep 111, loss: 0.16084066033363342\r\nStep 112, loss: 0.15732374787330627\r\nStep 113, loss: 0.1501660794019699\r\nStep 114, loss: 0.15066221356391907\r\nStep 115, loss: 0.15051837265491486\r\nStep 116, loss: 0.1540840119123459\r\nStep 117, loss: 0.15616631507873535\r\nStep 118, loss: 0.14677441120147705\r\nStep 119, loss: 0.15072251856327057\r\nSaved checkpoint at step 120\r\nStep 120, loss: 0.14081387221813202\r\nStep 121, loss: 0.1404900848865509\r\nStep 122, loss: 0.13735362887382507\r\nStep 123, loss: 0.14047737419605255\r\nStep 124, loss: 0.14097315073013306\r\nStep 125, loss: 0.13012643158435822\r\nStep 126, loss: 0.13319577276706696\r\nStep 127, loss: 0.12502308189868927\r\nStep 128, loss: 0.13217462599277496\r\nStep 129, loss: 0.12322020530700684\r\nSaved checkpoint at step 130\r\nStep 130, loss: 0.1309303492307663\r\nStep 131, loss: 0.12471090257167816\r\nStep 132, loss: 0.1259678304195404\r\nStep 133, loss: 0.11994366347789764\r\nStep 134, loss: 0.12396806478500366\r\nStep 135, loss: 0.12951606512069702\r\nStep 162, loss: 0.10980116575956345\r\nStep 163, loss: 0.10331201553344727\r\nStep 164, loss: 0.09606180340051651\r\nStep 165, loss: 0.10288108885288239\r\nStep 166, loss: 0.10788475722074509\r\nStep 167, loss: 0.11626137793064117\r\nStep 168, loss: 0.11271797865629196\r\nStep 169, loss: 0.10126489400863647\r\nSaved checkpoint at step 170\r\nStep 170, loss: 0.10251966863870621\r\nStep 171, loss: 0.09720180183649063\r\nStep 172, loss: 0.10119358450174332\r\nStep 173, loss: 0.10164882242679596\r\nStep 174, loss: 0.09646020829677582\r\nStep 175, loss: 0.09435637295246124\r\nStep 176, loss: 0.09778910130262375\r\nStep 177, loss: 0.09556873142719269\r\nStep 178, loss: 0.10166315734386444\r\nStep 179, loss: 0.09307191520929337\r\nSaved checkpoint at step 180\r\nStep 180, loss: 0.09001213312149048\r\nStep 181, loss: 0.09384429454803467\r\nStep 182, loss: 0.09797736257314682\r\nStep 183, loss: 0.09913866221904755\r\nStep 184, loss: 0.09100943058729172\r\nStep 185, loss: 0.09598495066165924\r\nStep 186, loss: 0.09197870641946793\r\nStep 187, loss: 0.09248428791761398\r\nStep 136, loss: 0.12501047551631927\r\nStep 137, loss: 0.12460850179195404\r\nStep 138, loss: 0.12312853336334229\r\nStep 139, loss: 0.11870359629392624\r\nSaved checkpoint at step 140\r\nStep 140, loss: 0.10972652584314346\r\nStep 141, loss: 0.11810508370399475\r\nStep 142, loss: 0.11488064378499985\r\nStep 143, loss: 0.10100933909416199\r\nStep 144, loss: 0.11190460622310638\r\nStep 145, loss: 0.10655320435762405\r\nStep 146, loss: 0.11590567231178284\r\nStep 147, loss: 0.11223439872264862\r\nStep 148, loss: 0.11101178079843521\r\nStep 149, loss: 0.11008752137422562\r\nSaved checkpoint at step 150\r\nStep 150, loss: 0.1229330375790596\r\nStep 151, loss: 0.12016700953245163\r\nStep 152, loss: 0.11195394396781921\r\nStep 153, loss: 0.11655457317829132\r\nStep 154, loss: 0.10912562161684036\r\nStep 155, loss: 0.12283866107463837\r\nStep 156, loss: 0.10622447729110718\r\nStep 157, loss: 0.10579507797956467\r\nStep 158, loss: 0.09965544939041138\r\nStep 159, loss: 0.10587912797927856\r\nSaved checkpoint at step 160\r\nStep 160, loss: 0.11324921250343323\r\nStep 161, loss: 0.11093202978372574\r\nStep 188, loss: 0.0881965383887291\r\nStep 189, loss: 0.08620244264602661\r\nSaved checkpoint at step 190\r\nStep 190, loss: 0.0936160534620285\r\nStep 191, loss: 0.08297702670097351\r\nStep 192, loss: 0.08549270033836365\r\nStep 193, loss: 0.09282968193292618\r\nStep 194, loss: 0.09096062183380127\r\nStep 195, loss: 0.08280432224273682\r\nStep 196, loss: 0.08810421824455261\r\nStep 197, loss: 0.08211309462785721\r\nStep 198, loss: 0.07822512835264206\r\nStep 199, loss: 0.08037614822387695\r\nSaved checkpoint at step 200\r\nStep 200, loss: 0.08436369150876999\r\nStep 201, loss: 0.0888621136546135\r\nStep 202, loss: 0.08680012822151184\r\nStep 203, loss: 0.08875889331102371\r\nStep 204, loss: 0.08150099217891693\r\nStep 205, loss: 0.08258917927742004\r\nStep 206, loss: 0.07716099172830582\r\nStep 207, loss: 0.07317893952131271\r\nStep 208, loss: 0.07029818743467331\r\nStep 209, loss: 0.07618950307369232\r\nSaved checkpoint at step 210\r\nStep 210, loss: 0.07858575135469437\r\nStep 211, loss: 0.07911297678947449\r\nStep 212, loss: 0.09101250022649765\r\nStep 213, loss: 0.07573471963405609\r\nStep 162, loss: 0.10980116575956345\r\nStep 163, loss: 0.10331201553344727\r\nStep 164, loss: 0.09606180340051651\r\nStep 165, loss: 0.10288108885288239\r\nStep 166, loss: 0.10788475722074509\r\nStep 167, loss: 0.11626137793064117\r\nStep 168, loss: 0.11271797865629196\r\nStep 169, loss: 0.10126489400863647\r\nSaved checkpoint at step 170\r\nStep 170, loss: 0.10251966863870621\r\nStep 171, loss: 0.09720180183649063\r\nStep 172, loss: 0.10119358450174332\r\nStep 173, loss: 0.10164882242679596\r\nStep 174, loss: 0.09646020829677582\r\nStep 175, loss: 0.09435637295246124\r\nStep 176, loss: 0.09778910130262375\r\nStep 177, loss: 0.09556873142719269\r\nStep 178, loss: 0.10166315734386444\r\nStep 179, loss: 0.09307191520929337\r\nSaved checkpoint at step 180\r\nStep 180, loss: 0.09001213312149048\r\nStep 181, loss: 0.09384429454803467\r\nStep 182, loss: 0.09797736257314682\r\nStep 183, loss: 0.09913866221904755\r\nStep 184, loss: 0.09100943058729172\r\nStep 185, loss: 0.09598495066165924\r\nStep 186, loss: 0.09197870641946793\r\nStep 187, loss: 0.09248428791761398\r\nStep 214, loss: 0.07784590125083923\r\nStep 188, loss: 0.0881965383887291\r\nStep 189, loss: 0.08620244264602661\r\nSaved checkpoint at step 190\r\nStep 190, loss: 0.0936160534620285\r\nStep 191, loss: 0.08297702670097351\r\nStep 192, loss: 0.08549270033836365\r\nStep 193, loss: 0.09282968193292618\r\nStep 194, loss: 0.09096062183380127\r\nStep 195, loss: 0.08280432224273682\r\nStep 196, loss: 0.08810421824455261\r\nStep 197, loss: 0.08211309462785721\r\nStep 198, loss: 0.07822512835264206\r\nStep 199, loss: 0.08037614822387695\r\nSaved checkpoint at step 200\r\nStep 200, loss: 0.08436369150876999\r\nStep 201, loss: 0.0888621136546135\r\nStep 202, loss: 0.08680012822151184\r\nStep 203, loss: 0.08875889331102371\r\nStep 204, loss: 0.08150099217891693\r\nStep 205, loss: 0.08258917927742004\r\nStep 206, loss: 0.07716099172830582\r\nStep 207, loss: 0.07317893952131271\r\nStep 208, loss: 0.07029818743467331\r\nStep 209, loss: 0.07618950307369232\r\nSaved checkpoint at step 210\r\nStep 210, loss: 0.07858575135469437\r\nStep 211, loss: 0.07911297678947449\r\nStep 212, loss: 0.09101250022649765\r\nStep 213, loss: 0.07573471963405609\r\nStep 214, loss: 0.07784590125083923\r\nSaved checkpoint at step 110\r\nStep 110, loss: 0.15965776145458221\r\nStep 111, loss: 0.16084066033363342\r\nStep 112, loss: 0.15732374787330627\r\nStep 113, loss: 0.1501660794019699\r\nStep 114, loss: 0.15066221356391907\r\nStep 115, loss: 0.15051837265491486\r\nStep 116, loss: 0.1540840119123459\r\nStep 117, loss: 0.15616631507873535\r\nStep 118, loss: 0.14677441120147705\r\nStep 119, loss: 0.15072251856327057\r\nSaved checkpoint at step 120\r\nStep 120, loss: 0.14081387221813202\r\nStep 121, loss: 0.1404900848865509\r\nStep 122, loss: 0.13735362887382507\r\nStep 123, loss: 0.14047737419605255\r\nStep 124, loss: 0.14097315073013306\r\nStep 125, loss: 0.13012643158435822\r\nStep 126, loss: 0.13319577276706696\r\nStep 127, loss: 0.12502308189868927\r\nStep 128, loss: 0.13217462599277496\r\nStep 129, loss: 0.12322020530700684\r\nSaved checkpoint at step 130\r\nStep 130, loss: 0.1309303492307663\r\nStep 131, loss: 0.12471090257167816\r\nStep 132, loss: 0.1259678304195404\r\nStep 133, loss: 0.11994366347789764\r\nStep 134, loss: 0.12396806478500366\r\nStep 135, loss: 0.12951606512069702\r\nStep 136, loss: 0.12501047551631927\r\nStep 137, loss: 0.12460850179195404\r\nStep 138, loss: 0.12312853336334229\r\nStep 139, loss: 0.11870359629392624\r\nSaved checkpoint at step 140\r\nStep 140, loss: 0.10972652584314346\r\nStep 141, loss: 0.11810508370399475\r\nStep 142, loss: 0.11488064378499985\r\nStep 143, loss: 0.10100933909416199\r\nStep 144, loss: 0.11190460622310638\r\nStep 145, loss: 0.10655320435762405\r\nStep 146, loss: 0.11590567231178284\r\nStep 147, loss: 0.11223439872264862\r\nStep 148, loss: 0.11101178079843521\r\nStep 149, loss: 0.11008752137422562\r\nSaved checkpoint at step 150\r\nStep 150, loss: 0.1229330375790596\r\nStep 151, loss: 0.12016700953245163\r\nStep 152, loss: 0.11195394396781921\r\nStep 153, loss: 0.11655457317829132\r\nStep 154, loss: 0.10912562161684036\r\nStep 155, loss: 0.12283866107463837\r\nStep 156, loss: 0.10622447729110718\r\nStep 157, loss: 0.10579507797956467\r\nStep 158, loss: 0.09965544939041138\r\nStep 159, loss: 0.10587912797927856\r\nSaved checkpoint at step 160\r\nStep 160, loss: 0.11324921250343323\r\nStep 161, loss: 0.11093202978372574\r\nStep 162, loss: 0.10980116575956345\r\nStep 163, loss: 0.10331201553344727\r\nStep 164, loss: 0.09606180340051651\r\nStep 165, loss: 0.10288108885288239\r\nStep 166, loss: 0.10788475722074509\r\nStep 167, loss: 0.11626137793064117\r\nStep 168, loss: 0.11271797865629196\r\nStep 169, loss: 0.10126489400863647\r\nSaved checkpoint at step 170\r\nStep 170, loss: 0.10251966863870621\r\nStep 171, loss: 0.09720180183649063\r\nStep 172, loss: 0.10119358450174332\r\nStep 173, loss: 0.10164882242679596\r\nStep 174, loss: 0.09646020829677582\r\nStep 175, loss: 0.09435637295246124\r\nStep 176, loss: 0.09778910130262375\r\nStep 177, loss: 0.09556873142719269\r\nStep 178, loss: 0.10166315734386444\r\nStep 179, loss: 0.09307191520929337\r\nSaved checkpoint at step 180\r\nStep 180, loss: 0.09001213312149048\r\nStep 181, loss: 0.09384429454803467\r\nStep 182, loss: 0.09797736257314682\r\nStep 183, loss: 0.09913866221904755\r\nStep 184, loss: 0.09100943058729172\r\nStep 185, loss: 0.09598495066165924\r\nStep 186, loss: 0.09197870641946793\r\nStep 187, loss: 0.09248428791761398\r\nStep 188, loss: 0.0881965383887291\r\nStep 189, loss: 0.08620244264602661\r\nSaved checkpoint at step 190\r\nStep 190, loss: 0.0936160534620285\r\nStep 191, loss: 0.08297702670097351\r\nStep 192, loss: 0.08549270033836365\r\nStep 193, loss: 0.09282968193292618\r\nStep 194, loss: 0.09096062183380127\r\nStep 195, loss: 0.08280432224273682\r\nStep 196, loss: 0.08810421824455261\r\nStep 197, loss: 0.08211309462785721\r\nStep 198, loss: 0.07822512835264206\r\nStep 199, loss: 0.08037614822387695\r\nSaved checkpoint at step 200\r\nStep 200, loss: 0.08436369150876999\r\nStep 201, loss: 0.0888621136546135\r\nStep 202, loss: 0.08680012822151184\r\nStep 203, loss: 0.08875889331102371\r\nStep 204, loss: 0.08150099217891693\r\nStep 205, loss: 0.08258917927742004\r\nStep 206, loss: 0.07716099172830582\r\nStep 207, loss: 0.07317893952131271\r\nStep 208, loss: 0.07029818743467331\r\nStep 209, loss: 0.07618950307369232\r\nSaved checkpoint at step 210\r\nStep 210, loss: 0.07858575135469437\r\nStep 211, loss: 0.07911297678947449\r\nStep 212, loss: 0.09101250022649765\r\nStep 213, loss: 0.07573471963405609\r\nStep 0, loss: 0.33292484283447266\r\nStep 1, loss: 0.33973613381385803\r\nStep 2, loss: 0.3459185063838959\r\nStep 3, loss: 0.33234596252441406\r\nStep 4, loss: 0.3411635756492615\r\nStep 5, loss: 0.3499658405780792\r\nStep 6, loss: 0.3499256670475006\r\nStep 7, loss: 0.34394073486328125\r\nStep 8, loss: 0.3490101397037506\r\nStep 9, loss: 0.3501622974872589\r\nSaved checkpoint at step 10\r\nStep 10, loss: 0.3454197645187378\r\nStep 11, loss: 0.3307967782020569\r\nStep 12, loss: 0.3328101336956024\r\nStep 13, loss: 0.34072747826576233\r\nStep 14, loss: 0.3415687382221222\r\nStep 15, loss: 0.3421204090118408\r\nStep 16, loss: 0.34998252987861633\r\nStep 17, loss: 0.3370188772678375\r\nStep 18, loss: 0.3368116021156311\r\nStep 19, loss: 0.3397890627384186\r\nSaved checkpoint at step 20\r\nStep 20, loss: 0.33846819400787354\r\nStep 21, loss: 0.328185111284256\r\nStep 22, loss: 0.33603933453559875\r\nStep 23, loss: 0.3364322781562805\r\nStep 24, loss: 0.3406607210636139\r\nStep 25, loss: 0.3376489579677582\r\nStep 26, loss: 0.3372756540775299\r\nStep 27, loss: 0.3242914080619812\r\nStep 214, loss: 0.07784590125083923\r\nStep 0, loss: 0.33292484283447266\r\nStep 1, loss: 0.33973613381385803\r\nStep 2, loss: 0.3459185063838959\r\nStep 3, loss: 0.33234596252441406\r\nStep 4, loss: 0.3411635756492615\r\nStep 5, loss: 0.3499658405780792\r\nStep 6, loss: 0.3499256670475006\r\nStep 7, loss: 0.34394073486328125\r\nStep 8, loss: 0.3490101397037506\r\nStep 9, loss: 0.3501622974872589\r\nSaved checkpoint at step 10\r\nStep 10, loss: 0.3454197645187378\r\nStep 11, loss: 0.3307967782020569\r\nStep 12, loss: 0.3328101336956024\r\nStep 13, loss: 0.34072747826576233\r\nStep 14, loss: 0.3415687382221222\r\nStep 15, loss: 0.3421204090118408\r\nStep 16, loss: 0.34998252987861633\r\nStep 17, loss: 0.3370188772678375\r\nStep 18, loss: 0.3368116021156311\r\nStep 19, loss: 0.3397890627384186\r\nSaved checkpoint at step 20\r\nStep 20, loss: 0.33846819400787354\r\nStep 21, loss: 0.328185111284256\r\nStep 22, loss: 0.33603933453559875\r\nStep 23, loss: 0.3364322781562805\r\nStep 24, loss: 0.3406607210636139\r\nStep 25, loss: 0.3376489579677582\r\nStep 26, loss: 0.3372756540775299\r\nStep 27, loss: 0.3242914080619812\r\nStep 28, loss: 0.32896149158477783\r\nStep 29, loss: 0.32680416107177734\r\nSaved checkpoint at step 30\r\nStep 30, loss: 0.31527817249298096\r\nStep 31, loss: 0.32453596591949463\r\nStep 32, loss: 0.32291722297668457\r\nStep 33, loss: 0.3086308240890503\r\nStep 34, loss: 0.3201759457588196\r\nStep 35, loss: 0.32361075282096863\r\nStep 36, loss: 0.3231964409351349\r\nStep 37, loss: 0.32022643089294434\r\nStep 38, loss: 0.31821075081825256\r\nStep 39, loss: 0.3172142505645752\r\nSaved checkpoint at step 40\r\nStep 40, loss: 0.32197457551956177\r\nStep 41, loss: 0.3097772002220154\r\nStep 42, loss: 0.3095897436141968\r\nStep 43, loss: 0.30802375078201294\r\nStep 44, loss: 0.3085505962371826\r\nStep 45, loss: 0.30728694796562195\r\nStep 46, loss: 0.30426204204559326\r\nStep 47, loss: 0.31653067469596863\r\nStep 48, loss: 0.29350337386131287\r\nStep 49, loss: 0.2900897264480591\r\nSaved checkpoint at step 50\r\nStep 50, loss: 0.3072112500667572\r\nStep 51, loss: 0.29866600036621094\r\nStep 52, loss: 0.2933558225631714\r\nStep 53, loss: 0.28990650177001953\r\nStep 54, loss: 0.2917138338088989\r\nStep 28, loss: 0.32896149158477783\r\nStep 29, loss: 0.32680416107177734\r\nSaved checkpoint at step 30\r\nStep 30, loss: 0.31527817249298096\r\nStep 31, loss: 0.32453596591949463\r\nStep 32, loss: 0.32291722297668457\r\nStep 33, loss: 0.3086308240890503\r\nStep 34, loss: 0.3201759457588196\r\nStep 35, loss: 0.32361075282096863\r\nStep 36, loss: 0.3231964409351349\r\nStep 37, loss: 0.32022643089294434\r\nStep 38, loss: 0.31821075081825256\r\nStep 39, loss: 0.3172142505645752\r\nSaved checkpoint at step 40\r\nStep 40, loss: 0.32197457551956177\r\nStep 41, loss: 0.3097772002220154\r\nStep 42, loss: 0.3095897436141968\r\nStep 43, loss: 0.30802375078201294\r\nStep 44, loss: 0.3085505962371826\r\nStep 45, loss: 0.30728694796562195\r\nStep 46, loss: 0.30426204204559326\r\nStep 47, loss: 0.31653067469596863\r\nStep 48, loss: 0.29350337386131287\r\nStep 49, loss: 0.2900897264480591\r\nSaved checkpoint at step 50\r\nStep 50, loss: 0.3072112500667572\r\nStep 51, loss: 0.29866600036621094\r\nStep 52, loss: 0.2933558225631714\r\nStep 53, loss: 0.28990650177001953\r\nStep 54, loss: 0.2917138338088989\r\nStep 55, loss: 0.29527750611305237\r\nStep 56, loss: 0.2855745553970337\r\nStep 57, loss: 0.29226770997047424\r\nStep 58, loss: 0.28528568148612976\r\nStep 59, loss: 0.2834138870239258\r\nSaved checkpoint at step 60\r\nStep 60, loss: 0.2725016474723816\r\nStep 61, loss: 0.28022363781929016\r\nStep 62, loss: 0.27179959416389465\r\nStep 63, loss: 0.2702350914478302\r\nStep 64, loss: 0.2684958875179291\r\nStep 65, loss: 0.26719269156455994\r\nStep 66, loss: 0.25736168026924133\r\nStep 67, loss: 0.26021021604537964\r\nStep 68, loss: 0.2600043714046478\r\nStep 69, loss: 0.2554631233215332\r\nSaved checkpoint at step 70\r\nStep 70, loss: 0.25915834307670593\r\nStep 71, loss: 0.2538088262081146\r\nStep 72, loss: 0.2527981102466583\r\nStep 73, loss: 0.24763624370098114\r\nStep 74, loss: 0.23982031643390656\r\nStep 75, loss: 0.2431134581565857\r\nStep 76, loss: 0.24418558180332184\r\nStep 77, loss: 0.23978117108345032\r\nStep 78, loss: 0.23628409206867218\r\nStep 79, loss: 0.22659100592136383\r\nSaved checkpoint at step 80\r\nStep 80, loss: 0.23002658784389496\r\nStep 81, loss: 0.2237771600484848\r\nStep 55, loss: 0.29527750611305237\r\nStep 56, loss: 0.2855745553970337\r\nStep 57, loss: 0.29226770997047424\r\nStep 58, loss: 0.28528568148612976\r\nStep 59, loss: 0.2834138870239258\r\nSaved checkpoint at step 60\r\nStep 60, loss: 0.2725016474723816\r\nStep 61, loss: 0.28022363781929016\r\nStep 62, loss: 0.27179959416389465\r\nStep 63, loss: 0.2702350914478302\r\nStep 64, loss: 0.2684958875179291\r\nStep 65, loss: 0.26719269156455994\r\nStep 66, loss: 0.25736168026924133\r\nStep 67, loss: 0.26021021604537964\r\nStep 68, loss: 0.2600043714046478\r\nStep 69, loss: 0.2554631233215332\r\nSaved checkpoint at step 70\r\nStep 70, loss: 0.25915834307670593\r\nStep 71, loss: 0.2538088262081146\r\nStep 72, loss: 0.2527981102466583\r\nStep 73, loss: 0.24763624370098114\r\nStep 74, loss: 0.23982031643390656\r\nStep 75, loss: 0.2431134581565857\r\nStep 76, loss: 0.24418558180332184\r\nStep 77, loss: 0.23978117108345032\r\nStep 78, loss: 0.23628409206867218\r\nStep 79, loss: 0.22659100592136383\r\nSaved checkpoint at step 80\r\nStep 80, loss: 0.23002658784389496\r\nStep 81, loss: 0.2237771600484848\r\nStep 82, loss: 0.21777285635471344\r\nStep 83, loss: 0.21955353021621704\r\nStep 84, loss: 0.2197633534669876\r\nStep 85, loss: 0.21693949401378632\r\nStep 86, loss: 0.2156858891248703\r\nStep 87, loss: 0.20907285809516907\r\nStep 88, loss: 0.20304414629936218\r\nStep 89, loss: 0.20950746536254883\r\nSaved checkpoint at step 90\r\nStep 90, loss: 0.20157180726528168\r\nStep 91, loss: 0.19864089787006378\r\nStep 92, loss: 0.1961870640516281\r\nStep 93, loss: 0.1970020979642868\r\nStep 94, loss: 0.19751623272895813\r\nStep 95, loss: 0.1926620602607727\r\nStep 96, loss: 0.1907958984375\r\nStep 97, loss: 0.18797270953655243\r\nStep 98, loss: 0.18371835350990295\r\nStep 99, loss: 0.19347549974918365\r\nSaved checkpoint at step 100\r\nStep 100, loss: 0.17737938463687897\r\nStep 101, loss: 0.18215568363666534\r\nStep 102, loss: 0.17679424583911896\r\nStep 103, loss: 0.17569106817245483\r\nStep 104, loss: 0.17612037062644958\r\nStep 105, loss: 0.1708865910768509\r\nStep 106, loss: 0.17582806944847107\r\nStep 107, loss: 0.16020779311656952\r\nStep 108, loss: 0.16374938189983368\r\nStep 82, loss: 0.21777285635471344\r\nStep 83, loss: 0.21955353021621704\r\nStep 84, loss: 0.2197633534669876\r\nStep 85, loss: 0.21693949401378632\r\nStep 86, loss: 0.2156858891248703\r\nStep 87, loss: 0.20907285809516907\r\nStep 88, loss: 0.20304414629936218\r\nStep 89, loss: 0.20950746536254883\r\nSaved checkpoint at step 90\r\nStep 90, loss: 0.20157180726528168\r\nStep 91, loss: 0.19864089787006378\r\nStep 92, loss: 0.1961870640516281\r\nStep 93, loss: 0.1970020979642868\r\nStep 94, loss: 0.19751623272895813\r\nStep 95, loss: 0.1926620602607727\r\nStep 96, loss: 0.1907958984375\r\nStep 97, loss: 0.18797270953655243\r\nStep 98, loss: 0.18371835350990295\r\nStep 99, loss: 0.19347549974918365\r\nSaved checkpoint at step 100\r\nStep 100, loss: 0.17737938463687897\r\nStep 101, loss: 0.18215568363666534\r\nStep 102, loss: 0.17679424583911896\r\nStep 103, loss: 0.17569106817245483\r\nStep 104, loss: 0.17612037062644958\r\nStep 105, loss: 0.1708865910768509\r\nStep 106, loss: 0.17582806944847107\r\nStep 107, loss: 0.16020779311656952\r\nStep 108, loss: 0.16374938189983368\r\nStep 109, loss: 0.16316014528274536\r\nStep 109, loss: 0.16316014528274536\r\nSaved checkpoint at step 110\r\nStep 110, loss: 0.15965776145458221\r\nStep 111, loss: 0.16084066033363342\r\nStep 112, loss: 0.15732374787330627\r\nStep 113, loss: 0.1501660794019699\r\nStep 114, loss: 0.15066221356391907\r\nStep 115, loss: 0.15051837265491486\r\nStep 116, loss: 0.1540840119123459\r\nStep 117, loss: 0.15616631507873535\r\nStep 118, loss: 0.14677441120147705\r\nStep 119, loss: 0.15072251856327057\r\nSaved checkpoint at step 120\r\nStep 120, loss: 0.14081387221813202\r\nStep 121, loss: 0.1404900848865509\r\nStep 122, loss: 0.13735362887382507\r\nStep 123, loss: 0.14047737419605255\r\nStep 124, loss: 0.14097315073013306\r\nStep 125, loss: 0.13012643158435822\r\nStep 126, loss: 0.13319577276706696\r\nStep 127, loss: 0.12502308189868927\r\nStep 128, loss: 0.13217462599277496\r\nStep 129, loss: 0.12322020530700684\r\nSaved checkpoint at step 130\r\nStep 130, loss: 0.1309303492307663\r\nStep 131, loss: 0.12471090257167816\r\nStep 132, loss: 0.1259678304195404\r\nStep 133, loss: 0.11994366347789764\r\nStep 134, loss: 0.12396806478500366\r\nStep 135, loss: 0.12951606512069702\r\nSaved checkpoint at step 110\r\nStep 110, loss: 0.15965776145458221\r\nStep 111, loss: 0.16084066033363342\r\nStep 112, loss: 0.15732374787330627\r\nStep 113, loss: 0.1501660794019699\r\nStep 114, loss: 0.15066221356391907\r\nStep 115, loss: 0.15051837265491486\r\nStep 116, loss: 0.1540840119123459\r\nStep 117, loss: 0.15616631507873535\r\nStep 118, loss: 0.14677441120147705\r\nStep 119, loss: 0.15072251856327057\r\nSaved checkpoint at step 120\r\nStep 120, loss: 0.14081387221813202\r\nStep 121, loss: 0.1404900848865509\r\nStep 122, loss: 0.13735362887382507\r\nStep 123, loss: 0.14047737419605255\r\nStep 124, loss: 0.14097315073013306\r\nStep 125, loss: 0.13012643158435822\r\nStep 126, loss: 0.13319577276706696\r\nStep 127, loss: 0.12502308189868927\r\nStep 128, loss: 0.13217462599277496\r\nStep 129, loss: 0.12322020530700684\r\nSaved checkpoint at step 130\r\nStep 130, loss: 0.1309303492307663\r\nStep 131, loss: 0.12471090257167816\r\nStep 132, loss: 0.1259678304195404\r\nStep 133, loss: 0.11994366347789764\r\nStep 134, loss: 0.12396806478500366\r\nStep 135, loss: 0.12951606512069702\r\nStep 136, loss: 0.12501047551631927\r\nStep 137, loss: 0.12460850179195404\r\nStep 138, loss: 0.12312853336334229\r\nStep 139, loss: 0.11870359629392624\r\nSaved checkpoint at step 140\r\nStep 140, loss: 0.10972652584314346\r\nStep 141, loss: 0.11810508370399475\r\nStep 142, loss: 0.11488064378499985\r\nStep 143, loss: 0.10100933909416199\r\nStep 144, loss: 0.11190460622310638\r\nStep 145, loss: 0.10655320435762405\r\nStep 146, loss: 0.11590567231178284\r\nStep 147, loss: 0.11223439872264862\r\nStep 148, loss: 0.11101178079843521\r\nStep 149, loss: 0.11008752137422562\r\nSaved checkpoint at step 150\r\nStep 150, loss: 0.1229330375790596\r\nStep 151, loss: 0.12016700953245163\r\nStep 152, loss: 0.11195394396781921\r\nStep 153, loss: 0.11655457317829132\r\nStep 154, loss: 0.10912562161684036\r\nStep 155, loss: 0.12283866107463837\r\nStep 156, loss: 0.10622447729110718\r\nStep 157, loss: 0.10579507797956467\r\nStep 158, loss: 0.09965544939041138\r\nStep 159, loss: 0.10587912797927856\r\nSaved checkpoint at step 160\r\nStep 160, loss: 0.11324921250343323\r\nStep 161, loss: 0.11093202978372574\r\nStep 136, loss: 0.12501047551631927\r\nStep 137, loss: 0.12460850179195404\r\nStep 138, loss: 0.12312853336334229\r\nStep 139, loss: 0.11870359629392624\r\nSaved checkpoint at step 140\r\nStep 140, loss: 0.10972652584314346\r\nStep 141, loss: 0.11810508370399475\r\nStep 142, loss: 0.11488064378499985\r\nStep 143, loss: 0.10100933909416199\r\nStep 144, loss: 0.11190460622310638\r\nStep 145, loss: 0.10655320435762405\r\nStep 146, loss: 0.11590567231178284\r\nStep 147, loss: 0.11223439872264862\r\nStep 148, loss: 0.11101178079843521\r\nStep 149, loss: 0.11008752137422562\r\nSaved checkpoint at step 150\r\nStep 150, loss: 0.1229330375790596\r\nStep 151, loss: 0.12016700953245163\r\nStep 152, loss: 0.11195394396781921\r\nStep 153, loss: 0.11655457317829132\r\nStep 154, loss: 0.10912562161684036\r\nStep 155, loss: 0.12283866107463837\r\nStep 156, loss: 0.10622447729110718\r\nStep 157, loss: 0.10579507797956467\r\nStep 158, loss: 0.09965544939041138\r\nStep 159, loss: 0.10587912797927856\r\nSaved checkpoint at step 160\r\nStep 160, loss: 0.11324921250343323\r\nStep 161, loss: 0.11093202978372574\r\nStep 162, loss: 0.10980116575956345\r\nStep 163, loss: 0.10331201553344727\r\nStep 164, loss: 0.09606180340051651\r\nStep 165, loss: 0.10288108885288239\r\nStep 166, loss: 0.10788475722074509\r\nStep 167, loss: 0.11626137793064117\r\nStep 168, loss: 0.11271797865629196\r\nStep 169, loss: 0.10126489400863647\r\nSaved checkpoint at step 170\r\nStep 170, loss: 0.10251966863870621\r\nStep 171, loss: 0.09720180183649063\r\nStep 172, loss: 0.10119358450174332\r\nStep 173, loss: 0.10164882242679596\r\nStep 174, loss: 0.09646020829677582\r\nStep 175, loss: 0.09435637295246124\r\nStep 176, loss: 0.09778910130262375\r\nStep 177, loss: 0.09556873142719269\r\nStep 178, loss: 0.10166315734386444\r\nStep 179, loss: 0.09307191520929337\r\nSaved checkpoint at step 180\r\nStep 180, loss: 0.09001213312149048\r\nStep 181, loss: 0.09384429454803467\r\nStep 182, loss: 0.09797736257314682\r\nStep 183, loss: 0.09913866221904755\r\nStep 184, loss: 0.09100943058729172\r\nStep 185, loss: 0.09598495066165924\r\nStep 186, loss: 0.09197870641946793\r\nStep 187, loss: 0.09248428791761398\r\nStep 162, loss: 0.10980116575956345\r\nStep 163, loss: 0.10331201553344727\r\nStep 164, loss: 0.09606180340051651\r\nStep 165, loss: 0.10288108885288239\r\nStep 166, loss: 0.10788475722074509\r\nStep 167, loss: 0.11626137793064117\r\nStep 168, loss: 0.11271797865629196\r\nStep 169, loss: 0.10126489400863647\r\nSaved checkpoint at step 170\r\nStep 170, loss: 0.10251966863870621\r\nStep 171, loss: 0.09720180183649063\r\nStep 172, loss: 0.10119358450174332\r\nStep 173, loss: 0.10164882242679596\r\nStep 174, loss: 0.09646020829677582\r\nStep 175, loss: 0.09435637295246124\r\nStep 176, loss: 0.09778910130262375\r\nStep 177, loss: 0.09556873142719269\r\nStep 178, loss: 0.10166315734386444\r\nStep 179, loss: 0.09307191520929337\r\nSaved checkpoint at step 180\r\nStep 180, loss: 0.09001213312149048\r\nStep 181, loss: 0.09384429454803467\r\nStep 182, loss: 0.09797736257314682\r\nStep 183, loss: 0.09913866221904755\r\nStep 184, loss: 0.09100943058729172\r\nStep 185, loss: 0.09598495066165924\r\nStep 186, loss: 0.09197870641946793\r\nStep 187, loss: 0.09248428791761398\r\nStep 188, loss: 0.0881965383887291\r\nStep 189, loss: 0.08620244264602661\r\nSaved checkpoint at step 190\r\nStep 190, loss: 0.0936160534620285\r\nStep 191, loss: 0.08297702670097351\r\nStep 192, loss: 0.08549270033836365\r\nStep 193, loss: 0.09282968193292618\r\nStep 194, loss: 0.09096062183380127\r\nStep 195, loss: 0.08280432224273682\r\nStep 196, loss: 0.08810421824455261\r\nStep 197, loss: 0.08211309462785721\r\nStep 198, loss: 0.07822512835264206\r\nStep 199, loss: 0.08037614822387695\r\nSaved checkpoint at step 200\r\nStep 200, loss: 0.08436369150876999\r\nStep 201, loss: 0.0888621136546135\r\nStep 202, loss: 0.08680012822151184\r\nStep 203, loss: 0.08875889331102371\r\nStep 204, loss: 0.08150099217891693\r\nStep 205, loss: 0.08258917927742004\r\nStep 206, loss: 0.07716099172830582\r\nStep 207, loss: 0.07317893952131271\r\nStep 208, loss: 0.07029818743467331\r\nStep 209, loss: 0.07618950307369232\r\nSaved checkpoint at step 210\r\nStep 210, loss: 0.07858575135469437\r\nStep 211, loss: 0.07911297678947449\r\nStep 212, loss: 0.09101250022649765\r\nStep 213, loss: 0.07573471963405609\r\nStep 188, loss: 0.0881965383887291\r\nStep 189, loss: 0.08620244264602661\r\nSaved checkpoint at step 190\r\nStep 190, loss: 0.0936160534620285\r\nStep 191, loss: 0.08297702670097351\r\nStep 192, loss: 0.08549270033836365\r\nStep 193, loss: 0.09282968193292618\r\nStep 194, loss: 0.09096062183380127\r\nStep 195, loss: 0.08280432224273682\r\nStep 196, loss: 0.08810421824455261\r\nStep 197, loss: 0.08211309462785721\r\nStep 198, loss: 0.07822512835264206\r\nStep 199, loss: 0.08037614822387695\r\nSaved checkpoint at step 200\r\nStep 200, loss: 0.08436369150876999\r\nStep 201, loss: 0.0888621136546135\r\nStep 202, loss: 0.08680012822151184\r\nStep 203, loss: 0.08875889331102371\r\nStep 204, loss: 0.08150099217891693\r\nStep 205, loss: 0.08258917927742004\r\nStep 206, loss: 0.07716099172830582\r\nStep 207, loss: 0.07317893952131271\r\nStep 208, loss: 0.07029818743467331\r\nStep 209, loss: 0.07618950307369232\r\nSaved checkpoint at step 210\r\nStep 210, loss: 0.07858575135469437\r\nStep 211, loss: 0.07911297678947449\r\nStep 212, loss: 0.09101250022649765\r\nStep 213, loss: 0.07573471963405609\r\nStep 214, loss: 0.07784590125083923\r\nStep 214, loss: 0.07784590125083923\r\n",,terminal_output +3475,6755869,"TERMINAL",0,0,"bash",,terminal_focus +3476,6757781,"TERMINAL",0,0,"cd /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/interactive/3335335",,terminal_command +3477,6759074,"TERMINAL",0,0,"ls",,terminal_command +3478,6759089,"TERMINAL",0,0,"]633;E;2025-07-10 18:10:53 ls;d0af5434-9a5b-4b8f-9412-c82c37ee36e4]633;C000230 000240 000250\r\n]0;tum_cte0515@hkn1993:/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/interactive/3335335]633;D;0",,terminal_output +3479,6766286,"TERMINAL",0,0,"srun",,terminal_focus +3480,6767110,"TERMINAL",0,0,"^Csrun: interrupt (one more within 1 sec to abort)\r\nsrun: StepId=3335335.2 tasks 0-7: running\r\n",,terminal_output +3481,6767321,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.2\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-1:\r\nStep 215, loss: 0.07533489167690277\r\nStep 216, loss: 0.07364197075366974\r\nStep 217, loss: 0.07628205418586731\r\nStep 218, loss: 0.08951474726200104\r\nStep 219, loss: 0.07684393227100372\r\nSaved checkpoint at step 220\r\nStep 220, loss: 0.07174445688724518\r\nStep 221, loss: 0.07478570193052292\r\nStep 222, loss: 0.07103043049573898\r\nStep 223, loss: 0.07417479902505875\r\nStep 224, loss: 0.07142274081707001\r\nStep 225, loss: 0.07241654396057129\r\nStep 226, loss: 0.07215618342161179\r\nStep 227, loss: 0.07136914879083633\r\nStep 228, loss: 0.07855178415775299\r\nStep 229, loss: 0.07575874030590057\r\nSaved checkpoint at step 230\r\nStep 230, loss: 0.07756376266479492\r\nStep 231, loss: 0.07309763133525848\r\nStep 232, loss: 0.07378087192773819\r\nStep 233, loss: 0.07816367596387863\r\nStep 234, loss: 0.07020875811576843\r\nStep 235, loss: 0.06862890720367432\r\nStep 236, loss: 0.06681060791015625\r\nStep 237, loss: 0.07396500557661057\r\nStep 238, loss: 0.06766857206821442\r\nStep 239, loss: 0.06940465420484543\r\nSaved checkpoint at step 240\r\nStep 240, loss: 0.06836851686239243\r\nStep 241, loss: 0.06678757816553116\r\nStep 242, loss: 0.0684138759970665\r\nStep 243, loss: 0.07096099853515625\r\nStep 244, loss: 0.07954933494329453\r\nStep 245, loss: 0.0722353458404541\r\nStep 246, loss: 0.06548785418272018\r\nStep 247, loss: 0.07080160826444626\r\nStep 248, loss: 0.07977914065122604\r\nStep 249, loss: 0.07421938329935074\r\nSaved checkpoint at step 250\r\nStep 250, loss: 0.06870173662900925\r\nStep 251, loss: 0.06417525559663773\r\nStep 252, loss: 0.0673845112323761\r\nStep 253, loss: 0.06332477182149887\r\nStep 254, loss: 0.0709465965628624\r\nStep 255, loss: 0.07682745158672333\r\nStep 256, loss: 0.06788769364356995\r\nStep 257, loss: 0.06887709349393845\r\nStep 258, loss: 0.07298767566680908\r\nStep 259, loss: 0.06821200996637344\r\nProcess SpawnProcess-5:\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_tokenizer.py"", line 295, in \r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-3:\r\nsrun: forcing job termination\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-3:\r\nStep 215, loss: 0.07533489167690277\r\nStep 216, loss: 0.07364197075366974\r\nStep 217, loss: 0.07628205418586731\r\nStep 218, loss: 0.08951474726200104\r\nStep 219, loss: 0.07684393227100372\r\nSaved checkpoint at step 220\r\nStep 220, loss: 0.07174445688724518\r\nStep 221, loss: 0.07478570193052292\r\nStep 222, loss: 0.07103043049573898\r\nStep 223, loss: 0.07417479902505875\r\nStep 224, loss: 0.07142274081707001\r\nStep 225, loss: 0.07241654396057129\r\nStep 226, loss: 0.07215618342161179\r\nStep 227, loss: 0.07136914879083633\r\nStep 228, loss: 0.07855178415775299\r\nStep 229, loss: 0.07575874030590057\r\nSaved checkpoint at step 230\r\nStep 230, loss: 0.07756376266479492\r\nStep 231, loss: 0.07309763133525848\r\nStep 232, loss: 0.07378087192773819\r\nStep 233, loss: 0.07816367596387863\r\nStep 234, loss: 0.07020875811576843\r\nStep 235, loss: 0.06862890720367432\r\nStep 236, loss: 0.06681060791015625\r\nStep 237, loss: 0.07396500557661057\r\nStep 238, loss: 0.06766857206821442\r\nStep 239, loss: 0.06940465420484543\r\nSaved checkpoint at step 240\r\nStep 240, loss: 0.06836851686239243\r\nStep 241, loss: 0.06678757816553116\r\nStep 242, loss: 0.0684138759970665\r\nStep 243, loss: 0.07096099853515625\r\nStep 244, loss: 0.07954933494329453\r\nStep 245, loss: 0.0722353458404541\r\nStep 246, loss: 0.06548785418272018\r\nStep 247, loss: 0.07080160826444626\r\nStep 248, loss: 0.07977914065122604\r\nStep 249, loss: 0.07421938329935074\r\nSaved checkpoint at step 250\r\nStep 250, loss: 0.06870173662900925\r\nStep 251, loss: 0.06417525559663773\r\nStep 252, loss: 0.0673845112323761\r\nStep 253, loss: 0.06332477182149887\r\nStep 254, loss: 0.0709465965628624\r\nStep 255, loss: 0.07682745158672333\r\nStep 256, loss: 0.06788769364356995\r\nStep 257, loss: 0.06887709349393845\r\nStep 258, loss: 0.07298767566680908\r\nStep 259, loss: 0.06821200996637344\r\nStep 215, loss: 0.07533489167690277\r\nStep 216, loss: 0.07364197075366974\r\nStep 217, loss: 0.07628205418586731\r\nStep 218, loss: 0.08951474726200104\r\nStep 219, loss: 0.07684393227100372\r\nSaved checkpoint at step 220\r\nStep 220, loss: 0.07174445688724518\r\nStep 221, loss: 0.07478570193052292\r\nStep 222, loss: 0.07103043049573898\r\nStep 223, loss: 0.07417479902505875\r\nStep 224, loss: 0.07142274081707001\r\nStep 225, loss: 0.07241654396057129\r\nStep 226, loss: 0.07215618342161179\r\nStep 227, loss: 0.07136914879083633\r\nStep 228, loss: 0.07855178415775299\r\nStep 229, loss: 0.07575874030590057\r\nSaved checkpoint at step 230\r\nStep 230, loss: 0.07756376266479492\r\nStep 231, loss: 0.07309763133525848\r\nStep 232, loss: 0.07378087192773819\r\nStep 233, loss: 0.07816367596387863\r\nStep 234, loss: 0.07020875811576843\r\nStep 235, loss: 0.06862890720367432\r\nStep 236, loss: 0.06681060791015625\r\nStep 237, loss: 0.07396500557661057\r\nStep 238, loss: 0.06766857206821442\r\nStep 239, loss: 0.06940465420484543\r\nSaved checkpoint at step 240\r\nStep 240, loss: 0.06836851686239243\r\nProcess SpawnProcess-3:\r\nTraceback (most recent call last):\r\nStep 241, loss: 0.06678757816553116\r\nStep 242, loss: 0.0684138759970665\r\nStep 243, loss: 0.07096099853515625\r\nStep 244, loss: 0.07954933494329453\r\nStep 245, loss: 0.0722353458404541\r\nStep 246, loss: 0.06548785418272018\r\nStep 247, loss: 0.07080160826444626\r\nStep 248, loss: 0.07977914065122604\r\nStep 249, loss: 0.07421938329935074\r\nSaved checkpoint at step 250\r\nStep 250, loss: 0.06870173662900925\r\nStep 251, loss: 0.06417525559663773\r\nStep 252, loss: 0.0673845112323761\r\nStep 253, loss: 0.06332477182149887\r\nStep 254, loss: 0.0709465965628624\r\nStep 255, loss: 0.07682745158672333\r\nStep 256, loss: 0.06788769364356995\r\nStep 257, loss: 0.06887709349393845\r\nStep 258, loss: 0.07298767566680908\r\nStep 259, loss: 0.06821200996637344\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_tokenizer.py"", line 295, in \r\nProcess SpawnProcess-3:\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_tokenizer.py"", line 295, in \r\nProcess SpawnProcess-5:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-7:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-1:\r\nProcess SpawnProcess-5:\r\nStep 215, loss: 0.07533489167690277\r\nStep 216, loss: 0.07364197075366974\r\nStep 217, loss: 0.07628205418586731\r\nStep 218, loss: 0.08951474726200104\r\nStep 219, loss: 0.07684393227100372\r\nSaved checkpoint at step 220\r\nStep 220, loss: 0.07174445688724518\r\nStep 221, loss: 0.07478570193052292\r\nStep 222, loss: 0.07103043049573898\r\nStep 223, loss: 0.07417479902505875\r\nStep 224, loss: 0.07142274081707001\r\nStep 225, loss: 0.07241654396057129\r\nStep 226, loss: 0.07215618342161179\r\nStep 227, loss: 0.07136914879083633\r\nStep 228, loss: 0.07855178415775299\r\nStep 229, loss: 0.07575874030590057\r\nSaved checkpoint at step 230\r\nStep 230, loss: 0.07756376266479492\r\nStep 231, loss: 0.07309763133525848\r\nStep 232, loss: 0.07378087192773819\r\nStep 233, loss: 0.07816367596387863\r\nStep 234, loss: 0.07020875811576843\r\nStep 235, loss: 0.06862890720367432\r\nStep 236, loss: 0.06681060791015625\r\nStep 237, loss: 0.07396500557661057\r\nStep 238, loss: 0.06766857206821442\r\nStep 239, loss: 0.06940465420484543\r\nSaved checkpoint at step 240\r\nStep 240, loss: 0.06836851686239243\r\nProcess SpawnProcess-7:\r\nStep 215, loss: 0.07533489167690277\r\nStep 216, loss: 0.07364197075366974\r\nStep 217, loss: 0.07628205418586731\r\nStep 218, loss: 0.08951474726200104\r\nStep 219, loss: 0.07684393227100372\r\nSaved checkpoint at step 220\r\nStep 220, loss: 0.07174445688724518\r\nStep 221, loss: 0.07478570193052292\r\nStep 222, loss: 0.07103043049573898\r\nStep 223, loss: 0.07417479902505875\r\nStep 224, loss: 0.07142274081707001\r\nStep 225, loss: 0.07241654396057129\r\nStep 226, loss: 0.07215618342161179\r\nStep 227, loss: 0.07136914879083633\r\nStep 228, loss: 0.07855178415775299\r\nStep 229, loss: 0.07575874030590057\r\nSaved checkpoint at step 230\r\nStep 230, loss: 0.07756376266479492\r\nStep 231, loss: 0.07309763133525848\r\nStep 232, loss: 0.07378087192773819\r\nStep 233, loss: 0.07816367596387863\r\nStep 234, loss: 0.07020875811576843\r\nStep 235, loss: 0.06862890720367432\r\nStep 236, loss: 0.06681060791015625\r\nStep 237, loss: 0.07396500557661057\r\nStep 238, loss: 0.06766857206821442\r\nStep 239, loss: 0.06940465420484543\r\nSaved checkpoint at step 240\r\nStep 240, loss: 0.06836851686239243\r\nStep 241, loss: 0.06678757816553116\r\nStep 242, loss: 0.0684138759970665\r\nStep 243, loss: 0.07096099853515625\r\nStep 244, loss: 0.07954933494329453\r\nStep 245, loss: 0.0722353458404541\r\nStep 246, loss: 0.06548785418272018\r\nStep 247, loss: 0.07080160826444626\r\nStep 248, loss: 0.07977914065122604\r\nStep 249, loss: 0.07421938329935074\r\nSaved checkpoint at step 250\r\nStep 250, loss: 0.06870173662900925\r\nStep 251, loss: 0.06417525559663773\r\nStep 252, loss: 0.0673845112323761\r\nStep 253, loss: 0.06332477182149887\r\nStep 254, loss: 0.0709465965628624\r\nStep 255, loss: 0.07682745158672333\r\nStep 256, loss: 0.06788769364356995\r\nStep 257, loss: 0.06887709349393845\r\nStep 258, loss: 0.07298767566680908\r\nStep 259, loss: 0.06821200996637344\r\nProcess SpawnProcess-6:\r\nStep 241, loss: 0.06678757816553116\r\nStep 242, loss: 0.0684138759970665\r\nStep 243, loss: 0.07096099853515625\r\nStep 244, loss: 0.07954933494329453\r\nStep 245, loss: 0.0722353458404541\r\nStep 246, loss: 0.06548785418272018\r\nStep 247, loss: 0.07080160826444626\r\nStep 248, loss: 0.07977914065122604\r\nStep 249, loss: 0.07421938329935074\r\nSaved checkpoint at step 250\r\nStep 250, loss: 0.06870173662900925\r\nStep 251, loss: 0.06417525559663773\r\nStep 252, loss: 0.0673845112323761\r\nStep 253, loss: 0.06332477182149887\r\nStep 254, loss: 0.0709465965628624\r\nStep 255, loss: 0.07682745158672333\r\nStep 256, loss: 0.06788769364356995\r\nStep 257, loss: 0.06887709349393845\r\nStep 258, loss: 0.07298767566680908\r\nStep 259, loss: 0.06821200996637344\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-7:\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_tokenizer.py"", line 295, in \r\nProcess SpawnProcess-1:\r\nStep 215, loss: 0.07533489167690277\r\nStep 216, loss: 0.07364197075366974\r\nStep 217, loss: 0.07628205418586731\r\nStep 218, loss: 0.08951474726200104\r\nStep 219, loss: 0.07684393227100372\r\nSaved checkpoint at step 220\r\nStep 220, loss: 0.07174445688724518\r\nStep 221, loss: 0.07478570193052292\r\nStep 222, loss: 0.07103043049573898\r\nStep 223, loss: 0.07417479902505875\r\nStep 224, loss: 0.07142274081707001\r\nStep 225, loss: 0.07241654396057129\r\nStep 226, loss: 0.07215618342161179\r\nStep 227, loss: 0.07136914879083633\r\nStep 228, loss: 0.07855178415775299\r\nStep 229, loss: 0.07575874030590057\r\nSaved checkpoint at step 230\r\nStep 230, loss: 0.07756376266479492\r\nStep 231, loss: 0.07309763133525848\r\nStep 232, loss: 0.07378087192773819\r\nStep 233, loss: 0.07816367596387863\r\nStep 234, loss: 0.07020875811576843\r\nStep 235, loss: 0.06862890720367432\r\nStep 236, loss: 0.06681060791015625\r\nStep 237, loss: 0.07396500557661057\r\nStep 238, loss: 0.06766857206821442\r\nStep 239, loss: 0.06940465420484543\r\nSaved checkpoint at step 240\r\nStep 240, loss: 0.06836851686239243\r\nProcess SpawnProcess-5:\r\nStep 241, loss: 0.06678757816553116\r\nStep 242, loss: 0.0684138759970665\r\nStep 243, loss: 0.07096099853515625\r\nStep 244, loss: 0.07954933494329453\r\nStep 245, loss: 0.0722353458404541\r\nStep 246, loss: 0.06548785418272018\r\nStep 247, loss: 0.07080160826444626\r\nStep 248, loss: 0.07977914065122604\r\nStep 249, loss: 0.07421938329935074\r\nSaved checkpoint at step 250\r\nStep 250, loss: 0.06870173662900925\r\nStep 251, loss: 0.06417525559663773\r\nStep 252, loss: 0.0673845112323761\r\nStep 253, loss: 0.06332477182149887\r\nStep 254, loss: 0.0709465965628624\r\nStep 255, loss: 0.07682745158672333\r\nStep 256, loss: 0.06788769364356995\r\nStep 257, loss: 0.06887709349393845\r\nStep 258, loss: 0.07298767566680908\r\nStep 259, loss: 0.06821200996637344\r\nProcess SpawnProcess-2:\r\nProcess SpawnProcess-6:\r\nProcess SpawnProcess-4:\r\nProcess SpawnProcess-8:\r\nProcess SpawnProcess-3:\r\nProcess SpawnProcess-4:\r\nStep 215, loss: 0.07533489167690277\r\nStep 216, loss: 0.07364197075366974\r\nStep 217, loss: 0.07628205418586731\r\nStep 218, loss: 0.08951474726200104\r\nStep 219, loss: 0.07684393227100372\r\nSaved checkpoint at step 220\r\nStep 220, loss: 0.07174445688724518\r\nStep 221, loss: 0.07478570193052292\r\nStep 222, loss: 0.07103043049573898\r\nStep 223, loss: 0.07417479902505875\r\nStep 224, loss: 0.07142274081707001\r\nStep 225, loss: 0.07241654396057129\r\nStep 226, loss: 0.07215618342161179\r\nStep 227, loss: 0.07136914879083633\r\nStep 228, loss: 0.07855178415775299\r\nStep 229, loss: 0.07575874030590057\r\nSaved checkpoint at step 230\r\nStep 230, loss: 0.07756376266479492\r\nStep 231, loss: 0.07309763133525848\r\nStep 232, loss: 0.07378087192773819\r\nStep 233, loss: 0.07816367596387863\r\nStep 234, loss: 0.07020875811576843\r\nStep 235, loss: 0.06862890720367432\r\nStep 236, loss: 0.06681060791015625\r\nStep 237, loss: 0.07396500557661057\r\nStep 238, loss: 0.06766857206821442\r\nStep 239, loss: 0.06940465420484543\r\nSaved checkpoint at step 240\r\nStep 240, loss: 0.06836851686239243\r\nStep 241, loss: 0.06678757816553116\r\nStep 242, loss: 0.0684138759970665\r\nStep 243, loss: 0.07096099853515625\r\nStep 244, loss: 0.07954933494329453\r\nStep 245, loss: 0.0722353458404541\r\nStep 246, loss: 0.06548785418272018\r\nStep 247, loss: 0.07080160826444626\r\nStep 248, loss: 0.07977914065122604\r\nStep 249, loss: 0.07421938329935074\r\nSaved checkpoint at step 250\r\nStep 250, loss: 0.06870173662900925\r\nStep 251, loss: 0.06417525559663773\r\nStep 252, loss: 0.0673845112323761\r\nStep 253, loss: 0.06332477182149887\r\nStep 254, loss: 0.0709465965628624\r\nStep 255, loss: 0.07682745158672333\r\nStep 256, loss: 0.06788769364356995\r\nStep 257, loss: 0.06887709349393845\r\nStep 258, loss: 0.07298767566680908\r\nStep 259, loss: 0.06821200996637344\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_tokenizer.py"", line 295, in \r\n checkpoint_manager.save(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/checkpoint_manager.py"", line 1412, in save\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/train_tokenizer.py"", line 295, in \r\nProcess SpawnProcess-5:\r\n checkpoint_manager.save(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/checkpoint_manager.py"", line 1412, in save\r\n checkpoint_manager.save(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/checkpoint_manager.py"", line 1412, in save\r\n checkpoint_manager.save(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/checkpoint_manager.py"", line 1412, in save\r\n checkpoint_manager.save(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/checkpoint_manager.py"", line 1412, in save\r\n checkpoint_manager.save(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/orbax/checkpoint/checkpoint_manager.py"", line 1412, in save\r\nsrun: Job step aborted: Waiting up to 32 seconds for job step to finish.\r\nslurmstepd: error: *** STEP 3335335.2 ON hkn0425 CANCELLED AT 2025-07-10T18:11:01 ***\r\n",,terminal_output +3482,6767447,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.2\r\nsrun: job abort in progress\r\n",,terminal_output +3483,6767645,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.2\r\n",,terminal_output +3484,6767858,"TERMINAL",0,0,"^Csrun: sending Ctrl-C to StepId=3335335.2\r\nsrun: job abort in progress\r\n",,terminal_output +3485,6768545,"TERMINAL",0,0,"^C]0;tum_cte0515@hkn0425:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +3486,6768731,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0425:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +3487,6769836,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +3488,6769837,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1387,0,"",shellscript,selection_mouse +3489,6770357,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1302,0,"",shellscript,selection_mouse +3490,6770359,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1301,0,"",shellscript,selection_command +3491,6770882,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1302,0,"",shellscript,selection_mouse +3492,6770883,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1301,0,"",shellscript,selection_command +3493,6771833,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1468,0,"",shellscript,selection_mouse +3494,6771834,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1467,0,"",shellscript,selection_command +3495,6772315,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1439,0,"",shellscript,selection_mouse +3496,6778159,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +3497,6779473,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1038,0,"",shellscript,selection_mouse +3498,6780455,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1038,0,"0",shellscript,content +3499,6780457,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1039,0,"",shellscript,selection_keyboard +3500,6780585,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1039,0,"0",shellscript,content +3501,6780586,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1040,0,"",shellscript,selection_keyboard +3502,6781195,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1072,0,"",shellscript,selection_command +3503,6782288,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1075,0,"0",shellscript,content +3504,6782289,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1076,0,"",shellscript,selection_keyboard +3505,6782423,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1076,0,"0",shellscript,content +3506,6782424,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1077,0,"",shellscript,selection_keyboard +3507,6784309,"TERMINAL",0,0,"\r(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +3508,6790252,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +3509,6790380,"TERMINAL",0,0,"\r(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +3510,6818912,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",844,0,"",shellscript,selection_mouse +3511,6838586,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",823,0,"",shellscript,selection_mouse +3512,6839205,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",872,0,"",shellscript,selection_mouse +3513,6839684,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",817,0,"",shellscript,selection_mouse +3514,6839822,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",812,7,"Running",shellscript,selection_mouse +3515,6839932,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",807,65,"echo Running dynamics model overfit run. Slurm id: $slurm_job_id\n",shellscript,selection_mouse +3516,6840498,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",817,0,"",shellscript,selection_mouse +3517,6840612,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",812,7,"Running",shellscript,selection_mouse +3518,6840839,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",812,8,"Running ",shellscript,selection_mouse +3519,6840858,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",812,16,"Running dynamics",shellscript,selection_mouse +3520,6840917,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",812,17,"Running dynamics ",shellscript,selection_mouse +3521,6840936,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",812,22,"Running dynamics model",shellscript,selection_mouse +3522,6841002,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",812,30,"Running dynamics model overfit",shellscript,selection_mouse +3523,6841168,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",812,31,"Running dynamics model overfit ",shellscript,selection_mouse +3524,6841226,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",812,34,"Running dynamics model overfit run",shellscript,selection_mouse +3525,6841996,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",812,34,"",shellscript,content +3526,6842581,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",812,0,"T",shellscript,content +3527,6842582,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",813,0,"",shellscript,selection_keyboard +3528,6842779,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",813,0,"r",shellscript,content +3529,6842780,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",814,0,"",shellscript,selection_keyboard +3530,6842966,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",814,0,"a",shellscript,content +3531,6842967,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",815,0,"",shellscript,selection_keyboard +3532,6843048,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",815,0,"i",shellscript,content +3533,6843049,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",816,0,"",shellscript,selection_keyboard +3534,6843125,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",816,0,"n",shellscript,content +3535,6843126,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",817,0,"",shellscript,selection_keyboard +3536,6843258,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",817,0,"i",shellscript,content +3537,6843259,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",818,0,"",shellscript,selection_keyboard +3538,6843320,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",818,0,"n",shellscript,content +3539,6843321,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",819,0,"",shellscript,selection_keyboard +3540,6843455,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",819,0,"g",shellscript,content +3541,6843456,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",820,0,"",shellscript,selection_keyboard +3542,6843512,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",820,0," ",shellscript,content +3543,6843513,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",821,0,"",shellscript,selection_keyboard +3544,6843771,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",821,0,"d",shellscript,content +3545,6843772,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",822,0,"",shellscript,selection_keyboard +3546,6843950,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",822,0,"y",shellscript,content +3547,6843951,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",823,0,"",shellscript,selection_keyboard +3548,6844033,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",823,0,"n",shellscript,content +3549,6844034,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",824,0,"",shellscript,selection_keyboard +3550,6844223,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",824,0,"a",shellscript,content +3551,6844224,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",825,0,"",shellscript,selection_keyboard +3552,6844297,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",825,0,"m",shellscript,content +3553,6844297,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",826,0,"",shellscript,selection_keyboard +3554,6844458,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",826,0,"i",shellscript,content +3555,6844459,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",827,0,"",shellscript,selection_keyboard +3556,6844550,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",827,0,"c",shellscript,content +3557,6844551,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",828,0,"",shellscript,selection_keyboard +3558,6844621,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",828,0,"s",shellscript,content +3559,6844622,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",829,0,"",shellscript,selection_keyboard +3560,6844750,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",829,0," ",shellscript,content +3561,6844751,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",830,0,"",shellscript,selection_keyboard +3562,6844885,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",830,0,"m",shellscript,content +3563,6844886,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",831,0,"",shellscript,selection_keyboard +3564,6845028,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",831,0,"o",shellscript,content +3565,6845029,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",832,0,"",shellscript,selection_keyboard +3566,6845135,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",832,0,"d",shellscript,content +3567,6845136,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",833,0,"",shellscript,selection_keyboard +3568,6845196,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",833,0,"e",shellscript,content +3569,6845197,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",834,0,"",shellscript,selection_keyboard +3570,6845295,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",834,0,"l",shellscript,content +3571,6845296,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",835,0,"",shellscript,selection_keyboard +3572,6845363,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",835,0," ",shellscript,content +3573,6845364,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",836,0,"",shellscript,selection_keyboard +3574,6846286,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",835,1,"",shellscript,content +3575,6851234,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1385,0,"",shellscript,selection_mouse +3576,6852366,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1291,0,"",shellscript,selection_mouse +3577,6852936,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1504,0,"",shellscript,selection_mouse +3578,6853487,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1309,0,"",shellscript,selection_mouse +3579,6854012,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1291,0,"",shellscript,selection_mouse +3580,6854531,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1242,0,"",shellscript,selection_mouse +3581,6855018,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1219,0,"",shellscript,selection_mouse +3582,6855581,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1221,0,"",shellscript,selection_mouse +3583,6855809,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1220,1,"6",shellscript,selection_mouse +3584,6855869,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1219,2,"-6",shellscript,selection_mouse +3585,6856067,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1218,3,"e-6",shellscript,selection_mouse +3586,6856129,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1217,4,"5e-6",shellscript,selection_mouse +3587,6856825,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1217,4,"0",shellscript,content +3588,6856826,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1218,0,"",shellscript,selection_keyboard +3589,6857693,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1202,0,"",shellscript,selection_mouse +3590,6857841,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1201,1," ",shellscript,selection_mouse +3591,6857842,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1200,2,"7 ",shellscript,selection_mouse +3592,6857849,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1199,3,"-7 ",shellscript,selection_mouse +3593,6857864,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1198,4,"e-7 ",shellscript,selection_mouse +3594,6857883,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1197,5,"5e-7 ",shellscript,selection_mouse +3595,6857957,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1196,6,"=5e-7 ",shellscript,selection_mouse +3596,6858467,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1196,0,"",shellscript,selection_mouse +3597,6858886,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1198,0,"",shellscript,selection_mouse +3598,6859195,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1198,1,"e",shellscript,selection_mouse +3599,6859215,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1198,2,"e-",shellscript,selection_mouse +3600,6859245,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1198,3,"e-7",shellscript,selection_mouse +3601,6859603,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1201,0,"",shellscript,selection_mouse +3602,6859747,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1200,1,"7",shellscript,selection_mouse +3603,6859987,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1199,2,"-7",shellscript,selection_mouse +3604,6860117,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1197,4,"5e-7",shellscript,selection_mouse +3605,6860137,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1200,18,"7 \\n --max_lr=0",shellscript,selection_mouse +3606,6860476,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1197,4,"5e-7",shellscript,selection_mouse +3607,6862320,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1181,0,"",shellscript,selection_mouse +3608,6863080,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1180,1,"",shellscript,content +3609,6865705,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1180,0,"9",shellscript,content +3610,6865706,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1181,0,"",shellscript,selection_keyboard +3611,6865868,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1181,0,"6",shellscript,content +3612,6865869,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1182,0,"",shellscript,selection_keyboard +3613,6867285,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1203,0,"",shellscript,selection_mouse +3614,6867435,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1202,1," ",shellscript,selection_mouse +3615,6867492,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1201,2,"7 ",shellscript,selection_mouse +3616,6867573,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1200,3,"-7 ",shellscript,selection_mouse +3617,6867691,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1199,4,"e-7 ",shellscript,selection_mouse +3618,6868030,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1198,5,"5e-7 ",shellscript,selection_mouse +3619,6876120,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +3620,6877600,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1137,0,"",shellscript,selection_mouse +3621,6877741,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1136,2,"96",shellscript,selection_mouse +3622,6878670,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1171,0,"",shellscript,selection_mouse +3623,6878837,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1171,1,"1",shellscript,selection_mouse +3624,6878857,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1171,2,"1.",shellscript,selection_mouse +3625,6878881,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1171,3,"1.5",shellscript,selection_mouse +3626,6878898,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1171,4,"1.5e",shellscript,selection_mouse +3627,6878912,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1171,5,"1.5e-",shellscript,selection_mouse +3628,6878927,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1171,6,"1.5e-5",shellscript,selection_mouse +3629,6881548,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +3630,6883249,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1198,5,"",shellscript,content +3631,6883859,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1198,0,"1.5e-5",shellscript,content +3632,6884247,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1204,0," ",shellscript,content +3633,6884248,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1205,0,"",shellscript,selection_keyboard +3634,6884756,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1204,0,"",shellscript,selection_command +3635,6989108,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1268,0,"",shellscript,selection_mouse +3636,6990163,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1246,0,"",shellscript,selection_mouse +3637,6990740,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1243,0,"",shellscript,selection_mouse +3638,6995574,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"",shellscript,tab +3639,6999412,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1498,0,"",shellscript,selection_mouse +3640,6999413,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1497,0,"",shellscript,selection_command +3641,6999564,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1497,1,"\",shellscript,selection_mouse +3642,6999565,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1498,0,"",shellscript,selection_command +3643,6999639,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1496,2," \",shellscript,selection_mouse +3644,6999653,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1444,54,"kenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3645,6999653,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1413,85,"\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3646,6999669,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1388,110,"v \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3647,6999684,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1354,144," yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3648,6999740,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1308,190,"cs-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3649,6999740,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1268,230,"kpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3650,6999741,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1237,261,"ent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3651,6999796,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1223,275,"\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3652,6999798,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1191,307,"mage_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3653,6999892,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1169,329,"r=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3654,6999893,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1168,330,"lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3655,6999893,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1151,347,"lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3656,6999902,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1128,370,"ch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3657,6999917,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1127,371,"tch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3658,6999972,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1093,405,"kpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3659,7000033,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1092,406,"ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3660,7000089,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1091,407,"-ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3661,7000101,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1090,408,"--ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3662,7000162,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1071,427," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3663,7000168,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1070,428," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3664,7000338,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1069,429," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3665,7000516,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1068,430," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3666,7003289,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1498,0,"",shellscript,selection_mouse +3667,7003289,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1497,0,"",shellscript,selection_command +3668,7003468,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1497,1,"\",shellscript,selection_mouse +3669,7003469,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1497,2,"\\n",shellscript,selection_mouse +3670,7003469,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1498,0,"",shellscript,selection_command +3671,7003482,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1498,1,"\n",shellscript,selection_mouse +3672,7003546,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1498,0,"",shellscript,selection_mouse +3673,7003546,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1497,0,"",shellscript,selection_command +3674,7003554,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1443,54,"okenizer_ckpt_dir \\n --data_dir $array_records_dir ",shellscript,selection_mouse +3675,7003557,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1443,55,"okenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_command +3676,7003576,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1313,185,"lorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3677,7003591,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1132,366,"ize=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3678,7003620,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",671,827,"job_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3679,7003636,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",461,1037,"module unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3680,7003651,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",453,1045,"cat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3681,7003667,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",428,1070,"\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3682,7003869,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",429,1069,"# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3683,7003895,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",453,1045,"cat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3684,7003953,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",460,1038,"\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3685,7004127,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",453,1045,"cat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3686,7004382,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",460,1038,"\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3687,7004442,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",461,1037,"module unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \",shellscript,selection_mouse +3688,7006194,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",500,0,"",shellscript,selection_mouse +3689,7006899,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1499,0,"",shellscript,selection_mouse +3690,7007054,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1498,1,"\n",shellscript,selection_mouse +3691,7007073,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1391,108,"\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3692,7007096,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1106,393,"KPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3693,7007110,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",463,1036,"dule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3694,7007124,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",269,1230,"#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3695,7007139,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",123,1376,"#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3696,7007155,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",91,1408,"#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3697,7007209,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",67,1432,"#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3698,7007210,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",39,1460,"#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3699,7007272,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",21,1478,"#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3700,7007273,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",20,1479,"\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3701,7007274,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,1499,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3702,7009988,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +3703,7011164,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1596,0,"",shellscript,selection_mouse +3704,7011166,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1595,0,"",shellscript,selection_command +3705,7011280,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1595,1,"r",shellscript,selection_mouse +3706,7011281,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1596,0,"",shellscript,selection_command +3707,7011339,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1241,355,"s=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3708,7011346,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",807,789,"echo Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3709,7011360,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",651,945,"job_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3710,7011416,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",507,1089,"source .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3711,7011416,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",477,1119,"module unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3712,7011417,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",447,1149,"module unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3713,7011425,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",446,1150,"\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3714,7011482,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",439,1157,"cat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3715,7011483,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",415,1181,"# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3716,7011498,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",414,1182,"\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3717,7011512,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",372,1224,"#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3718,7011569,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",309,1287,"#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_36M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3719,7011570,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",216,1380,"#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_36M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3720,7011570,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",170,1426,"#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_36M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3721,7011578,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",123,1473,"#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_36M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3722,7011596,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",91,1505,"#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_36M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3723,7011656,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",39,1557,"#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_36M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Training dynamics model. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=1.5e-5 \\n --max_lr=0 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-36M-$slurm_job_id \\n --tags dynamics model-size-scaling 36M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir",shellscript,selection_mouse +3724,7013446,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",39,1557,"",shellscript,content +3725,7013474,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",21,0,"",shellscript,selection_command +3726,7013629,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",20,0,"",shellscript,selection_command +3727,7013759,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,selection_command +3728,7014112,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,21,"",shellscript,content +3729,7014291,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",18,0,"",shellscript,selection_command +3730,7014774,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,18,"",shellscript,content +3731,7017947,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,content +3732,7019824,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1442,57,"tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3733,7019839,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1133,366,"ze=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3734,7019858,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",461,1038,"module unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3735,7019869,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",453,1046,"cat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3736,7019895,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",39,1460,"#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3737,7019911,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,1499,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-yolorun-$slurm_job_id \\n --tags dynamics yolo-run \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3738,7021049,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,selection_command +3739,7042245,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",367,0,"#SBATCH --job-name=train_dynamics_modelsize_scaling_36M_2_node\n",shellscript,content +3740,7042247,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",430,61,"",shellscript,content +3741,7043856,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1293,0," --name=dynamics-modelsize-scaling-36M-$slurm_job_id \\n",shellscript,content +3742,7043878,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1351,0," --tags dynamics modelsize-scaling 36M \\n",shellscript,content +3743,7043936,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1395,75,"",shellscript,content +3744,7046396,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1247,0,"",shellscript,selection_mouse +3745,7047261,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1246,0,"",shellscript,selection_command +3746,7047715,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1226,30,"",shellscript,content +3747,7047730,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1230,0,"",shellscript,selection_command +3748,7050318,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +3749,7055989,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +3750,7064482,"TERMINAL",0,0,"\r(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +3751,7066489,"TERMINAL",0,0,"\r(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +3752,7067386,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1213,0,"",shellscript,selection_mouse +3753,7067388,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1212,0,"",shellscript,selection_command +3754,7068194,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1497,0,"",shellscript,selection_mouse +3755,7068195,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1496,0,"",shellscript,selection_command +3756,7078922,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_110M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-110M-$slurm_job_id \\n --tags dynamics model-size-scaling 110M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=12\n",shellscript,tab +3757,7080504,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",390,0,"",shellscript,selection_mouse +3758,7080505,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",389,0,"",shellscript,selection_command +3759,7081410,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",373,18,"",shellscript,content +3760,7083116,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +3761,7083915,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +3762,7091723,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1594,0,"",shellscript,selection_mouse +3763,7091728,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1593,0,"",shellscript,selection_command +3764,7092612,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1376,0,"",shellscript,selection_mouse +3765,7093317,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1318,0,"",shellscript,selection_mouse +3766,7094211,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +3767,7096917,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +3768,7099630,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +3769,7105154,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +3770,7106275,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1667,0,"",shellscript,selection_mouse +3771,7106471,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1475,192,"\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=12\n",shellscript,selection_mouse +3772,7106492,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",988,679,"dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-110M-$slurm_job_id \\n --tags dynamics model-size-scaling 110M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=12\n",shellscript,selection_mouse +3773,7106574,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",778,889,"grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-110M-$slurm_job_id \\n --tags dynamics model-size-scaling 110M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=12\n",shellscript,selection_mouse +3774,7106574,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",398,1269,"# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-110M-$slurm_job_id \\n --tags dynamics model-size-scaling 110M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=12\n",shellscript,selection_mouse +3775,7106574,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",39,1628,"#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_110M\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-110M-$slurm_job_id \\n --tags dynamics model-size-scaling 110M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=12\n",shellscript,selection_mouse +3776,7106575,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,1667,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_110M\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-110M-$slurm_job_id \\n --tags dynamics model-size-scaling 110M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=12\n",shellscript,selection_mouse +3777,7108178,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,selection_command +3778,7126299,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",21,0,"#SBATCH --nodes=2\n",shellscript,content +3779,7126319,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",39,0,"#SBATCH --ntasks-per-node=4\n",shellscript,content +3780,7126375,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",67,0,"#SBATCH --time=36:00:00\n",shellscript,content +3781,7126406,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",91,0,"#SBATCH --partition=accelerated\n",shellscript,content +3782,7126462,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",123,102,"",shellscript,content +3783,7126463,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",149,0,"#SBATCH --gres=gpu:4\n",shellscript,content +3784,7126759,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",170,0,"#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n",shellscript,content +3785,7126873,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",269,0,"#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n",shellscript,content +3786,7127080,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",367,0,"#SBATCH --job-name=train_dynamics_modelsize_scaling_110M_2_node\n",shellscript,content +3787,7127081,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",431,0,"\n",shellscript,content +3788,7127092,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",432,0,"# Log the sbatch script\n",shellscript,content +3789,7127149,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",432,273,"",shellscript,content +3790,7127522,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",551,0,"array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n",shellscript,content +3791,7127523,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",673,116,"",shellscript,content +3792,7127752,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",727,0,"CHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\n",shellscript,content +3793,7127791,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",839,59,"",shellscript,content +3794,7128080,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",865,0,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n",shellscript,content +3795,7128081,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1020,17,"",shellscript,content +3796,7128121,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1021,0,"env | grep SLURM\n\n",shellscript,content +3797,7128122,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1039,0,"srun python train_dynamics.py \\n",shellscript,content +3798,7128161,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1071,0," --save_ckpt \\n",shellscript,content +3799,7128263,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1089,334,"",shellscript,content +3800,7128337,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1122,0," --batch_size=96 \\n",shellscript,content +3801,7128356,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1144,0," --min_lr=0 \\n",shellscript,content +3802,7128504,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1161,0," --max_lr=1.5e-5 \\n",shellscript,content +3803,7128535,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1183,0," --log_image_interval=1000 \\n",shellscript,content +3804,7128591,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1215,0," --log \\n",shellscript,content +3805,7128592,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1227,0," --log_checkpoint_interval=1000 \\n",shellscript,content +3806,7128658,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1264,0," --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n",shellscript,content +3807,7128760,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1323,0," --tags dynamics modelsize-scaling 110M \\n",shellscript,content +3808,7128809,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1368,0," --entity instant-uv \\n",shellscript,content +3809,7128855,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1394,0," --project jafar \\n",shellscript,content +3810,7128967,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1416,319,"",shellscript,content +3811,7129097,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1465,0," --data_dir $array_records_dir \\n",shellscript,content +3812,7129097,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1501,142,"",shellscript,content +3813,7145047,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1465,35," --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=12",shellscript,content +3814,7145047,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1227,140," --name=dynamics-model-size-scaling-110M-$slurm_job_id \\n --tags dynamics model-size-scaling 110M \\n --log_checkpoint_interval=500 \",shellscript,content +3815,7145048,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1122,92," --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \",shellscript,content +3816,7145048,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1071,18,"",shellscript,content +3817,7145048,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1039,0,"echo Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\n",shellscript,content +3818,7145048,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",865,156,"",shellscript,content +3819,7145048,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",727,111,"CHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id",shellscript,content +3820,7145048,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",551,121,"tf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'",shellscript,content +3821,7145048,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",149,281,"#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_110M\n#SBATCH --mail-type=ALL",shellscript,content +3822,7145048,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",21,69,"#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00",shellscript,content +3823,7168303,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",21,0,"#SBATCH --nodes=2\n",shellscript,content +3824,7168330,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",39,0,"#SBATCH --ntasks-per-node=4\n",shellscript,content +3825,7168388,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",67,0,"#SBATCH --time=36:00:00\n",shellscript,content +3826,7168401,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",91,0,"#SBATCH --partition=accelerated\n",shellscript,content +3827,7168462,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",123,102,"",shellscript,content +3828,7168468,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",149,0,"#SBATCH --gres=gpu:4\n",shellscript,content +3829,7168658,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",170,0,"#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n",shellscript,content +3830,7168764,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",269,0,"#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n",shellscript,content +3831,7169094,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",367,0,"#SBATCH --job-name=train_dynamics_modelsize_scaling_110M_2_node\n",shellscript,content +3832,7169094,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",431,0,"\n",shellscript,content +3833,7169248,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",432,0,"# Log the sbatch script\n",shellscript,content +3834,7169249,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",432,273,"",shellscript,content +3835,7169362,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",551,0,"array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n",shellscript,content +3836,7169362,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",673,116,"",shellscript,content +3837,7169776,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",727,0,"CHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\n",shellscript,content +3838,7169828,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",839,59,"",shellscript,content +3839,7170110,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",865,0,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n",shellscript,content +3840,7170111,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1020,17,"",shellscript,content +3841,7170146,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1021,0,"env | grep SLURM\n\n",shellscript,content +3842,7170184,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1039,0,"srun python train_dynamics.py \\n",shellscript,content +3843,7170185,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1071,0," --save_ckpt \\n",shellscript,content +3844,7170252,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1089,334,"",shellscript,content +3845,7170308,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1122,0," --batch_size=96 \\n",shellscript,content +3846,7170322,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1144,0," --min_lr=0 \\n",shellscript,content +3847,7170414,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1161,0," --max_lr=1.5e-5 \\n",shellscript,content +3848,7170441,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1183,0," --log_image_interval=1000 \\n",shellscript,content +3849,7170463,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1215,0," --log \\n",shellscript,content +3850,7170548,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1227,0," --log_checkpoint_interval=1000 \\n",shellscript,content +3851,7170629,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1264,0," --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n",shellscript,content +3852,7170732,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1323,0," --tags dynamics modelsize-scaling 110M \\n",shellscript,content +3853,7170764,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1368,0," --entity instant-uv \\n",shellscript,content +3854,7170765,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1394,0," --project jafar \\n",shellscript,content +3855,7170828,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1416,319,"",shellscript,content +3856,7170863,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1465,0," --data_dir $array_records_dir \\n",shellscript,content +3857,7170919,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1501,70,"",shellscript,content +3858,7180847,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +3859,7181830,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +3860,7220899,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1038,0,"",shellscript,selection_mouse +3861,7441908,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +3862,7442639,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1548,0,"",shellscript,selection_mouse +3863,7442640,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1547,0,"",shellscript,selection_command +3864,7442799,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1548,0,"",shellscript,selection_mouse +3865,7442800,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1547,0,"",shellscript,selection_command +3866,7442904,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1547,1,"\",shellscript,selection_mouse +3867,7442907,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1548,0,"",shellscript,selection_command +3868,7442936,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1521,27,"\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3869,7442956,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1500,48,"\n --dyna_dim=768 \\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3870,7443015,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1415,133,"\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3871,7443016,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1353,195,"scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3872,7443016,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1233,315,"log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3873,7443027,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1189,359,"log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3874,7443043,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1182,366,"\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3875,7443102,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1160,388,"\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3876,7443103,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1142,406,"\\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3877,7443103,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1141,407," \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3878,7443109,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1139,409,"96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3879,7443125,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1105,443,"CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3880,7443188,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1088,460,"\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \",shellscript,selection_mouse +3881,7443330,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1088,0,"",shellscript,selection_mouse +3882,7443347,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1087,0,"",shellscript,selection_command +3883,7443514,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1088,0,"",shellscript,selection_mouse +3884,7443525,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1087,0,"",shellscript,selection_command +3885,7443687,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1071,18," --save_ckpt \\n",shellscript,selection_mouse +3886,7443697,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,17," --save_ckpt \\n",shellscript,selection_command +3887,7443816,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,0,"",shellscript,selection_mouse +3888,7443829,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,50," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n",shellscript,selection_mouse +3889,7443845,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,111," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n",shellscript,selection_mouse +3890,7443866,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,143," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n",shellscript,selection_mouse +3891,7443929,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,192," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n",shellscript,selection_mouse +3892,7443930,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,251," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n",shellscript,selection_mouse +3893,7443930,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,296," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n",shellscript,selection_mouse +3894,7443930,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,322," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n",shellscript,selection_mouse +3895,7443989,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,344," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n",shellscript,selection_mouse +3896,7443989,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,393," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n",shellscript,selection_mouse +3897,7443993,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,429," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,selection_mouse +3898,7444049,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,450," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n",shellscript,selection_mouse +3899,7444059,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,477," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \\n",shellscript,selection_mouse +3900,7444142,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1072,501," --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=12\n",shellscript,selection_mouse +3901,7445365,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1573,0,"",shellscript,selection_mouse +3902,7448267,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",915,0,"",shellscript,selection_mouse +3903,7449036,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1458,0,"",shellscript,selection_mouse +3904,7454524,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_180M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-180M-$slurm_job_id \\n --tags dynamics model-size-scaling 180M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16\n",shellscript,tab +3905,7455920,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1686,0,"",shellscript,selection_mouse +3906,7456109,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1209,477,"5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-180M-$slurm_job_id \\n --tags dynamics model-size-scaling 180M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16\n",shellscript,selection_mouse +3907,7456131,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",808,878,"echo Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-180M-$slurm_job_id \\n --tags dynamics model-size-scaling 180M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16\n",shellscript,selection_mouse +3908,7456163,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",651,1035,"\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-180M-$slurm_job_id \\n --tags dynamics model-size-scaling 180M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16\n",shellscript,selection_mouse +3909,7456179,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",415,1271,"\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-180M-$slurm_job_id \\n --tags dynamics model-size-scaling 180M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16\n",shellscript,selection_mouse +3910,7456192,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",67,1619,"#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_180M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-180M-$slurm_job_id \\n --tags dynamics model-size-scaling 180M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16\n",shellscript,selection_mouse +3911,7456204,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",0,1686,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_180M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-180M-$slurm_job_id \\n --tags dynamics model-size-scaling 180M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16\n",shellscript,selection_mouse +3912,7458809,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",0,0,"",shellscript,selection_command +3913,7475044,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",21,0,"#SBATCH --nodes=2\n",shellscript,content +3914,7475116,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",39,0,"#SBATCH --ntasks-per-node=4\n",shellscript,content +3915,7475131,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",67,0,"#SBATCH --time=36:00:00\n",shellscript,content +3916,7475164,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",91,0,"#SBATCH --partition=accelerated\n",shellscript,content +3917,7475220,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",123,102,"",shellscript,content +3918,7475226,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",149,0,"#SBATCH --gres=gpu:4\n",shellscript,content +3919,7475442,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",170,0,"#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n",shellscript,content +3920,7475621,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",269,0,"#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n",shellscript,content +3921,7475744,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",367,0,"#SBATCH --job-name=train_dynamics_modelsize_scaling_180M_2_node\n",shellscript,content +3922,7475745,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",431,0,"\n",shellscript,content +3923,7475762,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",432,0,"# Log the sbatch script\n",shellscript,content +3924,7475823,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",432,291,"",shellscript,content +3925,7476137,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",551,0,"array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n",shellscript,content +3926,7476138,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",673,116,"",shellscript,content +3927,7476453,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",727,0,"CHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\n",shellscript,content +3928,7476472,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",839,59,"",shellscript,content +3929,7476901,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",865,0,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n",shellscript,content +3930,7476901,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1020,17,"",shellscript,content +3931,7476948,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1021,0,"env | grep SLURM\n\n",shellscript,content +3932,7476981,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1039,0,"srun python train_dynamics.py \\n",shellscript,content +3933,7476982,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1071,0," --save_ckpt \\n",shellscript,content +3934,7477059,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1089,334,"",shellscript,content +3935,7477104,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1122,0," --batch_size=96 \\n",shellscript,content +3936,7477168,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1144,0," --min_lr=0 \\n",shellscript,content +3937,7477226,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1161,0," --max_lr=1.5e-5 \\n",shellscript,content +3938,7477372,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1183,0," --log_image_interval=1000 \\n",shellscript,content +3939,7477387,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1215,0," --log \\n",shellscript,content +3940,7477421,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1227,0," --log_checkpoint_interval=1000 \\n",shellscript,content +3941,7477490,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1264,0," --name=dynamics-modelsize-scaling-180M-$slurm_job_id \\n",shellscript,content +3942,7477548,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1323,0," --tags dynamics modelsize-scaling 180M \\n",shellscript,content +3943,7477616,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1368,0," --entity instant-uv \\n",shellscript,content +3944,7477624,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1394,0," --project jafar \\n",shellscript,content +3945,7477683,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1416,319,"",shellscript,content +3946,7477787,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1465,0," --data_dir $array_records_dir \\n",shellscript,content +3947,7477804,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1501,70,"",shellscript,content +3948,7493303,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_270M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-270M-$slurm_job_id \\n --tags dynamics model-size-scaling 270M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=16\n",shellscript,tab +3949,7494794,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1686,0,"",shellscript,selection_mouse +3950,7494922,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1655,31,"s=24 \\n --dyna_num_heads=16\n",shellscript,selection_mouse +3951,7494943,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",813,873,"Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-270M-$slurm_job_id \\n --tags dynamics model-size-scaling 270M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=16\n",shellscript,selection_mouse +3952,7495000,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",651,1035,"\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-270M-$slurm_job_id \\n --tags dynamics model-size-scaling 270M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=16\n",shellscript,selection_mouse +3953,7495001,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",415,1271,"\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-270M-$slurm_job_id \\n --tags dynamics model-size-scaling 270M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=16\n",shellscript,selection_mouse +3954,7495002,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",67,1619,"#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_270M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-270M-$slurm_job_id \\n --tags dynamics model-size-scaling 270M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=16\n",shellscript,selection_mouse +3955,7495011,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",0,1686,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_270M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-270M-$slurm_job_id \\n --tags dynamics model-size-scaling 270M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=16\n",shellscript,selection_mouse +3956,7496593,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",0,0,"",shellscript,selection_command +3957,7506974,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",21,0,"#SBATCH --nodes=2\n",shellscript,content +3958,7507016,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",39,0,"#SBATCH --ntasks-per-node=4\n",shellscript,content +3959,7507032,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",67,0,"#SBATCH --time=36:00:00\n",shellscript,content +3960,7507032,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",91,0,"#SBATCH --partition=accelerated\n",shellscript,content +3961,7507063,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",123,102,"",shellscript,content +3962,7507084,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",149,0,"#SBATCH --gres=gpu:4\n",shellscript,content +3963,7507286,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",170,0,"#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n",shellscript,content +3964,7507829,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",269,0,"#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n",shellscript,content +3965,7507983,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",367,0,"#SBATCH --job-name=train_dynamics_modelsize_scaling_270M_2_node\n",shellscript,content +3966,7507984,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",431,0,"\n",shellscript,content +3967,7508080,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",432,0,"# Log the sbatch script\n",shellscript,content +3968,7508118,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",432,291,"",shellscript,content +3969,7508374,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",551,0,"array_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n",shellscript,content +3970,7508375,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",673,116,"",shellscript,content +3971,7508591,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",727,0,"CHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/checkpoints/$job_name/$slurm_job_id\n",shellscript,content +3972,7508642,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",839,59,"",shellscript,content +3973,7508919,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",865,0,"tokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n",shellscript,content +3974,7508920,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1020,17,"",shellscript,content +3975,7508990,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1021,0,"env | grep SLURM\n\n",shellscript,content +3976,7508992,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1039,0,"srun python train_dynamics.py \\n",shellscript,content +3977,7509017,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1071,0," --save_ckpt \\n",shellscript,content +3978,7509043,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1089,334,"",shellscript,content +3979,7509100,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1122,0," --batch_size=96 \\n",shellscript,content +3980,7509245,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1144,0," --min_lr=0 \\n",shellscript,content +3981,7509246,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1161,0," --max_lr=1.5e-5 \\n",shellscript,content +3982,7509309,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1183,0," --log_image_interval=1000 \\n",shellscript,content +3983,7509310,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1215,0," --log \\n",shellscript,content +3984,7509380,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1227,0," --log_checkpoint_interval=1000 \\n",shellscript,content +3985,7509473,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1264,0," --name=dynamics-modelsize-scaling-270M-$slurm_job_id \\n",shellscript,content +3986,7509505,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1323,0," --tags dynamics modelsize-scaling 270M \\n",shellscript,content +3987,7509527,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1368,0," --entity instant-uv \\n",shellscript,content +3988,7509554,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1394,0," --project jafar \\n",shellscript,content +3989,7509654,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1416,319,"",shellscript,content +3990,7509677,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1465,0," --data_dir $array_records_dir \\n",shellscript,content +3991,7509737,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",1501,70,"",shellscript,content +3992,7548377,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_500M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-500M-$slurm_job_id \\n --tags dynamics model-size-scaling 500M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1536 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=24\n",shellscript,tab +3993,7549740,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",415,0,"",shellscript,selection_mouse +3994,7549857,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",414,1,"\n",shellscript,selection_mouse +3995,7549901,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",412,3,"LL\n",shellscript,selection_mouse +3996,7549924,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",409,6,"e=ALL\n",shellscript,selection_mouse +3997,7549939,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",405,10,"-type=ALL\n",shellscript,selection_mouse +3998,7549954,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",403,12,"il-type=ALL\n",shellscript,selection_mouse +3999,7549970,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",401,14,"mail-type=ALL\n",shellscript,selection_mouse +4000,7549997,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",398,17," --mail-type=ALL\n",shellscript,selection_mouse +4001,7550010,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",397,18,"H --mail-type=ALL\n",shellscript,selection_mouse +4002,7550025,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",396,19,"CH --mail-type=ALL\n",shellscript,selection_mouse +4003,7550041,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",378,37,"CH --mem=50G\n#SBATCH --mail-type=ALL\n",shellscript,selection_mouse +4004,7550098,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",377,38,"TCH --mem=50G\n#SBATCH --mail-type=ALL\n",shellscript,selection_mouse +4005,7550983,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",409,0,"",shellscript,selection_mouse +4006,7550984,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",406,4,"type",shellscript,selection_mouse +4007,7551127,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",391,24,"#SBATCH --mail-type=ALL\n",shellscript,selection_mouse +4008,7614614,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",437,0,"",shellscript,selection_mouse +4009,7616835,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",0,0,"",shellscript,tab +4010,7619170,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,0,"",shellscript,tab +4011,7621454,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",1686,0,"",shellscript,selection_mouse +4012,7621631,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",1658,28,"4 \\n --dyna_num_heads=24\n",shellscript,selection_mouse +4013,7621650,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",1252,434,"ps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-500M-$slurm_job_id \\n --tags dynamics model-size-scaling 500M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1536 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=24\n",shellscript,selection_mouse +4014,7621665,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",807,879,"\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-500M-$slurm_job_id \\n --tags dynamics model-size-scaling 500M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1536 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=24\n",shellscript,selection_mouse +4015,7621693,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",705,981,"CHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-500M-$slurm_job_id \\n --tags dynamics model-size-scaling 500M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1536 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=24\n",shellscript,selection_mouse +4016,7621748,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",447,1239,"\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-500M-$slurm_job_id \\n --tags dynamics model-size-scaling 500M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1536 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=24\n",shellscript,selection_mouse +4017,7621749,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",149,1537,"#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_500M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-500M-$slurm_job_id \\n --tags dynamics model-size-scaling 500M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1536 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=24\n",shellscript,selection_mouse +4018,7621770,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,1686,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_500M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=$ws_dir/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-500M-$slurm_job_id \\n --tags dynamics model-size-scaling 500M \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=1536 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=24\n",shellscript,selection_mouse +4019,7622826,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,0,"",shellscript,selection_command +4020,7626998,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,0,"#SBATCH --nodes=2\n",shellscript,content +4021,7627083,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",39,0,"#SBATCH --ntasks-per-node=4\n",shellscript,content +4022,7627116,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",67,0,"#SBATCH --time=36:00:00\n",shellscript,content +4023,7630153,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,70,"",shellscript,content +4024,7637479,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,0,"#SBATCH --nodes=2\n",shellscript,content +4025,7637511,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",39,0,"#SBATCH --ntasks-per-node=4\n",shellscript,content +4026,7637570,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",67,0,"#SBATCH --time=36:00:00\n",shellscript,content +4027,7637632,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",91,0,"#SBATCH --partition=accelerated\n",shellscript,content +4028,7637689,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",123,102,"",shellscript,content +4029,7637826,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",149,0,"#SBATCH --gres=gpu:4\n",shellscript,content +4030,7637939,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",170,0,"#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n",shellscript,content +4031,7638199,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",269,0,"#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n",shellscript,content +4032,7638311,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",367,0,"#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node\n",shellscript,content +4033,7638314,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",0,0,"",shellscript,tab +4034,7640612,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,0,"",shellscript,tab +4035,7640670,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",1532,0," --data_dir $array_records_dir \\n",shellscript,content +4036,7640708,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",1568,70,"",shellscript,content +4037,7665633,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",1049,0,"",shellscript,selection_mouse +4038,7666910,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",1020,67,"",shellscript,content +4039,7669452,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +4040,7674182,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +4041,7676592,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +4042,7678310,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +4043,7681864,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",0,0,"",shellscript,tab +4044,7683752,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",0,0,"",shellscript,tab +4045,7685767,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,0,"",shellscript,tab +4046,7688775,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",0,0,"",shellscript,tab +4047,7689800,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",0,0,"",shellscript,tab +4048,7691317,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +4049,7693581,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +4050,7714174,"TERMINAL",0,0,"bash",,terminal_focus +4051,7876392,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1159,0,"",shellscript,selection_mouse +4052,7876393,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1158,0,"",shellscript,selection_command +4053,7877003,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1497,0,"",shellscript,selection_mouse +4054,7877004,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1496,0,"",shellscript,selection_command +4055,7880408,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +4056,7882942,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",0,0,"",shellscript,tab +4057,7883369,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",0,0,"",shellscript,tab +4058,7883760,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,0,"",shellscript,tab +4059,7884897,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +4060,7886756,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,0,"",shellscript,tab +4061,7891084,"TERMINAL",0,0,"srun",,terminal_focus +4062,7891292,"TERMINAL",0,0,"\r(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +4063,7896085,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +4064,7896220,"TERMINAL",0,0,"\r(jafar) [tum_cte0515@hkn0425 jafar]$ ",,terminal_output +4065,7901167,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +4066,7903930,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +4067,7922806,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",66,0,"",shellscript,selection_mouse +4068,7922807,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",65,0,"",shellscript,selection_command +4069,7923758,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",66,0,"",shellscript,selection_command +4070,7923875,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",65,1,"",shellscript,content +4071,7924231,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",65,0,"6",shellscript,content +4072,7924232,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",66,0,"",shellscript,selection_keyboard +4073,7925782,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",0,0,"",shellscript,tab +4074,8020604,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",0,0,"",shellscript,tab +4075,8020605,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1261,0,"",shellscript,selection_mouse +4076,8044084,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1317,0,"",shellscript,selection_mouse +4077,8053578,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +4078,8055194,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",0,0,"",shellscript,tab +4079,8079961,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",62,0,"",shellscript,selection_mouse +4080,8080719,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",66,0,"",shellscript,selection_mouse +4081,8080723,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",65,0,"",shellscript,selection_command +4082,8082092,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",37,0,"",shellscript,selection_command +4083,8082411,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",38,0,"",shellscript,selection_command +4084,8083096,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",37,1,"",shellscript,content +4085,8083402,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",37,0,"1",shellscript,content +4086,8083403,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",38,0,"",shellscript,selection_keyboard +4087,8083466,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",38,0,"2",shellscript,content +4088,8083467,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",39,0,"",shellscript,selection_keyboard +4089,8085709,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",0,0,"",shellscript,tab +4090,8132934,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",66,0,"",shellscript,selection_mouse +4091,8132941,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",65,0,"",shellscript,selection_command +4092,8134443,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",66,0,"",shellscript,selection_command +4093,8134588,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",65,1,"",shellscript,content +4094,8135101,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",65,0,"1",shellscript,content +4095,8135101,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",66,0,"",shellscript,selection_keyboard +4096,8135169,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",66,0,"6",shellscript,content +4097,8135170,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",67,0,"",shellscript,selection_keyboard +4098,8138603,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,0,"",shellscript,tab +4099,8183993,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",90,0,"",shellscript,selection_mouse +4100,8183998,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",89,0,"",shellscript,selection_command +4101,8184819,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",66,0,"",shellscript,selection_mouse +4102,8184820,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",65,0,"",shellscript,selection_command +4103,8186230,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",38,0,"",shellscript,selection_mouse +4104,8186231,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",37,0,"",shellscript,selection_command +4105,8187258,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",37,0,"3",shellscript,content +4106,8187259,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",38,0,"",shellscript,selection_keyboard +4107,8188720,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",85,0,"",shellscript,selection_mouse +4108,8188861,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",83,2,"36",shellscript,selection_mouse +4109,8189078,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",83,5,"36:00",shellscript,selection_mouse +4110,8189092,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",83,6,"36:00:",shellscript,selection_mouse +4111,8189150,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",83,8,"36:00:00",shellscript,selection_mouse +4112,8189344,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",91,0,"",shellscript,selection_mouse +4113,8189531,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",89,2,"00",shellscript,selection_mouse +4114,8189727,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",88,3,":00",shellscript,selection_mouse +4115,8189758,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",56,35,"-per-node=4\n#SBATCH --time=36:00:00",shellscript,selection_mouse +4116,8189763,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",50,41,"ntasks-per-node=4\n#SBATCH --time=36:00:00",shellscript,selection_mouse +4117,8189822,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",49,42,"-ntasks-per-node=4\n#SBATCH --time=36:00:00",shellscript,selection_mouse +4118,8189826,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",48,43,"--ntasks-per-node=4\n#SBATCH --time=36:00:00",shellscript,selection_mouse +4119,8189843,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",47,44," --ntasks-per-node=4\n#SBATCH --time=36:00:00",shellscript,selection_mouse +4120,8189868,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",28,63," --nodes=32\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00",shellscript,selection_mouse +4121,8189926,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",22,69,"SBATCH --nodes=32\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00",shellscript,selection_mouse +4122,8190496,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",23,0,"",shellscript,selection_mouse +4123,8190497,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",22,6,"SBATCH",shellscript,selection_mouse +4124,8190782,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,19,"#SBATCH --nodes=32\n",shellscript,selection_mouse +4125,8190972,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,47,"#SBATCH --nodes=32\n#SBATCH --ntasks-per-node=4\n",shellscript,selection_mouse +4126,8191033,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,71,"#SBATCH --nodes=32\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n",shellscript,selection_mouse +4127,8191033,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,129,"#SBATCH --nodes=32\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n",shellscript,selection_mouse +4128,8191045,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,150,"#SBATCH --nodes=32\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n",shellscript,selection_mouse +4129,8191103,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,249,"#SBATCH --nodes=32\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n",shellscript,selection_mouse +4130,8191160,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,347,"#SBATCH --nodes=32\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n",shellscript,selection_mouse +4131,8191220,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,411,"#SBATCH --nodes=32\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node\n",shellscript,selection_mouse +4132,8191276,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,412,"#SBATCH --nodes=32\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node\n\n",shellscript,selection_mouse +4133,8191333,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",21,411,"#SBATCH --nodes=32\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=36:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node\n",shellscript,selection_mouse +4134,8191616,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",431,0,"",shellscript,selection_mouse +4135,8192000,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",387,44,"train_dynamics_modelsize_scaling_500M_2_node",shellscript,selection_mouse +4136,8192228,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",307,124,"scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node",shellscript,selection_mouse +4137,8192289,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",297,134,"workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node",shellscript,selection_mouse +4138,8192394,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",296,135,"/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node",shellscript,selection_mouse +4139,8192412,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",194,237,"work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node",shellscript,selection_mouse +4140,8192469,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",193,238,"/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node",shellscript,selection_mouse +4141,8192470,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",189,242,"hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node",shellscript,selection_mouse +4142,8192470,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",168,263,":4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node",shellscript,selection_mouse +4143,8192477,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",165,266,"gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node",shellscript,selection_mouse +4144,8192495,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",134,297,"cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node",shellscript,selection_mouse +4145,8192556,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",102,329,"partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node",shellscript,selection_mouse +4146,8192561,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",101,330,"-partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_2_node",shellscript,selection_mouse +4147,8192919,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",101,0,"",shellscript,selection_mouse +4148,8193674,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",431,0,"",shellscript,selection_mouse +4149,8193772,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",430,1,"e",shellscript,selection_mouse +4150,8193787,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",428,3,"ode",shellscript,selection_mouse +4151,8193801,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",427,4,"node",shellscript,selection_mouse +4152,8193856,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",426,5,"_node",shellscript,selection_mouse +4153,8193913,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",425,6,"2_node",shellscript,selection_mouse +4154,8193969,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",424,7,"_2_node",shellscript,selection_mouse +4155,8195684,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",425,0,"",shellscript,selection_command +4156,8196491,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",426,0,"",shellscript,selection_command +4157,8200061,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",425,0,"",shellscript,selection_command +4158,8200278,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",425,0,"3",shellscript,content +4159,8200279,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",426,0,"",shellscript,selection_keyboard +4160,8201587,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",0,0,"",shellscript,tab +4161,8202943,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",426,0,"",shellscript,selection_mouse +4162,8208024,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",67,0,"",shellscript,selection_mouse +4163,8209765,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",66,1,"",shellscript,content +4164,8209901,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",65,1,"",shellscript,content +4165,8210009,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",65,0,"4",shellscript,content +4166,8210010,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",66,0,"",shellscript,selection_keyboard +4167,8210137,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",38,0,"",shellscript,selection_command +4168,8210489,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",37,1,"",shellscript,content +4169,8210711,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",37,0,"1",shellscript,content +4170,8210712,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",38,0,"",shellscript,selection_keyboard +4171,8211546,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",38,0,"6",shellscript,content +4172,8211547,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",39,0,"",shellscript,selection_keyboard +4173,8213713,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,0,"",shellscript,tab +4174,8215522,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",0,0,"",shellscript,tab +4175,8217653,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",427,0,"",shellscript,selection_mouse +4176,8218418,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",426,0,"",shellscript,selection_command +4177,8218940,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",425,1,"",shellscript,content +4178,8219123,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",425,0,"1",shellscript,content +4179,8219123,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",426,0,"",shellscript,selection_keyboard +4180,8219381,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",426,0,"6",shellscript,content +4181,8219382,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",427,0,"",shellscript,selection_keyboard +4182,8224816,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",0,0,"",shellscript,tab +4183,8228584,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",425,0,"",shellscript,selection_mouse +4184,8229202,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",425,0,"1",shellscript,content +4185,8229203,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",426,0,"",shellscript,selection_keyboard +4186,8230542,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"",shellscript,tab +4187,8233348,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",65,1,"",shellscript,content +4188,8233498,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",65,0,"2",shellscript,content +4189,8233499,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",66,0,"",shellscript,selection_keyboard +4190,8233796,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",38,0,"",shellscript,selection_command +4191,8234848,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",37,1,"",shellscript,content +4192,8234944,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",37,0,"6",shellscript,content +4193,8234945,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",38,0,"",shellscript,selection_keyboard +4194,8235625,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",56,0,"",shellscript,selection_command +4195,8235990,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",66,0,"",shellscript,selection_mouse +4196,8236628,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",65,1,"",shellscript,content +4197,8236685,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",65,0,"4",shellscript,content +4198,8236686,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",66,0,"",shellscript,selection_keyboard +4199,8238142,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",425,0,"",shellscript,selection_mouse +4200,8239019,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",424,1,"",shellscript,content +4201,8239079,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",424,0,"6",shellscript,content +4202,8239080,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",425,0,"",shellscript,selection_keyboard +4203,8240781,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"",shellscript,tab +4204,8243066,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",0,0,"",shellscript,tab +4205,8244948,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",0,0,"",shellscript,tab +4206,8246457,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,0,"",shellscript,tab diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c2f47c74-4812-499a-abea-eafb3b83db691752315370022-2025_07_12-12.17.39.814/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c2f47c74-4812-499a-abea-eafb3b83db691752315370022-2025_07_12-12.17.39.814/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..9d497980eae73516ede25e25123071861da5fa5e --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-c2f47c74-4812-499a-abea-eafb3b83db691752315370022-2025_07_12-12.17.39.814/source.csv @@ -0,0 +1,1472 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +2,744,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"12:17:39 PM [info] Activating crowd-code\n12:17:39 PM [info] Recording started\n12:17:39 PM [info] Initializing git provider using file system watchers...\n12:17:40 PM [info] Git repository found\n12:17:40 PM [info] Git provider initialized successfully\n",Log,tab +3,970,"extension-output-pdoom-org.crowd-code-#1-crowd-code",250,0,"12:17:40 PM [info] Initial git state: [object Object]\n",Log,content +4,1751,"weekend-job-starter.sh",0,0,"sbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch\n\n\n\nsbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch\nsbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch\n\n\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch\n\n\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",shellscript,tab +5,4090,"TERMINAL",0,0,"/bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt",,terminal_command +6,4106,"TERMINAL",0,0,"]633;E;2025-07-12 12:17:43 /bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt;09e26284-9196-4e36-8df8-ec6826b3091e]633;C",,terminal_output +7,4139,"TERMINAL",0,0,"]0;tum_cte0515@hkn1991:/hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash]633;D;0",,terminal_output +8,5692,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=24\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-modelsize-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-modelsize-scaling/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_500M_32_node\n#SBATCH --mem=400G\n\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-500M-$slurm_job_id \\n --tags dynamics modelsize-scaling 500M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=1536 \\n --dyna_num_blocks=24 \\n --dyna_num_heads=24\n",shellscript,tab +9,24942,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer-lr-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer-lr-scaling/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_5e-5\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=5e-5 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-5e-5-$slurm_job_id \\n --tags tokenizer lr-sweep 5e-5 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +10,28316,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,tab +11,29924,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer-lr-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big_run/tokenizer-lr-scaling/%x_%j.log\n#SBATCH --job-name=train_tokenizer_lr_sweep_1e-4\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\nsrun python train_tokenizer.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log_checkpoint_interval=1000 \\n --log \\n --name=tokenizer-lr-sweep-1e-4-$slurm_job_id \\n --tags tokenizer lr-sweep 1e-4 \\n --entity instant-uv \\n --project jafar \\n --data_dir $array_records_dir\n",shellscript,tab +12,31579,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +13,41023,"train_tokenizer.py",0,0,"from dataclasses import dataclass, field\nimport os\n\nimport einops\nfrom flax.training import orbax_utils\nfrom flax.training.train_state import TrainState\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\n\nfrom models.tokenizer import TokenizerVQVAE\nfrom utils.dataloader import get_dataloader\nfrom utils.parameter_utils import count_parameters_by_component\n\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 300_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n vq_beta: float = 0.25\n batch_size: int = 48\n min_lr: float = 0.0\n max_lr: float = 3e-4\n warmup_steps: int = 10000\n # Tokenizer\n model_dim: int = 512\n latent_dim: int = 32\n num_latents: int = 1024\n patch_size: int = 4\n num_blocks: int = 8\n num_heads: int = 8\n dropout: float = 0.0\n codebook_dropout: float = 0.01\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_tokenizer""\n tags: list[str] = field(default_factory=lambda: [""tokenizer""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 10000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n\n\nargs = tyro.cli(Args)\n\n\ndef tokenizer_loss_fn(params, state, inputs):\n # --- Compute loss ---\n outputs = state.apply_fn(\n params,\n inputs,\n training=True,\n rngs={""params"": inputs[""rng""], ""dropout"": inputs[""dropout_rng""]},\n )\n mse = jnp.square(inputs[""videos""] - outputs[""recon""]).mean()\n q_loss = jnp.square(jax.lax.stop_gradient(outputs[""emb""]) - outputs[""z""]).mean()\n commitment_loss = jnp.square(\n outputs[""emb""] - jax.lax.stop_gradient(outputs[""z""])\n ).mean()\n loss = mse + q_loss + args.vq_beta * commitment_loss\n\n # --- Compute validation metrics ---\n gt = inputs[""videos""].clip(0, 1).reshape(-1, *inputs[""videos""].shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = pix.psnr(gt, recon).mean() # type: ignore\n ssim = pix.ssim(gt, recon).mean() # type: ignore\n _, index_counts = jnp.unique_counts(\n jnp.ravel(outputs[""indices""]), size=args.num_latents, fill_value=0\n )\n codebook_usage = (index_counts != 0).mean()\n metrics = dict(\n loss=loss,\n mse=mse,\n q_loss=q_loss,\n commitment_loss=commitment_loss,\n psnr=psnr,\n ssim=ssim,\n codebook_usage=codebook_usage,\n )\n return loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n grad_fn = jax.value_and_grad(tokenizer_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""encoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""encoder""]\n )\n metrics[""vq_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""vq""]\n )\n metrics[""decoder_gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""decoder""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n\n # --- Initialize model ---\n tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.model_dim,\n latent_dim=args.latent_dim,\n num_latents=args.num_latents,\n patch_size=args.patch_size,\n num_blocks=args.num_blocks,\n num_heads=args.num_heads,\n dropout=args.dropout,\n codebook_dropout=args.codebook_dropout,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_height, args.image_width, args.image_channels)\n inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=jnp.float32,\n ),\n )\n init_params = tokenizer.init(_rng, inputs)\n\n param_counts = count_parameters_by_component(init_params)\n\n if args.log and jax.process_index() == 0:\n wandb.init(\n entity=args.entity,\n project=args.project,\n name=args.name,\n tags=args.tags,\n group=""debug"",\n config=args,\n )\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = optax.warmup_cosine_decay_schedule(\n args.min_lr, args.max_lr, args.warmup_steps, args.num_steps\n )\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4)\n train_state = TrainState.create(apply_fn=tokenizer.apply, params=init_params, tx=tx)\n\n # FIXME: switch to create_hybrid_device_mesh for runs spanning multiple nodes\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(\n mesh, PartitionSpec(""data"", None, None, None, None)\n )\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add('model_state', ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler)\n handler_registry.add('model_state', ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler)\n handler_registry.add('dataloader_state', grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add('dataloader_state', grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler) # type: ignore\n \n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n \n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n \n # --- Restore checkpoint ---\n if args.restore_ckpt:\n abstract_train_state = jax.tree_util.tree_map(ocp.utils.to_shape_dtype_struct, train_state)\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_train_state),\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator),\n )\n )\n train_state = restored[""model_state""]\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n\n # --- TRAIN LOOP ---\n dataloader = (jax.make_array_from_process_local_data(videos_sharding, elem) for elem in grain_iterator) # type: ignore\n print(f""Starting training from step {step}..."")\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _rng_dropout = jax.random.split(rng, 3)\n\n inputs = dict(videos=videos, rng=_rng, dropout_rng=_rng_dropout)\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0]\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n # NOTE: Process-dependent control flow deliberately happens\n # after indexing operation since it must not contain code\n # sections that lead to cross-accelerator communication.\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[0])),\n recon=wandb.Image(np.asarray(recon_seq[0])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.StandardSave(train_state),\n dataloader_state=grain.checkpoint.CheckpointSave(grain_iterator),\n )\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()",python,tab +14,44707,"train_tokenizer.py",0,0,"",python,tab +15,51914,"train_tokenizer.py",835,0,"",python,selection_mouse +16,54372,"train_tokenizer.py",834,0,"",python,selection_mouse +17,55084,"train_tokenizer.py",7724,0,"",python,selection_mouse +18,55227,"train_tokenizer.py",7718,14,"grain_iterator",python,selection_mouse +19,56178,"train_tokenizer.py",7849,0,"",python,selection_mouse +20,84805,"train_tokenizer.py",8395,0,"",python,selection_mouse +21,84989,"train_tokenizer.py",8387,18,"checkpoint_manager",python,selection_mouse +22,161335,"train_tokenizer.py",7847,0,"",python,selection_mouse +23,162922,"train_tokenizer.py",838,0,"",python,selection_mouse +24,163105,"train_tokenizer.py",830,12,"restore_ckpt",python,selection_mouse +25,165844,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +26,167056,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",987,0,"",shellscript,selection_mouse +27,167059,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",986,0,"",shellscript,selection_command +28,168040,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",987,0,"\n ",shellscript,content +29,169305,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",992,0,"-",shellscript,content +30,169306,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",993,0,"",shellscript,selection_keyboard +31,169440,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",993,0,"-",shellscript,content +32,169441,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",994,0,"",shellscript,selection_keyboard +33,169706,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",994,0,"restore_ckpt",shellscript,content +34,170489,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1006,0," ",shellscript,content +35,170490,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1007,0,"",shellscript,selection_keyboard +36,170772,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1007,0,"\",shellscript,content +37,170773,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1008,0,"",shellscript,selection_keyboard +38,171572,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",1007,0,"",shellscript,selection_command +39,174177,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,tab +40,175892,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",986,0,"",shellscript,selection_mouse +41,176707,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",987,0,"\n --restore_ckpt \",shellscript,content +42,176740,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",992,0,"",shellscript,selection_command +43,185555,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1295,0,"",shellscript,selection_mouse +44,185588,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1294,0,"",shellscript,selection_command +45,186272,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1266,0,"",shellscript,selection_mouse +46,187422,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",1267,0,"",shellscript,selection_mouse +47,197643,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +48,198991,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",0,0,"",shellscript,tab +49,207925,"slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch",0,0,"",shellscript,tab +50,212230,"train_tokenizer.py",0,0,"",python,tab +51,226326,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",0,0,"",shellscript,tab +52,246791,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-batchsize-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-batchsize-scaling/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_1_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-batchsize-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=48 \\n --min_lr=0 \\n --max_lr=1e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-batch-size-scaling-1-node-$slurm_job_id \\n --tags dynamics batch-size-scaling 1-node \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +53,249709,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +54,251192,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1203,0,"",shellscript,selection_mouse +55,252195,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1207,0,"\n --restore_ckpt \",shellscript,content +56,252198,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",1212,0,"",shellscript,selection_command +57,253660,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-batchsize-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-batchsize-scaling/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-batchsize-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-batch-size-scaling-2-node-$slurm_job_id \\n --tags dynamics batch-size-scaling 2-node \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,tab +58,254920,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1203,0,"",shellscript,selection_mouse +59,255672,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1207,0,"\n --restore_ckpt \",shellscript,content +60,255688,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",1212,0,"",shellscript,selection_command +61,257374,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=4\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-batchsize-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-batchsize-scaling/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_4_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-batchsize-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=192 \\n --min_lr=0 \\n --max_lr=2.1e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-batch-size-scaling-4-node-$slurm_job_id \\n --tags dynamics batch-size-scaling 4-node \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +62,258236,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1198,0,"",shellscript,selection_mouse +63,259094,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1207,0,"\n --restore_ckpt \",shellscript,content +64,259105,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",1212,0,"",shellscript,selection_command +65,260974,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=8\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-batchsize-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-batchsize-scaling/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_8_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-batchsize-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=384 \\n --min_lr=0 \\n --max_lr=3e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-batch-size-scaling-8-node-$slurm_job_id \\n --tags dynamics batch-size-scaling 8-node \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +66,261823,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1202,0,"",shellscript,selection_mouse +67,262688,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1207,0,"\n --restore_ckpt \",shellscript,content +68,262694,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch",1212,0,"",shellscript,selection_command +69,264393,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=16\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-batchsize-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-batchsize-scaling/%x_%j.log\n#SBATCH --job-name=train_tokenizer_batch_size_scaling_16_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-batchsize-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=768 \\n --min_lr=0 \\n --max_lr=4.2e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-batch-size-scaling-16-node-$slurm_job_id \\n --tags dynamics batch-size-scaling 16-node \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +70,265205,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1201,0,"",shellscript,selection_mouse +71,265989,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1209,0,"\n --restore_ckpt \",shellscript,content +72,266040,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",1214,0,"",shellscript,selection_command +73,275508,"slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch",0,0,"",shellscript,tab +74,286629,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-actionspace-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-actionspace-scaling/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling_6\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-actionspace-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=6 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-6-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 6 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +75,287973,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1243,0,"",shellscript,selection_mouse +76,289105,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1225,0,"",shellscript,selection_command +77,289728,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1232,0,"\n --restore_ckpt \",shellscript,content +78,289738,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",1237,0,"",shellscript,selection_command +79,291407,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-actionspace-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-actionspace-scaling/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling_8\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-actionspace-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=8 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-8-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 8 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +80,292341,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1224,0,"",shellscript,selection_mouse +81,293623,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1232,0,"\n --restore_ckpt \",shellscript,content +82,293638,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",1237,0,"",shellscript,selection_command +83,295360,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-actionspace-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-actionspace-scaling/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling_12\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-actionspace-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=12 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-12-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 12 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +84,296689,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1229,0,"",shellscript,selection_mouse +85,297658,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1233,0,"\n --restore_ckpt \",shellscript,content +86,297659,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",1238,0,"",shellscript,selection_command +87,299394,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-actionspace-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-actionspace-scaling/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling_20\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-actionspace-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=20 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-20-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 20 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +88,300126,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1226,0,"",shellscript,selection_mouse +89,300872,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1233,0,"\n --restore_ckpt \",shellscript,content +90,300874,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",1238,0,"",shellscript,selection_command +91,302406,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-actionspace-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-actionspace-scaling/%x_%j.log\n#SBATCH --job-name=dynamics_cotraining_action_space_scaling_50\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-actionspace-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --num_latent_actions=50 \\n --log_checkpoint_interval=1000 \\n --name=dynamics-cotraining-action-space-scaling-50-$slurm_job_id \\n --tags dynamics cotraining action-space-scaling 50 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n",shellscript,tab +92,303120,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1226,0,"",shellscript,selection_mouse +93,303905,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1233,0,"\n --restore_ckpt \",shellscript,content +94,303907,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch",1238,0,"",shellscript,selection_command +95,311575,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch",0,0,"",shellscript,tab +96,312842,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch",0,0,"",shellscript,tab +97,313374,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",0,0,"",shellscript,tab +98,313911,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch",0,0,"",shellscript,tab +99,322124,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-modelsize-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-modelsize-scaling/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_36M_2_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-36M-$slurm_job_id \\n --tags dynamics modelsize-scaling 36M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n",shellscript,tab +100,323606,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1204,0,"",shellscript,selection_mouse +101,324324,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1209,0,"\n --restore_ckpt \",shellscript,content +102,324339,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch",1214,0,"",shellscript,selection_command +103,326959,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=6\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-modelsize-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-modelsize-scaling/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_110M_6_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-110M-$slurm_job_id \\n --tags dynamics modelsize-scaling 110M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=768 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=12\n",shellscript,tab +104,328091,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1204,0,"",shellscript,selection_mouse +105,328840,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1210,0,"\n --restore_ckpt \",shellscript,content +106,328842,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch",1215,0,"",shellscript,selection_command +107,331693,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=12\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-modelsize-scaling/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/big-runs/dynamics-cotraining-modelsize-scaling/%x_%j.log\n#SBATCH --job-name=train_dynamics_modelsize_scaling_180M_12_node\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/$job_name\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_tokenizer_batch_size_scaling_16_node/3321526/tokenizer_22000/\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --min_lr=0 \\n --max_lr=1.5e-5 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-modelsize-scaling-180M-$slurm_job_id \\n --tags dynamics modelsize-scaling 180M \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir \\n --dyna_dim=1024 \\n --dyna_num_blocks=16 \\n --dyna_num_heads=16\n",shellscript,tab +108,332506,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1208,0,"",shellscript,selection_mouse +109,333255,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1212,0,"\n --restore_ckpt \",shellscript,content +110,333257,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",1217,0,"",shellscript,selection_command +111,365080,"weekend-job-starter.sh",0,0,"",shellscript,tab +112,366847,"weekend-job-starter.sh",0,0,"",shellscript,tab +113,383942,"TERMINAL",0,0,"cp weekend-job-starter.sh weekend-job-requeuer.sh",,terminal_command +114,383988,"TERMINAL",0,0,"]633;E;2025-07-12 12:24:03 cp weekend-job-starter.sh weekend-job-requeuer.sh;cd331af2-3b46-4f78-aca4-9dda848374ca]633;C",,terminal_output +115,384105,"TERMINAL",0,0,"]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +116,385774,"weekend-job-requeuer.sh",0,0,"sbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch\n\n\n\nsbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch\nsbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch\n\n\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch\n\n\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",shellscript,tab +117,387309,"weekend-job-requeuer.sh",0,0,"",shellscript,tab +118,389212,"weekend-job-requeuer.sh",0,0,"\n",shellscript,content +119,389860,"weekend-job-requeuer.sh",0,0,"\n",shellscript,content +120,390038,"weekend-job-requeuer.sh",1,0,"\n",shellscript,content +121,391215,"weekend-job-requeuer.sh",2,0,"#",shellscript,content +122,391216,"weekend-job-requeuer.sh",3,0,"",shellscript,selection_keyboard +123,391522,"weekend-job-requeuer.sh",3,0," ",shellscript,content +124,391523,"weekend-job-requeuer.sh",4,0,"",shellscript,selection_keyboard +125,393357,"weekend-job-requeuer.sh",4,0,"a",shellscript,content +126,393359,"weekend-job-requeuer.sh",5,0,"",shellscript,selection_keyboard +127,393521,"weekend-job-requeuer.sh",5,0,"c",shellscript,content +128,393522,"weekend-job-requeuer.sh",6,0,"",shellscript,selection_keyboard +129,393745,"weekend-job-requeuer.sh",6,0,"t",shellscript,content +130,393746,"weekend-job-requeuer.sh",7,0,"",shellscript,selection_keyboard +131,393888,"weekend-job-requeuer.sh",7,0,"i",shellscript,content +132,393889,"weekend-job-requeuer.sh",8,0,"",shellscript,selection_keyboard +133,393991,"weekend-job-requeuer.sh",8,0,"o",shellscript,content +134,393992,"weekend-job-requeuer.sh",9,0,"",shellscript,selection_keyboard +135,394142,"weekend-job-requeuer.sh",9,0,"n",shellscript,content +136,394143,"weekend-job-requeuer.sh",10,0,"",shellscript,selection_keyboard +137,394239,"weekend-job-requeuer.sh",10,0,"s",shellscript,content +138,394240,"weekend-job-requeuer.sh",11,0,"",shellscript,selection_keyboard +139,394354,"weekend-job-requeuer.sh",11,0,"p",shellscript,content +140,394355,"weekend-job-requeuer.sh",12,0,"",shellscript,selection_keyboard +141,394455,"weekend-job-requeuer.sh",12,0,"a",shellscript,content +142,394456,"weekend-job-requeuer.sh",13,0,"",shellscript,selection_keyboard +143,394571,"weekend-job-requeuer.sh",13,0,"c",shellscript,content +144,394572,"weekend-job-requeuer.sh",14,0,"",shellscript,selection_keyboard +145,394688,"weekend-job-requeuer.sh",14,0,"e",shellscript,content +146,394689,"weekend-job-requeuer.sh",15,0,"",shellscript,selection_keyboard +147,394805,"weekend-job-requeuer.sh",15,0," ",shellscript,content +148,394806,"weekend-job-requeuer.sh",16,0,"",shellscript,selection_keyboard +149,395038,"weekend-job-requeuer.sh",16,0,"s",shellscript,content +150,395040,"weekend-job-requeuer.sh",17,0,"",shellscript,selection_keyboard +151,395171,"weekend-job-requeuer.sh",17,0,"c",shellscript,content +152,395172,"weekend-job-requeuer.sh",18,0,"",shellscript,selection_keyboard +153,395304,"weekend-job-requeuer.sh",18,0,"a",shellscript,content +154,395305,"weekend-job-requeuer.sh",19,0,"",shellscript,selection_keyboard +155,395404,"weekend-job-requeuer.sh",19,0,"l",shellscript,content +156,395405,"weekend-job-requeuer.sh",20,0,"",shellscript,selection_keyboard +157,395593,"weekend-job-requeuer.sh",20,0,"i",shellscript,content +158,395594,"weekend-job-requeuer.sh",21,0,"",shellscript,selection_keyboard +159,395673,"weekend-job-requeuer.sh",21,0,"n",shellscript,content +160,395674,"weekend-job-requeuer.sh",22,0,"",shellscript,selection_keyboard +161,395859,"weekend-job-requeuer.sh",22,0,"g",shellscript,content +162,395860,"weekend-job-requeuer.sh",23,0,"",shellscript,selection_keyboard +163,396488,"weekend-job-requeuer.sh",22,0,"",shellscript,selection_command +164,397405,"weekend-job-requeuer.sh",44,0,"",shellscript,selection_command +165,397554,"weekend-job-requeuer.sh",152,0,"",shellscript,selection_command +166,397704,"weekend-job-requeuer.sh",260,0,"",shellscript,selection_command +167,397839,"weekend-job-requeuer.sh",368,0,"",shellscript,selection_command +168,398005,"weekend-job-requeuer.sh",475,0,"",shellscript,selection_command +169,398154,"weekend-job-requeuer.sh",562,0,"",shellscript,selection_command +170,398293,"weekend-job-requeuer.sh",563,0,"",shellscript,selection_command +171,398422,"weekend-job-requeuer.sh",564,0,"",shellscript,selection_command +172,398754,"weekend-job-requeuer.sh",564,0,"\n# actionspace scaling",shellscript,content +173,398809,"weekend-job-requeuer.sh",565,0,"",shellscript,selection_command +174,399842,"weekend-job-requeuer.sh",586,0,"",shellscript,selection_command +175,410459,"weekend-job-requeuer.sh",578,0,"",shellscript,selection_command +176,411755,"weekend-job-requeuer.sh",579,7,"",shellscript,content +177,412004,"weekend-job-requeuer.sh",567,12,"",shellscript,content +178,412608,"weekend-job-requeuer.sh",567,0,"l",shellscript,content +179,412609,"weekend-job-requeuer.sh",568,0,"",shellscript,selection_keyboard +180,412691,"weekend-job-requeuer.sh",568,0,"r",shellscript,content +181,412692,"weekend-job-requeuer.sh",569,0,"",shellscript,selection_keyboard +182,412971,"weekend-job-requeuer.sh",569,0," ",shellscript,content +183,412972,"weekend-job-requeuer.sh",570,0,"",shellscript,selection_keyboard +184,413176,"weekend-job-requeuer.sh",570,0,"s",shellscript,content +185,413177,"weekend-job-requeuer.sh",571,0,"",shellscript,selection_keyboard +186,413608,"weekend-job-requeuer.sh",571,0,"w",shellscript,content +187,413609,"weekend-job-requeuer.sh",572,0,"",shellscript,selection_keyboard +188,413922,"weekend-job-requeuer.sh",572,0,"e",shellscript,content +189,413923,"weekend-job-requeuer.sh",573,0,"",shellscript,selection_keyboard +190,414105,"weekend-job-requeuer.sh",573,0,"e",shellscript,content +191,414105,"weekend-job-requeuer.sh",574,0,"",shellscript,selection_keyboard +192,414205,"weekend-job-requeuer.sh",574,0,"p",shellscript,content +193,414205,"weekend-job-requeuer.sh",575,0,"",shellscript,selection_keyboard +194,414605,"weekend-job-requeuer.sh",586,0,"",shellscript,selection_command +195,415126,"weekend-job-requeuer.sh",0,0,"",shellscript,selection_keyboard +196,415688,"weekend-job-requeuer.sh",1,0,"",shellscript,selection_command +197,415724,"weekend-job-requeuer.sh",2,0,"",shellscript,selection_command +198,415938,"weekend-job-requeuer.sh",24,0,"",shellscript,selection_command +199,416423,"weekend-job-requeuer.sh",132,0,"",shellscript,selection_command +200,416466,"weekend-job-requeuer.sh",240,0,"",shellscript,selection_command +201,416482,"weekend-job-requeuer.sh",348,0,"",shellscript,selection_command +202,416527,"weekend-job-requeuer.sh",455,0,"",shellscript,selection_command +203,416588,"weekend-job-requeuer.sh",562,0,"",shellscript,selection_command +204,416589,"weekend-job-requeuer.sh",563,0,"",shellscript,selection_command +205,416641,"weekend-job-requeuer.sh",564,0,"",shellscript,selection_command +206,416642,"weekend-job-requeuer.sh",565,0,"",shellscript,selection_command +207,416688,"weekend-job-requeuer.sh",576,0,"",shellscript,selection_command +208,416689,"weekend-job-requeuer.sh",658,0,"",shellscript,selection_command +209,416727,"weekend-job-requeuer.sh",740,0,"",shellscript,selection_command +210,416760,"weekend-job-requeuer.sh",741,0,"",shellscript,selection_command +211,417674,"weekend-job-requeuer.sh",741,0,"\n# actionspace scaling",shellscript,content +212,417710,"weekend-job-requeuer.sh",742,0,"",shellscript,selection_command +213,418273,"weekend-job-requeuer.sh",763,0,"",shellscript,selection_command +214,418705,"weekend-job-requeuer.sh",762,0,"",shellscript,selection_command +215,419272,"weekend-job-requeuer.sh",755,0,"",shellscript,selection_command +216,419504,"weekend-job-requeuer.sh",744,11,"",shellscript,content +217,420106,"weekend-job-requeuer.sh",744,0,"b",shellscript,content +218,420107,"weekend-job-requeuer.sh",745,0,"",shellscript,selection_keyboard +219,420221,"weekend-job-requeuer.sh",745,0,"a",shellscript,content +220,420222,"weekend-job-requeuer.sh",746,0,"",shellscript,selection_keyboard +221,420321,"weekend-job-requeuer.sh",746,0,"t",shellscript,content +222,420322,"weekend-job-requeuer.sh",747,0,"",shellscript,selection_keyboard +223,420439,"weekend-job-requeuer.sh",747,0,"c",shellscript,content +224,420439,"weekend-job-requeuer.sh",748,0,"",shellscript,selection_keyboard +225,420507,"weekend-job-requeuer.sh",748,0,"h",shellscript,content +226,420508,"weekend-job-requeuer.sh",749,0,"",shellscript,selection_keyboard +227,420708,"weekend-job-requeuer.sh",749,0,"i",shellscript,content +228,420709,"weekend-job-requeuer.sh",750,0,"",shellscript,selection_keyboard +229,420788,"weekend-job-requeuer.sh",750,0,"s",shellscript,content +230,420788,"weekend-job-requeuer.sh",751,0,"",shellscript,selection_keyboard +231,421361,"weekend-job-requeuer.sh",750,1,"",shellscript,content +232,421471,"weekend-job-requeuer.sh",749,1,"",shellscript,content +233,421609,"weekend-job-requeuer.sh",749,0,"s",shellscript,content +234,421610,"weekend-job-requeuer.sh",750,0,"",shellscript,selection_keyboard +235,421704,"weekend-job-requeuer.sh",750,0,"i",shellscript,content +236,421705,"weekend-job-requeuer.sh",751,0,"",shellscript,selection_keyboard +237,421756,"weekend-job-requeuer.sh",751,0,"z",shellscript,content +238,421756,"weekend-job-requeuer.sh",752,0,"",shellscript,selection_keyboard +239,421874,"weekend-job-requeuer.sh",752,0,"e",shellscript,content +240,421875,"weekend-job-requeuer.sh",753,0,"",shellscript,selection_keyboard +241,422238,"weekend-job-requeuer.sh",752,0,"",shellscript,selection_command +242,423210,"weekend-job-requeuer.sh",772,0,"",shellscript,selection_command +243,423374,"weekend-job-requeuer.sh",880,0,"",shellscript,selection_command +244,423504,"weekend-job-requeuer.sh",987,0,"",shellscript,selection_command +245,423661,"weekend-job-requeuer.sh",1094,0,"",shellscript,selection_command +246,423808,"weekend-job-requeuer.sh",1201,0,"",shellscript,selection_command +247,423975,"weekend-job-requeuer.sh",1298,0,"",shellscript,selection_command +248,424138,"weekend-job-requeuer.sh",1299,0,"",shellscript,selection_command +249,424321,"weekend-job-requeuer.sh",1299,0,"\n# batchsize scaling",shellscript,content +250,424371,"weekend-job-requeuer.sh",1300,0,"",shellscript,selection_command +251,425284,"weekend-job-requeuer.sh",1299,0,"",shellscript,selection_command +252,425689,"weekend-job-requeuer.sh",1299,1,"",shellscript,content +253,425788,"weekend-job-requeuer.sh",1319,0,"",shellscript,selection_command +254,426037,"weekend-job-requeuer.sh",1299,0,"",shellscript,selection_command +255,426538,"weekend-job-requeuer.sh",1300,0,"",shellscript,selection_command +256,426783,"weekend-job-requeuer.sh",1301,0,"",shellscript,selection_command +257,428828,"weekend-job-requeuer.sh",1301,10,"",shellscript,content +258,430476,"weekend-job-requeuer.sh",1301,0,"m",shellscript,content +259,430478,"weekend-job-requeuer.sh",1302,0,"",shellscript,selection_keyboard +260,430587,"weekend-job-requeuer.sh",1302,0,"o",shellscript,content +261,430588,"weekend-job-requeuer.sh",1303,0,"",shellscript,selection_keyboard +262,430691,"weekend-job-requeuer.sh",1303,0,"d",shellscript,content +263,430692,"weekend-job-requeuer.sh",1304,0,"",shellscript,selection_keyboard +264,430793,"weekend-job-requeuer.sh",1304,0,"e",shellscript,content +265,430794,"weekend-job-requeuer.sh",1305,0,"",shellscript,selection_keyboard +266,430856,"weekend-job-requeuer.sh",1305,0,"l",shellscript,content +267,430857,"weekend-job-requeuer.sh",1306,0,"",shellscript,selection_keyboard +268,431093,"weekend-job-requeuer.sh",1306,0,"s",shellscript,content +269,431094,"weekend-job-requeuer.sh",1307,0,"",shellscript,selection_keyboard +270,431158,"weekend-job-requeuer.sh",1307,0,"i",shellscript,content +271,431159,"weekend-job-requeuer.sh",1308,0,"",shellscript,selection_keyboard +272,431305,"weekend-job-requeuer.sh",1308,0,"z",shellscript,content +273,431306,"weekend-job-requeuer.sh",1309,0,"",shellscript,selection_keyboard +274,431324,"weekend-job-requeuer.sh",1309,0,"e",shellscript,content +275,431325,"weekend-job-requeuer.sh",1310,0,"",shellscript,selection_keyboard +276,431457,"weekend-job-requeuer.sh",1310,0," ",shellscript,content +277,431458,"weekend-job-requeuer.sh",1311,0,"",shellscript,selection_keyboard +278,431623,"weekend-job-requeuer.sh",1310,0,"",shellscript,selection_command +279,433254,"weekend-job-requeuer.sh",564,0,"",shellscript,selection_mouse +280,434255,"weekend-job-requeuer.sh",564,1,"",shellscript,content +281,434743,"weekend-job-requeuer.sh",563,0,"",shellscript,selection_command +282,435188,"weekend-job-requeuer.sh",563,1,"",shellscript,content +283,435309,"weekend-job-requeuer.sh",574,0,"",shellscript,selection_command +284,435505,"weekend-job-requeuer.sh",656,0,"",shellscript,selection_command +285,435611,"weekend-job-requeuer.sh",738,0,"",shellscript,selection_command +286,436055,"weekend-job-requeuer.sh",738,1,"",shellscript,content +287,446223,"weekend-job-requeuer.sh",1,0,"",shellscript,selection_mouse +288,455039,"weekend-job-requeuer.sh",1,0,"sleep 12h\n\n",shellscript,content +289,455042,"weekend-job-requeuer.sh",11,1,"",shellscript,content +290,460162,"TERMINAL",0,0,"bash",,terminal_focus +291,462805,"TERMINAL",0,0,"sleep 12h",,terminal_command +292,462854,"TERMINAL",0,0,"]633;E;2025-07-12 12:25:22 sleep 12h;16c08be6-3885-420f-ac53-f5272aed6e54]633;C",,terminal_output +293,464293,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;130]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +294,469639,"TERMINAL",0,0,"man sleep",,terminal_command +295,469688,"TERMINAL",0,0,"]633;E;2025-07-12 12:25:29 man sleep;16c08be6-3885-420f-ac53-f5272aed6e54]633;C",,terminal_output +296,469772,"TERMINAL",0,0,"[?1049h[?1h=\r",,terminal_output +297,469822,"TERMINAL",0,0,"SLEEP(1) User Commands SLEEP(1)\r\n\r\nNAME\r\n sleep - delay for a specified amount of time\r\n\r\nSYNOPSIS\r\n sleep NUMBER[SUFFIX]...\r\n sleep OPTION\r\n\r\nDESCRIPTION\r\n Pause for NUMBER seconds. SUFFIX may be 's' for seconds (the default), 'm' for minutes, 'h' for hours or 'd' for\r\n days. NUMBER need not be an integer. Given two or more arguments, pause for the amount of time specified by the\r\n sum of their values.\r\n\r\n --help display this help and exit\r\n\r\n --version\r\n output version information and exit\r\n\r\nAUTHOR\r\n Written by Jim Meyering and Paul Eggert.\r\n\r\nREPORTING BUGS\r\n GNU coreutils online help: \r\n Report any translation bugs to \r\n\r\nCOPYRIGHT\r\n Copyright © 2020 Free Software Foundation, Inc. License GPLv3+: GNU GPL version 3 or later .\r\n This is free software: you are free to change and redistribute it. There is NO WARRANTY, to the extent permitted\r\n by law.\r\n\r\nSEE ALSO\r\n sleep(3)\r\n\r\n Full documentation \r\n or available locally via: info '(coreutils) sleep invocation'\r\n\r\n Manual page sleep(1) line 1 (press h for help or q to quit)",,terminal_output +298,482304,"TERMINAL",0,0,"\r ESCESCOOBB\rGNU coreutils 8.32 January 2024 SLEEP(1)\r\n Manual page sleep(1) line 2 (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)\r ESCESCOOBB\r\r Manual page sleep(1) line 2/39 (END) (press h for help or q to quit)",,terminal_output +299,482839,"TERMINAL",0,0,"\r ESCESCOOAA\rMSLEEP(1) User Commands SLEEP(1)\r\n\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)\r ESCESCOOAA\r\r Manual page sleep(1) line 1/39 92% (press h for help or q to quit)",,terminal_output +300,484806,"TERMINAL",0,0,"SLEEP(1) User Commands SLEEP(1)\r\n\r\nNAME\r\n sleep - delay for a specified amount of time\r\n\r\nSYNOPSIS\r\n sleep NUMBER[SUFFIX]...\r\n sleep OPTION\r\n\r\nDESCRIPTION\r\n Pause for NUMBER seconds. SUFFIX may be 's' for seconds (the default), 'm' for minutes, 'h' for hours or 'd' for\r\n days. NUMBER need not be an integer. Given two or more arguments, pause for the amount of time specified by the\r\n sum of their values.\r\n\r\n --help display this help and exit\r\n\r\n --version\r\n output version information and exit\r\n\r\nAUTHOR\r\n Written by Jim Meyering and Paul Eggert.\r\n\r\nREPORTING BUGS\r\n GNU coreutils online help: \r\n Report any translation bugs to \r\n\r\nCOPYRIGHT\r\n Copyright © 2020 Free Software Foundation, Inc. License GPLv3+: GNU GPL version 3 or later .\r\n This is free software: you are free to change and redistribute it. There is NO WARRANTY, to the extent permitted\r\n by law.\r\n\r\nSEE ALSO\r\n sleep(3)\r\n\r\n Full documentation \r\n or available locally via: info '(coreutils) sleep invocation'\r\n\r\n Manual page sleep(1) line 1/39 92% (press h for help or q to quit)",,terminal_output +301,486723,"TERMINAL",0,0,"\r[?1l>[?1049l]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0]633;P;Cwd=/home/hk-project-p0023960/tum_cte0515/Projects/jafar",,terminal_output +302,487891,"weekend-job-requeuer.sh",0,0,"",shellscript,tab +303,488757,"weekend-job-requeuer.sh",9,0,"",shellscript,selection_mouse +304,490321,"weekend-job-requeuer.sh",0,0,"",shellscript,selection_mouse +305,490804,"weekend-job-requeuer.sh",0,0,"\n",shellscript,content +306,492176,"weekend-job-requeuer.sh",1,0,"e",shellscript,content +307,492179,"weekend-job-requeuer.sh",2,0,"",shellscript,selection_keyboard +308,492322,"weekend-job-requeuer.sh",2,0,"c",shellscript,content +309,492323,"weekend-job-requeuer.sh",3,0,"",shellscript,selection_keyboard +310,492372,"weekend-job-requeuer.sh",3,0,"h",shellscript,content +311,492373,"weekend-job-requeuer.sh",4,0,"",shellscript,selection_keyboard +312,492555,"weekend-job-requeuer.sh",4,0,"o",shellscript,content +313,492556,"weekend-job-requeuer.sh",5,0,"",shellscript,selection_keyboard +314,492638,"weekend-job-requeuer.sh",5,0," ",shellscript,content +315,492639,"weekend-job-requeuer.sh",6,0,"",shellscript,selection_keyboard +316,496160,"weekend-job-requeuer.sh",6,0,"w",shellscript,content +317,496162,"weekend-job-requeuer.sh",7,0,"",shellscript,selection_keyboard +318,496371,"weekend-job-requeuer.sh",7,0,"e",shellscript,content +319,496372,"weekend-job-requeuer.sh",8,0,"",shellscript,selection_keyboard +320,496539,"weekend-job-requeuer.sh",8,0,"e",shellscript,content +321,496540,"weekend-job-requeuer.sh",9,0,"",shellscript,selection_keyboard +322,496588,"weekend-job-requeuer.sh",9,0,"k",shellscript,content +323,496589,"weekend-job-requeuer.sh",10,0,"",shellscript,selection_keyboard +324,496761,"weekend-job-requeuer.sh",10,0,"e",shellscript,content +325,496762,"weekend-job-requeuer.sh",11,0,"",shellscript,selection_keyboard +326,496805,"weekend-job-requeuer.sh",11,0,"n",shellscript,content +327,496805,"weekend-job-requeuer.sh",12,0,"",shellscript,selection_keyboard +328,496940,"weekend-job-requeuer.sh",12,0,"d",shellscript,content +329,496941,"weekend-job-requeuer.sh",13,0,"",shellscript,selection_keyboard +330,496992,"weekend-job-requeuer.sh",13,0," ",shellscript,content +331,496993,"weekend-job-requeuer.sh",14,0,"",shellscript,selection_keyboard +332,497171,"weekend-job-requeuer.sh",14,0,"r",shellscript,content +333,497172,"weekend-job-requeuer.sh",15,0,"",shellscript,selection_keyboard +334,497421,"weekend-job-requeuer.sh",15,0,"q",shellscript,content +335,497422,"weekend-job-requeuer.sh",16,0,"",shellscript,selection_keyboard +336,498217,"weekend-job-requeuer.sh",15,1,"",shellscript,content +337,498954,"weekend-job-requeuer.sh",15,0,"e",shellscript,content +338,498955,"weekend-job-requeuer.sh",16,0,"",shellscript,selection_keyboard +339,499177,"weekend-job-requeuer.sh",16,0,"q",shellscript,content +340,499178,"weekend-job-requeuer.sh",17,0,"",shellscript,selection_keyboard +341,499288,"weekend-job-requeuer.sh",17,0,"u",shellscript,content +342,499289,"weekend-job-requeuer.sh",18,0,"",shellscript,selection_keyboard +343,499359,"weekend-job-requeuer.sh",18,0,"e",shellscript,content +344,499360,"weekend-job-requeuer.sh",19,0,"",shellscript,selection_keyboard +345,499421,"weekend-job-requeuer.sh",19,0,"u",shellscript,content +346,499422,"weekend-job-requeuer.sh",20,0,"",shellscript,selection_keyboard +347,499522,"weekend-job-requeuer.sh",20,0,"e",shellscript,content +348,499523,"weekend-job-requeuer.sh",21,0,"",shellscript,selection_keyboard +349,499656,"weekend-job-requeuer.sh",21,0,"r",shellscript,content +350,499657,"weekend-job-requeuer.sh",22,0,"",shellscript,selection_keyboard +351,501088,"weekend-job-requeuer.sh",22,0,"!",shellscript,content +352,501089,"weekend-job-requeuer.sh",23,0,"",shellscript,selection_keyboard +353,501774,"weekend-job-requeuer.sh",22,1,"",shellscript,content +354,502358,"weekend-job-requeuer.sh",22,0,"""",shellscript,content +355,502359,"weekend-job-requeuer.sh",23,0,"",shellscript,selection_keyboard +356,503038,"weekend-job-requeuer.sh",22,0,"",shellscript,selection_command +357,503871,"weekend-job-requeuer.sh",6,0,"""",shellscript,content +358,503872,"weekend-job-requeuer.sh",7,0,"",shellscript,selection_keyboard +359,505090,"weekend-job-requeuer.sh",23,0," ",shellscript,content +360,505091,"weekend-job-requeuer.sh",24,0,"",shellscript,selection_keyboard +361,505193,"weekend-job-requeuer.sh",24,0,"s",shellscript,content +362,505194,"weekend-job-requeuer.sh",25,0,"",shellscript,selection_keyboard +363,505325,"weekend-job-requeuer.sh",25,0,"t",shellscript,content +364,505326,"weekend-job-requeuer.sh",26,0,"",shellscript,selection_keyboard +365,505354,"weekend-job-requeuer.sh",26,0,"a",shellscript,content +366,505355,"weekend-job-requeuer.sh",27,0,"",shellscript,selection_keyboard +367,505555,"weekend-job-requeuer.sh",27,0,"r",shellscript,content +368,505556,"weekend-job-requeuer.sh",28,0,"",shellscript,selection_keyboard +369,505671,"weekend-job-requeuer.sh",28,0,"t",shellscript,content +370,505672,"weekend-job-requeuer.sh",29,0,"",shellscript,selection_keyboard +371,505754,"weekend-job-requeuer.sh",29,0,"e",shellscript,content +372,505755,"weekend-job-requeuer.sh",30,0,"",shellscript,selection_keyboard +373,506004,"weekend-job-requeuer.sh",30,0,"d",shellscript,content +374,506005,"weekend-job-requeuer.sh",31,0,"",shellscript,selection_keyboard +375,506521,"weekend-job-requeuer.sh",31,0,".",shellscript,content +376,506522,"weekend-job-requeuer.sh",32,0,"",shellscript,selection_keyboard +377,506704,"weekend-job-requeuer.sh",32,0,".",shellscript,content +378,506705,"weekend-job-requeuer.sh",33,0,"",shellscript,selection_keyboard +379,506837,"weekend-job-requeuer.sh",33,0,".",shellscript,content +380,506838,"weekend-job-requeuer.sh",34,0,"",shellscript,selection_keyboard +381,507391,"weekend-job-requeuer.sh",33,0,"",shellscript,selection_command +382,508054,"weekend-job-requeuer.sh",44,0,"",shellscript,selection_command +383,508204,"weekend-job-requeuer.sh",46,0,"",shellscript,selection_command +384,508621,"weekend-job-requeuer.sh",46,0,"\necho ""weekend requeuer started...""",shellscript,content +385,508624,"weekend-job-requeuer.sh",47,0,"",shellscript,selection_command +386,509171,"weekend-job-requeuer.sh",46,0,"",shellscript,selection_command +387,509804,"weekend-job-requeuer.sh",46,1,"",shellscript,content +388,510156,"weekend-job-requeuer.sh",36,0,"",shellscript,selection_command +389,510389,"weekend-job-requeuer.sh",37,0,"",shellscript,selection_command +390,510555,"weekend-job-requeuer.sh",47,0,"",shellscript,selection_command +391,510904,"weekend-job-requeuer.sh",37,0,"",shellscript,selection_command +392,511106,"weekend-job-requeuer.sh",38,0,"",shellscript,selection_command +393,511321,"weekend-job-requeuer.sh",48,0,"",shellscript,selection_command +394,511521,"weekend-job-requeuer.sh",49,0,"",shellscript,selection_command +395,511705,"weekend-job-requeuer.sh",50,0,"",shellscript,selection_command +396,511838,"weekend-job-requeuer.sh",51,0,"",shellscript,selection_command +397,512340,"weekend-job-requeuer.sh",51,1,"",shellscript,content +398,513005,"weekend-job-requeuer.sh",51,8,"",shellscript,content +399,513254,"weekend-job-requeuer.sh",51,9,"",shellscript,content +400,513877,"weekend-job-requeuer.sh",46,17,"",shellscript,content +401,513988,"weekend-job-requeuer.sh",48,0,"",shellscript,selection_command +402,514888,"weekend-job-requeuer.sh",46,0,"echo started...""\n",shellscript,content +403,514891,"weekend-job-requeuer.sh",51,0,"",shellscript,selection_command +404,515691,"weekend-job-requeuer.sh",51,7,"",shellscript,content +405,516805,"weekend-job-requeuer.sh",51,0,"""",shellscript,content +406,516808,"weekend-job-requeuer.sh",52,0,"",shellscript,selection_keyboard +407,517226,"weekend-job-requeuer.sh",52,0,"1",shellscript,content +408,517227,"weekend-job-requeuer.sh",53,0,"",shellscript,selection_keyboard +409,517321,"weekend-job-requeuer.sh",53,0,"2",shellscript,content +410,517321,"weekend-job-requeuer.sh",54,0,"",shellscript,selection_keyboard +411,517705,"weekend-job-requeuer.sh",54,0," ",shellscript,content +412,517706,"weekend-job-requeuer.sh",55,0,"",shellscript,selection_keyboard +413,518192,"weekend-job-requeuer.sh",55,0,"h",shellscript,content +414,518193,"weekend-job-requeuer.sh",56,0,"",shellscript,selection_keyboard +415,518354,"weekend-job-requeuer.sh",56,0,"o",shellscript,content +416,518355,"weekend-job-requeuer.sh",57,0,"",shellscript,selection_keyboard +417,518473,"weekend-job-requeuer.sh",57,0,"u",shellscript,content +418,518474,"weekend-job-requeuer.sh",58,0,"",shellscript,selection_keyboard +419,518477,"weekend-job-requeuer.sh",58,0,"r",shellscript,content +420,518478,"weekend-job-requeuer.sh",59,0,"",shellscript,selection_keyboard +421,518744,"weekend-job-requeuer.sh",59,0,"s",shellscript,content +422,518745,"weekend-job-requeuer.sh",60,0,"",shellscript,selection_keyboard +423,518804,"weekend-job-requeuer.sh",60,0," ",shellscript,content +424,518805,"weekend-job-requeuer.sh",61,0,"",shellscript,selection_keyboard +425,518942,"weekend-job-requeuer.sh",61,0,"p",shellscript,content +426,518942,"weekend-job-requeuer.sh",62,0,"",shellscript,selection_keyboard +427,518988,"weekend-job-requeuer.sh",62,0,"a",shellscript,content +428,518989,"weekend-job-requeuer.sh",63,0,"",shellscript,selection_keyboard +429,519108,"weekend-job-requeuer.sh",63,0,"s",shellscript,content +430,519109,"weekend-job-requeuer.sh",64,0,"",shellscript,selection_keyboard +431,519288,"weekend-job-requeuer.sh",64,0,"s",shellscript,content +432,519288,"weekend-job-requeuer.sh",65,0,"",shellscript,selection_keyboard +433,519521,"weekend-job-requeuer.sh",65,0,"e",shellscript,content +434,519522,"weekend-job-requeuer.sh",66,0,"",shellscript,selection_keyboard +435,519621,"weekend-job-requeuer.sh",66,0,"d",shellscript,content +436,519622,"weekend-job-requeuer.sh",67,0,"",shellscript,selection_keyboard +437,519940,"weekend-job-requeuer.sh",67,0,"!",shellscript,content +438,519941,"weekend-job-requeuer.sh",68,0,"",shellscript,selection_keyboard +439,520421,"weekend-job-requeuer.sh",68,1,"",shellscript,content +440,520594,"weekend-job-requeuer.sh",68,1,"",shellscript,content +441,520704,"weekend-job-requeuer.sh",68,1,"",shellscript,content +442,521326,"weekend-job-requeuer.sh",67,0,"",shellscript,selection_command +443,522771,"weekend-job-requeuer.sh",67,1,"",shellscript,content +444,523590,"weekend-job-requeuer.sh",89,0,"",shellscript,selection_command +445,523988,"weekend-job-requeuer.sh",90,0,"\necho ""12 hours passed""",shellscript,content +446,524004,"weekend-job-requeuer.sh",91,0,"",shellscript,selection_command +447,524554,"weekend-job-requeuer.sh",92,0,"",shellscript,selection_command +448,524721,"weekend-job-requeuer.sh",93,0,"",shellscript,selection_command +449,524855,"weekend-job-requeuer.sh",94,0,"",shellscript,selection_command +450,525004,"weekend-job-requeuer.sh",95,0,"",shellscript,selection_command +451,525137,"weekend-job-requeuer.sh",96,0,"",shellscript,selection_command +452,525305,"weekend-job-requeuer.sh",97,0,"",shellscript,selection_command +453,525524,"weekend-job-requeuer.sh",97,3,"",shellscript,content +454,525804,"weekend-job-requeuer.sh",97,6,"",shellscript,content +455,526071,"weekend-job-requeuer.sh",97,6,"",shellscript,content +456,526937,"weekend-job-requeuer.sh",97,0,"s",shellscript,content +457,526938,"weekend-job-requeuer.sh",98,0,"",shellscript,selection_keyboard +458,527873,"weekend-job-requeuer.sh",98,0,"t",shellscript,content +459,527874,"weekend-job-requeuer.sh",99,0,"",shellscript,selection_keyboard +460,528004,"weekend-job-requeuer.sh",99,0,"a",shellscript,content +461,528005,"weekend-job-requeuer.sh",100,0,"",shellscript,selection_keyboard +462,528204,"weekend-job-requeuer.sh",100,0,"r",shellscript,content +463,528205,"weekend-job-requeuer.sh",101,0,"",shellscript,selection_keyboard +464,528355,"weekend-job-requeuer.sh",101,0,"t",shellscript,content +465,528356,"weekend-job-requeuer.sh",102,0,"",shellscript,selection_keyboard +466,528540,"weekend-job-requeuer.sh",102,0,"i",shellscript,content +467,528541,"weekend-job-requeuer.sh",103,0,"",shellscript,selection_keyboard +468,528639,"weekend-job-requeuer.sh",103,0,"n",shellscript,content +469,528640,"weekend-job-requeuer.sh",104,0,"",shellscript,selection_keyboard +470,528755,"weekend-job-requeuer.sh",104,0,"g",shellscript,content +471,528756,"weekend-job-requeuer.sh",105,0,"",shellscript,selection_keyboard +472,528811,"weekend-job-requeuer.sh",105,0," ",shellscript,content +473,528812,"weekend-job-requeuer.sh",106,0,"",shellscript,selection_keyboard +474,528957,"weekend-job-requeuer.sh",106,0,"a",shellscript,content +475,528958,"weekend-job-requeuer.sh",107,0,"",shellscript,selection_keyboard +476,529071,"weekend-job-requeuer.sh",107,0,"c",shellscript,content +477,529072,"weekend-job-requeuer.sh",108,0,"",shellscript,selection_keyboard +478,529293,"weekend-job-requeuer.sh",108,0,"t",shellscript,content +479,529294,"weekend-job-requeuer.sh",109,0,"",shellscript,selection_keyboard +480,529408,"weekend-job-requeuer.sh",109,0,"i",shellscript,content +481,529409,"weekend-job-requeuer.sh",110,0,"",shellscript,selection_keyboard +482,529471,"weekend-job-requeuer.sh",110,0,"o",shellscript,content +483,529472,"weekend-job-requeuer.sh",111,0,"",shellscript,selection_keyboard +484,529592,"weekend-job-requeuer.sh",111,0,"n",shellscript,content +485,529593,"weekend-job-requeuer.sh",112,0,"",shellscript,selection_keyboard +486,529771,"weekend-job-requeuer.sh",112,0,"s",shellscript,content +487,529772,"weekend-job-requeuer.sh",113,0,"",shellscript,selection_keyboard +488,529854,"weekend-job-requeuer.sh",113,0,"p",shellscript,content +489,529855,"weekend-job-requeuer.sh",114,0,"",shellscript,selection_keyboard +490,530021,"weekend-job-requeuer.sh",114,0,"a",shellscript,content +491,530022,"weekend-job-requeuer.sh",115,0,"",shellscript,selection_keyboard +492,530221,"weekend-job-requeuer.sh",115,0,"c",shellscript,content +493,530222,"weekend-job-requeuer.sh",116,0,"",shellscript,selection_keyboard +494,530328,"weekend-job-requeuer.sh",116,0,"e",shellscript,content +495,530329,"weekend-job-requeuer.sh",117,0,"",shellscript,selection_keyboard +496,530404,"weekend-job-requeuer.sh",117,0," ",shellscript,content +497,530405,"weekend-job-requeuer.sh",118,0,"",shellscript,selection_keyboard +498,530588,"weekend-job-requeuer.sh",118,0,"s",shellscript,content +499,530589,"weekend-job-requeuer.sh",119,0,"",shellscript,selection_keyboard +500,530804,"weekend-job-requeuer.sh",119,0,"c",shellscript,content +501,530805,"weekend-job-requeuer.sh",120,0,"",shellscript,selection_keyboard +502,530971,"weekend-job-requeuer.sh",120,0,"a",shellscript,content +503,530972,"weekend-job-requeuer.sh",121,0,"",shellscript,selection_keyboard +504,531094,"weekend-job-requeuer.sh",121,0,"l",shellscript,content +505,531094,"weekend-job-requeuer.sh",122,0,"",shellscript,selection_keyboard +506,531287,"weekend-job-requeuer.sh",122,0,"i",shellscript,content +507,531288,"weekend-job-requeuer.sh",123,0,"",shellscript,selection_keyboard +508,531339,"weekend-job-requeuer.sh",123,0,"n",shellscript,content +509,531340,"weekend-job-requeuer.sh",124,0,"",shellscript,selection_keyboard +510,531405,"weekend-job-requeuer.sh",124,0,"g",shellscript,content +511,531406,"weekend-job-requeuer.sh",125,0,"",shellscript,selection_keyboard +512,531510,"weekend-job-requeuer.sh",125,0," ",shellscript,content +513,531511,"weekend-job-requeuer.sh",126,0,"",shellscript,selection_keyboard +514,531938,"weekend-job-requeuer.sh",126,0,"j",shellscript,content +515,531939,"weekend-job-requeuer.sh",127,0,"",shellscript,selection_keyboard +516,532041,"weekend-job-requeuer.sh",127,0,"o",shellscript,content +517,532042,"weekend-job-requeuer.sh",128,0,"",shellscript,selection_keyboard +518,532189,"weekend-job-requeuer.sh",128,0,"b",shellscript,content +519,532190,"weekend-job-requeuer.sh",129,0,"",shellscript,selection_keyboard +520,532255,"weekend-job-requeuer.sh",129,0,"s",shellscript,content +521,532257,"weekend-job-requeuer.sh",130,0,"",shellscript,selection_keyboard +522,533804,"weekend-job-requeuer.sh",97,9,"",shellscript,content +523,534474,"weekend-job-requeuer.sh",97,0,"r",shellscript,content +524,534475,"weekend-job-requeuer.sh",98,0,"",shellscript,selection_keyboard +525,534655,"weekend-job-requeuer.sh",98,0,"e",shellscript,content +526,534656,"weekend-job-requeuer.sh",99,0,"",shellscript,selection_keyboard +527,535054,"weekend-job-requeuer.sh",99,0,"q",shellscript,content +528,535055,"weekend-job-requeuer.sh",100,0,"",shellscript,selection_keyboard +529,535176,"weekend-job-requeuer.sh",100,0,"u",shellscript,content +530,535177,"weekend-job-requeuer.sh",101,0,"",shellscript,selection_keyboard +531,535291,"weekend-job-requeuer.sh",101,0,"e",shellscript,content +532,535292,"weekend-job-requeuer.sh",102,0,"",shellscript,selection_keyboard +533,535357,"weekend-job-requeuer.sh",102,0,"u",shellscript,content +534,535358,"weekend-job-requeuer.sh",103,0,"",shellscript,selection_keyboard +535,535457,"weekend-job-requeuer.sh",103,0,"e",shellscript,content +536,535458,"weekend-job-requeuer.sh",104,0,"",shellscript,selection_keyboard +537,535687,"weekend-job-requeuer.sh",104,0,"i",shellscript,content +538,535688,"weekend-job-requeuer.sh",105,0,"",shellscript,selection_keyboard +539,535754,"weekend-job-requeuer.sh",105,0,"n",shellscript,content +540,535755,"weekend-job-requeuer.sh",106,0,"",shellscript,selection_keyboard +541,535873,"weekend-job-requeuer.sh",106,0,"g",shellscript,content +542,535874,"weekend-job-requeuer.sh",107,0,"",shellscript,selection_keyboard +543,535974,"weekend-job-requeuer.sh",107,0," ",shellscript,content +544,535975,"weekend-job-requeuer.sh",108,0,"",shellscript,selection_keyboard +545,536771,"weekend-job-requeuer.sh",107,0,"",shellscript,selection_command +546,561244,"weekend-job-requeuer.sh",60,0,"",shellscript,selection_mouse +547,562005,"weekend-job-requeuer.sh",105,0,"",shellscript,selection_mouse +548,563488,"weekend-job-requeuer.sh",148,0,"",shellscript,selection_command +549,563656,"weekend-job-requeuer.sh",256,0,"",shellscript,selection_command +550,563805,"weekend-job-requeuer.sh",364,0,"",shellscript,selection_command +551,563939,"weekend-job-requeuer.sh",472,0,"",shellscript,selection_command +552,564105,"weekend-job-requeuer.sh",579,0,"",shellscript,selection_command +553,564259,"weekend-job-requeuer.sh",672,0,"",shellscript,selection_command +554,564454,"weekend-job-requeuer.sh",682,0,"",shellscript,selection_command +555,564722,"weekend-job-requeuer.sh",683,0,"\necho ""requeueing actionspace scaling jobs""",shellscript,content +556,564728,"weekend-job-requeuer.sh",684,0,"",shellscript,selection_command +557,565222,"weekend-job-requeuer.sh",685,0,"",shellscript,selection_command +558,565406,"weekend-job-requeuer.sh",686,0,"",shellscript,selection_command +559,565538,"weekend-job-requeuer.sh",687,0,"",shellscript,selection_command +560,566037,"weekend-job-requeuer.sh",688,0,"",shellscript,selection_command +561,566055,"weekend-job-requeuer.sh",689,0,"",shellscript,selection_command +562,566077,"weekend-job-requeuer.sh",690,0,"",shellscript,selection_command +563,566122,"weekend-job-requeuer.sh",691,0,"",shellscript,selection_command +564,566171,"weekend-job-requeuer.sh",692,0,"",shellscript,selection_command +565,566172,"weekend-job-requeuer.sh",693,0,"",shellscript,selection_command +566,566224,"weekend-job-requeuer.sh",694,0,"",shellscript,selection_command +567,566260,"weekend-job-requeuer.sh",695,0,"",shellscript,selection_command +568,566261,"weekend-job-requeuer.sh",696,0,"",shellscript,selection_command +569,566321,"weekend-job-requeuer.sh",697,0,"",shellscript,selection_command +570,566439,"weekend-job-requeuer.sh",698,0,"",shellscript,selection_command +571,566555,"weekend-job-requeuer.sh",699,0,"",shellscript,selection_command +572,566704,"weekend-job-requeuer.sh",700,0,"",shellscript,selection_command +573,566837,"weekend-job-requeuer.sh",701,0,"",shellscript,selection_command +574,567071,"weekend-job-requeuer.sh",701,12,"",shellscript,content +575,567407,"weekend-job-requeuer.sh",701,8,"",shellscript,content +576,567788,"weekend-job-requeuer.sh",701,4,"",shellscript,content +577,568258,"weekend-job-requeuer.sh",701,0,"l",shellscript,content +578,568259,"weekend-job-requeuer.sh",702,0,"",shellscript,selection_keyboard +579,568805,"weekend-job-requeuer.sh",702,0,"r",shellscript,content +580,568806,"weekend-job-requeuer.sh",703,0,"",shellscript,selection_keyboard +581,568893,"weekend-job-requeuer.sh",703,0," ",shellscript,content +582,568895,"weekend-job-requeuer.sh",704,0,"",shellscript,selection_keyboard +583,569204,"weekend-job-requeuer.sh",704,0,"s",shellscript,content +584,569205,"weekend-job-requeuer.sh",705,0,"",shellscript,selection_keyboard +585,569389,"weekend-job-requeuer.sh",705,0,"w",shellscript,content +586,569390,"weekend-job-requeuer.sh",706,0,"",shellscript,selection_keyboard +587,569604,"weekend-job-requeuer.sh",706,0,"e",shellscript,content +588,569605,"weekend-job-requeuer.sh",707,0,"",shellscript,selection_keyboard +589,569804,"weekend-job-requeuer.sh",707,0,"e",shellscript,content +590,569805,"weekend-job-requeuer.sh",708,0,"",shellscript,selection_keyboard +591,569889,"weekend-job-requeuer.sh",708,0,"p",shellscript,content +592,569890,"weekend-job-requeuer.sh",709,0,"",shellscript,selection_keyboard +593,570004,"weekend-job-requeuer.sh",709,0," ",shellscript,content +594,570005,"weekend-job-requeuer.sh",710,0,"",shellscript,selection_keyboard +595,570154,"weekend-job-requeuer.sh",710,0,"j",shellscript,content +596,570155,"weekend-job-requeuer.sh",711,0,"",shellscript,selection_keyboard +597,570321,"weekend-job-requeuer.sh",711,0,"o",shellscript,content +598,570322,"weekend-job-requeuer.sh",712,0,"",shellscript,selection_keyboard +599,570490,"weekend-job-requeuer.sh",712,0,"b",shellscript,content +600,570491,"weekend-job-requeuer.sh",713,0,"",shellscript,selection_keyboard +601,570604,"weekend-job-requeuer.sh",713,0,"s",shellscript,content +602,570606,"weekend-job-requeuer.sh",714,0,"",shellscript,selection_keyboard +603,571304,"weekend-job-requeuer.sh",713,0,"",shellscript,selection_command +604,603839,"weekend-job-requeuer.sh",708,0,"",shellscript,selection_mouse +605,605711,"weekend-job-requeuer.sh",740,0,"",shellscript,selection_command +606,605841,"weekend-job-requeuer.sh",822,0,"",shellscript,selection_command +607,605988,"weekend-job-requeuer.sh",880,0,"",shellscript,selection_command +608,606223,"weekend-job-requeuer.sh",899,0,"",shellscript,selection_command +609,606422,"weekend-job-requeuer.sh",900,0,"\necho ""requeueing lr sweep jobs""",shellscript,content +610,606437,"weekend-job-requeuer.sh",901,0,"",shellscript,selection_command +611,606991,"weekend-job-requeuer.sh",902,0,"",shellscript,selection_command +612,607524,"weekend-job-requeuer.sh",903,0,"",shellscript,selection_command +613,607524,"weekend-job-requeuer.sh",904,0,"",shellscript,selection_command +614,607537,"weekend-job-requeuer.sh",905,0,"",shellscript,selection_command +615,607577,"weekend-job-requeuer.sh",906,0,"",shellscript,selection_command +616,607624,"weekend-job-requeuer.sh",907,0,"",shellscript,selection_command +617,607738,"weekend-job-requeuer.sh",908,0,"",shellscript,selection_command +618,608372,"weekend-job-requeuer.sh",908,10,"",shellscript,content +619,609087,"weekend-job-requeuer.sh",908,0,"equeueing ",shellscript,content +620,609090,"weekend-job-requeuer.sh",908,0,"",shellscript,selection_command +621,609456,"weekend-job-requeuer.sh",909,0,"",shellscript,selection_command +622,609621,"weekend-job-requeuer.sh",910,0,"",shellscript,selection_command +623,609771,"weekend-job-requeuer.sh",911,0,"",shellscript,selection_command +624,609921,"weekend-job-requeuer.sh",912,0,"",shellscript,selection_command +625,610055,"weekend-job-requeuer.sh",913,0,"",shellscript,selection_command +626,610207,"weekend-job-requeuer.sh",914,0,"",shellscript,selection_command +627,610322,"weekend-job-requeuer.sh",915,0,"",shellscript,selection_command +628,610472,"weekend-job-requeuer.sh",916,0,"",shellscript,selection_command +629,610622,"weekend-job-requeuer.sh",917,0,"",shellscript,selection_command +630,610938,"weekend-job-requeuer.sh",918,0,"",shellscript,selection_command +631,611204,"weekend-job-requeuer.sh",918,3,"",shellscript,content +632,611438,"weekend-job-requeuer.sh",918,6,"",shellscript,content +633,611709,"weekend-job-requeuer.sh",918,4,"",shellscript,content +634,612256,"weekend-job-requeuer.sh",918,0,"s",shellscript,content +635,612256,"weekend-job-requeuer.sh",919,0,"",shellscript,selection_keyboard +636,612322,"weekend-job-requeuer.sh",919,0,"o",shellscript,content +637,612323,"weekend-job-requeuer.sh",920,0,"",shellscript,selection_keyboard +638,612470,"weekend-job-requeuer.sh",920,0,"m",shellscript,content +639,612471,"weekend-job-requeuer.sh",921,0,"",shellscript,selection_keyboard +640,612560,"weekend-job-requeuer.sh",921,0,"e",shellscript,content +641,612561,"weekend-job-requeuer.sh",922,0,"",shellscript,selection_keyboard +642,612638,"weekend-job-requeuer.sh",922,0," ",shellscript,content +643,612638,"weekend-job-requeuer.sh",923,0,"",shellscript,selection_keyboard +644,612804,"weekend-job-requeuer.sh",923,0,"b",shellscript,content +645,612805,"weekend-job-requeuer.sh",924,0,"",shellscript,selection_keyboard +646,612854,"weekend-job-requeuer.sh",924,0,"a",shellscript,content +647,612855,"weekend-job-requeuer.sh",925,0,"",shellscript,selection_keyboard +648,612971,"weekend-job-requeuer.sh",925,0,"c",shellscript,content +649,612972,"weekend-job-requeuer.sh",926,0,"",shellscript,selection_keyboard +650,613155,"weekend-job-requeuer.sh",926,0,"t",shellscript,content +651,613155,"weekend-job-requeuer.sh",927,0,"",shellscript,selection_keyboard +652,613776,"weekend-job-requeuer.sh",926,1,"",shellscript,content +653,614354,"weekend-job-requeuer.sh",926,0,"t",shellscript,content +654,614355,"weekend-job-requeuer.sh",927,0,"",shellscript,selection_keyboard +655,614621,"weekend-job-requeuer.sh",927,0,"c",shellscript,content +656,614622,"weekend-job-requeuer.sh",928,0,"",shellscript,selection_keyboard +657,614671,"weekend-job-requeuer.sh",928,0,"h",shellscript,content +658,614672,"weekend-job-requeuer.sh",929,0,"",shellscript,selection_keyboard +659,615371,"weekend-job-requeuer.sh",928,1,"",shellscript,content +660,615537,"weekend-job-requeuer.sh",927,1,"",shellscript,content +661,615654,"weekend-job-requeuer.sh",926,1,"",shellscript,content +662,615757,"weekend-job-requeuer.sh",925,1,"",shellscript,content +663,615854,"weekend-job-requeuer.sh",925,0,"t",shellscript,content +664,615855,"weekend-job-requeuer.sh",926,0,"",shellscript,selection_keyboard +665,616039,"weekend-job-requeuer.sh",926,0,"c",shellscript,content +666,616039,"weekend-job-requeuer.sh",927,0,"",shellscript,selection_keyboard +667,616157,"weekend-job-requeuer.sh",927,0,"h",shellscript,content +668,616158,"weekend-job-requeuer.sh",928,0,"",shellscript,selection_keyboard +669,616271,"weekend-job-requeuer.sh",928,0,"s",shellscript,content +670,616271,"weekend-job-requeuer.sh",929,0,"",shellscript,selection_keyboard +671,616360,"weekend-job-requeuer.sh",929,0,"i",shellscript,content +672,616361,"weekend-job-requeuer.sh",930,0,"",shellscript,selection_keyboard +673,616471,"weekend-job-requeuer.sh",930,0,"z",shellscript,content +674,616472,"weekend-job-requeuer.sh",931,0,"",shellscript,selection_keyboard +675,616576,"weekend-job-requeuer.sh",931,0,"e",shellscript,content +676,616577,"weekend-job-requeuer.sh",932,0,"",shellscript,selection_keyboard +677,616670,"weekend-job-requeuer.sh",932,0," ",shellscript,content +678,616671,"weekend-job-requeuer.sh",933,0,"",shellscript,selection_keyboard +679,616787,"weekend-job-requeuer.sh",933,0,"s",shellscript,content +680,616788,"weekend-job-requeuer.sh",934,0,"",shellscript,selection_keyboard +681,616839,"weekend-job-requeuer.sh",934,0,"c",shellscript,content +682,616842,"weekend-job-requeuer.sh",935,0,"",shellscript,selection_keyboard +683,617023,"weekend-job-requeuer.sh",935,0,"a",shellscript,content +684,617023,"weekend-job-requeuer.sh",936,0,"",shellscript,selection_keyboard +685,617104,"weekend-job-requeuer.sh",936,0,"l",shellscript,content +686,617105,"weekend-job-requeuer.sh",937,0,"",shellscript,selection_keyboard +687,617306,"weekend-job-requeuer.sh",937,0,"i",shellscript,content +688,617306,"weekend-job-requeuer.sh",938,0,"",shellscript,selection_keyboard +689,617341,"weekend-job-requeuer.sh",938,0,"n",shellscript,content +690,617341,"weekend-job-requeuer.sh",939,0,"",shellscript,selection_keyboard +691,617388,"weekend-job-requeuer.sh",939,0,"g",shellscript,content +692,617388,"weekend-job-requeuer.sh",940,0,"",shellscript,selection_keyboard +693,617522,"weekend-job-requeuer.sh",940,0," ",shellscript,content +694,617523,"weekend-job-requeuer.sh",941,0,"",shellscript,selection_keyboard +695,617642,"weekend-job-requeuer.sh",941,0,"r",shellscript,content +696,617642,"weekend-job-requeuer.sh",942,0,"",shellscript,selection_keyboard +697,617689,"weekend-job-requeuer.sh",942,0,"u",shellscript,content +698,617690,"weekend-job-requeuer.sh",943,0,"",shellscript,selection_keyboard +699,617787,"weekend-job-requeuer.sh",943,0,"n",shellscript,content +700,617788,"weekend-job-requeuer.sh",944,0,"",shellscript,selection_keyboard +701,617871,"weekend-job-requeuer.sh",944,0,"s",shellscript,content +702,617871,"weekend-job-requeuer.sh",945,0,"",shellscript,selection_keyboard +703,618311,"weekend-job-requeuer.sh",944,0,"",shellscript,selection_command +704,629354,"weekend-job-requeuer.sh",990,0,"",shellscript,selection_command +705,629454,"weekend-job-requeuer.sh",1098,0,"",shellscript,selection_command +706,647355,"weekend-job-requeuer.sh",1178,0,"",shellscript,selection_mouse +707,648221,"weekend-job-requeuer.sh",1483,0,"",shellscript,selection_mouse +708,648988,"weekend-job-requeuer.sh",1189,0,"",shellscript,selection_mouse +709,649588,"weekend-job-requeuer.sh",978,0,"",shellscript,selection_mouse +710,650792,"weekend-job-requeuer.sh",947,108,"",shellscript,content +711,650976,"weekend-job-requeuer.sh",1054,0,"",shellscript,selection_command +712,651221,"weekend-job-requeuer.sh",1161,0,"",shellscript,selection_command +713,651421,"weekend-job-requeuer.sh",1268,0,"",shellscript,selection_command +714,651694,"weekend-job-requeuer.sh",1374,0,"\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",shellscript,content +715,651725,"weekend-job-requeuer.sh",1375,0,"",shellscript,selection_command +716,652354,"weekend-job-requeuer.sh",978,0,"",shellscript,selection_mouse +717,652837,"weekend-job-requeuer.sh",947,107,"",shellscript,content +718,653705,"weekend-job-requeuer.sh",978,0,"",shellscript,selection_mouse +719,653939,"weekend-job-requeuer.sh",1039,0,"1_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_",shellscript,content +720,654239,"weekend-job-requeuer.sh",1085,0,"",shellscript,selection_command +721,654439,"weekend-job-requeuer.sh",1192,0,"",shellscript,selection_command +722,654540,"weekend-job-requeuer.sh",1299,0,"",shellscript,selection_command +723,654704,"weekend-job-requeuer.sh",1406,0,"",shellscript,selection_command +724,654887,"weekend-job-requeuer.sh",1483,0,"",shellscript,selection_command +725,655195,"weekend-job-requeuer.sh",1406,0,"",shellscript,selection_command +726,655587,"weekend-job-requeuer.sh",1375,108,"",shellscript,content +727,655640,"weekend-job-requeuer.sh",1268,0,"",shellscript,selection_command +728,655804,"weekend-job-requeuer.sh",1161,0,"",shellscript,selection_command +729,655971,"weekend-job-requeuer.sh",1054,0,"",shellscript,selection_command +730,656771,"weekend-job-requeuer.sh",1160,0,"\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch",shellscript,content +731,656809,"weekend-job-requeuer.sh",1161,0,"",shellscript,selection_command +732,657420,"weekend-job-requeuer.sh",1054,0,"",shellscript,selection_command +733,657805,"weekend-job-requeuer.sh",1160,0,"\n",shellscript,content +734,659955,"weekend-job-requeuer.sh",1054,0,"",shellscript,selection_command +735,660460,"weekend-job-requeuer.sh",947,0,"",shellscript,selection_command +736,660470,"weekend-job-requeuer.sh",901,0,"",shellscript,selection_command +737,660490,"weekend-job-requeuer.sh",881,0,"",shellscript,selection_command +738,660537,"weekend-job-requeuer.sh",880,0,"",shellscript,selection_command +739,660588,"weekend-job-requeuer.sh",798,0,"",shellscript,selection_command +740,660589,"weekend-job-requeuer.sh",716,0,"",shellscript,selection_command +741,660637,"weekend-job-requeuer.sh",684,0,"",shellscript,selection_command +742,660688,"weekend-job-requeuer.sh",673,0,"",shellscript,selection_command +743,660688,"weekend-job-requeuer.sh",672,0,"",shellscript,selection_command +744,660739,"weekend-job-requeuer.sh",565,0,"",shellscript,selection_command +745,660740,"weekend-job-requeuer.sh",458,0,"",shellscript,selection_command +746,660773,"weekend-job-requeuer.sh",350,0,"",shellscript,selection_command +747,660876,"weekend-job-requeuer.sh",242,0,"",shellscript,selection_command +748,660877,"weekend-job-requeuer.sh",134,0,"",shellscript,selection_command +749,660877,"weekend-job-requeuer.sh",91,0,"",shellscript,selection_command +750,660887,"weekend-job-requeuer.sh",69,0,"",shellscript,selection_command +751,660921,"weekend-job-requeuer.sh",46,0,"",shellscript,selection_command +752,660971,"weekend-job-requeuer.sh",36,0,"",shellscript,selection_command +753,661021,"weekend-job-requeuer.sh",1,0,"",shellscript,selection_command +754,661260,"weekend-job-requeuer.sh",36,0,"",shellscript,selection_command +755,661671,"weekend-job-requeuer.sh",46,0,"",shellscript,selection_command +756,662173,"weekend-job-requeuer.sh",69,0,"",shellscript,selection_command +757,662222,"weekend-job-requeuer.sh",91,0,"",shellscript,selection_command +758,662255,"weekend-job-requeuer.sh",134,0,"",shellscript,selection_command +759,662271,"weekend-job-requeuer.sh",242,0,"",shellscript,selection_command +760,662276,"weekend-job-requeuer.sh",350,0,"",shellscript,selection_command +761,662311,"weekend-job-requeuer.sh",458,0,"",shellscript,selection_command +762,662391,"weekend-job-requeuer.sh",565,0,"",shellscript,selection_command +763,662407,"weekend-job-requeuer.sh",672,0,"",shellscript,selection_command +764,662408,"weekend-job-requeuer.sh",673,0,"",shellscript,selection_command +765,662439,"weekend-job-requeuer.sh",684,0,"",shellscript,selection_command +766,662456,"weekend-job-requeuer.sh",716,0,"",shellscript,selection_command +767,662504,"weekend-job-requeuer.sh",798,0,"",shellscript,selection_command +768,662559,"weekend-job-requeuer.sh",880,0,"",shellscript,selection_command +769,662560,"weekend-job-requeuer.sh",881,0,"",shellscript,selection_command +770,662603,"weekend-job-requeuer.sh",901,0,"",shellscript,selection_command +771,662606,"weekend-job-requeuer.sh",947,0,"",shellscript,selection_command +772,662721,"weekend-job-requeuer.sh",1054,0,"",shellscript,selection_command +773,662821,"weekend-job-requeuer.sh",1161,0,"",shellscript,selection_command +774,663054,"weekend-job-requeuer.sh",1161,0,"\nsleep 12h",shellscript,content +775,663057,"weekend-job-requeuer.sh",1162,0,"",shellscript,selection_command +776,663622,"weekend-job-requeuer.sh",1163,0,"",shellscript,selection_command +777,664137,"weekend-job-requeuer.sh",1164,0,"",shellscript,selection_command +778,664138,"weekend-job-requeuer.sh",1165,0,"",shellscript,selection_command +779,664195,"weekend-job-requeuer.sh",1166,0,"",shellscript,selection_command +780,664221,"weekend-job-requeuer.sh",1167,0,"",shellscript,selection_command +781,664253,"weekend-job-requeuer.sh",1168,0,"",shellscript,selection_command +782,664270,"weekend-job-requeuer.sh",1169,0,"",shellscript,selection_command +783,664644,"weekend-job-requeuer.sh",1168,0,"",shellscript,selection_command +784,664804,"weekend-job-requeuer.sh",1168,1,"",shellscript,content +785,666120,"weekend-job-requeuer.sh",1177,0,"",shellscript,selection_command +786,666290,"weekend-job-requeuer.sh",1285,0,"",shellscript,selection_command +787,686342,"weekend-job-requeuer.sh",1214,0,"",shellscript,selection_mouse +788,687138,"weekend-job-requeuer.sh",1161,0,"",shellscript,selection_mouse +789,687754,"weekend-job-requeuer.sh",1170,0,"",shellscript,selection_mouse +790,687808,"weekend-job-requeuer.sh",1169,0,"",shellscript,selection_command +791,688825,"weekend-job-requeuer.sh",1178,0,"",shellscript,selection_command +792,689187,"weekend-job-requeuer.sh",1278,0,"\nsleep 2h",shellscript,content +793,689204,"weekend-job-requeuer.sh",1279,0,"",shellscript,selection_command +794,689724,"weekend-job-requeuer.sh",1280,0,"",shellscript,selection_command +795,689888,"weekend-job-requeuer.sh",1281,0,"",shellscript,selection_command +796,690054,"weekend-job-requeuer.sh",1282,0,"",shellscript,selection_command +797,690173,"weekend-job-requeuer.sh",1283,0,"",shellscript,selection_command +798,690320,"weekend-job-requeuer.sh",1284,0,"",shellscript,selection_command +799,690505,"weekend-job-requeuer.sh",1285,0,"",shellscript,selection_command +800,692821,"weekend-job-requeuer.sh",1285,1,"1",shellscript,content +801,714621,"weekend-job-requeuer.sh",1287,0,"",shellscript,selection_mouse +802,714671,"weekend-job-requeuer.sh",1286,0,"",shellscript,selection_command +803,715704,"weekend-job-requeuer.sh",1295,0,"",shellscript,selection_command +804,716172,"weekend-job-requeuer.sh",1394,0,"\nsleep 1h",shellscript,content +805,716187,"weekend-job-requeuer.sh",1395,0,"",shellscript,selection_command +806,716592,"weekend-job-requeuer.sh",1396,0,"",shellscript,selection_command +807,716721,"weekend-job-requeuer.sh",1397,0,"",shellscript,selection_command +808,716856,"weekend-job-requeuer.sh",1398,0,"",shellscript,selection_command +809,716988,"weekend-job-requeuer.sh",1399,0,"",shellscript,selection_command +810,717121,"weekend-job-requeuer.sh",1400,0,"",shellscript,selection_command +811,717254,"weekend-job-requeuer.sh",1401,0,"",shellscript,selection_command +812,718354,"weekend-job-requeuer.sh",1401,1,"5",shellscript,content +813,718771,"weekend-job-requeuer.sh",1294,0,"",shellscript,selection_command +814,718925,"weekend-job-requeuer.sh",1285,0,"",shellscript,selection_command +815,719088,"weekend-job-requeuer.sh",1177,0,"",shellscript,selection_command +816,719272,"weekend-job-requeuer.sh",1168,0,"",shellscript,selection_command +817,719389,"weekend-job-requeuer.sh",1161,0,"",shellscript,selection_command +818,719887,"weekend-job-requeuer.sh",1161,1,"",shellscript,content +819,720804,"weekend-job-requeuer.sh",1054,0,"",shellscript,selection_command +820,720954,"weekend-job-requeuer.sh",947,0,"",shellscript,selection_command +821,721088,"weekend-job-requeuer.sh",901,0,"",shellscript,selection_command +822,721693,"weekend-job-requeuer.sh",947,0,"",shellscript,selection_command +823,721891,"weekend-job-requeuer.sh",1054,0,"",shellscript,selection_command +824,722807,"weekend-job-requeuer.sh",1161,0,"",shellscript,selection_command +825,723359,"weekend-job-requeuer.sh",1169,0,"\necho ""requeueing some batchsize scaling runs""",shellscript,content +826,723372,"weekend-job-requeuer.sh",1170,0,"",shellscript,selection_command +827,723772,"weekend-job-requeuer.sh",1171,0,"",shellscript,selection_command +828,724239,"weekend-job-requeuer.sh",1172,0,"",shellscript,selection_command +829,724287,"weekend-job-requeuer.sh",1173,0,"",shellscript,selection_command +830,724338,"weekend-job-requeuer.sh",1174,0,"",shellscript,selection_command +831,724339,"weekend-job-requeuer.sh",1175,0,"",shellscript,selection_command +832,724388,"weekend-job-requeuer.sh",1176,0,"",shellscript,selection_command +833,724391,"weekend-job-requeuer.sh",1177,0,"",shellscript,selection_command +834,724443,"weekend-job-requeuer.sh",1178,0,"",shellscript,selection_command +835,724489,"weekend-job-requeuer.sh",1179,0,"",shellscript,selection_command +836,724490,"weekend-job-requeuer.sh",1180,0,"",shellscript,selection_command +837,724527,"weekend-job-requeuer.sh",1181,0,"",shellscript,selection_command +838,724571,"weekend-job-requeuer.sh",1182,0,"",shellscript,selection_command +839,724592,"weekend-job-requeuer.sh",1183,0,"",shellscript,selection_command +840,724620,"weekend-job-requeuer.sh",1184,0,"",shellscript,selection_command +841,724639,"weekend-job-requeuer.sh",1185,0,"",shellscript,selection_command +842,724673,"weekend-job-requeuer.sh",1186,0,"",shellscript,selection_command +843,724721,"weekend-job-requeuer.sh",1187,0,"",shellscript,selection_command +844,724738,"weekend-job-requeuer.sh",1188,0,"",shellscript,selection_command +845,724770,"weekend-job-requeuer.sh",1189,0,"",shellscript,selection_command +846,724787,"weekend-job-requeuer.sh",1190,0,"",shellscript,selection_command +847,724821,"weekend-job-requeuer.sh",1191,0,"",shellscript,selection_command +848,724871,"weekend-job-requeuer.sh",1192,0,"",shellscript,selection_command +849,724887,"weekend-job-requeuer.sh",1193,0,"",shellscript,selection_command +850,724921,"weekend-job-requeuer.sh",1194,0,"",shellscript,selection_command +851,724940,"weekend-job-requeuer.sh",1195,0,"",shellscript,selection_command +852,724970,"weekend-job-requeuer.sh",1196,0,"",shellscript,selection_command +853,725021,"weekend-job-requeuer.sh",1197,0,"",shellscript,selection_command +854,725039,"weekend-job-requeuer.sh",1198,0,"",shellscript,selection_command +855,725327,"weekend-job-requeuer.sh",1197,0,"",shellscript,selection_command +856,725493,"weekend-job-requeuer.sh",1196,0,"",shellscript,selection_command +857,725643,"weekend-job-requeuer.sh",1195,0,"",shellscript,selection_command +858,725806,"weekend-job-requeuer.sh",1194,0,"",shellscript,selection_command +859,725961,"weekend-job-requeuer.sh",1193,0,"",shellscript,selection_command +860,726139,"weekend-job-requeuer.sh",1192,0,"",shellscript,selection_command +861,726490,"weekend-job-requeuer.sh",1168,0,"",shellscript,selection_command +862,726642,"weekend-job-requeuer.sh",1076,0,"",shellscript,selection_command +863,726793,"weekend-job-requeuer.sh",969,0,"",shellscript,selection_command +864,726970,"weekend-job-requeuer.sh",923,0,"",shellscript,selection_command +865,727943,"weekend-job-requeuer.sh",922,0,"",shellscript,selection_command +866,728088,"weekend-job-requeuer.sh",921,0,"",shellscript,selection_command +867,728255,"weekend-job-requeuer.sh",920,0,"",shellscript,selection_command +868,728409,"weekend-job-requeuer.sh",919,0,"",shellscript,selection_command +869,728637,"weekend-job-requeuer.sh",918,0,"",shellscript,selection_command +870,728924,"weekend-job-requeuer.sh",918,5,"",shellscript,content +871,733504,"weekend-job-requeuer.sh",940,0," ",shellscript,content +872,733505,"weekend-job-requeuer.sh",941,0,"",shellscript,selection_keyboard +873,733738,"weekend-job-requeuer.sh",941,0,"()",shellscript,content +874,733738,"weekend-job-requeuer.sh",942,0,"",shellscript,selection_keyboard +875,734041,"weekend-job-requeuer.sh",942,0,"1",shellscript,content +876,734042,"weekend-job-requeuer.sh",943,0,"",shellscript,selection_keyboard +877,737154,"weekend-job-requeuer.sh",942,1,"",shellscript,content +878,738854,"weekend-job-requeuer.sh",941,2,"",shellscript,content +879,739709,"weekend-job-requeuer.sh",941,0,"[]",shellscript,content +880,739710,"weekend-job-requeuer.sh",942,0,"",shellscript,selection_keyboard +881,740021,"weekend-job-requeuer.sh",942,0,"1",shellscript,content +882,740022,"weekend-job-requeuer.sh",943,0,"",shellscript,selection_keyboard +883,740244,"weekend-job-requeuer.sh",943,0,",",shellscript,content +884,740245,"weekend-job-requeuer.sh",944,0,"",shellscript,selection_keyboard +885,741404,"weekend-job-requeuer.sh",944,0,"2",shellscript,content +886,741405,"weekend-job-requeuer.sh",945,0,"",shellscript,selection_keyboard +887,741637,"weekend-job-requeuer.sh",945,1,"]",shellscript,content +888,741638,"weekend-job-requeuer.sh",946,0,"",shellscript,selection_keyboard +889,742258,"weekend-job-requeuer.sh",946,0,"n",shellscript,content +890,742259,"weekend-job-requeuer.sh",947,0,"",shellscript,selection_keyboard +891,742439,"weekend-job-requeuer.sh",947,0,"o",shellscript,content +892,742440,"weekend-job-requeuer.sh",948,0,"",shellscript,selection_keyboard +893,742525,"weekend-job-requeuer.sh",948,0,"d",shellscript,content +894,742526,"weekend-job-requeuer.sh",949,0,"",shellscript,selection_keyboard +895,742725,"weekend-job-requeuer.sh",949,0,"e",shellscript,content +896,742726,"weekend-job-requeuer.sh",950,0,"",shellscript,selection_keyboard +897,743010,"weekend-job-requeuer.sh",950,0,"s",shellscript,content +898,743011,"weekend-job-requeuer.sh",951,0,"",shellscript,selection_keyboard +899,744790,"weekend-job-requeuer.sh",1265,0,"",shellscript,selection_mouse +900,754659,"weekend-job-requeuer.sh",1220,0,"",shellscript,selection_mouse +901,755301,"weekend-job-requeuer.sh",1220,0," ",shellscript,content +902,755302,"weekend-job-requeuer.sh",1221,0,"",shellscript,selection_keyboard +903,755821,"weekend-job-requeuer.sh",1220,1,"",shellscript,content +904,756191,"weekend-job-requeuer.sh",1220,0,":",shellscript,content +905,756191,"weekend-job-requeuer.sh",1221,0,"",shellscript,selection_keyboard +906,756404,"weekend-job-requeuer.sh",1221,0," ",shellscript,content +907,756405,"weekend-job-requeuer.sh",1222,0,"",shellscript,selection_keyboard +908,756606,"weekend-job-requeuer.sh",1222,0,"1",shellscript,content +909,756607,"weekend-job-requeuer.sh",1223,0,"",shellscript,selection_keyboard +910,757009,"weekend-job-requeuer.sh",1223,0,"6",shellscript,content +911,757010,"weekend-job-requeuer.sh",1224,0,"",shellscript,selection_keyboard +912,757473,"weekend-job-requeuer.sh",1224,0,"n",shellscript,content +913,757474,"weekend-job-requeuer.sh",1225,0,"",shellscript,selection_keyboard +914,757487,"weekend-job-requeuer.sh",1225,0,"o",shellscript,content +915,757488,"weekend-job-requeuer.sh",1226,0,"",shellscript,selection_keyboard +916,757539,"weekend-job-requeuer.sh",1226,0,"d",shellscript,content +917,757540,"weekend-job-requeuer.sh",1227,0,"",shellscript,selection_keyboard +918,757637,"weekend-job-requeuer.sh",1227,0,"e",shellscript,content +919,757638,"weekend-job-requeuer.sh",1228,0,"",shellscript,selection_keyboard +920,757825,"weekend-job-requeuer.sh",1228,0,"s",shellscript,content +921,757825,"weekend-job-requeuer.sh",1229,0,"",shellscript,selection_keyboard +922,759506,"weekend-job-requeuer.sh",940,0,"",shellscript,selection_mouse +923,761258,"weekend-job-requeuer.sh",940,0,":",shellscript,content +924,761259,"weekend-job-requeuer.sh",941,0,"",shellscript,selection_keyboard +925,762256,"weekend-job-requeuer.sh",1195,0,"",shellscript,selection_mouse +926,762410,"weekend-job-requeuer.sh",1194,4,"some",shellscript,selection_mouse +927,762773,"weekend-job-requeuer.sh",1194,4,"",shellscript,content +928,762973,"weekend-job-requeuer.sh",1193,1,"",shellscript,content +929,765207,"weekend-job-requeuer.sh",1343,0,"",shellscript,selection_mouse +930,765838,"weekend-job-requeuer.sh",1242,0,"",shellscript,selection_mouse +931,766337,"weekend-job-requeuer.sh",1241,0,"",shellscript,selection_command +932,766671,"weekend-job-requeuer.sh",1191,0,"",shellscript,selection_command +933,767254,"weekend-job-requeuer.sh",1241,0,"",shellscript,selection_command +934,767587,"weekend-job-requeuer.sh",1342,0,"",shellscript,selection_command +935,767772,"weekend-job-requeuer.sh",1343,0,"\necho ""requeueing batchsize scaling runs: 16nodes""",shellscript,content +936,767839,"weekend-job-requeuer.sh",1344,0,"",shellscript,selection_command +937,768371,"weekend-job-requeuer.sh",1345,0,"",shellscript,selection_command +938,768871,"weekend-job-requeuer.sh",1346,0,"",shellscript,selection_command +939,768921,"weekend-job-requeuer.sh",1347,0,"",shellscript,selection_command +940,768921,"weekend-job-requeuer.sh",1348,0,"",shellscript,selection_command +941,768971,"weekend-job-requeuer.sh",1349,0,"",shellscript,selection_command +942,769010,"weekend-job-requeuer.sh",1350,0,"",shellscript,selection_command +943,769021,"weekend-job-requeuer.sh",1351,0,"",shellscript,selection_command +944,769056,"weekend-job-requeuer.sh",1352,0,"",shellscript,selection_command +945,769071,"weekend-job-requeuer.sh",1353,0,"",shellscript,selection_command +946,769108,"weekend-job-requeuer.sh",1354,0,"",shellscript,selection_command +947,769139,"weekend-job-requeuer.sh",1355,0,"",shellscript,selection_command +948,769154,"weekend-job-requeuer.sh",1356,0,"",shellscript,selection_command +949,769187,"weekend-job-requeuer.sh",1357,0,"",shellscript,selection_command +950,769221,"weekend-job-requeuer.sh",1358,0,"",shellscript,selection_command +951,769259,"weekend-job-requeuer.sh",1359,0,"",shellscript,selection_command +952,769306,"weekend-job-requeuer.sh",1360,0,"",shellscript,selection_command +953,769337,"weekend-job-requeuer.sh",1361,0,"",shellscript,selection_command +954,769354,"weekend-job-requeuer.sh",1362,0,"",shellscript,selection_command +955,769387,"weekend-job-requeuer.sh",1363,0,"",shellscript,selection_command +956,769420,"weekend-job-requeuer.sh",1364,0,"",shellscript,selection_command +957,769438,"weekend-job-requeuer.sh",1365,0,"",shellscript,selection_command +958,769470,"weekend-job-requeuer.sh",1366,0,"",shellscript,selection_command +959,769521,"weekend-job-requeuer.sh",1367,0,"",shellscript,selection_command +960,769539,"weekend-job-requeuer.sh",1368,0,"",shellscript,selection_command +961,769572,"weekend-job-requeuer.sh",1369,0,"",shellscript,selection_command +962,769620,"weekend-job-requeuer.sh",1370,0,"",shellscript,selection_command +963,769621,"weekend-job-requeuer.sh",1371,0,"",shellscript,selection_command +964,769694,"weekend-job-requeuer.sh",1372,0,"",shellscript,selection_command +965,769694,"weekend-job-requeuer.sh",1373,0,"",shellscript,selection_command +966,769725,"weekend-job-requeuer.sh",1374,0,"",shellscript,selection_command +967,769725,"weekend-job-requeuer.sh",1375,0,"",shellscript,selection_command +968,769771,"weekend-job-requeuer.sh",1376,0,"",shellscript,selection_command +969,769820,"weekend-job-requeuer.sh",1377,0,"",shellscript,selection_command +970,769821,"weekend-job-requeuer.sh",1378,0,"",shellscript,selection_command +971,769871,"weekend-job-requeuer.sh",1379,0,"",shellscript,selection_command +972,769920,"weekend-job-requeuer.sh",1380,0,"",shellscript,selection_command +973,769921,"weekend-job-requeuer.sh",1381,0,"",shellscript,selection_command +974,769974,"weekend-job-requeuer.sh",1382,0,"",shellscript,selection_command +975,770071,"weekend-job-requeuer.sh",1383,0,"",shellscript,selection_command +976,770487,"weekend-job-requeuer.sh",1384,0,"",shellscript,selection_command +977,770687,"weekend-job-requeuer.sh",1385,0,"",shellscript,selection_command +978,774909,"weekend-job-requeuer.sh",1385,1,"",shellscript,content +979,775054,"weekend-job-requeuer.sh",1385,1,"",shellscript,content +980,775754,"weekend-job-requeuer.sh",1385,0,"4",shellscript,content +981,775755,"weekend-job-requeuer.sh",1386,0,"",shellscript,selection_keyboard +982,776460,"weekend-job-requeuer.sh",1385,0,"",shellscript,selection_command +983,778155,"weekend-job-requeuer.sh",1434,0,"",shellscript,selection_command +984,778337,"weekend-job-requeuer.sh",1507,0,"",shellscript,selection_command +985,778604,"weekend-job-requeuer.sh",1508,0,"\necho ""requeueing batchsize scaling runs: 4nodes""",shellscript,content +986,778605,"weekend-job-requeuer.sh",1509,0,"",shellscript,selection_command +987,778893,"weekend-job-requeuer.sh",1510,0,"",shellscript,selection_command +988,779394,"weekend-job-requeuer.sh",1511,0,"",shellscript,selection_command +989,779422,"weekend-job-requeuer.sh",1512,0,"",shellscript,selection_command +990,779441,"weekend-job-requeuer.sh",1513,0,"",shellscript,selection_command +991,779472,"weekend-job-requeuer.sh",1514,0,"",shellscript,selection_command +992,779537,"weekend-job-requeuer.sh",1515,0,"",shellscript,selection_command +993,779590,"weekend-job-requeuer.sh",1516,0,"",shellscript,selection_command +994,779591,"weekend-job-requeuer.sh",1517,0,"",shellscript,selection_command +995,779627,"weekend-job-requeuer.sh",1518,0,"",shellscript,selection_command +996,779628,"weekend-job-requeuer.sh",1519,0,"",shellscript,selection_command +997,779677,"weekend-job-requeuer.sh",1520,0,"",shellscript,selection_command +998,779721,"weekend-job-requeuer.sh",1521,0,"",shellscript,selection_command +999,779722,"weekend-job-requeuer.sh",1522,0,"",shellscript,selection_command +1000,779778,"weekend-job-requeuer.sh",1523,0,"",shellscript,selection_command +1001,779788,"weekend-job-requeuer.sh",1524,0,"",shellscript,selection_command +1002,779820,"weekend-job-requeuer.sh",1525,0,"",shellscript,selection_command +1003,779854,"weekend-job-requeuer.sh",1526,0,"",shellscript,selection_command +1004,779870,"weekend-job-requeuer.sh",1527,0,"",shellscript,selection_command +1005,779905,"weekend-job-requeuer.sh",1528,0,"",shellscript,selection_command +1006,779959,"weekend-job-requeuer.sh",1529,0,"",shellscript,selection_command +1007,779976,"weekend-job-requeuer.sh",1530,0,"",shellscript,selection_command +1008,780007,"weekend-job-requeuer.sh",1531,0,"",shellscript,selection_command +1009,780021,"weekend-job-requeuer.sh",1532,0,"",shellscript,selection_command +1010,780054,"weekend-job-requeuer.sh",1533,0,"",shellscript,selection_command +1011,780103,"weekend-job-requeuer.sh",1534,0,"",shellscript,selection_command +1012,780120,"weekend-job-requeuer.sh",1535,0,"",shellscript,selection_command +1013,780160,"weekend-job-requeuer.sh",1536,0,"",shellscript,selection_command +1014,780171,"weekend-job-requeuer.sh",1537,0,"",shellscript,selection_command +1015,780221,"weekend-job-requeuer.sh",1538,0,"",shellscript,selection_command +1016,780260,"weekend-job-requeuer.sh",1539,0,"",shellscript,selection_command +1017,780263,"weekend-job-requeuer.sh",1540,0,"",shellscript,selection_command +1018,780310,"weekend-job-requeuer.sh",1541,0,"",shellscript,selection_command +1019,780370,"weekend-job-requeuer.sh",1542,0,"",shellscript,selection_command +1020,780372,"weekend-job-requeuer.sh",1543,0,"",shellscript,selection_command +1021,780420,"weekend-job-requeuer.sh",1544,0,"",shellscript,selection_command +1022,780421,"weekend-job-requeuer.sh",1545,0,"",shellscript,selection_command +1023,780472,"weekend-job-requeuer.sh",1546,0,"",shellscript,selection_command +1024,780473,"weekend-job-requeuer.sh",1547,0,"",shellscript,selection_command +1025,780526,"weekend-job-requeuer.sh",1548,0,"",shellscript,selection_command +1026,780560,"weekend-job-requeuer.sh",1549,0,"",shellscript,selection_command +1027,780821,"weekend-job-requeuer.sh",1550,0,"",shellscript,selection_command +1028,781926,"weekend-job-requeuer.sh",1550,1,"8",shellscript,content +1029,887623,"weekend-job-requeuer.sh",44,0,"",shellscript,selection_mouse +1030,888975,"weekend-job-requeuer.sh",43,1,"",shellscript,content +1031,889073,"weekend-job-requeuer.sh",42,1,"",shellscript,content +1032,889260,"weekend-job-requeuer.sh",42,0,"9",shellscript,content +1033,889261,"weekend-job-requeuer.sh",43,0,"",shellscript,selection_keyboard +1034,890556,"weekend-job-requeuer.sh",53,0,"",shellscript,selection_mouse +1035,891155,"weekend-job-requeuer.sh",52,1,"",shellscript,content +1036,891188,"weekend-job-requeuer.sh",51,1,"",shellscript,content +1037,891955,"weekend-job-requeuer.sh",51,0,"9",shellscript,content +1038,891956,"weekend-job-requeuer.sh",52,0,"",shellscript,selection_keyboard +1039,903225,"weekend-job-requeuer.sh",1341,0,"",shellscript,selection_mouse +1040,904022,"weekend-job-requeuer.sh",1340,0,"",shellscript,selection_command +1041,904754,"weekend-job-requeuer.sh",1232,0,"",shellscript,selection_command +1042,904904,"weekend-job-requeuer.sh",1182,0,"",shellscript,selection_command +1043,905188,"weekend-job-requeuer.sh",1173,0,"",shellscript,selection_command +1044,905340,"weekend-job-requeuer.sh",1066,0,"",shellscript,selection_command +1045,905470,"weekend-job-requeuer.sh",959,0,"",shellscript,selection_command +1046,905603,"weekend-job-requeuer.sh",906,0,"",shellscript,selection_command +1047,905870,"weekend-job-requeuer.sh",886,0,"",shellscript,selection_command +1048,906020,"weekend-job-requeuer.sh",878,0,"",shellscript,selection_command +1049,906171,"weekend-job-requeuer.sh",803,0,"",shellscript,selection_command +1050,906308,"weekend-job-requeuer.sh",721,0,"",shellscript,selection_command +1051,906470,"weekend-job-requeuer.sh",689,0,"",shellscript,selection_command +1052,906672,"weekend-job-requeuer.sh",678,0,"",shellscript,selection_command +1053,907121,"weekend-job-requeuer.sh",681,0,"\nsleep 1h",shellscript,content +1054,907122,"weekend-job-requeuer.sh",682,0,"",shellscript,selection_command +1055,918188,"weekend-job-requeuer.sh",1183,0,"",shellscript,selection_mouse +1056,918189,"weekend-job-requeuer.sh",1182,0,"",shellscript,selection_command +1057,918487,"weekend-job-requeuer.sh",1182,1,"h",shellscript,selection_mouse +1058,918488,"weekend-job-requeuer.sh",1183,0,"",shellscript,selection_command +1059,918504,"weekend-job-requeuer.sh",1182,1,"h",shellscript,selection_mouse +1060,918520,"weekend-job-requeuer.sh",1181,2,"2h",shellscript,selection_mouse +1061,918570,"weekend-job-requeuer.sh",1180,3," 2h",shellscript,selection_mouse +1062,918571,"weekend-job-requeuer.sh",1179,4,"p 2h",shellscript,selection_mouse +1063,918654,"weekend-job-requeuer.sh",1178,5,"ep 2h",shellscript,selection_mouse +1064,918704,"weekend-job-requeuer.sh",1177,6,"eep 2h",shellscript,selection_mouse +1065,919339,"weekend-job-requeuer.sh",1177,0,"",shellscript,selection_mouse +1066,939808,"weekend-job-requeuer.sh",1182,0,"",shellscript,selection_mouse +1067,942087,"weekend-job-requeuer.sh",817,0,"",shellscript,selection_mouse +1068,942739,"weekend-job-requeuer.sh",688,0,"",shellscript,selection_mouse +1069,945722,"weekend-job-requeuer.sh",697,0,"",shellscript,selection_command +1070,945862,"weekend-job-requeuer.sh",729,0,"",shellscript,selection_command +1071,946021,"weekend-job-requeuer.sh",811,0,"",shellscript,selection_command +1072,946320,"weekend-job-requeuer.sh",887,0,"",shellscript,selection_command +1073,946471,"weekend-job-requeuer.sh",894,0,"",shellscript,selection_command +1074,946805,"weekend-job-requeuer.sh",907,0,"\nsleep 1h",shellscript,content +1075,946837,"weekend-job-requeuer.sh",908,0,"",shellscript,selection_command +1076,969823,"weekend-job-requeuer.sh",1190,0,"",shellscript,selection_mouse +1077,971404,"weekend-job-requeuer.sh",1190,1,"3",shellscript,content +1078,978906,"weekend-job-requeuer.sh",1444,0,"",shellscript,selection_mouse +1079,980409,"weekend-job-requeuer.sh",1395,0,"",shellscript,selection_command +1080,980574,"weekend-job-requeuer.sh",1358,0,"",shellscript,selection_command +1081,980720,"weekend-job-requeuer.sh",1278,0,"",shellscript,selection_command +1082,980870,"weekend-job-requeuer.sh",1228,0,"",shellscript,selection_command +1083,981021,"weekend-job-requeuer.sh",1191,0,"",shellscript,selection_command +1084,981309,"weekend-job-requeuer.sh",1228,0,"",shellscript,selection_command +1085,981504,"weekend-job-requeuer.sh",1229,0,"",shellscript,selection_command +1086,981687,"weekend-job-requeuer.sh",1230,0,"",shellscript,selection_command +1087,981803,"weekend-job-requeuer.sh",1231,0,"",shellscript,selection_command +1088,981953,"weekend-job-requeuer.sh",1232,0,"",shellscript,selection_command +1089,982087,"weekend-job-requeuer.sh",1233,0,"",shellscript,selection_command +1090,982254,"weekend-job-requeuer.sh",1234,0,"",shellscript,selection_command +1091,983887,"weekend-job-requeuer.sh",1234,0,"[",shellscript,content +1092,983889,"weekend-job-requeuer.sh",1235,0,"",shellscript,selection_keyboard +1093,984405,"weekend-job-requeuer.sh",1235,0,"4",shellscript,content +1094,984406,"weekend-job-requeuer.sh",1236,0,"",shellscript,selection_keyboard +1095,984539,"weekend-job-requeuer.sh",1236,0,",",shellscript,content +1096,984540,"weekend-job-requeuer.sh",1237,0,"",shellscript,selection_keyboard +1097,985537,"weekend-job-requeuer.sh",1243,0,"",shellscript,selection_command +1098,985721,"weekend-job-requeuer.sh",1242,0,"",shellscript,selection_command +1099,985873,"weekend-job-requeuer.sh",1241,0,"",shellscript,selection_command +1100,986020,"weekend-job-requeuer.sh",1240,0,"",shellscript,selection_command +1101,986170,"weekend-job-requeuer.sh",1239,0,"",shellscript,selection_command +1102,986958,"weekend-job-requeuer.sh",1239,0,"]",shellscript,content +1103,986959,"weekend-job-requeuer.sh",1240,0,"",shellscript,selection_keyboard +1104,987342,"weekend-job-requeuer.sh",1239,0,"",shellscript,selection_command +1105,988259,"weekend-job-requeuer.sh",1293,0,"",shellscript,selection_command +1106,988422,"weekend-job-requeuer.sh",1362,0,"",shellscript,selection_command +1107,988559,"weekend-job-requeuer.sh",1410,0,"",shellscript,selection_command +1108,988737,"weekend-job-requeuer.sh",1459,0,"",shellscript,selection_command +1109,989206,"weekend-job-requeuer.sh",1413,107,"",shellscript,content +1110,989287,"weekend-job-requeuer.sh",1364,0,"",shellscript,selection_command +1111,989454,"weekend-job-requeuer.sh",1355,0,"",shellscript,selection_command +1112,989754,"weekend-job-requeuer.sh",1247,0,"",shellscript,selection_command +1113,989959,"weekend-job-requeuer.sh",1354,0,"\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",shellscript,content +1114,989960,"weekend-job-requeuer.sh",1355,0,"",shellscript,selection_command +1115,990208,"weekend-job-requeuer.sh",1462,0,"",shellscript,selection_command +1116,990420,"weekend-job-requeuer.sh",1471,0,"",shellscript,selection_command +1117,990870,"weekend-job-requeuer.sh",1471,49,"",shellscript,content +1118,991123,"weekend-job-requeuer.sh",1462,0,"",shellscript,selection_command +1119,991543,"weekend-job-requeuer.sh",1462,9,"",shellscript,content +1120,1037838,"weekend-job-requeuer.sh",1469,0,"",shellscript,selection_mouse +1121,1039941,"weekend-job-requeuer.sh",1469,0,"\n",shellscript,content +1122,1040004,"weekend-job-requeuer.sh",1470,0,"6",shellscript,content +1123,1040005,"weekend-job-requeuer.sh",1471,0,"",shellscript,selection_keyboard +1124,1042187,"weekend-job-requeuer.sh",1470,1,"train_dynamics_cotraining_6_actions",shellscript,content +1125,1043806,"weekend-job-requeuer.sh",1470,35,"6",shellscript,content +1126,1044104,"weekend-job-requeuer.sh",1469,2,"",shellscript,content +1127,1046903,"weekend-job-requeuer.sh",1468,1,"",shellscript,content +1128,1046971,"weekend-job-requeuer.sh",1468,0,"6",shellscript,content +1129,1046972,"weekend-job-requeuer.sh",1469,0,"",shellscript,selection_keyboard +1130,1086407,"weekend-job-requeuer.sh",1365,0,"",shellscript,selection_mouse +1131,1086870,"weekend-job-requeuer.sh",1470,0,"",shellscript,selection_mouse +1132,1086976,"weekend-job-requeuer.sh",1469,0,"",shellscript,selection_command +1133,1087974,"weekend-job-requeuer.sh",1478,0,"",shellscript,selection_command +1134,1088137,"weekend-job-requeuer.sh",1527,0,"",shellscript,selection_command +1135,1088275,"weekend-job-requeuer.sh",1627,0,"",shellscript,selection_command +1136,1088421,"weekend-job-requeuer.sh",1635,0,"",shellscript,selection_command +1137,1088754,"weekend-job-requeuer.sh",1647,0,"\nsleep 6h",shellscript,content +1138,1088755,"weekend-job-requeuer.sh",1648,0,"",shellscript,selection_command +1139,1089137,"weekend-job-requeuer.sh",1649,0,"",shellscript,selection_command +1140,1089626,"weekend-job-requeuer.sh",1650,0,"",shellscript,selection_command +1141,1089673,"weekend-job-requeuer.sh",1651,0,"",shellscript,selection_command +1142,1089720,"weekend-job-requeuer.sh",1652,0,"",shellscript,selection_command +1143,1089721,"weekend-job-requeuer.sh",1653,0,"",shellscript,selection_command +1144,1089771,"weekend-job-requeuer.sh",1654,0,"",shellscript,selection_command +1145,1090871,"weekend-job-requeuer.sh",1654,1,"2",shellscript,content +1146,1091109,"weekend-job-requeuer.sh",1663,0,"",shellscript,selection_command +1147,1091353,"weekend-job-requeuer.sh",1744,0,"",shellscript,selection_command +1148,1091520,"weekend-job-requeuer.sh",1826,0,"",shellscript,selection_command +1149,1092089,"weekend-job-requeuer.sh",1908,0,"",shellscript,selection_command +1150,1092904,"weekend-job-requeuer.sh",1902,81,"sbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch",shellscript,selection_command +1151,1093191,"weekend-job-requeuer.sh",1902,163,"sbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",shellscript,selection_command +1152,1093504,"weekend-job-requeuer.sh",1902,0,"",shellscript,selection_command +1153,1094390,"weekend-job-requeuer.sh",1984,0,"#",shellscript,content +1154,1094391,"weekend-job-requeuer.sh",1902,0,"#",shellscript,content +1155,1094392,"weekend-job-requeuer.sh",1903,0,"",shellscript,selection_keyboard +1156,1094470,"weekend-job-requeuer.sh",1986,0," ",shellscript,content +1157,1094470,"weekend-job-requeuer.sh",1903,0," ",shellscript,content +1158,1094471,"weekend-job-requeuer.sh",1904,0,"",shellscript,selection_keyboard +1159,1095039,"weekend-job-requeuer.sh",1903,0,"",shellscript,selection_command +1160,1095304,"weekend-job-requeuer.sh",1821,0,"",shellscript,selection_command +1161,1095456,"weekend-job-requeuer.sh",1739,0,"",shellscript,selection_command +1162,1095607,"weekend-job-requeuer.sh",1658,0,"",shellscript,selection_command +1163,1095758,"weekend-job-requeuer.sh",1649,0,"",shellscript,selection_command +1164,1095887,"weekend-job-requeuer.sh",1629,0,"",shellscript,selection_command +1165,1096038,"weekend-job-requeuer.sh",1627,0,"",shellscript,selection_command +1166,1096587,"weekend-job-requeuer.sh",1629,0,"",shellscript,selection_command +1167,1096837,"weekend-job-requeuer.sh",1627,0,"",shellscript,selection_command +1168,1097020,"weekend-job-requeuer.sh",1521,0,"",shellscript,selection_command +1169,1097153,"weekend-job-requeuer.sh",1472,0,"",shellscript,selection_command +1170,1097804,"weekend-job-requeuer.sh",1463,0,"",shellscript,selection_command +1171,1097953,"weekend-job-requeuer.sh",1356,0,"",shellscript,selection_command +1172,1098120,"weekend-job-requeuer.sh",1248,0,"",shellscript,selection_command +1173,1098292,"weekend-job-requeuer.sh",1194,0,"",shellscript,selection_command +1174,1098421,"weekend-job-requeuer.sh",1248,0,"",shellscript,selection_command +1175,1098571,"weekend-job-requeuer.sh",1356,0,"",shellscript,selection_command +1176,1098753,"weekend-job-requeuer.sh",1463,0,"",shellscript,selection_command +1177,1098853,"weekend-job-requeuer.sh",1472,0,"",shellscript,selection_command +1178,1099403,"weekend-job-requeuer.sh",1521,0,"",shellscript,selection_command +1179,1099560,"weekend-job-requeuer.sh",1627,0,"",shellscript,selection_command +1180,1099706,"weekend-job-requeuer.sh",1629,0,"",shellscript,selection_command +1181,1099853,"weekend-job-requeuer.sh",1649,0,"",shellscript,selection_command +1182,1100122,"weekend-job-requeuer.sh",1656,0,"\necho ""requeueing batchsize scaling runs: 8nodes""",shellscript,content +1183,1100123,"weekend-job-requeuer.sh",1657,0,"",shellscript,selection_command +1184,1100460,"weekend-job-requeuer.sh",1658,0,"",shellscript,selection_command +1185,1100971,"weekend-job-requeuer.sh",1659,0,"",shellscript,selection_command +1186,1101003,"weekend-job-requeuer.sh",1660,0,"",shellscript,selection_command +1187,1101020,"weekend-job-requeuer.sh",1661,0,"",shellscript,selection_command +1188,1101053,"weekend-job-requeuer.sh",1662,0,"",shellscript,selection_command +1189,1101103,"weekend-job-requeuer.sh",1663,0,"",shellscript,selection_command +1190,1101142,"weekend-job-requeuer.sh",1664,0,"",shellscript,selection_command +1191,1101143,"weekend-job-requeuer.sh",1665,0,"",shellscript,selection_command +1192,1101193,"weekend-job-requeuer.sh",1666,0,"",shellscript,selection_command +1193,1101203,"weekend-job-requeuer.sh",1667,0,"",shellscript,selection_command +1194,1101239,"weekend-job-requeuer.sh",1668,0,"",shellscript,selection_command +1195,1101298,"weekend-job-requeuer.sh",1669,0,"",shellscript,selection_command +1196,1101303,"weekend-job-requeuer.sh",1670,0,"",shellscript,selection_command +1197,1101344,"weekend-job-requeuer.sh",1671,0,"",shellscript,selection_command +1198,1101389,"weekend-job-requeuer.sh",1672,0,"",shellscript,selection_command +1199,1101503,"weekend-job-requeuer.sh",1673,0,"",shellscript,selection_command +1200,1101639,"weekend-job-requeuer.sh",1674,0,"",shellscript,selection_command +1201,1102003,"weekend-job-requeuer.sh",1674,10,"",shellscript,content +1202,1102576,"weekend-job-requeuer.sh",1674,0,"m",shellscript,content +1203,1102577,"weekend-job-requeuer.sh",1675,0,"",shellscript,selection_keyboard +1204,1102804,"weekend-job-requeuer.sh",1675,0,"o",shellscript,content +1205,1102805,"weekend-job-requeuer.sh",1676,0,"",shellscript,selection_keyboard +1206,1102954,"weekend-job-requeuer.sh",1676,0,"d",shellscript,content +1207,1102954,"weekend-job-requeuer.sh",1677,0,"",shellscript,selection_keyboard +1208,1103157,"weekend-job-requeuer.sh",1677,0,"e",shellscript,content +1209,1103158,"weekend-job-requeuer.sh",1678,0,"",shellscript,selection_keyboard +1210,1103253,"weekend-job-requeuer.sh",1678,0,"l",shellscript,content +1211,1103254,"weekend-job-requeuer.sh",1679,0,"",shellscript,selection_keyboard +1212,1103620,"weekend-job-requeuer.sh",1679,0,"s",shellscript,content +1213,1103621,"weekend-job-requeuer.sh",1680,0,"",shellscript,selection_keyboard +1214,1103670,"weekend-job-requeuer.sh",1680,0,"i",shellscript,content +1215,1103671,"weekend-job-requeuer.sh",1681,0,"",shellscript,selection_keyboard +1216,1103857,"weekend-job-requeuer.sh",1681,0,"z",shellscript,content +1217,1103857,"weekend-job-requeuer.sh",1682,0,"",shellscript,selection_keyboard +1218,1103953,"weekend-job-requeuer.sh",1682,0,"e",shellscript,content +1219,1103954,"weekend-job-requeuer.sh",1683,0,"",shellscript,selection_keyboard +1220,1104054,"weekend-job-requeuer.sh",1683,0," ",shellscript,content +1221,1104055,"weekend-job-requeuer.sh",1684,0,"",shellscript,selection_keyboard +1222,1104570,"weekend-job-requeuer.sh",1683,0,"",shellscript,selection_command +1223,1106372,"weekend-job-requeuer.sh",1696,0,"",shellscript,selection_mouse +1224,1107338,"weekend-job-requeuer.sh",1696,2,"",shellscript,content +1225,1107654,"weekend-job-requeuer.sh",1696,6,"",shellscript,content +1226,1113770,"weekend-job-requeuer.sh",2110,0,"",shellscript,selection_mouse +1227,1113772,"weekend-job-requeuer.sh",2109,0,"",shellscript,selection_command +1228,1113937,"weekend-job-requeuer.sh",2109,1,"h",shellscript,selection_mouse +1229,1113938,"weekend-job-requeuer.sh",2110,0,"",shellscript,selection_command +1230,1114020,"weekend-job-requeuer.sh",1657,453,"echo ""requeueing modelsize scaling runs""\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch\n# sbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch\n# sbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",shellscript,selection_mouse +1231,1114070,"weekend-job-requeuer.sh",1077,1033,"sbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch\nsleep 3h\necho ""requeueing batchsize scaling runs: [4,16]nodes""\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch\nsleep 6h\necho ""requeueing batchsize scaling runs: 8nodes""\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch\n\n# modelsize scaling\nsleep 2h\necho ""requeueing modelsize scaling runs""\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch\n# sbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch\n# sbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",shellscript,selection_mouse +1232,1114071,"weekend-job-requeuer.sh",456,1654,"sbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch\n\n# lr sweep\nsleep 1h\necho ""requeueing lr sweep jobs""\nsbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch\nsbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch\n\n# batchsize scaling\nsleep 1h\necho ""requeueing batchsize scaling runs: [1,2]nodes""\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch\nsleep 3h\necho ""requeueing batchsize scaling runs: [4,16]nodes""\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch\nsleep 6h\necho ""requeueing batchsize scaling runs: 8nodes""\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch\n\n# modelsize scaling\nsleep 2h\necho ""requeueing modelsize scaling runs""\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch\n# sbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch\n# sbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",shellscript,selection_mouse +1233,1114103,"weekend-job-requeuer.sh",0,2110,"\necho ""weekend requeuer started...""\nsleep 9h\necho ""9 hours passed""\n# actionspace scaling\necho ""requeueing actionspace scaling jobs""\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch\n\n# lr sweep\nsleep 1h\necho ""requeueing lr sweep jobs""\nsbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch\nsbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch\n\n# batchsize scaling\nsleep 1h\necho ""requeueing batchsize scaling runs: [1,2]nodes""\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch\nsleep 3h\necho ""requeueing batchsize scaling runs: [4,16]nodes""\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch\nsleep 6h\necho ""requeueing batchsize scaling runs: 8nodes""\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch\n\n# modelsize scaling\nsleep 2h\necho ""requeueing modelsize scaling runs""\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch\n# sbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch\n# sbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/5_train_dyn_500M.sbatch",shellscript,selection_mouse +1234,1115271,"weekend-job-requeuer.sh",0,0,"",shellscript,selection_command +1235,1125108,"weekend-job-requeuer.sh",0,0,"\n",shellscript,content +1236,1125420,"weekend-job-requeuer.sh",1,0,"echo ""[INFO] weekend requeuer started at $(date)""",shellscript,content +1237,1125471,"weekend-job-requeuer.sh",51,35,"",shellscript,content +1238,1125687,"weekend-job-requeuer.sh",60,0,"echo ""[INFO] 9 hours passed at $(date)""\n",shellscript,content +1239,1125688,"weekend-job-requeuer.sh",100,0,"\n",shellscript,content +1240,1125720,"weekend-job-requeuer.sh",101,22,"",shellscript,content +1241,1126037,"weekend-job-requeuer.sh",123,0,"echo ""[INFO] Requeueing actionspace scaling jobs at $(date)""\n",shellscript,content +1242,1126170,"weekend-job-requeuer.sh",184,0,"for job in 12 20 50 6 8; do\n",shellscript,content +1243,1126620,"weekend-job-requeuer.sh",212,0," echo ""[INFO] Submitting actionspace scaling job: train_dynamics_cotraining_${job}_actions.sbatch""\n",shellscript,content +1244,1127187,"weekend-job-requeuer.sh",314,0," sbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_${job}_actions.sbatch\n",shellscript,content +1245,1127225,"weekend-job-requeuer.sh",430,0,"done\n",shellscript,content +1246,1127270,"weekend-job-requeuer.sh",435,0,"\n",shellscript,content +1247,1127271,"weekend-job-requeuer.sh",436,0,"# lr sweep\n",shellscript,content +1248,1127320,"weekend-job-requeuer.sh",436,593,"",shellscript,content +1249,1127653,"weekend-job-requeuer.sh",456,0,"echo ""[INFO] Requeueing lr sweep jobs at $(date)""\n",shellscript,content +1250,1127893,"weekend-job-requeuer.sh",506,0,"for lr in 1e-4 5e-5; do\n",shellscript,content +1251,1128771,"weekend-job-requeuer.sh",530,0," echo ""[INFO] Submitting lr sweep job: train_tokenizer_lr_${lr}.sbatch""\n",shellscript,content +1252,1129138,"weekend-job-requeuer.sh",605,0," sbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_${lr}.sbatch\n",shellscript,content +1253,1129254,"weekend-job-requeuer.sh",692,0,"done\n",shellscript,content +1254,1129255,"weekend-job-requeuer.sh",697,0,"\n",shellscript,content +1255,1129271,"weekend-job-requeuer.sh",698,197,"",shellscript,content +1256,1129704,"weekend-job-requeuer.sh",727,0,"echo ""[INFO] Requeueing batchsize scaling runs: [1,2]nodes at $(date)""\n",shellscript,content +1257,1129803,"weekend-job-requeuer.sh",798,0,"for nodes in 1 2; do\n",shellscript,content +1258,1130559,"weekend-job-requeuer.sh",819,0," echo ""[INFO] Submitting batchsize scaling job: train_dynamics_${nodes}_nodes.sbatch""\n",shellscript,content +1259,1133620,"weekend-job-requeuer.sh",908,0," sbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_${nodes}_nodes.sbatch\n",shellscript,content +1260,1133654,"weekend-job-requeuer.sh",1026,0,"done\n",shellscript,content +1261,1133655,"weekend-job-requeuer.sh",1031,0,"\n",shellscript,content +1262,1133688,"weekend-job-requeuer.sh",1032,0,"sleep 3h\n",shellscript,content +1263,1134287,"weekend-job-requeuer.sh",1041,0,"echo ""[INFO] Requeueing batchsize scaling runs: [4,16]nodes at $(date)""\n",shellscript,content +1264,1134389,"weekend-job-requeuer.sh",1113,0,"for nodes in 16 4; do\n",shellscript,content +1265,1134856,"weekend-job-requeuer.sh",1135,0," echo ""[INFO] Submitting batchsize scaling job: train_dynamics_${nodes}_nodes.sbatch""\n",shellscript,content +1266,1135737,"weekend-job-requeuer.sh",1224,0," sbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_${nodes}_nodes.sbatch\n",shellscript,content +1267,1135738,"weekend-job-requeuer.sh",1342,0,"done\n",shellscript,content +1268,1135739,"weekend-job-requeuer.sh",1347,0,"\n",shellscript,content +1269,1135753,"weekend-job-requeuer.sh",1348,0,"sleep 6h\n",shellscript,content +1270,1136005,"weekend-job-requeuer.sh",1357,0,"echo ""[INFO] Requeueing batchsize scaling runs: 8nodes at $(date)""\n",shellscript,content +1271,1136403,"weekend-job-requeuer.sh",1424,0,"echo ""[INFO] Submitting batchsize scaling job: train_dynamics_8_nodes.sbatch""\n",shellscript,content +1272,1136936,"weekend-job-requeuer.sh",1502,603,"",shellscript,content +1273,1137553,"weekend-job-requeuer.sh",1639,0,"echo ""[INFO] Requeueing modelsize scaling runs at $(date)""\n",shellscript,content +1274,1137890,"weekend-job-requeuer.sh",1698,0,"for model in 1_train_dyn_36M 2_train_dyn_110M 3_train_dyn_180M; do\n",shellscript,content +1275,1138192,"weekend-job-requeuer.sh",1765,0," echo ""[INFO] Submitting modelsize scaling job: ${model}.sbatch""\n",shellscript,content +1276,1138537,"weekend-job-requeuer.sh",1833,0," sbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/${model}.sbatch\n",shellscript,content +1277,1138538,"weekend-job-requeuer.sh",1911,0,"done\n",shellscript,content +1278,1138886,"weekend-job-requeuer.sh",1916,0,"# echo ""[INFO] Submitting modelsize scaling job: 4_train_dyn_270M.sbatch""\n",shellscript,content +1279,1139187,"weekend-job-requeuer.sh",1990,0,"# sbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/4_train_dyn_270M.sbatch\n",shellscript,content +1280,1139592,"weekend-job-requeuer.sh",2074,0,"# echo ""[INFO] Submitting modelsize scaling job: 5_train_dyn_500M.sbatch""\n",shellscript,content +1281,1139922,"weekend-job-requeuer.sh",2148,370,"",shellscript,content +1282,1148887,"weekend-job-requeuer.sh",0,0,"",shellscript,selection_command +1283,1150189,"weekend-job-requeuer.sh",2074,74,"",shellscript,content +1284,1150189,"weekend-job-requeuer.sh",1639,350,"echo ""requeueing modelsize scaling runs""\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch\nsbatch slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch",shellscript,content +1285,1150189,"weekend-job-requeuer.sh",1357,144,"echo ""requeueing batchsize scaling runs: 8nodes""",shellscript,content +1286,1150189,"weekend-job-requeuer.sh",1041,306,"echo ""requeueing batchsize scaling runs: [4,16]nodes""\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch",shellscript,content +1287,1150190,"weekend-job-requeuer.sh",727,304,"echo ""requeueing batchsize scaling runs: [1,2]nodes""\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch\nsbatch slurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch",shellscript,content +1288,1150190,"weekend-job-requeuer.sh",456,240,"echo ""requeueing lr sweep jobs""\nsbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch\nsbatch slurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch",shellscript,content +1289,1150190,"weekend-job-requeuer.sh",123,311,"echo ""requeueing actionspace scaling jobs""\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch\nsbatch slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch",shellscript,content +1290,1150190,"weekend-job-requeuer.sh",60,40,"echo ""9 hours passed""",shellscript,content +1291,1150190,"weekend-job-requeuer.sh",1,49,"echo ""weekend requeuer started...""",shellscript,content +1292,1151487,"weekend-job-requeuer.sh",0,0,"\n",shellscript,content +1293,1151572,"weekend-job-requeuer.sh",1,0,"echo ""[INFO] weekend requeuer started at $(date)""",shellscript,content +1294,1151604,"weekend-job-requeuer.sh",51,35,"",shellscript,content +1295,1151640,"weekend-job-requeuer.sh",60,0,"echo ""[INFO] 9 hours passed at $(date)""\n",shellscript,content +1296,1151641,"weekend-job-requeuer.sh",100,0,"\n",shellscript,content +1297,1151720,"weekend-job-requeuer.sh",101,22,"",shellscript,content +1298,1151776,"weekend-job-requeuer.sh",123,0,"echo ""[INFO] Requeueing actionspace scaling jobs at $(date)""\n",shellscript,content +1299,1151872,"weekend-job-requeuer.sh",184,0,"echo ""[INFO] Submitting: train_dynamics_cotraining_12_actions.sbatch""\n",shellscript,content +1300,1151973,"weekend-job-requeuer.sh",254,43,"",shellscript,content +1301,1152092,"weekend-job-requeuer.sh",362,0,"echo ""[INFO] Submitting: train_dynamics_cotraining_20_actions.sbatch""\n",shellscript,content +1302,1152439,"weekend-job-requeuer.sh",540,0,"echo ""[INFO] Submitting: train_dynamics_cotraining_50_actions.sbatch""\n",shellscript,content +1303,1152704,"weekend-job-requeuer.sh",718,0,"echo ""[INFO] Submitting: train_dynamics_cotraining_6_actions.sbatch""\n",shellscript,content +1304,1152990,"weekend-job-requeuer.sh",894,0,"echo ""[INFO] Submitting: train_dynamics_cotraining_8_actions.sbatch""\n",shellscript,content +1305,1153288,"weekend-job-requeuer.sh",1091,0,"echo ""[INFO] Requeueing lr sweep jobs at $(date)""\n",shellscript,content +1306,1153654,"weekend-job-requeuer.sh",1141,0,"echo ""[INFO] Submitting: train_tokenizer_lr_1e-4.sbatch""\n",shellscript,content +1307,1153803,"weekend-job-requeuer.sh",1198,32,"",shellscript,content +1308,1153993,"weekend-job-requeuer.sh",1280,0,"echo ""[INFO] Submitting: train_tokenizer_lr_5e-5.sbatch""\n",shellscript,content +1309,1154322,"weekend-job-requeuer.sh",1449,0,"echo ""[INFO] Requeueing batchsize scaling runs: [1,2]nodes at $(date)""\n",shellscript,content +1310,1154670,"weekend-job-requeuer.sh",1520,0,"echo ""[INFO] Submitting: train_dynamics_1_nodes.sbatch""\n",shellscript,content +1311,1154887,"weekend-job-requeuer.sh",1576,53,"",shellscript,content +1312,1154903,"weekend-job-requeuer.sh",1683,0,"echo ""[INFO] Submitting: train_dynamics_2_nodes.sbatch""\n",shellscript,content +1313,1155243,"weekend-job-requeuer.sh",1846,0,"\n",shellscript,content +1314,1155560,"weekend-job-requeuer.sh",1856,0,"echo ""[INFO] Requeueing batchsize scaling runs: [4,16]nodes at $(date)""\n",shellscript,content +1315,1155903,"weekend-job-requeuer.sh",1928,0,"echo ""[INFO] Submitting: train_dynamics_16_nodes.sbatch""\n",shellscript,content +1316,1156303,"weekend-job-requeuer.sh",1985,54,"",shellscript,content +1317,1156421,"weekend-job-requeuer.sh",2093,0,"echo ""[INFO] Submitting: train_dynamics_4_nodes.sbatch""\n",shellscript,content +1318,1156721,"weekend-job-requeuer.sh",2256,0,"\n",shellscript,content +1319,1156837,"weekend-job-requeuer.sh",2266,0,"echo ""[INFO] Requeueing batchsize scaling runs: 8nodes at $(date)""\n",shellscript,content +1320,1157160,"weekend-job-requeuer.sh",2333,0,"echo ""[INFO] Submitting: train_dynamics_8_nodes.sbatch""\n",shellscript,content +1321,1157453,"weekend-job-requeuer.sh",2389,49,"",shellscript,content +1322,1157753,"weekend-job-requeuer.sh",2526,0,"echo ""[INFO] Requeueing modelsize scaling runs at $(date)""\n",shellscript,content +1323,1157888,"weekend-job-requeuer.sh",2585,0,"echo ""[INFO] Submitting: 1_train_dyn_36M.sbatch""\n",shellscript,content +1324,1158092,"weekend-job-requeuer.sh",2634,41,"",shellscript,content +1325,1158370,"weekend-job-requeuer.sh",2715,0,"echo ""[INFO] Submitting: 2_train_dyn_110M.sbatch""\n",shellscript,content +1326,1158887,"weekend-job-requeuer.sh",2847,0,"echo ""[INFO] Submitting: 3_train_dyn_180M.sbatch""\n",shellscript,content +1327,1159337,"weekend-job-requeuer.sh",2979,0,"# echo ""[INFO] Submitting: 4_train_dyn_270M.sbatch""\n",shellscript,content +1328,1159774,"weekend-job-requeuer.sh",3115,0,"# echo ""[INFO] Submitting: 5_train_dyn_500M.sbatch""\n",shellscript,content +1329,1166154,"TERMINAL",0,0,"date",,terminal_command +1330,1166171,"TERMINAL",0,0,"]633;E;2025-07-12 12:37:05 date;16c08be6-3885-420f-ac53-f5272aed6e54]633;CSat Jul 12 12:37:05 PM CEST 2025\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output +1331,1167775,"weekend-job-requeuer.sh",0,0,"",shellscript,tab +1332,1170621,"weekend-job-requeuer.sh",100,0,"",shellscript,selection_mouse +1333,1170926,"weekend-job-requeuer.sh",100,1,"\n",shellscript,selection_mouse +1334,1171536,"weekend-job-requeuer.sh",62,0,"",shellscript,selection_mouse +1335,1171687,"weekend-job-requeuer.sh",60,4,"echo",shellscript,selection_mouse +1336,1171920,"weekend-job-requeuer.sh",60,6,"echo """,shellscript,selection_mouse +1337,1171937,"weekend-job-requeuer.sh",60,11,"echo ""[INFO",shellscript,selection_mouse +1338,1171944,"weekend-job-requeuer.sh",60,14,"echo ""[INFO] 9",shellscript,selection_mouse +1339,1171970,"weekend-job-requeuer.sh",60,20,"echo ""[INFO] 9 hours",shellscript,selection_mouse +1340,1172004,"weekend-job-requeuer.sh",60,27,"echo ""[INFO] 9 hours passed",shellscript,selection_mouse +1341,1172053,"weekend-job-requeuer.sh",60,30,"echo ""[INFO] 9 hours passed at",shellscript,selection_mouse +1342,1172103,"weekend-job-requeuer.sh",60,31,"echo ""[INFO] 9 hours passed at ",shellscript,selection_mouse +1343,1172105,"weekend-job-requeuer.sh",60,32,"echo ""[INFO] 9 hours passed at $",shellscript,selection_mouse +1344,1172154,"weekend-job-requeuer.sh",60,33,"echo ""[INFO] 9 hours passed at $(",shellscript,selection_mouse +1345,1172170,"weekend-job-requeuer.sh",60,37,"echo ""[INFO] 9 hours passed at $(date",shellscript,selection_mouse +1346,1172272,"weekend-job-requeuer.sh",60,38,"echo ""[INFO] 9 hours passed at $(date)",shellscript,selection_mouse +1347,1172310,"weekend-job-requeuer.sh",60,39,"echo ""[INFO] 9 hours passed at $(date)""",shellscript,selection_mouse +1348,1176320,"TERMINAL",0,0,"echo ""[INFO] 9 hours passed at $(date)""",,terminal_command +1349,1176387,"TERMINAL",0,0,"]633;E;2025-07-12 12:37:16 echo ""[INFO] 9 hours passed at $(date)"";16c08be6-3885-420f-ac53-f5272aed6e54]633;C[INFO] 9 hours passed at Sat Jul 12 12:37:16 PM CEST 2025\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar]633;D;0",,terminal_output +1350,1179772,"weekend-job-requeuer.sh",0,0,"",shellscript,tab +1351,1180659,"weekend-job-requeuer.sh",764,0,"",shellscript,selection_mouse +1352,1191154,"TERMINAL",0,0,"bash",,terminal_focus +1353,1195653,"TERMINAL",0,0,"runner",,terminal_command +1354,1197603,"TERMINAL",0,0,"sync-runner",,terminal_command +1355,1197655,"TERMINAL",0,0,"]633;E;2025-07-12 12:37:37 sync-runner;cd331af2-3b46-4f78-aca4-9dda848374ca]633;Csending incremental file list\r\n",,terminal_output +1356,1199941,"TERMINAL",0,0,"./\r\nweekend-job-requeuer.sh\r\n",,terminal_output +1357,1200771,"TERMINAL",0,0,"slurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_12_actions.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_20_actions.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_50_actions.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_6_actions.sbatch\r\nslurm/jobs/mihir/horeka/action_space_scaling/co_training/train_dynamics_cotraining_8_actions.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_16_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_1_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_2_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_4_nodes.sbatch\r\nslurm/jobs/mihir/horeka/batchsize_scaling/dynamics_cotraining/sqrt_lr/train_dynamics_8_nodes.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_1e-4.sbatch\r\nslurm/jobs/mihir/horeka/lr_tuning/tokenizer/train_tokenizer_lr_5e-5.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/1_train_dyn_36M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/2_train_dyn_110M.sbatch\r\nslurm/jobs/mihir/horeka/modelsize_scaling/dynamics/3_train_dyn_180M.sbatch\r\n",,terminal_output +1358,1201076,"TERMINAL",0,0,"\r\nsent 47,894 bytes received 428 bytes 13,806.29 bytes/sec\r\ntotal size is 129,691,309 speedup is 2,683.90\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar_jobs]633;D;0",,terminal_output +1359,1209406,"weekend-job-requeuer.sh",0,0,"",shellscript,tab +1360,1212320,"weekend-job-requeuer.sh",1,0,"",shellscript,selection_mouse +1361,1212470,"weekend-job-requeuer.sh",1,4,"echo",shellscript,selection_mouse +1362,1212625,"weekend-job-requeuer.sh",1,55,"echo ""[INFO] weekend requeuer started at $(date)""\nsleep",shellscript,selection_mouse +1363,1212655,"weekend-job-requeuer.sh",1,58,"echo ""[INFO] weekend requeuer started at $(date)""\nsleep 9h",shellscript,selection_mouse +1364,1212786,"weekend-job-requeuer.sh",1,86,"echo ""[INFO] weekend requeuer started at $(date)""\nsleep 9h\necho ""[INFO] 9 hours passed",shellscript,selection_mouse +1365,1212789,"weekend-job-requeuer.sh",1,87,"echo ""[INFO] weekend requeuer started at $(date)""\nsleep 9h\necho ""[INFO] 9 hours passed ",shellscript,selection_mouse +1366,1212819,"weekend-job-requeuer.sh",1,89,"echo ""[INFO] weekend requeuer started at $(date)""\nsleep 9h\necho ""[INFO] 9 hours passed at",shellscript,selection_mouse +1367,1212853,"weekend-job-requeuer.sh",1,90,"echo ""[INFO] weekend requeuer started at $(date)""\nsleep 9h\necho ""[INFO] 9 hours passed at ",shellscript,selection_mouse +1368,1212855,"weekend-job-requeuer.sh",1,91,"echo ""[INFO] weekend requeuer started at $(date)""\nsleep 9h\necho ""[INFO] 9 hours passed at $",shellscript,selection_mouse +1369,1212903,"weekend-job-requeuer.sh",1,92,"echo ""[INFO] weekend requeuer started at $(date)""\nsleep 9h\necho ""[INFO] 9 hours passed at $(",shellscript,selection_mouse +1370,1212904,"weekend-job-requeuer.sh",1,96,"echo ""[INFO] weekend requeuer started at $(date)""\nsleep 9h\necho ""[INFO] 9 hours passed at $(date",shellscript,selection_mouse +1371,1213005,"weekend-job-requeuer.sh",1,98,"echo ""[INFO] weekend requeuer started at $(date)""\nsleep 9h\necho ""[INFO] 9 hours passed at $(date)""",shellscript,selection_mouse +1372,1213169,"weekend-job-requeuer.sh",99,0,"",shellscript,selection_mouse +1373,1213173,"weekend-job-requeuer.sh",98,0,"",shellscript,selection_command +1374,1216904,"TERMINAL",0,0,"tmux",,terminal_command +1375,1217006,"TERMINAL",0,0,"]633;E;2025-07-12 12:37:56 tmux;cd331af2-3b46-4f78-aca4-9dda848374ca]633;C[?1049h[?1h=[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[>c[>q[?25l\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[0] 0:bash* ""hkn1991.localdomain"" 12:37 12-Jul-25(B[?12l[?25h(B[?12l[?25h[?1006l[?1000l[?1002l[?1003l[?2004l[?25l\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n[0] 0:bash* ""hkn1991.localdomain"" 12:37 12-Jul-25(B[?12l[?25h",,terminal_output +1376,1217070,"TERMINAL",0,0,"[?7727h",,terminal_output +1377,1218707,"TERMINAL",0,0,"jafar[tum_cte0515@hkn1991 jafar_jobs]$ [?2004h",,terminal_output +1378,1220020,"TERMINAL",0,0,"s",,terminal_output +1379,1220172,"TERMINAL",0,0,"o",,terminal_output +1380,1220286,"TERMINAL",0,0,"u",,terminal_output +1381,1220407,"TERMINAL",0,0,"r",,terminal_output +1382,1220569,"TERMINAL",0,0,"c",,terminal_output +1383,1220675,"TERMINAL",0,0,"e",,terminal_output +1384,1220823,"TERMINAL",0,0," ",,terminal_output +1385,1220939,"TERMINAL",0,0,".",,terminal_output +1386,1221054,"TERMINAL",0,0,"v",,terminal_output +1387,1221437,"TERMINAL",0,0,"e",,terminal_output +1388,1221636,"TERMINAL",0,0,"",,terminal_output +1389,1222471,"TERMINAL",0,0,"n",,terminal_output +1390,1223020,"TERMINAL",0,0,"v",,terminal_output +1391,1223637,"TERMINAL",0,0,"",,terminal_output +1392,1224503,"TERMINAL",0,0,"",,terminal_output +1393,1224820,"TERMINAL",0,0,"",,terminal_output +1394,1225520,"TERMINAL",0,0,"",,terminal_output +1395,1225720,"TERMINAL",0,0,"",,terminal_output +1396,1226310,"TERMINAL",0,0,"c",,terminal_output +1397,1226486,"TERMINAL",0,0,"d",,terminal_output +1398,1226624,"TERMINAL",0,0," ",,terminal_output +1399,1226753,"TERMINAL",0,0,".",,terminal_output +1400,1226869,"TERMINAL",0,0,".",,terminal_output +1401,1227453,"TERMINAL",0,0,"/",,terminal_output +1402,1228170,"TERMINAL",0,0,"j",,terminal_output +1403,1228270,"TERMINAL",0,0,"a",,terminal_output +1404,1228553,"TERMINAL",0,0,"f",,terminal_output +1405,1228860,"TERMINAL",0,0,"ar",,terminal_output +1406,1229186,"TERMINAL",0,0,"\r\n[?2004l[?2004hjafar[tum_cte0515@hkn1991 jafar]$ ",,terminal_output +1407,1229721,"TERMINAL",0,0,"s",,terminal_output +1408,1229853,"TERMINAL",0,0,"o",,terminal_output +1409,1229959,"TERMINAL",0,0,"ur",,terminal_output +1410,1230220,"TERMINAL",0,0,"c",,terminal_output +1411,1230371,"TERMINAL",0,0,"e",,terminal_output +1412,1230454,"TERMINAL",0,0," ",,terminal_output +1413,1230603,"TERMINAL",0,0,".v",,terminal_output +1414,1231123,"TERMINAL",0,0,"env/",,terminal_output +1415,1231486,"TERMINAL",0,0,"b",,terminal_output +1416,1231603,"TERMINAL",0,0,"in/",,terminal_output +1417,1231853,"TERMINAL",0,0,"a",,terminal_output +1418,1232008,"TERMINAL",0,0,"c[?25l[0] 0:bash* ""hkn1991.localdomain"" 12:38 12-Jul-25(B[?12l[?25h",,terminal_output +1419,1232153,"TERMINAL",0,0,"tivate",,terminal_output +1420,1232486,"TERMINAL",0,0,"\r\n[?2004l[?2004h(jafar) jafar[tum_cte0515@hkn1991 jafar]$ ",,terminal_output +1421,1232937,"TERMINAL",0,0,"r",,terminal_output +1422,1233169,"TERMINAL",0,0,"u",,terminal_output +1423,1233269,"TERMINAL",0,0,"n",,terminal_output +1424,1233469,"TERMINAL",0,0,"n",,terminal_output +1425,1233523,"TERMINAL",0,0,"e",,terminal_output +1426,1233621,"TERMINAL",0,0,"r",,terminal_output +1427,1233721,"TERMINAL",0,0,"\r\n[?2004l(jafar) jafar[tum_cte0515@hkn1991 jafar_jobs]$ [?2004h",,terminal_output +1428,1235021,"TERMINAL",0,0,"s",,terminal_output +1429,1235470,"TERMINAL",0,0,"y",,terminal_output +1430,1235720,"TERMINAL",0,0,"n",,terminal_output +1431,1235788,"TERMINAL",0,0,"c",,terminal_output +1432,1236019,"TERMINAL",0,0,"-",,terminal_output +1433,1236286,"TERMINAL",0,0,"r",,terminal_output +1434,1236387,"TERMINAL",0,0,"u",,terminal_output +1435,1236507,"TERMINAL",0,0,"n",,terminal_output +1436,1236636,"TERMINAL",0,0,"n",,terminal_output +1437,1236707,"TERMINAL",0,0,"er",,terminal_output +1438,1237341,"TERMINAL",0,0,"\r\n[?2004lsending incremental file list\r\n",,terminal_output +1439,1237589,"TERMINAL",0,0,"\r\nsent 19,342 bytes received 107 bytes 38,898.00 bytes/sec\r\ntotal size is 129,691,309 speedup is 6,668.28\r\n[?25l[0] 0:sh* ""hkn1991.localdomain"" 12:38 12-Jul-25(B[?12l[?25h(jafar) jafar[tum_cte0515@hkn1991 jafar_jobs]$ [?2004h",,terminal_output +1440,1240136,"TERMINAL",0,0,"[?25l[0] 0:bash* ""hkn1991.localdomain"" 12:38 12-Jul-25(B[?12l[?25hc",,terminal_output +1441,1240238,"TERMINAL",0,0,"l",,terminal_output +1442,1240371,"TERMINAL",0,0,"e",,terminal_output +1443,1240537,"TERMINAL",0,0,"a",,terminal_output +1444,1240653,"TERMINAL",0,0,"r",,terminal_output +1445,1240888,"TERMINAL",0,0,"\r\n[?2004l[?2004h(jafar) jafar[tum_cte0515@hkn1991 jafar_jobs]$ ",,terminal_output +1446,1243570,"TERMINAL",0,0,"s",,terminal_output +1447,1243841,"TERMINAL",0,0,"h ",,terminal_output +1448,1244069,"TERMINAL",0,0,"w",,terminal_output +1449,1244220,"TERMINAL",0,0,"e",,terminal_output +1450,1244409,"TERMINAL",0,0,"ekend-job-",,terminal_output +1451,1245486,"TERMINAL",0,0,"r",,terminal_output +1452,1245736,"TERMINAL",0,0,"equeuer.sh ",,terminal_output +1453,1247186,"TERMINAL",0,0,">",,terminal_output +1454,1247438,"TERMINAL",0,0," ",,terminal_output +1455,1248603,"TERMINAL",0,0,"",,terminal_output +1456,1249087,"TERMINAL",0,0," ",,terminal_output +1457,1249404,"TERMINAL",0,0,"r",,terminal_output +1458,1249619,"TERMINAL",0,0,"e",,terminal_output +1459,1249836,"TERMINAL",0,0,"q",,terminal_output +1460,1249970,"TERMINAL",0,0,"u",,terminal_output +1461,1250038,"TERMINAL",0,0,"e",,terminal_output +1462,1250137,"TERMINAL",0,0,"u",,terminal_output +1463,1250207,"TERMINAL",0,0,"e",,terminal_output +1464,1250333,"TERMINAL",0,0,"r",,terminal_output +1465,1251371,"TERMINAL",0,0,".",,terminal_output +1466,1251603,"TERMINAL",0,0,"l",,terminal_output +1467,1251741,"TERMINAL",0,0,"o",,terminal_output +1468,1251870,"TERMINAL",0,0,"g",,terminal_output +1469,1292020,"TERMINAL",0,0,"[?25l[0] 0:bash* ""hkn1991.localdomain"" 12:39 12-Jul-25(B[?12l[?25h",,terminal_output +1470,1305003,"TERMINAL",0,0,"\r\n[?2004l",,terminal_output +1471,1315837,"TERMINAL",0,0,"(B[?1l>[?12l[?25h[?1000l[?1002l[?1003l[?1006l[?1005l[?7727l[?1004l[?1049l[detached (from session 0)]\r\n]0;tum_cte0515@hkn1991:~/Projects/jafar_jobs]633;D;0",,terminal_output +1472,1339311,"weekend-job-requeuer.sh",0,0,"",shellscript,tab diff --git a/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-e11dd94b-65ba-41d4-a0d1-6151e7930d741752738204303-2025_07_17-09.44.02.899/source.csv b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-e11dd94b-65ba-41d4-a0d1-6151e7930d741752738204303-2025_07_17-09.44.02.899/source.csv new file mode 100644 index 0000000000000000000000000000000000000000..8563620a8663f1ec1f9e35c8ea1ab836effb2f63 --- /dev/null +++ b/927a8af5474e5654810c00ce2e09fd2de87d3e5722f33fa1090d867db114e403/crowd-code-e11dd94b-65ba-41d4-a0d1-6151e7930d741752738204303-2025_07_17-09.44.02.899/source.csv @@ -0,0 +1,3517 @@ +Sequence,Time,File,RangeOffset,RangeLength,Text,Language,Type +1,5,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun_new_arch\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n# source .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-new-arch-run-$slurm_job_id \\n --tags dynamics \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n ",shellscript,tab +2,780,"extension-output-pdoom-org.crowd-code-#1-crowd-code",0,0,"9:44:02 AM [info] Activating crowd-code\n9:44:02 AM [info] Recording started\n9:44:02 AM [info] Initializing git provider using file system watchers...\n9:44:03 AM [info] Git repository found\n9:44:03 AM [info] Git provider initialized successfully\n9:44:03 AM [info] Initial git state: [object Object]\n",Log,tab +3,3975,"TERMINAL",0,0,"/bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt",,terminal_command +4,4021,"TERMINAL",0,0,"]633;E;2025-07-17 09:44:06 /bin/python3 /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/printEnvVariablesToFile.py /hkfs/home/project/hk-project-p0023960/tum_cte0515/.cursor-server/extensions/ms-python.python-2024.12.3-linux-x64/python_files/deactivate/bash/envVars.txt;93375418-5fef-411f-acff-e36f5d8f8e16]633;C",,terminal_output +5,7025,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"",shellscript,tab +6,25708,"genie.py",0,0,"from typing import Dict, Any\n\nimport optax\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nfrom flax.training.train_state import TrainState\nimport orbax.checkpoint as ocp\n\nfrom models.dynamics import DynamicsMaskGIT, DynamicsAutoregressive\nfrom models.lam import LatentActionModel\nfrom models.tokenizer import TokenizerVQVAE\n\nimport os\nimport grain\n\n\nclass Genie(nn.Module):\n """"""Genie model""""""\n\n # --- Tokenizer ---\n in_dim: int\n tokenizer_dim: int\n latent_patch_dim: int\n num_patch_latents: int\n patch_size: int\n tokenizer_num_blocks: int\n tokenizer_num_heads: int\n # --- LAM ---\n lam_dim: int\n latent_action_dim: int\n num_latent_actions: int\n lam_patch_size: int\n lam_num_blocks: int\n lam_num_heads: int\n lam_co_train: bool\n # --- Dynamics ---\n dyna_dim: int\n dyna_num_blocks: int\n dyna_num_heads: int\n use_maskgit: bool\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n dropout: float = 0.0\n mask_limit: float = 0.0\n\n def setup(self):\n self.tokenizer = TokenizerVQVAE(\n in_dim=self.in_dim,\n model_dim=self.tokenizer_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_patch_latents,\n patch_size=self.patch_size,\n num_blocks=self.tokenizer_num_blocks,\n num_heads=self.tokenizer_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n self.lam = LatentActionModel(\n in_dim=self.in_dim,\n model_dim=self.lam_dim,\n latent_dim=self.latent_patch_dim,\n num_latents=self.num_latent_actions,\n patch_size=self.lam_patch_size,\n num_blocks=self.lam_num_blocks,\n num_heads=self.lam_num_heads,\n dropout=0.0,\n codebook_dropout=0.0,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n if self.use_maskgit:\n self.dynamics = DynamicsMaskGIT(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n mask_limit=self.mask_limit,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n ) \n else:\n self.dynamics = DynamicsAutoregressive(\n model_dim=self.dyna_dim,\n num_latents=self.num_patch_latents,\n num_blocks=self.dyna_num_blocks,\n num_heads=self.dyna_num_heads,\n dropout=self.dropout,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\n lam_outputs = self.lam.vq_encode(batch[""videos""], training=False)\n latent_actions = jax.lax.cond(\n self.lam_co_train,\n lambda: lam_outputs[""z_q""],\n lambda: jax.lax.stop_gradient(lam_outputs[""z_q""])\n )\n outputs = dict(\n video_tokens=jax.lax.stop_gradient(tokenizer_outputs[""indices""]),\n latent_actions=latent_actions,\n )\n outputs[""mask_rng""] = batch[""mask_rng""]\n dyna_outputs = self.dynamics(outputs, training)\n outputs.update(dyna_outputs)\n mle_indices = jnp.argmax(outputs[""token_logits""], axis=-1)\n outputs[""recon""] = self.tokenizer.decode(\n mle_indices, batch[""videos""].shape[2:4]\n )\n outputs[""lam_indices""] = lam_outputs[""indices""]\n return outputs\n\n\n def sample_causal(\n self,\n batch: Dict[str, Any],\n seq_len: int,\n temperature: float = 1,\n sample_argmax: bool = False,\n ):\n """"""\n Autoregressively samples up to `seq_len` future frames using the causal transformer backend.\n\n - Input frames are tokenized once.\n - Future frames are generated one at a time, each conditioned on all previous frames.\n - All frames are detokenized in a single pass at the end.\n\n Args:\n batch: Dict with at least ""videos"" (B, T, H, W, C)\n seq_len: total number of frames to generate (including context)\n temperature: sampling temperature\n sample_argmax: if True, use argmax instead of sampling\n\n Returns:\n Generated video frames (B, seq_len, H, W, C)\n """"""\n # --- Encode context frames ---\n tokenizer_out = self.tokenizer.vq_encode(batch[""videos""], training=False)\n token_idxs = tokenizer_out[""indices""] # (B, T, N)\n B, T, N = token_idxs.shape\n\n # --- Prepare initial token sequence ---\n # Pad with zeros for future frames\n pad_shape = (B, seq_len - T, N)\n token_idxs_full = jnp.concatenate(\n [token_idxs, jnp.zeros(pad_shape, dtype=token_idxs.dtype)], axis=1\n ) # (B, seq_len, N)\n\n # --- Prepare latent actions ---\n # If you have action tokens, use them; otherwise, use zeros\n if ""latent_actions"" in batch:\n action_tokens = self.lam.vq.get_codes(batch[""latent_actions""])\n else:\n action_tokens = jnp.zeros((B, seq_len, self.num_latent_actions), dtype=jnp.int32)\n\n # --- Autoregressive generation loop ---\n rng = batch.get(""rng"", None)\n for t in range(T, seq_len):\n # Feed all tokens up to t (i.e., frames 0..t-1) to the dynamics model\n dyna_inputs = {\n ""video_tokens"": token_idxs_full[:, :t, :], # (B, t, N)\n ""latent_actions"": action_tokens[:, :t, ...], # (B, t, ...)\n ""mask_rng"": batch.get(""mask_rng"", None),\n }\n dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(\n step_rng, next_token_logits / temperature, axis=-1\n ) # (B, N)\n\n # Insert the generated tokens into the sequence\n token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n\n # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )\n return final_frames\n\n\n @nn.compact\n def sample_maskgit(\n self,\n batch: Dict[str, Any],\n seq_len: int,\n steps: int = 25,\n temperature: float = 1,\n sample_argmax: bool = False,\n ) -> Any:\n """"""\n Autoregressively samples up to `seq_len` future frames, following Figure 8 of the paper.\n\n - Input frames are tokenized once.\n - Future frames are generated autoregressively in token space.\n - All frames are detokenized in a single pass.\n\n Note:\n - For interactive or step-wise sampling, detokenization should occur after each action.\n - To maintain consistent tensor shapes across timesteps, all current and future frames are decoded at every step.\n - Temporal causal structure is preserved by \n a) reapplying the mask before each decoding step.\n b) a temporal causal mask is applied within each ST-transformer block.\n\n Dimension keys:\n B: batch size \n T: number of input (conditioning) frames \n N: patches per frame \n S: sequence length \n A: action space \n D: model latent dimension\n """"""\n # --- Encode videos and actions ---\n tokenizer_out = self.tokenizer.vq_encode(batch[""videos""], training=False)\n token_idxs = tokenizer_out[""indices""] # (B, T, N)\n B, T, N = token_idxs.shape\n pad_shape = (B, seq_len - T, N)\n pad = jnp.zeros(pad_shape, dtype=token_idxs.dtype)\n token_idxs = jnp.concatenate([token_idxs, pad], axis=1) # (B, S, N)\n action_tokens = self.lam.vq.get_codes(batch[""latent_actions""])\n\n MaskGITLoop = nn.scan(\n MaskGITStep,\n variable_broadcast=""params"",\n split_rngs={""params"": False},\n in_axes=0,\n out_axes=0,\n length=steps,\n )\n \n loop_fn = MaskGITLoop(\n dynamics=self.dynamics,\n tokenizer=self.tokenizer,\n temperature=temperature,\n sample_argmax=sample_argmax,\n steps=steps,\n )\n\n def generation_step_fn(carry, step_t):\n rng, current_token_idxs = carry\n rng, step_rng = jax.random.split(rng)\n\n # Mask current and future frames (i.e., t >= step_t)\n mask = jnp.arange(seq_len) >= step_t # (S,)\n mask = jnp.broadcast_to(mask[None, :, None], (B, seq_len, N)) # (B, S, N)\n mask = mask.astype(bool)\n masked_token_idxs = current_token_idxs * ~mask\n\n # --- Initialize and run MaskGIT loop ---\n init_carry_maskgit = (\n step_rng,\n masked_token_idxs,\n mask,\n action_tokens,\n )\n final_carry_maskgit, _ = loop_fn(init_carry_maskgit, jnp.arange(steps))\n updated_token_idxs = final_carry_maskgit[1]\n new_carry = (rng, updated_token_idxs)\n return new_carry, None\n\n # --- Run the autoregressive generation using scan ---\n initial_carry = (batch[""rng""], token_idxs)\n timesteps_to_scan = jnp.arange(T, seq_len)\n final_carry, _ = jax.lax.scan(\n generation_step_fn,\n initial_carry,\n timesteps_to_scan\n )\n final_token_idxs = final_carry[1]\n\n # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n final_token_idxs,\n video_hw=batch[""videos""].shape[2:4],\n )\n return final_frames\n\n def vq_encode(self, batch, training) -> Dict[str, Any]:\n # --- Preprocess videos ---\n lam_output = self.lam.vq_encode(batch[""videos""], training=training)\n return lam_output[""indices""]\n\n\nclass MaskGITStep(nn.Module):\n dynamics: nn.Module\n tokenizer: nn.Module\n temperature: float\n sample_argmax: bool\n steps: int\n\n @nn.compact\n def __call__(self, carry, x):\n rng, token_idxs, mask, action_tokens = carry\n step = x\n N = token_idxs.shape[2]\n\n # --- Construct + encode video ---\n vid_embed = self.dynamics.patch_embed(token_idxs) # (B, S, N, D)\n mask_token = self.dynamics.mask_token # (1, 1, 1, D,)\n mask_expanded = mask[..., None] # (B, S, N, 1) \n vid_embed = jnp.where(mask_expanded, mask_token, vid_embed)\n\n # --- Predict transition ---\n act_embed = self.dynamics.action_up(action_tokens)\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n unmasked_ratio = jnp.cos(jnp.pi * (step + 1) / (self.steps * 2))\n step_temp = self.temperature * (1.0 - unmasked_ratio)\n final_logits = self.dynamics.dynamics(vid_embed) / step_temp\n\n # --- Sample new tokens for final frame ---\n if self.sample_argmax:\n sampled_token_idxs = jnp.argmax(final_logits, axis=-1)\n else:\n rng, _rng = jax.random.split(rng)\n sampled_token_idxs = jax.random.categorical(_rng, final_logits)\n gather_fn = jax.vmap(jax.vmap(jax.vmap(lambda x, y: x[y])))\n final_token_probs = gather_fn(jax.nn.softmax(final_logits), sampled_token_idxs)\n final_token_probs += ~mask\n # Update masked tokens only\n token_idxs = jnp.where(mask, sampled_token_idxs, token_idxs)\n\n # --- Update mask ---\n num_unmasked_tokens = jnp.round(N * (1.0 - unmasked_ratio)).astype(int)\n idx_mask = jnp.arange(final_token_probs.shape[-1]) > num_unmasked_tokens\n sorted_idxs = jnp.argsort(final_token_probs, axis=-1, descending=True)\n mask_update_fn = jax.vmap(lambda msk, ids: msk.at[ids].set(idx_mask))\n new_mask = mask_update_fn(mask, sorted_idxs)\n\n new_carry = (rng, token_idxs, new_mask, action_tokens)\n return new_carry, None\n\ndef restore_genie_components(\n train_state: TrainState,\n sharding: jax.sharding.NamedSharding,\n grain_iterator: grain.DataLoaderIterator,\n inputs: Dict[str, jax.Array],\n rng: jax.Array,\n args,\n):\n """"""Restore pre-trained Genie components""""""\n rng, _rng = jax.random.split(rng)\n\n # dummy values since we only use tx to initialize the dummy train states\n dummy_tx = optax.adamw(\n learning_rate=optax.constant_schedule(args.max_lr),\n b1=0.9,\n b2=0.9,\n weight_decay=1e-4,\n )\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add('model_state', ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler)\n handler_registry.add('dataloader_state', grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler)\n \n\n checkpoint_options = ocp.CheckpointManagerOptions(\n step_format_fixed_length=6,\n )\n tokenizer_checkpoint_manager = ocp.CheckpointManager(\n directory=args.tokenizer_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_tokenizer = TokenizerVQVAE(\n in_dim=args.image_channels,\n model_dim=args.tokenizer_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n num_blocks=args.tokenizer_num_blocks,\n num_heads=args.tokenizer_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n )\n tokenizer_init_params = dummy_tokenizer.init(_rng, inputs)\n dummy_tokenizer_train_state = TrainState.create(\n apply_fn=dummy_tokenizer.apply, params=tokenizer_init_params, tx=dummy_tx\n )\n abstract_sharded_tokenizer_state = _create_abstract_sharded_pytree(\n dummy_tokenizer_train_state, sharding\n )\n restored_tokenizer = tokenizer_checkpoint_manager.restore(\n step=tokenizer_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_sharded_tokenizer_state),\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator),\n ),\n )[""model_state""]\n restored_tokenizer_params = restored_tokenizer.params[""params""]\n train_state.params[""params""][""tokenizer""].update(restored_tokenizer_params)\n tokenizer_checkpoint_manager.close()\n\n if args.lam_checkpoint:\n lam_checkpoint_manager = ocp.CheckpointManager(\n directory=args.lam_checkpoint,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n dummy_lam = LatentActionModel(\n in_dim=args.image_channels,\n model_dim=args.lam_dim,\n latent_dim=args.latent_patch_dim,\n num_latents=args.num_latent_actions,\n patch_size=args.lam_patch_size,\n num_blocks=args.lam_num_blocks,\n num_heads=args.lam_num_heads,\n dropout=args.dropout,\n codebook_dropout=args.dropout,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n )\n lam_init_params = dummy_lam.init(_rng, inputs)\n dummy_lam_train_state = TrainState.create(\n apply_fn=dummy_lam.apply, params=lam_init_params, tx=dummy_tx\n )\n abstract_sharded_lam_state = _create_abstract_sharded_pytree(\n dummy_lam_train_state, sharding\n )\n restored_lam = lam_checkpoint_manager.restore(\n step=lam_checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_sharded_lam_state),\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator),\n ),\n )[""model_state""]\n restored_lam_params = restored_lam.params[""params""]\n # Genie does not initialize all LAM modules, thus we omit those extra modules during restoration\n # (f.srambical) FIXME: Currently, this is a small HBM memory crunch since the LAM's decoder is loaded into HBM and immediately dicarded.\n # A workaround would be to restore to host memory first, and only move the weights to HBM after pruning the decoder\n restored_lam_params = {\n k: v\n for k, v in restored_lam_params.items()\n if k in train_state.params[""params""][""lam""]\n }\n train_state.params[""params""][""lam""].update(restored_lam_params)\n lam_checkpoint_manager.close()\n\n return train_state\n\ndef _create_abstract_sharded_pytree(pytree_template, sharding_spec):\n """"""Replaces arrays in a pytree with ShapeDtypeStructs having the given sharding.""""""\n\n def map_fn(leaf_template):\n if hasattr(leaf_template, ""shape"") and hasattr(leaf_template, ""dtype""):\n return jax.ShapeDtypeStruct(\n leaf_template.shape, leaf_template.dtype, sharding=sharding_spec\n )\n return leaf_template\n\n return jax.tree_util.tree_map(map_fn, pytree_template)",python,tab +7,30708,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py",0,0,"from typing import Dict, Any\n\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nimport einops\n\nfrom utils.nn import STTransformer\n\n\nclass DynamicsMaskGIT(nn.Module):\n """"""MaskGIT dynamics model""""""\n\n model_dim: int\n num_latents: int\n num_blocks: int\n num_heads: int\n dropout: float\n mask_limit: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n def setup(self):\n self.dynamics = STTransformer(\n self.model_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n spacial_bert=True,\n )\n self.patch_embed = nn.Embed(self.num_latents, self.model_dim)\n self.mask_token = self.param(\n ""mask_token"",\n nn.initializers.lecun_uniform(),\n (1, 1, 1, self.model_dim),\n )\n self.action_up = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n # --- Mask videos ---\n vid_embed = self.patch_embed(batch[""video_tokens""])\n if training:\n rng1, rng2 = jax.random.split(batch[""mask_rng""])\n mask_prob = jax.random.uniform(rng1, minval=self.mask_limit)\n mask = jax.random.bernoulli(rng2, mask_prob, vid_embed.shape[:-1])\n mask = mask.at[:, 0].set(False)\n vid_embed = jnp.where(jnp.expand_dims(mask, -1), self.mask_token, vid_embed)\n else:\n mask = None\n\n # --- Predict transition ---\n act_embed = self.action_up(batch[""latent_actions""])\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n logits = self.dynamics(vid_embed)\n return dict(token_logits=logits, mask=mask)\n\n\nclass DynamicsAutoregressive(nn.Module):\n """"""Autoregressive (causal) dynamics model""""""\n\n model_dim: int\n num_latents: int\n num_blocks: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n def setup(self):\n self.dynamics = STTransformer(\n self.model_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n spacial_bert=False,\n )\n self.patch_embed = nn.Embed(self.num_latents, self.model_dim)\n self.action_up = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n vid_embed = self.patch_embed(batch[""video_tokens""])\n act_embed = self.action_up(batch[""latent_actions""])\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n logits = self.dynamics(vid_embed)\n mask = jnp.ones(vid_embed.shape[:-1])\n return dict(token_logits=logits, mask=mask)",python,tab +8,36439,"train_dynamics.py",0,0,"from dataclasses import dataclass, field\nimport os\n\nimport einops\nfrom flax.training.train_state import TrainState\nfrom jax.sharding import Mesh, PartitionSpec, NamedSharding\nfrom jax.experimental.mesh_utils import create_device_mesh\nimport optax\nimport orbax.checkpoint as ocp\nimport numpy as np\nimport dm_pix as pix\nimport jax\nimport jax.numpy as jnp\nimport tyro\nimport wandb\nimport grain\n\nfrom genie import Genie, restore_genie_components\nfrom utils.dataloader import get_dataloader\nfrom utils.lr_utils import get_lr_schedule\nfrom utils.parameter_utils import count_parameters_by_component\n\n@dataclass\nclass Args:\n # Experiment\n num_steps: int = 200_000\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = """"\n save_ckpt: bool = False\n restore_ckpt: bool = False\n # Optimization\n batch_size: int = 36\n init_lr: float = 0.0\n max_lr: float = 3e-5\n decay_end: float = 0.0\n wsd_decay_steps: int = 10000 # NOTE: wsd_decay_steps will only be used when using a wsd-schedule\n warmup_steps: int = 5000\n lr_schedule : str = ""wsd"" # supported options: wsd, cos\n # Tokenizer\n tokenizer_dim: int = 512\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 8\n tokenizer_num_heads: int = 8\n tokenizer_checkpoint: str = """"\n # LAM\n lam_dim: int = 512\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 8\n lam_num_heads: int = 8\n lam_checkpoint: str = """"\n # Dynamics\n dyna_dim: int = 512\n dyna_num_blocks: int = 12\n dyna_num_heads: int = 8\n dropout: float = 0.0\n mask_limit: float = 0.5\n use_maskgit: bool = False\n param_dtype: jnp.dtype = jnp.float32\n dtype: jnp.dtype = jnp.bfloat16\n # Logging\n log: bool = False\n entity: str = """"\n project: str = """"\n name: str = ""train_dynamics""\n tags: list[str] = field(default_factory=lambda: [""dynamics""])\n log_interval: int = 5\n log_image_interval: int = 250\n ckpt_dir: str = """"\n log_checkpoint_interval: int = 25000\n log_checkpoint_keep_period: int = 20000\n log_gradients: bool = False\n wandb_id: str = """"\n\n\nargs = tyro.cli(Args)\n\n\ndef dynamics_loss_fn(params, state, inputs):\n """"""Compute masked dynamics loss""""""\n inputs[""videos""] = inputs[""videos""].astype(args.dtype) / 255.0\n outputs = state.apply_fn(\n params,\n inputs,\n training=True,\n rngs={""params"": inputs[""rng""], ""dropout"": inputs[""dropout_rng""]},\n )\n mask = outputs[""mask""]\n outputs[""token_logits""] = outputs[""token_logits""].astype(jnp.float32)\n logits = outputs[""token_logits""]\n targets = outputs[""video_tokens""]\n\n if not args.use_maskgit:\n logits = outputs[""token_logits""][:, :, :-1]\n targets = outputs[""video_tokens""][:, :, 1:]\n mask = outputs[""mask""][:, :, 1:] \n ce_loss = optax.softmax_cross_entropy_with_integer_labels(\n logits, targets\n )\n ce_loss = (mask * ce_loss).sum() / mask.sum()\n acc = logits.argmax(-1) == targets\n acc = (mask * acc).sum() / mask.sum()\n select_probs = jax.nn.softmax(logits)\n gt = inputs[""videos""].clip(0, 1).reshape(-1, *inputs[""videos""].shape[2:])\n recon = outputs[""recon""].clip(0, 1).reshape(-1, *outputs[""recon""].shape[2:])\n psnr = pix.psnr(gt, recon).mean() # type: ignore\n ssim = pix.ssim(gt, recon).mean() # type: ignore\n _, index_counts_lam = jnp.unique_counts(\n jnp.ravel(outputs[""lam_indices""]), size=args.num_latent_actions, fill_value=0\n )\n _, index_counts_tokenizer = jnp.unique_counts(\n jnp.ravel(outputs[""video_tokens""]), size=args.num_patch_latents, fill_value=0\n )\n codebook_usage_lam = (index_counts_lam != 0).mean()\n codebook_usage_tokenizer = (index_counts_tokenizer != 0).mean()\n metrics = dict(\n cross_entropy_loss=ce_loss,\n masked_token_accuracy=acc,\n select_logit=logits.max(-1).mean(),\n select_p=select_probs.max(-1).mean(),\n entropy=jax.scipy.special.entr(select_probs).sum(-1).mean(),\n psnr=psnr,\n ssim=ssim,\n codebook_usage_lam=codebook_usage_lam,\n codebook_usage_tokenizer=codebook_usage_tokenizer,\n )\n return ce_loss, (outputs[""recon""], metrics)\n\n\n@jax.jit\ndef train_step(state, inputs):\n """"""Update state and compute metrics""""""\n grad_fn = jax.value_and_grad(dynamics_loss_fn, has_aux=True, allow_int=True)\n (loss, (recon, metrics)), grads = grad_fn(state.params, state, inputs)\n state = state.apply_gradients(grads=grads)\n if args.log_gradients:\n metrics[""gradients_std/""] = jax.tree.map(\n lambda x: x.std(), grads[""params""][""dynamics""]\n )\n return state, loss, recon, metrics\n\n\nif __name__ == ""__main__"":\n jax.distributed.initialize()\n num_devices = jax.device_count()\n if num_devices == 0:\n raise ValueError(""No JAX devices found."")\n print(f""Running on {num_devices} devices."")\n\n if args.batch_size % num_devices != 0:\n raise ValueError(\n f""Global batch size {args.batch_size} must be divisible by ""\n f""number of devices {num_devices}.""\n )\n\n per_device_batch_size_for_init = args.batch_size // num_devices\n\n rng = jax.random.PRNGKey(args.seed)\n\n # --- Initialize model ---\n genie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=not args.lam_checkpoint,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n dropout=args.dropout,\n mask_limit=args.mask_limit,\n use_maskgit=args.use_maskgit,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n )\n rng, _rng = jax.random.split(rng)\n image_shape = (args.image_height, args.image_width, args.image_channels)\n dummy_inputs = dict(\n videos=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len, *image_shape),\n dtype=args.dtype,\n ),\n action=jnp.zeros(\n (per_device_batch_size_for_init, args.seq_len), dtype=args.dtype\n ),\n mask_rng=_rng,\n )\n rng, _rng = jax.random.split(rng)\n init_params = genie.init(_rng, dummy_inputs)\n\n param_counts = count_parameters_by_component(init_params)\n\n if args.log and jax.process_index() == 0:\n wandb_init_kwargs = {\n ""entity"": args.entity,\n ""project"": args.project,\n ""name"": args.name,\n ""tags"": args.tags,\n ""group"": ""debug"",\n ""config"": args,\n }\n\n if args.wandb_id:\n wandb_init_kwargs.update(\n {\n ""id"": args.wandb_id,\n ""resume"": ""allow"",\n }\n )\n wandb.init(**wandb_init_kwargs)\n\n wandb.config.update({""model_param_count"": param_counts})\n\n print(""Parameter counts:"")\n print(param_counts)\n\n # --- Initialize optimizer ---\n lr_schedule = get_lr_schedule(args.lr_schedule, \n args.init_lr, \n args.max_lr, \n args.decay_end, \n args.num_steps, \n args.warmup_steps, \n args.wsd_decay_steps)\n tx = optax.adamw(learning_rate=lr_schedule, b1=0.9, b2=0.9, weight_decay=1e-4, mu_dtype=args.dtype)\n train_state = TrainState.create(apply_fn=genie.apply, params=init_params, tx=tx)\n\n device_mesh_arr = create_device_mesh((num_devices,))\n mesh = Mesh(devices=device_mesh_arr, axis_names=(""data"",))\n\n replicated_sharding = NamedSharding(mesh, PartitionSpec())\n videos_sharding = NamedSharding(mesh, PartitionSpec(""data"", None, None, None, None))\n train_state = jax.device_put(train_state, replicated_sharding)\n\n # --- Initialize checkpoint manager ---\n step = 0\n handler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\n handler_registry.add(\n ""model_state"", ocp.args.StandardSave, ocp.handlers.StandardCheckpointHandler\n )\n handler_registry.add(\n ""model_state"", ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler\n )\n handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointSave, grain.checkpoint.CheckpointHandler) # type: ignore\n handler_registry.add(""dataloader_state"", grain.checkpoint.CheckpointRestore, grain.checkpoint.CheckpointHandler) # type: ignore\n\n checkpoint_options = ocp.CheckpointManagerOptions(\n save_interval_steps=args.log_checkpoint_interval,\n max_to_keep=3,\n keep_period=args.log_checkpoint_keep_period,\n step_format_fixed_length=6,\n cleanup_tmp_directories=True,\n )\n\n checkpoint_manager = ocp.CheckpointManager(\n args.ckpt_dir,\n options=checkpoint_options,\n handler_registry=handler_registry,\n )\n\n # --- Create DataLoaderIterator from dataloader ---\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=8,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n\n # --- Restore checkpoint ---\n if args.restore_ckpt:\n # Restore full dynamics model\n abstract_train_state = jax.tree_util.tree_map(\n ocp.utils.to_shape_dtype_struct, train_state\n )\n restored = checkpoint_manager.restore(\n checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_train_state),\n dataloader_state=grain.checkpoint.CheckpointRestore(grain_iterator),\n ),\n )\n train_state = restored[""model_state""]\n grain_iterator = restored[""dataloader_state""]\n step = checkpoint_manager.latest_step() or 0\n print(f""Restored dataloader and model state from step {step}"")\n else:\n # Restore from pre-trained tokenizer (and LAM)\n train_state = restore_genie_components(\n train_state, replicated_sharding, grain_iterator, dummy_inputs, rng, args\n )\n\n # --- TRAIN LOOP ---\n dataloader = (jax.make_array_from_process_local_data(videos_sharding, elem) for elem in grain_iterator) # type: ignore\n while step < args.num_steps:\n for videos in dataloader:\n # --- Train step ---\n rng, _rng, _rng_dropout, _rng_mask = jax.random.split(rng, 4)\n\n inputs = dict(\n videos=videos,\n rng=_rng,\n dropout_rng=_rng_dropout,\n mask_rng=_rng_mask,\n )\n train_state, loss, recon, metrics = train_step(train_state, inputs)\n metrics[""lr""] = lr_schedule(step)\n print(f""Step {step}, loss: {loss}"")\n step += 1\n\n # --- Logging ---\n if args.log:\n if step % args.log_interval == 0 and jax.process_index() == 0:\n wandb.log(\n {\n ""loss"": loss,\n ""step"": step,\n **metrics,\n }\n )\n if step % args.log_image_interval == 0:\n gt_seq = inputs[""videos""][0].astype(jnp.float32) / 255.0\n recon_seq = recon[0].clip(0, 1)\n comparison_seq = jnp.concatenate((gt_seq, recon_seq), axis=1)\n comparison_seq = einops.rearrange(\n comparison_seq * 255, ""t h w c -> h (t w) c""\n )\n if jax.process_index() == 0:\n log_images = dict(\n image=wandb.Image(np.asarray(gt_seq[args.seq_len - 1])),\n recon=wandb.Image(np.asarray(recon_seq[args.seq_len - 1])),\n true_vs_recon=wandb.Image(\n np.asarray(comparison_seq.astype(np.uint8))\n ),\n )\n wandb.log(log_images)\n # --- Checkpointing ---\n if args.save_ckpt and step % args.log_checkpoint_interval == 0:\n checkpoint_manager.save(\n step,\n args=ocp.args.Composite(\n model_state=ocp.args.StandardSave(train_state),\n dataloader_state=grain.checkpoint.CheckpointSave(\n grain_iterator\n ),\n ),\n )\n print(f""Saved checkpoint at step {step}"")\n if step >= args.num_steps:\n break\n\n checkpoint_manager.close()\n",python,tab +9,37974,"genie.py",0,0,"",python,tab +10,46539,"TERMINAL",0,0,"",,terminal_focus +11,53309,"TERMINAL",0,0,"idling",,terminal_command +12,53374,"TERMINAL",0,0,"]633;E;2025-07-17 09:44:56 idling;fcc07419-26ce-4c93-b0b0-b7e76ab3be27]633;C[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn1990.localdomain: Thu Jul 17 09:44:56 2025Partition dev_cpuonly: 11 nodes idle\rPartition cpuonly: 12 nodes idle\rPartition dev_accelerated:\t 3 nodes idle\rPartition accelerated: 53 nodes idle\rPartition dev_accelerated-h100 :\t 1 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 8 nodes idle",,terminal_output +13,54403,"TERMINAL",0,0,"7\t ",,terminal_output +14,54778,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;0",,terminal_output +15,58427,"TERMINAL",0,0,"salloc --time=10:00:00 --partition=accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5",,terminal_command +16,58495,"TERMINAL",0,0,"]633;E;2025-07-17 09:45:01 salloc --time=10:00:00 --partition=accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5;fcc07419-26ce-4c93-b0b0-b7e76ab3be27]633;Csalloc: Granted job allocation 3352994\r\n",,terminal_output +17,58647,"TERMINAL",0,0,"salloc: Waiting for resource configuration\r\n",,terminal_output +18,60959,"TERMINAL",0,0,"",,terminal_focus +19,75216,"TERMINAL",0,0,"salloc --time=10:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5",,terminal_command +20,75334,"TERMINAL",0,0,"]633;E;2025-07-17 09:45:18 salloc --time=10:00:00 --partition=accelerated --nodes=1 --ntasks-per-node=1 --gres=gpu:1 --cpus-per-task=5;c200d8ee-3924-40d0-a474-739536917d72]633;Csalloc: Granted job allocation 3352996\r\n",,terminal_output +21,75414,"TERMINAL",0,0,"salloc: Waiting for resource configuration\r\n",,terminal_output +22,85665,"TERMINAL",0,0,"salloc: Nodes hkn[0605,0614] are ready for job\r\n",,terminal_output +23,86509,"TERMINAL",0,0,"]0;tum_cte0515@hkn0605:~/Projects/jafar[?2004h[tum_cte0515@hkn0605 jafar]$ ",,terminal_output +24,95029,"TERMINAL",0,0,"srun",,terminal_focus +25,96448,"TERMINAL",0,0,"[?25lso[?25h",,terminal_output +26,96529,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +27,96625,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +28,96679,"TERMINAL",0,0,"[?25lr[?25h",,terminal_output +29,96848,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +30,97032,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +31,97118,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +32,97263,"TERMINAL",0,0,"[?25l.v[?25h",,terminal_output +33,97662,"TERMINAL",0,0,"env/",,terminal_output +34,98234,"TERMINAL",0,0,"",,terminal_output +35,98695,"TERMINAL",0,0,"[?25lv[?25h",,terminal_output +36,98888,"TERMINAL",0,0,"",,terminal_output +37,99775,"TERMINAL",0,0,"[?25lb[?25h",,terminal_output +38,99875,"TERMINAL",0,0,"in/",,terminal_output +39,100138,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +40,100226,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +41,100375,"TERMINAL",0,0,"tivate",,terminal_output +42,100628,"TERMINAL",0,0,"[?25l[?2004l\r]0;tum_cte0515@hkn0605:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0605 jafar]$ [?25h",,terminal_output +43,102441,"TERMINAL",0,0,"salloc: Nodes hkn0804 are ready for job\r\n",,terminal_output +44,103275,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h[tum_cte0515@hkn0804 jafar]$ ",,terminal_output +45,128966,"TERMINAL",0,0,"srun",,terminal_focus +46,130579,"TERMINAL",0,0,"[?25lso[?25h",,terminal_output +47,130656,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +48,130752,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +49,130806,"TERMINAL",0,0,"[?25lr[?25h",,terminal_output +50,131127,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +51,131248,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +52,131387,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +53,131570,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +54,131638,"TERMINAL",0,0,"[?25lv[?25h",,terminal_output +55,131851,"TERMINAL",0,0,"env/",,terminal_output +56,132235,"TERMINAL",0,0,"[?25lbin/[?25h",,terminal_output +57,132916,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +58,133013,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +59,133199,"TERMINAL",0,0,"tivate",,terminal_output +60,133577,"TERMINAL",0,0,"[?25l[?2004l\r[?25h]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +61,134458,"TERMINAL",0,0,"srun",,terminal_focus +62,136112,"TERMINAL",0,0,"srun",,terminal_focus +63,141857,"TERMINAL",0,0,"\r(reverse-i-search)`': ",,terminal_output +64,142069,"TERMINAL",0,0,"s': source .venv/bin/activate\r",,terminal_output +65,142340,"TERMINAL",0,0,"[?25lsm': smi[?25h",,terminal_output +66,143268,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +67,143719,"TERMINAL",0,0,"\r(reverse-i-search)`': ",,terminal_output +68,143982,"TERMINAL",0,0,"s': source .venv/bin/activate\r",,terminal_output +69,144208,"TERMINAL",0,0,"[?25lsa': salloc --time=10:00:00 --partition=accelerated --nodes=2 --ntasks-per-node=4 --gres=gpu:4 --cpus-per-task=5\r[?25h",,terminal_output +70,144346,"TERMINAL",0,0,"[?25lsm': git add sample.py[?25h",,terminal_output +71,144620,"TERMINAL",0,0,"[?25lasp': git add sample.py[?25h[?25lsl': git add sample.py[?25h",,terminal_output +72,144677,"TERMINAL",0,0,"e': git add sample.py",,terminal_output +73,144916,"TERMINAL",0,0,"[?25ls.': git add sample.py[?25h",,terminal_output +74,146383,"TERMINAL",0,0,"python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_500M_32_node --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked --dyna_dim=1536 --dyna_num_blocks=24 --dyna_num_heads=24\r",,terminal_output +75,147615,"TERMINAL",0,0,"\r\n\r\n\r",,terminal_output +76,149137,"TERMINAL",0,0,"\rjafar) [tum_cte0515@hkn0804 jafar]$ python sample.py --checkpoint /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/dynamics-cotraining-modelsize-scaling/train_dynamics_modelsize_scaling_500M_32_node --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +77,149828,"TERMINAL",0,0,"",,terminal_output +78,150015,"TERMINAL",0,0,"",,terminal_output +79,150128,"TERMINAL",0,0,"",,terminal_output +80,151337,"TERMINAL",0,0,"",,terminal_output +81,151834,"TERMINAL",0,0,"",,terminal_output +82,151974,"TERMINAL",0,0,"",,terminal_output +83,152207,"TERMINAL",0,0,"",,terminal_output +84,152346,"TERMINAL",0,0,"",,terminal_output +85,152507,"TERMINAL",0,0,"",,terminal_output +86,152699,"TERMINAL",0,0,"",,terminal_output +87,152884,"TERMINAL",0,0,"",,terminal_output +88,153069,"TERMINAL",0,0,"",,terminal_output +89,153248,"TERMINAL",0,0,"",,terminal_output +90,153428,"TERMINAL",0,0,"",,terminal_output +91,153688,"TERMINAL",0,0,"",,terminal_output +92,153832,"TERMINAL",0,0,"",,terminal_output +93,153954,"TERMINAL",0,0,"\r\n\r",,terminal_output +94,154235,"TERMINAL",0,0,"",,terminal_output +95,154422,"TERMINAL",0,0,"",,terminal_output +96,154611,"TERMINAL",0,0,"",,terminal_output +97,154795,"TERMINAL",0,0,"",,terminal_output +98,154981,"TERMINAL",0,0,"",,terminal_output +99,155225,"TERMINAL",0,0,"",,terminal_output +100,155417,"TERMINAL",0,0,"",,terminal_output +101,155825,"TERMINAL",0,0,"",,terminal_output +102,156361,"TERMINAL",0,0,"",,terminal_output +103,157691,"TERMINAL",0,0," --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r\n\r",,terminal_output +104,173066,"TERMINAL",0,0,"""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\r",,terminal_output +105,178041,"TERMINAL",0,0,"[?25l""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115""\r\n[?2004l\r[?25h",,terminal_output +106,193583,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +107,194204,"TERMINAL",0,0,"ERROR:2025-07-17 09:47:16,952:jax._src.xla_bridge:444: Jax plugin configuration error: Exception when calling jax_plugins.xla_cuda12.initialize()\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax_plugins/xla_cuda12/__init__.py"", line 197, in _version_check\r\n version = get_version()\r\nRuntimeError: jaxlib/cuda/versions_helpers.cc:81: operation cusparseGetProperty(MAJOR_VERSION, &major) failed: The cuSPARSE library was not found.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 442, in discover_pjrt_plugins\r\n plugin_module.initialize()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax_plugins/xla_cuda12/__init__.py"", line 324, in initialize\r\n _check_cuda_versions(raise_on_first_error=True)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax_plugins/xla_cuda12/__init__.py"", line 262, in _check_cuda_versions\r\n _version_check(""cuSPARSE"", cuda_versions.cusparse_get_version,\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax_plugins/xla_cuda12/__init__.py"", line 201, in _version_check\r\n raise RuntimeError(err_msg) from e\r\nRuntimeError: Unable to load cuSPARSE. Is it installed?\r\nERROR:jax._src.xla_bridge:Jax plugin configuration error: Exception when calling jax_plugins.xla_cuda12.initialize()\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax_plugins/xla_cuda12/__init__.py"", line 197, in _version_check\r\n version = get_version()\r\nRuntimeError: jaxlib/cuda/versions_helpers.cc:81: operation cusparseGetProperty(MAJOR_VERSION, &major) failed: The cuSPARSE library was not found.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/xla_bridge.py"", line 442, in discover_pjrt_plugins\r\n plugin_module.initialize()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax_plugins/xla_cuda12/__init__.py"", line 324, in initialize\r\n _check_cuda_versions(raise_on_first_error=True)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax_plugins/xla_cuda12/__init__.py"", line 262, in _check_cuda_versions\r\n _version_check(""cuSPARSE"", cuda_versions.cusparse_get_version,\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax_plugins/xla_cuda12/__init__.py"", line 201, in _version_check\r\n raise RuntimeError(err_msg) from e\r\nRuntimeError: Unable to load cuSPARSE. Is it installed?\r\nWARNING:2025-07-17 09:47:16,961:jax._src.xla_bridge:794: An NVIDIA GPU may be present on this machine, but a CUDA-enabled jaxlib is not installed. Falling back to cpu.\r\nWARNING:jax._src.xla_bridge:An NVIDIA GPU may be present on this machine, but a CUDA-enabled jaxlib is not installed. Falling back to cpu.\r\n",,terminal_output +108,199785,"TERMINAL",0,0,"^C",,terminal_output +109,199978,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 99, in \r\n params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 98, in __call__\r\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/tokenizer.py"", line 66, in vq_encode\r\n x = self.encoder(x) # (B, T, N, E)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 227, in __call__\r\n x = STBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 154, in __call__\r\n z = nn.MultiHeadAttention(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/attention.py"", line 676, in __call__\r\n out = DenseGeneral(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/linear.py"", line 164, in __call__\r\n kernel = self.param(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/linear.py"", line 154, in kernel_init_wrap\r\n return jnp.reshape(kernel, shape)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/lax_numpy.py"", line 2016, in reshape\r\n return a.reshape(shape, order=order) # type: ignore[call-overload,union-attr]\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 1089, in meth\r\n return getattr(self.aval, name).fun(self, *args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 315, in _reshape\r\n return lax.reshape(self, newshape, None, out_sharding=out_sharding)\r\njax._src.source_info_util.JaxStackTraceBeforeTransformation: KeyboardInterrupt\r\n\r\nThe preceding stack trace is the source of the JAX operation that, once transformed by JAX, triggered the following exception.\r\n\r\n--------------------\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 99, in \r\n params = genie.init(_rng, dummy_inputs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/traceback_util.py"", line 182, in reraise_with_filtered_traceback\r\n return fun(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/module.py"", line 2452, in init\r\n _, v_out = self.init_with_output(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/traceback_util.py"", line 182, in reraise_with_filtered_traceback\r\n return fun(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/module.py"", line 2304, in init_with_output\r\n return init_with_output(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/core/scope.py"", line 1115, in wrapper\r\n return apply(fn, mutable=mutable, flags=init_flags)(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/core/scope.py"", line 1079, in wrapper\r\n y = fn(root, *args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/module.py"", line 3093, in scope_fn\r\n return fn(module.clone(parent=scope, _deep_clone=True), *args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/module.py"", line 699, in wrapped_module_method\r\n return self._call_wrapped_method(fun, args, kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/module.py"", line 1216, in _call_wrapped_method\r\n y = run_fun(self, *args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 98, in __call__\r\n tokenizer_outputs = self.tokenizer.vq_encode(batch[""videos""], training=False)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/module.py"", line 699, in wrapped_module_method\r\n return self._call_wrapped_method(fun, args, kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/module.py"", line 1216, in _call_wrapped_method\r\n y = run_fun(self, *args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/tokenizer.py"", line 66, in vq_encode\r\n x = self.encoder(x) # (B, T, N, E)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/module.py"", line 699, in wrapped_module_method\r\n return self._call_wrapped_method(fun, args, kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/module.py"", line 1216, in _call_wrapped_method\r\n y = run_fun(self, *args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/utils/nn.py"", line 227, in __call__\r\n x = STBlock(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/linen/transforms.py"", line 433, in wrapped_fn\r\n return trafo_fn(module_scopes, *args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/core/lift.py"", line 319, in wrapper\r\n y, out_variable_groups_xs_t = fn(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/flax/core/lift.py"", line 1474, in inner\r\n return rematted(variable_groups, rng_groups, *args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/traceback_util.py"", line 182, in reraise_with_filtered_traceback\r\n return fun(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/ad_checkpoint.py"", line 333, in fun_remat\r\n out_flat = remat_p.bind(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/core.py"", line 536, in bind\r\n return self._true_bind(*args, **params)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/core.py"", line 552, in _true_bind\r\n return self.bind_with_trace(prev_trace, args, params)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/core.py"", line 562, in bind_with_trace\r\n return trace.process_primitive(self, args, params)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/core.py"", line 1066, in process_primitive\r\n return primitive.impl(*args, **params)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/ad_checkpoint.py"", line 514, in remat_impl\r\n return core.eval_jaxpr(jaxpr, (), *args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/core.py"", line 630, in eval_jaxpr\r\n ans = eqn.primitive.bind(*subfuns, *map(read, eqn.invars), **bind_params)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/core.py"", line 536, in bind\r\n return self._true_bind(*args, **params)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/core.py"", line 552, in _true_bind\r\n return self.bind_with_trace(prev_trace, args, params)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/core.py"", line 562, in bind_with_trace\r\n return trace.process_primitive(self, args, params)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/core.py"", line 1066, in process_primitive\r\n return primitive.impl(*args, **params)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/dispatch.py"", line 91, in apply_primitive\r\n outs = fun(*args)\r\nKeyboardInterrupt\r\n",,terminal_output +110,200051,"TERMINAL",0,0,"^CException ignored in: .remove at 0x14f0fc91cc10>\r\nTraceback (most recent call last):\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/weakref.py"", line 370, in remove\r\n def remove(k, selfref=ref(self)):\r\nKeyboardInterrupt: \r\n",,terminal_output +111,200257,"TERMINAL",0,0,"^CException ignored in atexit callback: \r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/api.py"", line 3189, in clean_up\r\n clear_caches()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/api.py"", line 3210, in clear_caches\r\n util.clear_all_weakref_lru_caches()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/util.py"", line 337, in clear_all_weakref_lru_caches\r\n cached_call.cache_clear()\r\nKeyboardInterrupt: \r\n",,terminal_output +112,200578,"TERMINAL",0,0,"^C\r\n]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +113,200738,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +114,200975,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +115,303404,"TERMINAL",0,0,"\r(reverse-i-search)`': ",,terminal_output +116,303666,"TERMINAL",0,0,"m': python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +117,303807,"TERMINAL",0,0,"\ro': git checkout -b ""causal-transformer-dynamics-model""\r\n\r",,terminal_output +118,304050,"TERMINAL",0,0,"\rd': git checkout -b ""causal-transformer-dynamics-model""\ru': module unload devel/cuda/12.4\r",,terminal_output +119,304365,"TERMINAL",0,0,"[?25lom\r[8@failed reverse-i-search)`modue': modu[1@l[?25h",,terminal_output +120,305938,"TERMINAL",0,0,"[?25lom[1@r[1@q[?25h",,terminal_output +121,306989,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +122,307170,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +123,307520,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +124,308493,"TERMINAL",0,0,"^C[?2004l\r[?2004h[?2004l\r\r\n]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +125,309383,"TERMINAL",0,0,"\r(reverse-i-search)`': ",,terminal_output +126,309537,"TERMINAL",0,0,"m': python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +127,309707,"TERMINAL",0,0,"\ro': git checkout -b ""causal-transformer-dynamics-model""\r\n\r",,terminal_output +128,309813,"TERMINAL",0,0,"\rd': git checkout -b ""causal-transformer-dynamics-model""",,terminal_output +129,310090,"TERMINAL",0,0,"[?25lm\ru': module unload devel/cuda/12.4\r[?25h",,terminal_output +130,310231,"TERMINAL",0,0,"[?25lom[1@l': modul[?25h",,terminal_output +131,310305,"TERMINAL",0,0,"[1@e': module",,terminal_output +132,311464,"TERMINAL",0,0,"mpi/openmpi/5.0",,terminal_output +133,312031,"TERMINAL",0,0,"[?25l\r[9@jafar) [tum_cte0515@hkn0804 jafar]$ module\r\n[?2004l\r[?25h",,terminal_output +134,312192,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +135,313152,"TERMINAL",0,0,"\r(reverse-i-search)`': ",,terminal_output +136,313398,"TERMINAL",0,0,"m': module unload mpi/openmpi/5.0",,terminal_output +137,313663,"TERMINAL",0,0,"[?25lm\ro': module unload mpi/openmpi/5.0\r[?25h",,terminal_output +138,313810,"TERMINAL",0,0,"[?25lm[1@d': mod[?25h",,terminal_output +139,313910,"TERMINAL",0,0,"[?25lm[1@u': modu[?25h",,terminal_output +140,314047,"TERMINAL",0,0,"[?25lm[1@l': modul[?25h",,terminal_output +141,314184,"TERMINAL",0,0,"[?25lm[1@e': module[?25h",,terminal_output +142,315781,"TERMINAL",0,0,"devel/cuda/12.4",,terminal_output +143,316491,"TERMINAL",0,0,"\r[9@jafar) [tum_cte0515@hkn0804 jafar]$ module\r\n[?2004l\r]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +144,316974,"TERMINAL",0,0,"module unload devel/cuda/12.4",,terminal_output +145,317081,"TERMINAL",0,0,"mpi/openmpi/5.0",,terminal_output +146,317627,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +147,318194,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +148,319142,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +149,321453,"TERMINAL",0,0,"2025-07-17 09:49:24.246221: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +150,335339,"TERMINAL",0,0,"2025-07-17 09:49:38.114612: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +151,350537,"TERMINAL",0,0,"2025-07-17 09:49:53.330365: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +152,354298,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 33000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/033000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 32000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/032000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 34000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/034000/metrics/metrics not found.\r\n",,terminal_output +153,395872,"TERMINAL",0,0,"jax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 178, in \r\n vid = _autoreg_sample(rng, video_batch, action_batch)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 141, in _autoreg_sample\r\n generated_vid = sampling_fn(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 133, in _sampling_wrapper\r\n return module.sample_causal(batch, args.seq_len, args.temperature, args.sample_argmax)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 171, in sample_causal\r\n dyna_outputs = self.dynamics(dyna_inputs, training=False)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 1083, in op\r\n return getattr(self.aval, f""_{name}"")(self, *args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 583, in deferring_binary_op\r\n return binary_op(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufunc_api.py"", line 182, in __call__\r\n return call(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufuncs.py"", line 1237, in add\r\n out = lax.add(x, y)\r\nTypeError: add got incompatible shapes for broadcasting: (1, 2, 920, 512), (1, 3, 1, 512).\r\n",,terminal_output +154,397530,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +155,546285,"genie.py",0,0,"",python,tab +156,551671,"genie.py",5425,0,"",python,selection_mouse +157,552728,"genie.py",5358,108,"",python,content +158,554660,"genie.py",5358,1,"",python,content +159,554696,"genie.py",5366,0,"",python,selection_command +160,554899,"genie.py",5291,0,"",python,selection_command +161,555103,"genie.py",5253,0,"",python,selection_command +162,555655,"genie.py",5245,38,"",python,content +163,555698,"genie.py",5257,0,"",python,selection_command +164,556602,"genie.py",5253,4,"",python,content +165,557003,"genie.py",5252,0,"",python,selection_command +166,557159,"genie.py",5184,0,"",python,selection_command +167,557536,"genie.py",5177,68,"",python,content +168,557582,"genie.py",5185,0,"",python,selection_command +169,558166,"genie.py",5256,0,"",python,selection_command +170,558441,"genie.py",5185,0,"",python,selection_command +171,558948,"genie.py",5144,0,"",python,selection_command +172,558993,"genie.py",5135,0,"",python,selection_command +173,559130,"genie.py",5114,0,"",python,selection_command +174,559259,"genie.py",5035,0,"",python,selection_command +175,559437,"genie.py",4992,0,"",python,selection_command +176,559938,"genie.py",4952,0,"",python,selection_command +177,560004,"genie.py",4909,0,"",python,selection_command +178,560026,"genie.py",4860,0,"",python,selection_command +179,560063,"genie.py",4851,0,"",python,selection_command +180,560090,"genie.py",4824,0,"",python,selection_command +181,560152,"genie.py",4765,0,"",python,selection_command +182,560181,"genie.py",4683,0,"",python,selection_command +183,560182,"genie.py",4643,0,"",python,selection_command +184,560199,"genie.py",4631,0,"",python,selection_command +185,560240,"genie.py",4574,0,"",python,selection_command +186,560302,"genie.py",4557,0,"",python,selection_command +187,560303,"genie.py",4548,0,"",python,selection_command +188,561368,"genie.py",4557,0,"",python,selection_command +189,561848,"genie.py",4574,0,"",python,selection_command +190,561891,"genie.py",4631,0,"",python,selection_command +191,561955,"genie.py",4643,0,"",python,selection_command +192,561992,"genie.py",4683,0,"",python,selection_command +193,561993,"genie.py",4765,0,"",python,selection_command +194,562028,"genie.py",4824,0,"",python,selection_command +195,562040,"genie.py",4851,0,"",python,selection_command +196,562069,"genie.py",4860,0,"",python,selection_command +197,562086,"genie.py",4909,0,"",python,selection_command +198,562148,"genie.py",4952,0,"",python,selection_command +199,562169,"genie.py",4992,0,"",python,selection_command +200,562232,"genie.py",5035,0,"",python,selection_command +201,562304,"genie.py",5114,0,"",python,selection_command +202,562304,"genie.py",5135,0,"",python,selection_command +203,562356,"genie.py",5144,0,"",python,selection_command +204,562357,"genie.py",5185,0,"",python,selection_command +205,562399,"genie.py",5256,0,"",python,selection_command +206,562399,"genie.py",5305,0,"",python,selection_command +207,565062,"genie.py",5342,0,"",python,selection_command +208,565693,"genie.py",5378,0,"",python,selection_command +209,565694,"genie.py",5460,0,"",python,selection_command +210,565695,"genie.py",5488,0,"",python,selection_command +211,565734,"genie.py",5560,0,"",python,selection_command +212,565897,"genie.py",5636,0,"",python,selection_command +213,566411,"genie.py",5693,0,"",python,selection_command +214,566706,"genie.py",5707,0,"",python,selection_command +215,567968,"genie.py",5693,0,"",python,selection_command +216,579341,"genie.py",5169,0,"",python,selection_mouse +217,581540,"genie.py",5723,0,"",python,selection_mouse +218,619129,"genie.py",5698,0,"",python,selection_mouse +219,619149,"genie.py",5697,0,"",python,selection_command +220,619976,"genie.py",5736,0,"",python,selection_mouse +221,710460,"genie.py",3806,0,"",python,selection_command +222,711139,"genie.py",3811,0,"",python,selection_mouse +223,711340,"genie.py",3806,13,"sample_causal",python,selection_mouse +224,712216,"genie.py",4115,0,"",python,selection_mouse +225,943263,"genie.py",5135,0,"",python,selection_mouse +226,943887,"genie.py",5167,0,"",python,selection_mouse +227,944542,"genie.py",5369,0,"",python,selection_mouse +228,944547,"genie.py",5368,0,"",python,selection_command +229,945169,"genie.py",5221,0,"",python,selection_mouse +230,948216,"genie.py",5735,0,"",python,selection_mouse +231,948768,"genie.py",5698,0,"",python,selection_mouse +232,948784,"genie.py",5697,0,"",python,selection_command +233,949615,"genie.py",5698,0,"\n ",python,content +234,950006,"genie.py",5711,0,"j",python,content +235,950007,"genie.py",5712,0,"",python,selection_keyboard +236,950122,"genie.py",5712,0,"a",python,content +237,950124,"genie.py",5713,0,"",python,selection_keyboard +238,950251,"genie.py",5713,0,"x",python,content +239,950253,"genie.py",5714,0,"",python,selection_keyboard +240,950453,"genie.py",5714,0,".",python,content +241,950455,"genie.py",5715,0,"",python,selection_keyboard +242,950903,"genie.py",5715,0,"d",python,content +243,950905,"genie.py",5716,0,"",python,selection_keyboard +244,951137,"genie.py",5716,0,"e",python,content +245,951139,"genie.py",5717,0,"",python,selection_keyboard +246,951208,"genie.py",5717,0,"b",python,content +247,951208,"genie.py",5718,0,"",python,selection_keyboard +248,952089,"genie.py",5715,3,"debug",python,content +249,952410,"genie.py",5720,0,".",python,content +250,952412,"genie.py",5721,0,"",python,selection_keyboard +251,953219,"genie.py",5721,0,"b",python,content +252,953221,"genie.py",5722,0,"",python,selection_keyboard +253,953326,"genie.py",5722,0,"r",python,content +254,953327,"genie.py",5723,0,"",python,selection_keyboard +255,953547,"genie.py",5723,0,"e",python,content +256,953548,"genie.py",5724,0,"",python,selection_keyboard +257,953692,"genie.py",5724,0,"a",python,content +258,953694,"genie.py",5725,0,"",python,selection_keyboard +259,953831,"genie.py",5725,0,"k",python,content +260,953834,"genie.py",5726,0,"",python,selection_keyboard +261,954874,"genie.py",5721,5,"breakpoint",python,content +262,955668,"genie.py",5731,0,"()",python,content +263,955669,"genie.py",5732,0,"",python,selection_keyboard +264,955744,"genie.py",5732,1,")",python,content +265,955745,"genie.py",5733,0,"",python,selection_keyboard +266,955944,"genie.py",5732,0,"",python,selection_command +267,957082,"TERMINAL",0,0,"srun",,terminal_focus +268,957776,"TERMINAL",0,0,"srun",,terminal_focus +269,958874,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +270,959121,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +271,963465,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +272,965628,"TERMINAL",0,0,"2025-07-17 10:00:08.394391: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +273,966049,"genie.py",0,0,"",python,tab +274,979212,"TERMINAL",0,0,"2025-07-17 10:00:22.001500: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +275,989389,"genie.py",0,0,"",python,tab +276,994271,"TERMINAL",0,0,"2025-07-17 10:00:37.060069: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +277,997899,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 32000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/032000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 33000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/033000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 34000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/034000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\n",,terminal_output +278,1039418,"TERMINAL",0,0,"jax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 178, in \r\n vid = _autoreg_sample(rng, video_batch, action_batch)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 141, in _autoreg_sample\r\n generated_vid = sampling_fn(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 133, in _sampling_wrapper\r\n return module.sample_causal(batch, args.seq_len, args.temperature, args.sample_argmax)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 167, in sample_causal\r\n dyna_outputs = self.dynamics(dyna_inputs, training=False)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 1083, in op\r\n return getattr(self.aval, f""_{name}"")(self, *args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 583, in deferring_binary_op\r\n return binary_op(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufunc_api.py"", line 182, in __call__\r\n return call(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufuncs.py"", line 1237, in add\r\n out = lax.add(x, y)\r\nTypeError: add got incompatible shapes for broadcasting: (1, 2, 920, 512), (1, 3, 1, 512).\r\n",,terminal_output +279,1040786,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +280,1062397,"TERMINAL",0,0,"srun",,terminal_focus +281,1064299,"genie.py",0,0,"",python,tab +282,1066755,"genie.py",5767,0,"",python,selection_command +283,1067420,"genie.py",5768,0,"",python,selection_command +284,1068037,"genie.py",5769,0,"",python,selection_command +285,1068539,"genie.py",5732,0,"",python,selection_command +286,1069020,"genie.py",5731,0,"",python,selection_command +287,1069517,"genie.py",5730,0,"",python,selection_command +288,1069572,"genie.py",5729,0,"",python,selection_command +289,1069606,"genie.py",5728,0,"",python,selection_command +290,1069638,"genie.py",5727,0,"",python,selection_command +291,1069661,"genie.py",5726,0,"",python,selection_command +292,1069686,"genie.py",5725,0,"",python,selection_command +293,1069711,"genie.py",5724,0,"",python,selection_command +294,1069735,"genie.py",5723,0,"",python,selection_command +295,1069760,"genie.py",5722,0,"",python,selection_command +296,1069794,"genie.py",5721,0,"",python,selection_command +297,1069910,"genie.py",5720,0,"",python,selection_command +298,1070238,"genie.py",5721,0,"",python,selection_command +299,1070686,"genie.py",5721,10,"",python,content +300,1071250,"genie.py",5721,0,"p",python,content +301,1071251,"genie.py",5722,0,"",python,selection_keyboard +302,1071394,"genie.py",5722,0,"r",python,content +303,1071395,"genie.py",5723,0,"",python,selection_keyboard +304,1071483,"genie.py",5723,0,"i",python,content +305,1071484,"genie.py",5724,0,"",python,selection_keyboard +306,1071560,"genie.py",5724,0,"n",python,content +307,1071561,"genie.py",5725,0,"",python,selection_keyboard +308,1071612,"genie.py",5725,0,"t",python,content +309,1071614,"genie.py",5726,0,"",python,selection_keyboard +310,1071992,"genie.py",5727,0,"",python,selection_command +311,1073812,"genie.py",5717,0,"",python,selection_mouse +312,1073985,"genie.py",5715,5,"debug",python,selection_mouse +313,1074132,"genie.py",5699,30," jax.debug.print()\n",python,selection_mouse +314,1075013,"genie.py",5711,0,"",python,selection_command +315,1098182,"genie.py",5699,0,"",python,selection_command +316,1100749,"genie.py",5699,0," jax.debug.print(""token_idxs_full shape: {}"", token_idxs_full.shape)\n",python,content +317,1101766,"genie.py",5779,0," jax.debug.print(""action_tokens shape: {}"", action_tokens.shape)\n",python,content +318,1102918,"genie.py",5855,0," jax.debug.print(""mask_rng: {}"", batch.get(""mask_rng"", None))\n",python,content +319,1102920,"genie.py",5928,30,"",python,content +320,1107666,"genie.py",5979,0,"",python,selection_mouse +321,1108225,"genie.py",5906,0,"",python,selection_mouse +322,1109241,"genie.py",5905,0,"",python,selection_command +323,1110430,"genie.py",4875,0,"",python,selection_mouse +324,1112007,"genie.py",5732,0,"",python,selection_mouse +325,1115189,"genie.py",4851,0,"",python,selection_mouse +326,1117314,"genie.py",4851,0,"\n jax.debug.print(""token_idxs_full shape: {}"", token_idxs_full.shape)",python,content +327,1117384,"genie.py",4864,0,"",python,selection_command +328,1118345,"genie.py",4860,4,"",python,content +329,1118669,"genie.py",4859,0,"",python,selection_command +330,1118845,"genie.py",4860,0,"",python,selection_command +331,1119288,"genie.py",4861,0,"",python,selection_command +332,1119335,"genie.py",4862,0,"",python,selection_command +333,1119361,"genie.py",4863,0,"",python,selection_command +334,1119407,"genie.py",4864,0,"",python,selection_command +335,1119473,"genie.py",4865,0,"",python,selection_command +336,1119474,"genie.py",4866,0,"",python,selection_command +337,1119494,"genie.py",4867,0,"",python,selection_command +338,1119523,"genie.py",4868,0,"",python,selection_command +339,1119668,"genie.py",4869,0,"",python,selection_command +340,1119669,"genie.py",4870,0,"",python,selection_command +341,1119771,"genie.py",4871,0,"",python,selection_command +342,1119772,"genie.py",4872,0,"",python,selection_command +343,1119773,"genie.py",4873,0,"",python,selection_command +344,1119811,"genie.py",4874,0,"",python,selection_command +345,1119812,"genie.py",4875,0,"",python,selection_command +346,1119812,"genie.py",4876,0,"",python,selection_command +347,1119847,"genie.py",4877,0,"",python,selection_command +348,1119848,"genie.py",4878,0,"",python,selection_command +349,1119872,"genie.py",4879,0,"",python,selection_command +350,1119904,"genie.py",4880,0,"",python,selection_command +351,1119975,"genie.py",4881,0,"",python,selection_command +352,1119995,"genie.py",4882,0,"",python,selection_command +353,1119996,"genie.py",4883,0,"",python,selection_command +354,1120024,"genie.py",4884,0,"",python,selection_command +355,1120041,"genie.py",4885,0,"",python,selection_command +356,1120077,"genie.py",4886,0,"",python,selection_command +357,1120113,"genie.py",4887,0,"",python,selection_command +358,1120141,"genie.py",4888,0,"",python,selection_command +359,1120175,"genie.py",4889,0,"",python,selection_command +360,1120187,"genie.py",4890,0,"",python,selection_command +361,1120220,"genie.py",4891,0,"",python,selection_command +362,1120266,"genie.py",4892,0,"",python,selection_command +363,1120313,"genie.py",4893,0,"",python,selection_command +364,1120323,"genie.py",4894,0,"",python,selection_command +365,1120362,"genie.py",4895,0,"",python,selection_command +366,1120369,"genie.py",4896,0,"",python,selection_command +367,1120407,"genie.py",4897,0,"",python,selection_command +368,1120414,"genie.py",4898,0,"",python,selection_command +369,1120452,"genie.py",4899,0,"",python,selection_command +370,1120504,"genie.py",4900,0,"",python,selection_command +371,1120539,"genie.py",4901,0,"",python,selection_command +372,1120555,"genie.py",4902,0,"",python,selection_command +373,1120780,"genie.py",4901,0,"",python,selection_command +374,1121268,"genie.py",4900,0,"",python,selection_command +375,1121323,"genie.py",4899,0,"",python,selection_command +376,1121359,"genie.py",4898,0,"",python,selection_command +377,1121373,"genie.py",4897,0,"",python,selection_command +378,1121394,"genie.py",4896,0,"",python,selection_command +379,1121444,"genie.py",4895,0,"",python,selection_command +380,1121451,"genie.py",4894,0,"",python,selection_command +381,1121482,"genie.py",4893,0,"",python,selection_command +382,1121525,"genie.py",4892,0,"",python,selection_command +383,1121534,"genie.py",4891,0,"",python,selection_command +384,1121589,"genie.py",4890,0,"",python,selection_command +385,1121607,"genie.py",4889,0,"",python,selection_command +386,1121712,"genie.py",4888,0,"",python,selection_command +387,1122099,"genie.py",4888,5,"",python,content +388,1122371,"genie.py",4887,0,"",python,selection_command +389,1122712,"genie.py",4887,1,"",python,content +390,1122951,"genie.py",4888,0,"",python,selection_command +391,1123407,"genie.py",4887,0,"",python,selection_command +392,1123768,"genie.py",4887,0," ",python,content +393,1123769,"genie.py",4888,0,"",python,selection_keyboard +394,1124308,"genie.py",4887,0,"",python,selection_command +395,1124413,"genie.py",4888,0,"",python,selection_command +396,1124900,"genie.py",4889,0,"",python,selection_command +397,1124926,"genie.py",4890,0,"",python,selection_command +398,1124967,"genie.py",4891,0,"",python,selection_command +399,1124974,"genie.py",4892,0,"",python,selection_command +400,1125021,"genie.py",4893,0,"",python,selection_command +401,1125078,"genie.py",4894,0,"",python,selection_command +402,1125149,"genie.py",4895,0,"",python,selection_command +403,1125275,"genie.py",4896,0,"",python,selection_command +404,1125276,"genie.py",4897,0,"",python,selection_command +405,1125325,"genie.py",4898,0,"",python,selection_command +406,1125326,"genie.py",4899,0,"",python,selection_command +407,1125326,"genie.py",4900,0,"",python,selection_command +408,1125327,"genie.py",4901,0,"",python,selection_command +409,1125327,"genie.py",4902,0,"",python,selection_command +410,1125379,"genie.py",4903,0,"",python,selection_command +411,1125380,"genie.py",4904,0,"",python,selection_command +412,1125408,"genie.py",4905,0,"",python,selection_command +413,1125441,"genie.py",4906,0,"",python,selection_command +414,1125466,"genie.py",4907,0,"",python,selection_command +415,1125497,"genie.py",4908,0,"",python,selection_command +416,1125512,"genie.py",4909,0,"",python,selection_command +417,1125559,"genie.py",4910,0,"",python,selection_command +418,1125588,"genie.py",4911,0,"",python,selection_command +419,1125606,"genie.py",4912,0,"",python,selection_command +420,1125857,"genie.py",4911,0,"",python,selection_command +421,1126012,"genie.py",4910,0,"",python,selection_command +422,1126361,"genie.py",4910,5,"",python,content +423,1184148,"genie.py",5009,0,"",python,selection_mouse +424,1184157,"genie.py",5008,0,"",python,selection_command +425,1808149,"genie.py",5009,0,"",python,selection_mouse +426,1808165,"genie.py",5008,0,"",python,selection_command +427,1889656,"genie.py",3798,0,"",python,selection_command +428,1890209,"genie.py",4040,0,"",python,selection_mouse +429,1903192,"genie.py",5092,0,"",python,selection_mouse +430,1903221,"genie.py",5091,0,"",python,selection_command +431,1903919,"genie.py",5092,0,"",python,selection_mouse +432,1903922,"genie.py",5091,0,"",python,selection_command +433,1907754,"TERMINAL",0,0,"srun",,terminal_focus +434,1913080,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +435,1913359,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +436,1914846,"genie.py",0,0,"",python,tab +437,1915948,"genie.py",5049,0,"",python,selection_mouse +438,1915953,"genie.py",5048,0,"",python,selection_command +439,1916502,"genie.py",5049,0,"",python,selection_mouse +440,1916520,"genie.py",5048,0,"",python,selection_command +441,1916753,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +442,1918822,"TERMINAL",0,0,"2025-07-17 10:16:01.571686: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +443,1925017,"genie.py",5005,0,"",python,selection_command +444,1925369,"genie.py",4956,0,"",python,selection_command +445,1925744,"genie.py",4890,0,"",python,selection_command +446,1925972,"genie.py",4851,0,"",python,selection_command +447,1926151,"genie.py",4890,0,"",python,selection_command +448,1926367,"genie.py",4956,0,"",python,selection_command +449,1926651,"genie.py",4890,0,"",python,selection_command +450,1926836,"genie.py",4956,0,"",python,selection_command +451,1927133,"genie.py",4890,0,"",python,selection_command +452,1927535,"genie.py",4851,0,"",python,selection_command +453,1927866,"genie.py",4890,0,"",python,selection_command +454,1932394,"TERMINAL",0,0,"2025-07-17 10:16:15.167351: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +455,1947300,"TERMINAL",0,0,"2025-07-17 10:16:30.026963: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +456,1951096,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 33000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/033000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 34000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/034000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\n",,terminal_output +457,1951151,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 35000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/035000/metrics/metrics not found.\r\n",,terminal_output +458,1992151,"TERMINAL",0,0,"jax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 178, in \r\n vid = _autoreg_sample(rng, video_batch, action_batch)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 141, in _autoreg_sample\r\n generated_vid = sampling_fn(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 133, in _sampling_wrapper\r\n return module.sample_causal(batch, args.seq_len, args.temperature, args.sample_argmax)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py"", line 170, in sample_causal\r\n dyna_outputs = self.dynamics(dyna_inputs, training=False)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/models/dynamics.py"", line 97, in __call__\r\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 1083, in op\r\n return getattr(self.aval, f""_{name}"")(self, *args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/array_methods.py"", line 583, in deferring_binary_op\r\n return binary_op(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufunc_api.py"", line 182, in __call__\r\n return call(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/numpy/ufuncs.py"", line 1237, in add\r\n out = lax.add(x, y)\r\nTypeError: add got incompatible shapes for broadcasting: (1, 2, 920, 512), (1, 3, 1, 512).\r\n",,terminal_output +459,1993499,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +460,2021457,"TERMINAL",0,0,"srun",,terminal_focus +461,2023260,"genie.py",0,0,"",python,tab +462,2028656,"genie.py",4789,0,"",python,selection_mouse +463,2029219,"genie.py",4838,0,"",python,selection_mouse +464,2030763,"genie.py",4827,0,"",python,selection_mouse +465,2031413,"genie.py",4830,0,"",python,selection_mouse +466,2034177,"genie.py",5038,0,"",python,selection_mouse +467,2034726,"genie.py",5047,0,"",python,selection_mouse +468,2036283,"genie.py",5194,0,"",python,selection_mouse +469,2036678,"genie.py",5201,0,"",python,selection_mouse +470,2045012,"genie.py",8712,0,"",python,selection_mouse +471,2050701,"genie.py",5259,0,"",python,selection_mouse +472,2050843,"genie.py",5251,13,"action_tokens",python,selection_mouse +473,2051856,"genie.py",5270,0,"",python,selection_mouse +474,2052041,"genie.py",5267,4,"self",python,selection_mouse +475,2052878,"genie.py",5285,0,"",python,selection_mouse +476,2053030,"genie.py",5279,9,"get_codes",python,selection_mouse +477,2054266,"genie.py",5259,0,"",python,selection_mouse +478,2054387,"genie.py",5251,13,"action_tokens",python,selection_mouse +479,2080227,"genie.py",5785,0,"",python,selection_mouse +480,2080704,"genie.py",5764,0,"",python,selection_mouse +481,2080734,"genie.py",5763,0,"",python,selection_command +482,2090469,"genie.py",5536,0,"",python,selection_mouse +483,2090614,"genie.py",5530,11,"dyna_inputs",python,selection_mouse +484,2091737,"genie.py",6032,0,"",python,selection_mouse +485,2091885,"genie.py",6026,8,"dynamics",python,selection_mouse +486,2092673,"genie.py",6023,0,"",python,selection_mouse +487,2094788,"genie.py",6031,0,"",python,selection_mouse +488,2096767,"genie.py",0,0,"",python,tab +489,2096768,"genie.py",2479,0,"",python,selection_mouse +490,2097414,"models/dynamics.py",0,0,"from typing import Dict, Any\n\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nimport einops\n\nfrom utils.nn import STTransformer\n\n\nclass DynamicsMaskGIT(nn.Module):\n """"""MaskGIT dynamics model""""""\n\n model_dim: int\n num_latents: int\n num_blocks: int\n num_heads: int\n dropout: float\n mask_limit: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n def setup(self):\n self.dynamics = STTransformer(\n self.model_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n spacial_bert=True,\n )\n self.patch_embed = nn.Embed(self.num_latents, self.model_dim)\n self.mask_token = self.param(\n ""mask_token"",\n nn.initializers.lecun_uniform(),\n (1, 1, 1, self.model_dim),\n )\n self.action_up = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n # --- Mask videos ---\n vid_embed = self.patch_embed(batch[""video_tokens""])\n if training:\n rng1, rng2 = jax.random.split(batch[""mask_rng""])\n mask_prob = jax.random.uniform(rng1, minval=self.mask_limit)\n mask = jax.random.bernoulli(rng2, mask_prob, vid_embed.shape[:-1])\n mask = mask.at[:, 0].set(False)\n vid_embed = jnp.where(jnp.expand_dims(mask, -1), self.mask_token, vid_embed)\n else:\n mask = None\n\n # --- Predict transition ---\n act_embed = self.action_up(batch[""latent_actions""])\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n logits = self.dynamics(vid_embed)\n return dict(token_logits=logits, mask=mask)\n\n\nclass DynamicsAutoregressive(nn.Module):\n """"""Autoregressive (causal) dynamics model""""""\n\n model_dim: int\n num_latents: int\n num_blocks: int\n num_heads: int\n dropout: float\n param_dtype: jnp.dtype\n dtype: jnp.dtype\n\n def setup(self):\n self.dynamics = STTransformer(\n self.model_dim,\n self.num_latents,\n self.num_blocks,\n self.num_heads,\n self.dropout,\n self.param_dtype,\n self.dtype,\n spacial_bert=False,\n )\n self.patch_embed = nn.Embed(self.num_latents, self.model_dim)\n self.action_up = nn.Dense(\n self.model_dim,\n param_dtype=self.param_dtype,\n dtype=self.dtype,\n )\n\n def __call__(self, batch: Dict[str, Any], training: bool = True) -> Dict[str, Any]:\n vid_embed = self.patch_embed(batch[""video_tokens""])\n act_embed = self.action_up(batch[""latent_actions""])\n vid_embed += jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))\n logits = self.dynamics(vid_embed)\n mask = jnp.ones(vid_embed.shape[:-1])\n return dict(token_logits=logits, mask=mask)",python,tab +491,2100856,"models/dynamics.py",2750,0,"",python,selection_mouse +492,2101474,"models/dynamics.py",2764,0,"",python,selection_mouse +493,2101631,"models/dynamics.py",2763,11,"patch_embed",python,selection_mouse +494,2102513,"models/dynamics.py",2789,0,"",python,selection_mouse +495,2102693,"models/dynamics.py",2782,12,"video_tokens",python,selection_mouse +496,2102979,"models/dynamics.py",2781,13,"""video_tokens",python,selection_mouse +497,2102979,"models/dynamics.py",2780,14,"[""video_tokens",python,selection_mouse +498,2102980,"models/dynamics.py",2775,19,"batch[""video_tokens",python,selection_mouse +499,2103475,"models/dynamics.py",2775,0,"",python,selection_mouse +500,2103476,"models/dynamics.py",2775,5,"batch",python,selection_mouse +501,2103735,"models/dynamics.py",2775,6,"batch[",python,selection_mouse +502,2103736,"models/dynamics.py",2775,7,"batch[""",python,selection_mouse +503,2103764,"models/dynamics.py",2775,19,"batch[""video_tokens",python,selection_mouse +504,2104387,"models/dynamics.py",2775,20,"batch[""video_tokens""",python,selection_mouse +505,2108211,"models/dynamics.py",2916,0,"",python,selection_mouse +506,2108853,"models/dynamics.py",2855,0,"",python,selection_mouse +507,2108995,"models/dynamics.py",2854,3,"""])",python,selection_mouse +508,2109164,"models/dynamics.py",2854,3,"""])",python,selection_mouse +509,2109209,"models/dynamics.py",2840,17,"latent_actions""])",python,selection_mouse +510,2109355,"models/dynamics.py",2839,18,"""latent_actions""])",python,selection_mouse +511,2109395,"models/dynamics.py",2838,19,"[""latent_actions""])",python,selection_mouse +512,2109443,"models/dynamics.py",2833,24,"batch[""latent_actions""])",python,selection_mouse +513,2110021,"models/dynamics.py",2837,0,"",python,selection_mouse +514,2110021,"models/dynamics.py",2833,5,"batch",python,selection_mouse +515,2110209,"models/dynamics.py",2833,6,"batch[",python,selection_mouse +516,2110237,"models/dynamics.py",2833,21,"batch[""latent_actions",python,selection_mouse +517,2110392,"models/dynamics.py",2833,22,"batch[""latent_actions""",python,selection_mouse +518,2110419,"models/dynamics.py",2833,23,"batch[""latent_actions""]",python,selection_mouse +519,2110420,"models/dynamics.py",2833,24,"batch[""latent_actions""])",python,selection_mouse +520,2110739,"models/dynamics.py",2833,23,"batch[""latent_actions""]",python,selection_mouse +521,2110816,"models/dynamics.py",2833,22,"batch[""latent_actions""",python,selection_mouse +522,2115396,"genie.py",0,0,"",python,tab +523,2116296,"genie.py",6101,0,"",python,selection_mouse +524,2116821,"genie.py",6033,0,"",python,selection_mouse +525,2117440,"genie.py",5964,0,"",python,selection_mouse +526,2119491,"genie.py",6030,0,"",python,selection_mouse +527,2134317,"genie.py",5994,69," dyna_outputs = self.dynamics(dyna_inputs, training=False)",python,selection_command +528,2134540,"genie.py",5994,135," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)",python,selection_command +529,2135076,"genie.py",5994,216," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)",python,selection_command +530,2135129,"genie.py",5994,312," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)",python,selection_command +531,2135130,"genie.py",5994,313," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n",python,selection_command +532,2135170,"genie.py",5994,359," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch",python,selection_command +533,2135305,"genie.py",5994,389," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:",python,selection_command +534,2135468,"genie.py",5994,467," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)",python,selection_command +535,2135469,"genie.py",5994,485," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:",python,selection_command +536,2135469,"genie.py",5994,521," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:",python,selection_command +537,2135469,"genie.py",5994,579," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)",python,selection_command +538,2135470,"genie.py",5994,601," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:",python,selection_command +539,2135630,"genie.py",5994,654," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)",python,selection_command +540,2135631,"genie.py",5994,707," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(",python,selection_command +541,2135631,"genie.py",5994,778," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(\n step_rng, next_token_logits / temperature, axis=-1",python,selection_command +542,2135631,"genie.py",5994,806," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(\n step_rng, next_token_logits / temperature, axis=-1\n ) # (B, N)",python,selection_command +543,2135632,"genie.py",5994,807," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(\n step_rng, next_token_logits / temperature, axis=-1\n ) # (B, N)\n",python,selection_command +544,2135632,"genie.py",5994,867," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(\n step_rng, next_token_logits / temperature, axis=-1\n ) # (B, N)\n\n # Insert the generated tokens into the sequence",python,selection_command +545,2135690,"genie.py",5994,941," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(\n step_rng, next_token_logits / temperature, axis=-1\n ) # (B, N)\n\n # Insert the generated tokens into the sequence\n token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)",python,selection_command +546,2135690,"genie.py",5994,942," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(\n step_rng, next_token_logits / temperature, axis=-1\n ) # (B, N)\n\n # Insert the generated tokens into the sequence\n token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n",python,selection_command +547,2135691,"genie.py",5994,997," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(\n step_rng, next_token_logits / temperature, axis=-1\n ) # (B, N)\n\n # Insert the generated tokens into the sequence\n token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n\n # --- Decode all tokens at once at the end ---",python,selection_command +548,2135691,"genie.py",5994,1043," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(\n step_rng, next_token_logits / temperature, axis=-1\n ) # (B, N)\n\n # Insert the generated tokens into the sequence\n token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n\n # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(",python,selection_command +549,2135787,"genie.py",5994,1108," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(\n step_rng, next_token_logits / temperature, axis=-1\n ) # (B, N)\n\n # Insert the generated tokens into the sequence\n token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n\n # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]",python,selection_command +550,2135940,"genie.py",5994,1118," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(\n step_rng, next_token_logits / temperature, axis=-1\n ) # (B, N)\n\n # Insert the generated tokens into the sequence\n token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n\n # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )",python,selection_command +551,2136385,"genie.py",5994,1146," dyna_outputs = self.dynamics(dyna_inputs, training=False)\n # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\n # We want the logits for the last time step (frame t-1 predicting t)\n next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\n\n # Sample or argmax for each patch\n if sample_argmax:\n next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n else:\n if rng is not None:\n rng, step_rng = jax.random.split(rng)\n else:\n step_rng = jax.random.PRNGKey(0)\n next_token = jax.random.categorical(\n step_rng, next_token_logits / temperature, axis=-1\n ) # (B, N)\n\n # Insert the generated tokens into the sequence\n token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n\n # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )\n return final_frames",python,selection_command +552,2136571,"genie.py",6006,0,"",python,selection_command +553,2137422,"genie.py",7121,0,"#",python,content +554,2137422,"genie.py",7111,0,"#",python,content +555,2137422,"genie.py",7050,0,"#",python,content +556,2137422,"genie.py",7000,0,"#",python,content +557,2137422,"genie.py",6945,0,"#",python,content +558,2137422,"genie.py",6874,0,"#",python,content +559,2137423,"genie.py",6814,0,"#",python,content +560,2137423,"genie.py",6789,0,"#",python,content +561,2137423,"genie.py",6722,0,"#",python,content +562,2137423,"genie.py",6665,0,"#",python,content +563,2137423,"genie.py",6616,0,"#",python,content +564,2137423,"genie.py",6590,0,"#",python,content +565,2137423,"genie.py",6536,0,"#",python,content +566,2137423,"genie.py",6496,0,"#",python,content +567,2137423,"genie.py",6474,0,"#",python,content +568,2137423,"genie.py",6400,0,"#",python,content +569,2137423,"genie.py",6366,0,"#",python,content +570,2137423,"genie.py",6320,0,"#",python,content +571,2137423,"genie.py",6223,0,"#",python,content +572,2137424,"genie.py",6142,0,"#",python,content +573,2137424,"genie.py",6076,0,"#",python,content +574,2137424,"genie.py",6006,0,"#",python,content +575,2137425,"genie.py",6007,0,"",python,selection_keyboard +576,2137479,"genie.py",7143,0," ",python,content +577,2137479,"genie.py",7132,0," ",python,content +578,2137479,"genie.py",7070,0," ",python,content +579,2137479,"genie.py",7019,0," ",python,content +580,2137479,"genie.py",6963,0," ",python,content +581,2137479,"genie.py",6891,0," ",python,content +582,2137479,"genie.py",6830,0," ",python,content +583,2137480,"genie.py",6804,0," ",python,content +584,2137480,"genie.py",6736,0," ",python,content +585,2137480,"genie.py",6678,0," ",python,content +586,2137480,"genie.py",6628,0," ",python,content +587,2137480,"genie.py",6601,0," ",python,content +588,2137480,"genie.py",6546,0," ",python,content +589,2137480,"genie.py",6505,0," ",python,content +590,2137480,"genie.py",6482,0," ",python,content +591,2137480,"genie.py",6407,0," ",python,content +592,2137480,"genie.py",6372,0," ",python,content +593,2137480,"genie.py",6325,0," ",python,content +594,2137480,"genie.py",6227,0," ",python,content +595,2137480,"genie.py",6145,0," ",python,content +596,2137481,"genie.py",6078,0," ",python,content +597,2137481,"genie.py",6007,0," ",python,content +598,2137482,"genie.py",6008,0,"",python,selection_keyboard +599,2138298,"genie.py",6007,0,"",python,selection_command +600,2142532,"genie.py",5539,0,"",python,selection_mouse +601,2143330,"genie.py",5788,0,"",python,selection_mouse +602,2144106,"genie.py",6021,0,"",python,selection_mouse +603,2144849,"genie.py",6018,0,"",python,selection_mouse +604,2145577,"genie.py",5941,0,"",python,selection_mouse +605,2158386,"TERMINAL",0,0,"srun",,terminal_focus +606,2159609,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +607,2160149,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +608,2162889,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +609,2164940,"TERMINAL",0,0,"2025-07-17 10:20:07.697362: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +610,2178224,"TERMINAL",0,0,"2025-07-17 10:20:20.990720: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +611,2193382,"TERMINAL",0,0,"2025-07-17 10:20:36.168656: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +612,2197054,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 33000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/033000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 34000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/034000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 35000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/035000/metrics/metrics not found.\r\n",,terminal_output +613,2237728,"TERMINAL",0,0,"action_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs shape: (Array(1, dtype=int32), Array(1, dtype=int32), Array(920, dtype=int32))\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nmask_rng: None\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 179, in \r\n gt = video_batch[:, : vid.shape[1]].clip(0, 1).reshape(-1, *video_batch.shape[2:])\r\nAttributeError: 'NoneType' object has no attribute 'shape'\r\n",,terminal_output +614,2239077,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +615,2273124,"genie.py",0,0,"",python,tab +616,2282812,"genie.py",8777,0,"",python,selection_mouse +617,2282993,"genie.py",8776,0,"",python,selection_mouse +618,2292244,"genie.py",9982,0,"",python,selection_mouse +619,2292402,"genie.py",9980,7,"loop_fn",python,selection_mouse +620,2294109,"genie.py",9970,0,"",python,selection_mouse +621,2294266,"genie.py",9955,19,"final_carry_maskgit",python,selection_mouse +622,2295746,"genie.py",9983,0,"",python,selection_mouse +623,2295910,"genie.py",9980,7,"loop_fn",python,selection_mouse +624,2297913,"genie.py",9088,0,"",python,selection_mouse +625,2298037,"genie.py",9079,11,"MaskGITLoop",python,selection_mouse +626,2298955,"genie.py",8883,0,"",python,selection_mouse +627,2305360,"genie.py",11666,0,"",python,selection_mouse +628,2305500,"genie.py",11665,3,"pad",python,selection_mouse +629,2306537,"genie.py",11671,0,"",python,selection_mouse +630,2306662,"genie.py",11669,9,"act_embed",python,selection_mouse +631,2308833,"genie.py",11602,0,"",python,selection_mouse +632,2308971,"genie.py",11601,4,"self",python,selection_mouse +633,2309164,"genie.py",11601,13,"self.dynamics",python,selection_mouse +634,2309235,"genie.py",11601,14,"self.dynamics.",python,selection_mouse +635,2309273,"genie.py",11601,23,"self.dynamics.action_up",python,selection_mouse +636,2309420,"genie.py",11601,24,"self.dynamics.action_up(",python,selection_mouse +637,2309420,"genie.py",11601,37,"self.dynamics.action_up(action_tokens",python,selection_mouse +638,2309669,"genie.py",11601,38,"self.dynamics.action_up(action_tokens)",python,selection_mouse +639,2318582,"models/dynamics.py",0,0,"",python,tab +640,2337727,"models/dynamics.py",2857,0,"",python,selection_mouse +641,2337759,"models/dynamics.py",2856,0,"",python,selection_command +642,2373086,"genie.py",0,0,"",python,tab +643,2375615,"genie.py",8738,0,"",python,selection_mouse +644,2376195,"genie.py",8818,0,"",python,selection_mouse +645,2377365,"genie.py",8832,0,"",python,selection_command +646,2378215,"genie.py",8832,0," ",python,content +647,2378216,"genie.py",8833,0,"",python,selection_keyboard +648,2378886,"genie.py",8833,0,"#",python,content +649,2378887,"genie.py",8834,0,"",python,selection_keyboard +650,2379029,"genie.py",8834,0," ",python,content +651,2379030,"genie.py",8835,0,"",python,selection_keyboard +652,2380659,"genie.py",8835,0,"()",python,content +653,2380660,"genie.py",8836,0,"",python,selection_keyboard +654,2380975,"genie.py",8836,0,"B",python,content +655,2380976,"genie.py",8837,0,"",python,selection_keyboard +656,2382652,"genie.py",8837,0,",",python,content +657,2382653,"genie.py",8838,0,"",python,selection_keyboard +658,2382754,"genie.py",8838,0," ",python,content +659,2382755,"genie.py",8839,0,"",python,selection_keyboard +660,2383139,"genie.py",8839,0,"S",python,content +661,2383139,"genie.py",8840,0,"",python,selection_keyboard +662,2383650,"genie.py",8840,0,"-",python,content +663,2383652,"genie.py",8841,0,"",python,selection_keyboard +664,2383686,"genie.py",8841,0,"1",python,content +665,2383687,"genie.py",8842,0,"",python,selection_keyboard +666,2389058,"genie.py",8842,0,",",python,content +667,2389059,"genie.py",8843,0,"",python,selection_keyboard +668,2389116,"genie.py",8843,0," ",python,content +669,2389117,"genie.py",8844,0,"",python,selection_keyboard +670,2396874,"genie.py",8840,0,"",python,selection_command +671,2420699,"genie.py",8845,0,"",python,selection_mouse +672,2421002,"genie.py",8844,1,")",python,selection_mouse +673,2421003,"genie.py",8842,3,", )",python,selection_mouse +674,2421005,"genie.py",8839,6,"S-1, )",python,selection_mouse +675,2421006,"genie.py",8838,7," S-1, )",python,selection_mouse +676,2421006,"genie.py",8837,8,", S-1, )",python,selection_mouse +677,2421006,"genie.py",8836,9,"B, S-1, )",python,selection_mouse +678,2421058,"genie.py",8835,10,"(B, S-1, )",python,selection_mouse +679,2421165,"genie.py",8834,11," (B, S-1, )",python,selection_mouse +680,2421229,"genie.py",8833,12,"# (B, S-1, )",python,selection_mouse +681,2423683,"genie.py",8833,12,"",python,content +682,2423685,"genie.py",8833,0,"",python,selection_keyboard +683,2425133,"genie.py",8407,0,"",python,selection_mouse +684,2425747,"genie.py",8309,0,"",python,selection_mouse +685,2426966,"genie.py",8301,0,"",python,selection_mouse +686,2427117,"genie.py",8300,1,"A",python,selection_mouse +687,2433331,"genie.py",6065,0,"",python,selection_mouse +688,2433843,"genie.py",5993,0,"",python,selection_mouse +689,2434146,"genie.py",5921,73," jax.debug.print(""mask_rng: {}"", batch.get(""mask_rng"", None))\n",python,selection_mouse +690,2434704,"genie.py",5921,73,"",python,content +691,2436022,"genie.py",5764,0,"",python,selection_mouse +692,2436651,"genie.py",5750,0,"",python,selection_mouse +693,2451733,"genie.py",5309,0,"",python,selection_mouse +694,2452230,"genie.py",5313,0,"",python,selection_mouse +695,2452979,"genie.py",5313,0," ",python,content +696,2452981,"genie.py",5314,0,"",python,selection_keyboard +697,2453677,"genie.py",5314,0,"# (B, S-1, )",python,content +698,2454754,"genie.py",5325,0,"",python,selection_mouse +699,2460216,"genie.py",6005,0,"",python,selection_mouse +700,2461076,"genie.py",5933,0,"",python,selection_mouse +701,2461835,"genie.py",5933,0,"\n ",python,content +702,2462612,"genie.py",5946,0,"j",python,content +703,2462613,"genie.py",5947,0,"",python,selection_keyboard +704,2462665,"genie.py",5947,0,"a",python,content +705,2462666,"genie.py",5948,0,"",python,selection_keyboard +706,2462831,"genie.py",5948,0,"x",python,content +707,2462832,"genie.py",5949,0,"",python,selection_keyboard +708,2462902,"genie.py",5949,0,".",python,content +709,2462903,"genie.py",5950,0,"",python,selection_keyboard +710,2463638,"genie.py",5950,0,"d",python,content +711,2463639,"genie.py",5951,0,"",python,selection_keyboard +712,2464271,"genie.py",5950,1,"debug",python,content +713,2464837,"genie.py",5955,0,"-",python,content +714,2464838,"genie.py",5956,0,"",python,selection_keyboard +715,2465405,"genie.py",5955,1,"",python,content +716,2465835,"genie.py",5955,0,".",python,content +717,2465837,"genie.py",5956,0,"",python,selection_keyboard +718,2466116,"genie.py",5956,0,"b",python,content +719,2466118,"genie.py",5957,0,"",python,selection_keyboard +720,2466189,"genie.py",5957,0,"r",python,content +721,2466191,"genie.py",5958,0,"",python,selection_keyboard +722,2466353,"genie.py",5958,0,"e",python,content +723,2466354,"genie.py",5959,0,"",python,selection_keyboard +724,2466755,"genie.py",5956,3,"breakpoint",python,content +725,2467596,"genie.py",5966,0,"()",python,content +726,2467597,"genie.py",5967,0,"",python,selection_keyboard +727,2467691,"genie.py",5967,1,")",python,content +728,2467692,"genie.py",5968,0,"",python,selection_keyboard +729,2468811,"genie.py",5967,0,"",python,selection_command +730,2473508,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +731,2473934,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +732,2475701,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +733,2477870,"TERMINAL",0,0,"2025-07-17 10:25:20.649234: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +734,2479323,"genie.py",0,0,"",python,tab +735,2483061,"genie.py",5198,0,"",python,selection_mouse +736,2483231,"genie.py",5198,1,"N",python,selection_mouse +737,2484087,"genie.py",5187,0,"",python,selection_mouse +738,2484241,"genie.py",5186,1,"B",python,selection_mouse +739,2485502,"genie.py",5070,0,"",python,selection_mouse +740,2485637,"genie.py",5058,15,"token_idxs_full",python,selection_mouse +741,2486843,"genie.py",5321,0,"",python,selection_mouse +742,2487995,"genie.py",5320,0,"",python,selection_mouse +743,2488867,"genie.py",5321,0,"",python,selection_mouse +744,2489058,"genie.py",5320,1,"S",python,selection_mouse +745,2489313,"genie.py",5320,2,"S-",python,selection_mouse +746,2489451,"genie.py",5320,3,"S-1",python,selection_mouse +747,2490212,"genie.py",5323,0,"",python,selection_mouse +748,2490881,"genie.py",5375,0,"",python,selection_mouse +749,2490917,"genie.py",5374,0,"",python,selection_command +750,2491030,"genie.py",5375,0,"",python,selection_mouse +751,2491035,"genie.py",5374,0,"",python,selection_command +752,2491200,"genie.py",5374,1,"-",python,selection_mouse +753,2491206,"genie.py",5375,0,"",python,selection_command +754,2491324,"genie.py",5321,54,"-1, )\n # --- Autoregressive generation loop ---",python,selection_mouse +755,2491548,"TERMINAL",0,0,"2025-07-17 10:25:34.339470: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +756,2491857,"genie.py",5321,0,"",python,selection_mouse +757,2492084,"genie.py",5320,1,"S",python,selection_mouse +758,2492831,"genie.py",5321,0,"",python,selection_mouse +759,2493571,"genie.py",5321,1,"-",python,selection_mouse +760,2494110,"genie.py",5322,0,"",python,selection_mouse +761,2506671,"TERMINAL",0,0,"2025-07-17 10:25:49.460547: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +762,2510687,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 34000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/034000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 33000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/033000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 35000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/035000/metrics/metrics not found.\r\n",,terminal_output +763,2520021,"genie.py",0,0,"",python,tab +764,2522062,"genie.py",5674,0,"",python,selection_mouse +765,2522179,"genie.py",5665,13,"action_tokens",python,selection_mouse +766,2525979,"genie.py",8753,0,"",python,selection_mouse +767,2526114,"genie.py",8745,13,"action_tokens",python,selection_mouse +768,2546746,"genie.py",8860,0,"",python,selection_mouse +769,2549341,"genie.py",11633,0,"",python,selection_mouse +770,2549532,"genie.py",11624,9,"vid_embed",python,selection_mouse +771,2552170,"genie.py",11572,0,"",python,selection_mouse +772,2552446,"TERMINAL",0,0,"2025-07-17 10:26:35.217316: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 10:26:35.217761: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +773,2559120,"TERMINAL",0,0,"token_idxs shape: (Array(1, dtype=int32), Array(1, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\nEntering jdb:\r\n(jdb) ",,terminal_output +774,2576543,"TERMINAL",0,0,"a",,terminal_output +775,2576739,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +776,2576956,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +777,2577066,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +778,2577174,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +779,2577604,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +780,2577964,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +781,2578226,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +782,2578282,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +783,2578371,"TERMINAL",0,0,"[?25lk[?25h",,terminal_output +784,2578487,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +785,2578622,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +786,2578735,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +787,2578839,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +788,2579021,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +789,2579164,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +790,2579309,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +791,2579409,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +792,2579570,"TERMINAL",0,0,"[?25le[?25h\r\n(1, 15, 1, 32)\r\n(jdb) ",,terminal_output +793,2582046,"TERMINAL",0,0,"t",,terminal_output +794,2582273,"TERMINAL",0,0,"[?25lk[?25h",,terminal_output +795,2582788,"TERMINAL",0,0,"ó",,terminal_output +796,2582841,"TERMINAL",0,0,"k",,terminal_output +797,2583605,"TERMINAL",0,0,"[?25lk\r[?25h",,terminal_output +798,2583901,"TERMINAL",0,0,"[?25ló\r[?25h",,terminal_output +799,2584010,"TERMINAL",0,0,"[?25lk\r[?25h",,terminal_output +800,2584262,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +801,2584322,"TERMINAL",0,0,"[?25lk[?25h",,terminal_output +802,2584531,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +803,2584946,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +804,2585319,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +805,2585640,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +806,2585750,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +807,2586886,"TERMINAL",0,0,"[?25lx[?25h",,terminal_output +808,2587013,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +809,2587635,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +810,2587893,"TERMINAL",0,0,"[?25lf[?25h",,terminal_output +811,2588033,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +812,2588410,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +813,2588554,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +814,2588953,"TERMINAL",0,0,"[?25l-[?25h",,terminal_output +815,2589163,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +816,2589279,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +817,2589537,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +818,2589675,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +819,2589810,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +820,2590772,"TERMINAL",0,0,"[?25le\r[?25h",,terminal_output +821,2590883,"TERMINAL",0,0,"[?25lp\r[?25h",,terminal_output +822,2591086,"TERMINAL",0,0,"[?25lha\r[?25h",,terminal_output +823,2591193,"TERMINAL",0,0,"\r",,terminal_output +824,2591305,"TERMINAL",0,0,"[?25ls\r[?25h",,terminal_output +825,2591884,"TERMINAL",0,0,"[?25l-\r[?25h",,terminal_output +826,2592093,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +827,2592284,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +828,2592745,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +829,2592893,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +830,2592998,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +831,2593190,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +832,2593306,"TERMINAL",0,0,"\r\n(1, 16, 920)\r\n(jdb) ",,terminal_output +833,2603594,"genie.py",0,0,"",python,tab +834,2604826,"genie.py",11603,0,"",python,selection_mouse +835,2605816,"genie.py",11636,0,"",python,selection_mouse +836,2606346,"genie.py",11638,0,"",python,selection_mouse +837,2606487,"genie.py",11637,3,"jnp",python,selection_mouse +838,2606727,"genie.py",11637,4,"jnp.",python,selection_mouse +839,2606727,"genie.py",11637,8,"jnp.pad(",python,selection_mouse +840,2606728,"genie.py",11637,17,"jnp.pad(act_embed",python,selection_mouse +841,2606820,"genie.py",11637,18,"jnp.pad(act_embed,",python,selection_mouse +842,2606821,"genie.py",11637,20,"jnp.pad(act_embed, (",python,selection_mouse +843,2606875,"genie.py",11637,21,"jnp.pad(act_embed, ((",python,selection_mouse +844,2606877,"genie.py",11600,40,"(action_tokens)\n vid_embed += jnp",python,selection_mouse +845,2606877,"genie.py",11601,39,"action_tokens)\n vid_embed += jnp",python,selection_mouse +846,2606978,"genie.py",11637,34,"jnp.pad(act_embed, ((0, 0), (1, 0)",python,selection_mouse +847,2607003,"genie.py",11637,35,"jnp.pad(act_embed, ((0, 0), (1, 0),",python,selection_mouse +848,2607022,"genie.py",11637,36,"jnp.pad(act_embed, ((0, 0), (1, 0), ",python,selection_mouse +849,2607055,"genie.py",11637,37,"jnp.pad(act_embed, ((0, 0), (1, 0), (",python,selection_mouse +850,2607075,"genie.py",11637,39,"jnp.pad(act_embed, ((0, 0), (1, 0), (0,",python,selection_mouse +851,2607110,"genie.py",11637,40,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, ",python,selection_mouse +852,2607139,"genie.py",11637,42,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0)",python,selection_mouse +853,2607230,"genie.py",11637,44,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), ",python,selection_mouse +854,2607231,"genie.py",11637,45,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (",python,selection_mouse +855,2607241,"genie.py",11637,47,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0,",python,selection_mouse +856,2607289,"genie.py",11637,48,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, ",python,selection_mouse +857,2607315,"genie.py",11637,49,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0",python,selection_mouse +858,2607338,"genie.py",11637,50,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)",python,selection_mouse +859,2607383,"genie.py",11637,51,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0))",python,selection_mouse +860,2607408,"genie.py",11637,52,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +861,2607730,"genie.py",11689,0,"",python,selection_mouse +862,2607749,"genie.py",11688,0,"",python,selection_command +863,2608100,"genie.py",11689,0,"",python,selection_mouse +864,2608115,"genie.py",11688,0,"",python,selection_command +865,2608344,"genie.py",11688,1,")",python,selection_mouse +866,2608345,"genie.py",11687,1,")",python,selection_mouse +867,2608345,"genie.py",11683,5,", 0))",python,selection_mouse +868,2608346,"genie.py",11678,10,"), (0, 0))",python,selection_mouse +869,2608346,"genie.py",11675,13,", 0), (0, 0))",python,selection_mouse +870,2608346,"genie.py",11671,17,", (0, 0), (0, 0))",python,selection_mouse +871,2608370,"genie.py",11689,0,"",python,selection_command +872,2608370,"genie.py",11667,22,", 0), (0, 0), (0, 0)))",python,selection_mouse +873,2608422,"genie.py",11664,25," (1, 0), (0, 0), (0, 0)))",python,selection_mouse +874,2608466,"genie.py",11659,30,", 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +875,2608492,"genie.py",11656,33,"((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +876,2608517,"genie.py",11655,34," ((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +877,2608547,"genie.py",11645,44,"act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +878,2608625,"genie.py",11689,34,"\n unmasked_ratio = jnp.cos(",python,selection_mouse +879,2608667,"genie.py",11689,33,"\n unmasked_ratio = jnp.cos",python,selection_mouse +880,2608761,"genie.py",11689,30,"\n unmasked_ratio = jnp.",python,selection_mouse +881,2608763,"genie.py",11689,29,"\n unmasked_ratio = jnp",python,selection_mouse +882,2608907,"genie.py",11689,26,"\n unmasked_ratio = ",python,selection_mouse +883,2608981,"genie.py",11689,25,"\n unmasked_ratio =",python,selection_mouse +884,2609032,"genie.py",11640,49,".pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +885,2609187,"genie.py",11637,52,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +886,2609726,"genie.py",11639,0,"",python,selection_mouse +887,2609727,"genie.py",11637,3,"jnp",python,selection_mouse +888,2609959,"genie.py",11637,4,"jnp.",python,selection_mouse +889,2609960,"genie.py",11637,7,"jnp.pad",python,selection_mouse +890,2609960,"genie.py",11637,8,"jnp.pad(",python,selection_mouse +891,2609960,"genie.py",11637,17,"jnp.pad(act_embed",python,selection_mouse +892,2610033,"genie.py",11637,21,"jnp.pad(act_embed, ((",python,selection_mouse +893,2610062,"genie.py",11637,23,"jnp.pad(act_embed, ((0,",python,selection_mouse +894,2610091,"genie.py",11637,25,"jnp.pad(act_embed, ((0, 0",python,selection_mouse +895,2610119,"genie.py",11637,27,"jnp.pad(act_embed, ((0, 0),",python,selection_mouse +896,2610148,"genie.py",11637,28,"jnp.pad(act_embed, ((0, 0), ",python,selection_mouse +897,2610173,"genie.py",11637,30,"jnp.pad(act_embed, ((0, 0), (1",python,selection_mouse +898,2610224,"genie.py",11637,33,"jnp.pad(act_embed, ((0, 0), (1, 0",python,selection_mouse +899,2610267,"genie.py",11637,35,"jnp.pad(act_embed, ((0, 0), (1, 0),",python,selection_mouse +900,2610270,"genie.py",11637,37,"jnp.pad(act_embed, ((0, 0), (1, 0), (",python,selection_mouse +901,2610302,"genie.py",11637,38,"jnp.pad(act_embed, ((0, 0), (1, 0), (0",python,selection_mouse +902,2610325,"genie.py",11637,39,"jnp.pad(act_embed, ((0, 0), (1, 0), (0,",python,selection_mouse +903,2610356,"genie.py",11637,41,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0",python,selection_mouse +904,2610386,"genie.py",11637,42,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0)",python,selection_mouse +905,2610417,"genie.py",11637,43,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0),",python,selection_mouse +906,2610448,"genie.py",11637,45,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (",python,selection_mouse +907,2610475,"genie.py",11637,46,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0",python,selection_mouse +908,2610502,"genie.py",11637,47,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0,",python,selection_mouse +909,2610551,"genie.py",11637,48,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, ",python,selection_mouse +910,2610602,"genie.py",11637,49,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0",python,selection_mouse +911,2610643,"genie.py",11637,50,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)",python,selection_mouse +912,2610665,"genie.py",11637,51,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0))",python,selection_mouse +913,2610706,"genie.py",11637,52,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +914,2611050,"genie.py",11689,0,"",python,selection_mouse +915,2611059,"genie.py",11688,0,"",python,selection_command +916,2611353,"genie.py",11689,0,"",python,selection_mouse +917,2611368,"genie.py",11688,0,"",python,selection_command +918,2611517,"genie.py",11688,1,")",python,selection_mouse +919,2611521,"genie.py",11689,0,"",python,selection_command +920,2611721,"genie.py",11686,3,")))",python,selection_mouse +921,2611721,"genie.py",11677,12,"0), (0, 0)))",python,selection_mouse +922,2611722,"genie.py",11674,15,"0, 0), (0, 0)))",python,selection_mouse +923,2611722,"genie.py",11671,18,", (0, 0), (0, 0)))",python,selection_mouse +924,2611722,"genie.py",11668,21," 0), (0, 0), (0, 0)))",python,selection_mouse +925,2611722,"genie.py",11665,24,"(1, 0), (0, 0), (0, 0)))",python,selection_mouse +926,2611723,"genie.py",11664,25," (1, 0), (0, 0), (0, 0)))",python,selection_mouse +927,2611723,"genie.py",11662,27,"), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +928,2611723,"genie.py",11661,28,"0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +929,2611754,"genie.py",11660,29," 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +930,2611783,"genie.py",11659,30,", 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +931,2611812,"genie.py",11658,31,"0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +932,2611849,"genie.py",11656,33,"((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +933,2611904,"genie.py",11645,44,"act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +934,2612049,"genie.py",11644,45,"(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +935,2612085,"genie.py",11641,48,"pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +936,2612140,"genie.py",11640,49,".pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +937,2612207,"genie.py",11637,52,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +938,2612582,"genie.py",11639,0,"",python,selection_mouse +939,2612583,"genie.py",11637,3,"jnp",python,selection_mouse +940,2612776,"genie.py",11637,7,"jnp.pad",python,selection_mouse +941,2612776,"genie.py",11637,17,"jnp.pad(act_embed",python,selection_mouse +942,2612850,"genie.py",11637,18,"jnp.pad(act_embed,",python,selection_mouse +943,2612851,"genie.py",11637,25,"jnp.pad(act_embed, ((0, 0",python,selection_mouse +944,2612851,"genie.py",11637,28,"jnp.pad(act_embed, ((0, 0), ",python,selection_mouse +945,2612919,"genie.py",11637,32,"jnp.pad(act_embed, ((0, 0), (1, ",python,selection_mouse +946,2612919,"genie.py",11637,37,"jnp.pad(act_embed, ((0, 0), (1, 0), (",python,selection_mouse +947,2612920,"genie.py",11637,39,"jnp.pad(act_embed, ((0, 0), (1, 0), (0,",python,selection_mouse +948,2612949,"genie.py",11637,41,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0",python,selection_mouse +949,2612975,"genie.py",11637,44,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), ",python,selection_mouse +950,2613004,"genie.py",11637,45,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (",python,selection_mouse +951,2613032,"genie.py",11637,48,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, ",python,selection_mouse +952,2613062,"genie.py",11637,49,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0",python,selection_mouse +953,2613095,"genie.py",11637,51,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0))",python,selection_mouse +954,2613124,"genie.py",11637,52,"jnp.pad(act_embed, ((0, 0), (1, 0), (0, 0), (0, 0)))",python,selection_mouse +955,2613574,"genie.py",11689,0,"",python,selection_mouse +956,2613594,"genie.py",11688,0,"",python,selection_command +957,2617538,"models/dynamics.py",0,0,"",python,tab +958,2639495,"genie.py",0,0,"",python,tab +959,2640629,"train_dynamics.py",0,0,"",python,tab +960,2671820,"genie.py",0,0,"",python,tab +961,2673162,"train_dynamics.py",0,0,"",python,tab +962,2674023,"genie.py",0,0,"",python,tab +963,2680980,"genie.py",5796,0,"",python,selection_mouse +964,2681717,"genie.py",5777,0,"",python,selection_mouse +965,2681721,"genie.py",5776,0,"",python,selection_command +966,2683065,"genie.py",5606,0,"",python,selection_mouse +967,2683236,"genie.py",5606,1,"[",python,selection_mouse +968,2683236,"genie.py",5606,3,"[:,",python,selection_mouse +969,2683237,"genie.py",5606,5,"[:, :",python,selection_mouse +970,2683285,"genie.py",5606,7,"[:, :t,",python,selection_mouse +971,2683336,"genie.py",5606,8,"[:, :t, ",python,selection_mouse +972,2683511,"genie.py",5606,9,"[:, :t, :",python,selection_mouse +973,2684103,"genie.py",5606,10,"[:, :t, :]",python,selection_mouse +974,2685290,"genie.py",5606,10,"",python,content +975,2686692,"genie.py",5668,0,"",python,selection_mouse +976,2686913,"genie.py",5668,1,"[",python,selection_mouse +977,2686913,"genie.py",5668,2,"[:",python,selection_mouse +978,2686914,"genie.py",5668,4,"[:, ",python,selection_mouse +979,2686954,"genie.py",5668,5,"[:, :",python,selection_mouse +980,2686998,"genie.py",5668,7,"[:, :t,",python,selection_mouse +981,2687037,"genie.py",5668,8,"[:, :t, ",python,selection_mouse +982,2687065,"genie.py",5668,9,"[:, :t, .",python,selection_mouse +983,2687089,"genie.py",5668,10,"[:, :t, ..",python,selection_mouse +984,2687172,"genie.py",5668,11,"[:, :t, ...",python,selection_mouse +985,2687529,"genie.py",5668,12,"[:, :t, ...]",python,selection_mouse +986,2687978,"genie.py",5668,12,"",python,content +987,2692758,"genie.py",5982,0,"",python,selection_mouse +988,2694162,"genie.py",5959,0,"",python,selection_command +989,2694682,"genie.py",5959,1,"",python,content +990,2694831,"genie.py",5959,1,"",python,content +991,2695015,"genie.py",5958,0,"",python,selection_command +992,2695578,"genie.py",5923,0,"",python,selection_command +993,2696618,"genie.py",5912,35,"",python,content +994,2696692,"genie.py",5924,0,"",python,selection_command +995,2696715,"genie.py",5994,0,"",python,selection_command +996,2697012,"genie.py",6049,0,"\n jax.debug.breakpoint()",python,content +997,2697039,"genie.py",6062,0,"",python,selection_command +998,2697542,"genie.py",5994,0,"",python,selection_command +999,2697858,"genie.py",6062,0,"",python,selection_command +1000,2698241,"genie.py",6050,35,"",python,content +1001,2698344,"genie.py",6062,0,"",python,selection_command +1002,2698416,"genie.py",5994,0,"",python,selection_command +1003,2698642,"genie.py",5924,0,"",python,selection_command +1004,2698811,"genie.py",5981,0,"\n jax.debug.breakpoint()",python,content +1005,2698861,"genie.py",5994,0,"",python,selection_command +1006,2700088,"genie.py",6057,0,"",python,selection_mouse +1007,2700609,"genie.py",6016,0,"",python,selection_mouse +1008,2700665,"genie.py",6015,0,"",python,selection_command +1009,2719622,"TERMINAL",0,0,"--KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) ",,terminal_output +1010,2720170,"TERMINAL",0,0,"--KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) ",,terminal_output +1011,2720385,"TERMINAL",0,0,"--KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) ",,terminal_output +1012,2720514,"TERMINAL",0,0,"--KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) ",,terminal_output +1013,2720721,"TERMINAL",0,0,"--KeyboardInterrupt--\r\nEntering jdb:\r\n(jdb) ",,terminal_output +1014,2721317,"TERMINAL",0,0,"^DERROR:2025-07-17 10:29:24,036:jax._src.debugging:98: jax.debug.callback failed\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 96, in debug_callback_impl\r\n callback(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 336, in _flat_callback\r\n callback(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/core.py"", line 220, in _breakpoint_callback\r\n debugger(frames, thread_id, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 167, in run_debugger\r\n CliDebugger(frames, thread_id, **kwargs).run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 160, in run\r\n self.cmdloop()\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 138, in cmdloop\r\n stop = self.onecmd(line)\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 217, in onecmd\r\n return func(arg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 146, in do_quit\r\n sys.exit(0)\r\nSystemExit: 0\r\nERROR:jax._src.debugging:jax.debug.callback failed\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 96, in debug_callback_impl\r\n callback(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 336, in _flat_callback\r\n callback(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/core.py"", line 220, in _breakpoint_callback\r\n debugger(frames, thread_id, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 167, in run_debugger\r\n CliDebugger(frames, thread_id, **kwargs).run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 160, in run\r\n self.cmdloop()\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 138, in cmdloop\r\n stop = self.onecmd(line)\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 217, in onecmd\r\n return func(arg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 146, in do_quit\r\n sys.exit(0)\r\nSystemExit: 0\r\nE0717 10:29:24.060070 938302 pjrt_stream_executor_client.cc:2916] Execution of replica 0 failed: INTERNAL: CpuCallback error calling callback: Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 178, in \r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 141, in _autoreg_sample\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/traceback_util.py"", line 182, in reraise_with_filtered_traceback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 292, in cache_miss\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 153, in _python_pjit_helper\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 1877, in _pjit_call_impl_python\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/profiler.py"", line 354, in wrapper\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/interpreters/pxla.py"", line 1297, in __call__\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/callback.py"", line 782, in _wrapped_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 202, in _callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 99, in debug_callback_impl\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 336, in _flat_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/core.py"", line 220, in _breakpoint_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 167, in run_debugger\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 162, in run\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 145, in cmdloop\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 217, in onecmd\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 146, in do_quit\r\nSystemExit: 0\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 178, in \r\n vid = _autoreg_sample(rng, video_batch, action_batch)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 141, in _autoreg_sample\r\n generated_vid = sampling_fn(\r\njaxlib._jax.XlaRuntimeError: INTERNAL: CpuCallback error calling callback: Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 178, in \r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 141, in _autoreg_sample\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/traceback_util.py"", line 182, in reraise_with_filtered_traceback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 292, in cache_miss\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 153, in _python_pjit_helper\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 1877, in _pjit_call_impl_python\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/profiler.py"", line 354, in wrapper\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/interpreters/pxla.py"", line 1297, in __call__\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/callback.py"", line 782, in _wrapped_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 202, in _callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 99, in debug_callback_impl\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 336, in _flat_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/core.py"", line 220, in _breakpoint_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 167, in run_debugger\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 162, in run\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 145, in cmdloop\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 217, in onecmd\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 146, in do_quit\r\nSystemExit: 0\r\n",,terminal_output +1015,2722888,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +1016,2723531,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +1017,2724173,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +1018,2725845,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +1019,2727889,"TERMINAL",0,0,"2025-07-17 10:29:30.676504: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1020,2742028,"TERMINAL",0,0,"2025-07-17 10:29:44.804006: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1021,2757245,"TERMINAL",0,0,"2025-07-17 10:30:00.028841: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1022,2760985,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 35000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/035000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 34000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/034000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 36000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/036000/metrics/metrics not found.\r\n",,terminal_output +1023,2820043,"TERMINAL",0,0,"2025-07-17 10:31:02.834126: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 10:31:02.834634: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1024,2835900,"TERMINAL",0,0,"token_idxs shape: (Array(1, dtype=int32), Array(1, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\nEntering jdb:\r\n(jdb) ",,terminal_output +1025,2889748,"TERMINAL",0,0,"l",,terminal_output +1026,2889965,"TERMINAL",0,0,"\r\n> /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py(170)\r\n ""latent_actions"": action_tokens, # (B, t, ...)\r\n ""mask_rng"": batch.get(""mask_rng"", None),\r\n }\r\n jax.debug.print(""token_idxs_full shape: {}"", token_idxs_full.shape)\r\n jax.debug.print(""action_tokens shape: {}"", action_tokens.shape)\r\n dyna_outputs = self.dynamics(dyna_inputs, training=False)\r\n-> jax.debug.breakpoint()\r\n # # dyna_outputs[""token_logits""]: (B, t, N, vocab_size)\r\n # # We want the logits for the last time step (frame t-1 predicting t)\r\n # next_token_logits = dyna_outputs[""token_logits""][:, -1, :, :] # (B, N, vocab_size)\r\n \r\n # # Sample or argmax for each patch\r\n(jdb) ",,terminal_output +1027,2892796,"TERMINAL",0,0,"d",,terminal_output +1028,2893049,"TERMINAL",0,0,"[?25ly[?25h",,terminal_output +1029,2893173,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +1030,2893298,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1031,2893920,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +1032,2894260,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +1033,2894489,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +1034,2894744,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +1035,2895003,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +1036,2895144,"TERMINAL",0,0,"[?25lu[?25h",,terminal_output +1037,2895244,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +1038,2895490,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1039,2896457,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +1040,2896721,"TERMINAL",0,0,"[?25lk[?25h",,terminal_output +1041,2896794,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1042,2897040,"TERMINAL",0,0,"[?25ly[?25h",,terminal_output +1043,2897204,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1044,2897641,"TERMINAL",0,0,"[?25l([?25h[?25l)[?25h",,terminal_output +1045,2897953,"TERMINAL",0,0,"\r\ndict_keys(['mask', 'token_logits'])\r\n(jdb) ",,terminal_output +1046,2899281,"TERMINAL",0,0,"\rdyna_outputs.keys()",,terminal_output +1047,2900002,"TERMINAL",0,0,"[?25l)\r[?25h",,terminal_output +1048,2900131,"TERMINAL",0,0,"[?25l(\r[?25h",,terminal_output +1049,2900275,"TERMINAL",0,0,"[?25ls\r[?25h",,terminal_output +1050,2900374,"TERMINAL",0,0,"[?25ly\r[?25h",,terminal_output +1051,2900534,"TERMINAL",0,0,"[?25le\r[?25h",,terminal_output +1052,2900684,"TERMINAL",0,0,"[?25lk\r[?25h",,terminal_output +1053,2901024,"TERMINAL",0,0,"[?25l.\r[?25h",,terminal_output +1054,2901564,"TERMINAL",0,0,"[",,terminal_output +1055,2901934,"TERMINAL",0,0,"[?25l""[?25h",,terminal_output +1056,2902288,"TERMINAL",0,0,"[?25lt[?25h[?25lo[?25h",,terminal_output +1057,2902386,"TERMINAL",0,0,"[?25lk[?25h",,terminal_output +1058,2902496,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1059,2902651,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +1060,2902888,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +1061,2903167,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +1062,2903290,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +1063,2903456,"TERMINAL",0,0,"[?25lg[?25h",,terminal_output +1064,2903572,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +1065,2903811,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +1066,2904042,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1067,2904427,"TERMINAL",0,0,"[?25l""[?25h",,terminal_output +1068,2904800,"TERMINAL",0,0,"[?25l][?25h",,terminal_output +1069,2905161,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +1070,2905352,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1071,2905475,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +1072,2905633,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1073,2905717,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +1074,2905807,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1075,2905907,"TERMINAL",0,0,"\r\n(1, 16, 920, 1024)\r\n(jdb) ",,terminal_output +1076,2916265,"TERMINAL",0,0,"[?25lma[?25h",,terminal_output +1077,2916370,"TERMINAL",0,0,"[?25la[?25h[?25ls[?25h",,terminal_output +1078,2916499,"TERMINAL",0,0,"[?25lk[?25h",,terminal_output +1079,2917513,"TERMINAL",0,0,"[?25lk\r[?25h",,terminal_output +1080,2917662,"TERMINAL",0,0,"[?25ls\r[?25h",,terminal_output +1081,2917762,"TERMINAL",0,0,"[?25la\r[?25h",,terminal_output +1082,2917952,"TERMINAL",0,0,"[?25lm\r[?25h",,terminal_output +1083,2918092,"TERMINAL",0,0,"",,terminal_output +1084,2918287,"TERMINAL",0,0,"\rdyna_outputs[""token_logits""].shape",,terminal_output +1085,2919035,"TERMINAL",0,0,"[?25l[?25h",,terminal_output +1086,2919825,"TERMINAL",0,0,"[?25ls\r""[?25h",,terminal_output +1087,2919934,"TERMINAL",0,0,"[?25lt\r""[?25h",,terminal_output +1088,2920143,"TERMINAL",0,0,"[?25li\r""[?25h",,terminal_output +1089,2920501,"TERMINAL",0,0,"[?25lg\r""[?25h",,terminal_output +1090,2920697,"TERMINAL",0,0,"[?25lo\r""[?25h",,terminal_output +1091,2920886,"TERMINAL",0,0,"[?25ll\r""[?25h",,terminal_output +1092,2921107,"TERMINAL",0,0,"[?25l_\r""[?25h",,terminal_output +1093,2921222,"TERMINAL",0,0,"[?25ln\r""[?25h",,terminal_output +1094,2921438,"TERMINAL",0,0,"[?25le\r""[?25h",,terminal_output +1095,2921608,"TERMINAL",0,0,"[?25lk\r""[?25h",,terminal_output +1096,2921771,"TERMINAL",0,0,"[?25lo\r""[?25h",,terminal_output +1097,2922132,"TERMINAL",0,0,"[?25lt\r""[?25h",,terminal_output +1098,2922530,"TERMINAL",0,0,"[?25l]""\r[1@m""[?25h",,terminal_output +1099,2922595,"TERMINAL",0,0,"[?25l""\r[1@a""[?25h",,terminal_output +1100,2922691,"TERMINAL",0,0,"[?25l""\r[1@s""[?25h\r[1@k""",,terminal_output +1101,2923377,"TERMINAL",0,0,"\r\n(1, 16, 920)\r\n(jdb) ",,terminal_output +1102,2948593,"genie.py",0,0,"",python,tab +1103,2968524,"genie.py",5755,0,"",python,selection_mouse +1104,2968568,"genie.py",5754,0,"",python,selection_command +1105,2969467,"genie.py",5644,0,"",python,selection_mouse +1106,2969607,"genie.py",5638,14,"latent_actions",python,selection_mouse +1107,2970129,"genie.py",5708,0,"",python,selection_mouse +1108,2970276,"genie.py",5702,8,"mask_rng",python,selection_mouse +1109,2970885,"genie.py",5718,0,"",python,selection_mouse +1110,2973870,"models/dynamics.py",0,0,"",python,tab +1111,2977230,"models/dynamics.py",2835,0,"",python,selection_mouse +1112,2977421,"models/dynamics.py",2833,5,"batch",python,selection_mouse +1113,2984843,"genie.py",0,0,"",python,tab +1114,2987246,"genie.py",5714,0,"",python,selection_mouse +1115,2988203,"genie.py",5685,57,"",python,content +1116,2988240,"genie.py",5697,0,"",python,selection_command +1117,2988877,"genie.py",5633,0,"",python,selection_command +1118,2989063,"genie.py",5571,0,"",python,selection_command +1119,2989271,"genie.py",5543,0,"",python,selection_command +1120,2989499,"genie.py",5571,0,"",python,selection_command +1121,2989657,"genie.py",5543,0,"",python,selection_command +1122,2989742,"genie.py",5571,0,"",python,selection_command +1123,2989839,"genie.py",5543,0,"",python,selection_command +1124,2989940,"genie.py",5571,0,"",python,selection_command +1125,2990054,"genie.py",5543,0,"",python,selection_command +1126,2990160,"genie.py",5571,0,"",python,selection_command +1127,2990260,"genie.py",5543,0,"",python,selection_command +1128,2990360,"genie.py",5571,0,"",python,selection_command +1129,2990474,"genie.py",5543,0,"",python,selection_command +1130,2990546,"genie.py",5571,0,"",python,selection_command +1131,2990682,"genie.py",5543,0,"",python,selection_command +1132,2990733,"genie.py",5571,0,"",python,selection_command +1133,2990868,"genie.py",5543,0,"",python,selection_command +1134,2990951,"genie.py",5571,0,"",python,selection_command +1135,2991255,"genie.py",5633,0,"",python,selection_command +1136,2991440,"genie.py",5571,0,"",python,selection_command +1137,2991571,"genie.py",5633,0,"",python,selection_command +1138,2991719,"genie.py",5571,0,"",python,selection_command +1139,2991769,"genie.py",5633,0,"",python,selection_command +1140,2991942,"genie.py",5571,0,"",python,selection_command +1141,2991995,"genie.py",5633,0,"",python,selection_command +1142,2992149,"genie.py",5571,0,"",python,selection_command +1143,2992224,"genie.py",5633,0,"",python,selection_command +1144,2992354,"genie.py",5571,0,"",python,selection_command +1145,2992444,"genie.py",5633,0,"",python,selection_command +1146,2992586,"genie.py",5571,0,"",python,selection_command +1147,2992693,"genie.py",5633,0,"",python,selection_command +1148,2992824,"genie.py",5571,0,"",python,selection_command +1149,2992940,"genie.py",5633,0,"",python,selection_command +1150,2993090,"genie.py",5571,0,"",python,selection_command +1151,2993159,"genie.py",5633,0,"",python,selection_command +1152,2993296,"genie.py",5571,0,"",python,selection_command +1153,2993422,"genie.py",5633,0,"",python,selection_command +1154,2993520,"genie.py",5571,0,"",python,selection_command +1155,2993628,"genie.py",5633,0,"",python,selection_command +1156,2993844,"genie.py",5571,0,"",python,selection_command +1157,2993964,"genie.py",5633,0,"",python,selection_command +1158,2994188,"genie.py",5571,0,"",python,selection_command +1159,2994321,"genie.py",5633,0,"",python,selection_command +1160,2994475,"genie.py",5571,0,"",python,selection_command +1161,2994585,"genie.py",5633,0,"",python,selection_command +1162,2994767,"genie.py",5571,0,"",python,selection_command +1163,2994881,"genie.py",5633,0,"",python,selection_command +1164,2995226,"genie.py",5571,0,"",python,selection_command +1165,2995511,"genie.py",5620,0,"",python,selection_command +1166,2996547,"genie.py",5619,1,"",python,content +1167,2997259,"genie.py",5618,1,"",python,content +1168,2997427,"genie.py",5616,2,"",python,content +1169,2997608,"genie.py",5615,1,"",python,content +1170,2997780,"genie.py",5613,2,"",python,content +1171,2997955,"genie.py",5612,1,"",python,content +1172,2998101,"genie.py",5611,1,"",python,content +1173,2998305,"genie.py",5609,2,"",python,content +1174,2998578,"genie.py",5607,2,"",python,content +1175,2999095,"genie.py",5656,0,"",python,selection_command +1176,2999614,"genie.py",5656,2,"",python,content +1177,2999813,"genie.py",5656,1,"",python,content +1178,2999981,"genie.py",5656,2,"",python,content +1179,3000269,"genie.py",5656,1,"",python,content +1180,3000443,"genie.py",5656,1,"",python,content +1181,3000607,"genie.py",5656,2,"",python,content +1182,3000762,"genie.py",5656,1,"",python,content +1183,3000912,"genie.py",5656,5,"",python,content +1184,3001101,"genie.py",5656,13,"",python,content +1185,3001835,"genie.py",5656,0,"\n ",python,content +1186,3003288,"genie.py",5674,0,"",python,selection_mouse +1187,3003789,"genie.py",5656,0,"",python,selection_mouse +1188,3004024,"genie.py",5655,1,"",python,content +1189,3004602,"genie.py",5654,0,"",python,selection_command +1190,3004715,"genie.py",5605,0,"",python,selection_command +1191,3004875,"genie.py",5557,0,"",python,selection_command +1192,3005003,"genie.py",5495,0,"",python,selection_command +1193,3005438,"genie.py",5449,82,"",python,content +1194,3005508,"genie.py",5461,0,"",python,selection_command +1195,3005764,"genie.py",5425,0,"",python,selection_command +1196,3006312,"genie.py",5461,0,"",python,selection_command +1197,3006530,"genie.py",5489,0,"",python,selection_command +1198,3006707,"genie.py",5538,0,"",python,selection_command +1199,3006882,"genie.py",5586,0,"",python,selection_command +1200,3007042,"genie.py",5604,0,"",python,selection_command +1201,3007186,"genie.py",5684,0,"",python,selection_command +1202,3007369,"genie.py",5760,0,"",python,selection_command +1203,3012920,"TERMINAL",0,0,"bash",,terminal_focus +1204,3013479,"TERMINAL",0,0,"srun",,terminal_focus +1205,3014234,"TERMINAL",0,0,"p",,terminal_output +1206,3014747,"TERMINAL",0,0,"[?25lr[?25h",,terminal_output +1207,3014920,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +1208,3015031,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +1209,3015137,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +1210,3015771,"TERMINAL",0,0,"[?25l([?25h",,terminal_output +1211,3016012,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +1212,3016272,"TERMINAL",0,0,"[?25l)[?25h",,terminal_output +1213,3016524,"TERMINAL",0,0,"\r\n1\r\nNone\r\n(jdb) ",,terminal_output +1214,3019218,"TERMINAL",0,0,"t",,terminal_output +1215,3019374,"TERMINAL",0,0,"\r\n1\r\n(jdb) ",,terminal_output +1216,3025910,"TERMINAL",0,0,"[?25lto[?25h",,terminal_output +1217,3025967,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +1218,3026116,"TERMINAL",0,0,"[?25lk[?25h",,terminal_output +1219,3026226,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1220,3026325,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +1221,3027555,"TERMINAL",0,0,"\r",,terminal_output +1222,3030175,"TERMINAL",0,0,"\rt",,terminal_output +1223,3030344,"TERMINAL",0,0,"\rprint(t)",,terminal_output +1224,3030663,"TERMINAL",0,0,"\rdyna_outputs[""mask""].shape",,terminal_output +1225,3031141,"TERMINAL",0,0,"\r[8@token_logits",,terminal_output +1226,3031584,"TERMINAL",0,0,"\r.keys()",,terminal_output +1227,3032028,"TERMINAL",0,0,"\rl",,terminal_output +1228,3032553,"TERMINAL",0,0,"\rdyna_outputs.keys()",,terminal_output +1229,3032735,"TERMINAL",0,0,"\r[""token_logits""].shape",,terminal_output +1230,3032935,"TERMINAL",0,0,"\rmask",,terminal_output +1231,3033081,"TERMINAL",0,0,"\rprint(t)",,terminal_output +1232,3033211,"TERMINAL",0,0,"\rt",,terminal_output +1233,3033404,"TERMINAL",0,0,"\r",,terminal_output +1234,3033980,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +1235,3034047,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +1236,3034155,"TERMINAL",0,0,"[?25lk[?25h",,terminal_output +1237,3034262,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1238,3034342,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +1239,3034757,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +1240,3035071,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +1241,3035274,"TERMINAL",0,0,"[?25lo[?25h",,terminal_output +1242,3035421,"TERMINAL",0,0,"[?25lg[?25h",,terminal_output +1243,3035542,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +1244,3035999,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +1245,3036256,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1246,3038162,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +1247,3038665,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1248,3038736,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +1249,3038880,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1250,3039012,"TERMINAL",0,0,"[?25lp[?25h[?25le[?25h",,terminal_output +1251,3039128,"TERMINAL",0,0,"\r\n*** NameError: name 'token_logits' is not defined\r\n(jdb) ",,terminal_output +1252,3040791,"train_dynamics.py",0,0,"",python,tab +1253,3042314,"genie.py",0,0,"",python,tab +1254,3044164,"genie.py",5516,0,"",python,selection_mouse +1255,3044327,"genie.py",5509,15,"token_idxs_full",python,selection_mouse +1256,3047972,"TERMINAL",0,0,"token_idxs_full",,terminal_output +1257,3048299,"TERMINAL",0,0,"[?25l.[?25h",,terminal_output +1258,3048451,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1259,3048575,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +1260,3048638,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1261,3048778,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +1262,3048871,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1263,3048944,"TERMINAL",0,0,"\r\n(1, 16, 920)\r\n(jdb) ",,terminal_output +1264,3049548,"TERMINAL",0,0,"\rtoken_idxs_full.shape",,terminal_output +1265,3050197,"TERMINAL",0,0,"[?25le\r[?25h",,terminal_output +1266,3050414,"TERMINAL",0,0,"[?25lap\r[?25h",,terminal_output +1267,3050541,"TERMINAL",0,0,"\r",,terminal_output +1268,3050667,"TERMINAL",0,0,"[?25lh\r[?25h",,terminal_output +1269,3050777,"TERMINAL",0,0,"[?25ls\r[?25h",,terminal_output +1270,3051254,"TERMINAL",0,0,"[?25l.\r[?25h",,terminal_output +1271,3051971,"TERMINAL",0,0,"[",,terminal_output +1272,3052449,"TERMINAL",0,0,"[?25l:[?25h",,terminal_output +1273,3052642,"TERMINAL",0,0,"[?25l,[?25h",,terminal_output +1274,3052906,"TERMINAL",0,0,"[?25l0[?25h",,terminal_output +1275,3055527,"TERMINAL",0,0,"[?25l0\r[?25h",,terminal_output +1276,3055709,"TERMINAL",0,0,"[?25l,\r[?25h",,terminal_output +1277,3055853,"TERMINAL",0,0,"[?25l:\r[?25h",,terminal_output +1278,3056447,"TERMINAL",0,0,"[?25l0[?25h",,terminal_output +1279,3057059,"TERMINAL",0,0,"[?25l,[?25h",,terminal_output +1280,3057236,"TERMINAL",0,0,"[?25l:[?25h",,terminal_output +1281,3057789,"TERMINAL",0,0,"[?25l,[?25h",,terminal_output +1282,3058071,"TERMINAL",0,0,"[?25l0[?25h",,terminal_output +1283,3058379,"TERMINAL",0,0,"[?25l][?25h",,terminal_output +1284,3058714,"TERMINAL",0,0,"\r\nArray([148, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0], dtype=int32)\r\n(jdb) ",,terminal_output +1285,3068448,"train_dynamics.py",0,0,"",python,tab +1286,3075364,"train_dynamics.py",12226,0,"",python,selection_mouse +1287,3075367,"train_dynamics.py",12225,0,"",python,selection_command +1288,3076381,"train_dynamics.py",12291,0,"",python,selection_mouse +1289,3076402,"train_dynamics.py",12290,0,"",python,selection_command +1290,3076945,"train_dynamics.py",12313,0,"",python,selection_mouse +1291,3076962,"train_dynamics.py",12312,0,"",python,selection_command +1292,3078692,"genie.py",0,0,"",python,tab +1293,3081168,"genie.py",5852,0,"",python,selection_mouse +1294,3081222,"genie.py",5851,0,"",python,selection_command +1295,3081936,"genie.py",5813,0,"",python,selection_mouse +1296,3082464,"genie.py",5817,0,"",python,selection_mouse +1297,3082505,"genie.py",5816,0,"",python,selection_command +1298,3088228,"genie.py",5817,0,"",python,selection_mouse +1299,3088231,"genie.py",5816,0,"",python,selection_command +1300,3088851,"genie.py",5852,0,"",python,selection_mouse +1301,3088859,"genie.py",5851,0,"",python,selection_command +1302,3095521,"genie.py",6102,0,"",python,selection_mouse +1303,3096067,"genie.py",6072,0,"",python,selection_mouse +1304,3096881,"genie.py",6071,0,"",python,selection_mouse +1305,3102898,"genie.py",6029,0,"",python,selection_mouse +1306,3103606,"genie.py",6015,0,"",python,selection_mouse +1307,3104109,"genie.py",6021,0,"",python,selection_mouse +1308,3105176,"genie.py",6101,0,"\n ",python,content +1309,3106892,"genie.py",6114,0,"t",python,content +1310,3106894,"genie.py",6115,0,"",python,selection_keyboard +1311,3106936,"genie.py",6115,0,"o",python,content +1312,3106937,"genie.py",6116,0,"",python,selection_keyboard +1313,3106963,"genie.py",6116,0,"k",python,content +1314,3106964,"genie.py",6117,0,"",python,selection_keyboard +1315,3108305,"genie.py",6114,3,"token_idxs_full",python,content +1316,3109579,"genie.py",6129,0,".",python,content +1317,3109580,"genie.py",6130,0,"",python,selection_keyboard +1318,3109857,"genie.py",6130,0,"a",python,content +1319,3109858,"genie.py",6131,0,"",python,selection_keyboard +1320,3109979,"genie.py",6131,0,"t",python,content +1321,3109980,"genie.py",6132,0,"",python,selection_keyboard +1322,3110535,"genie.py",6132,0,"[]",python,content +1323,3110536,"genie.py",6133,0,"",python,selection_keyboard +1324,3116969,"genie.py",6133,0,":",python,content +1325,3116970,"genie.py",6134,0,"",python,selection_keyboard +1326,3117198,"genie.py",6134,0,",",python,content +1327,3117199,"genie.py",6135,0,"",python,selection_keyboard +1328,3119486,"genie.py",6134,0,"",python,selection_command +1329,3120195,"genie.py",6135,0,"",python,selection_command +1330,3120429,"genie.py",6135,0," ",python,content +1331,3120430,"genie.py",6136,0,"",python,selection_keyboard +1332,3121027,"genie.py",6135,1,"",python,content +1333,3121439,"genie.py",6134,1,"",python,content +1334,3121832,"genie.py",6133,1,"",python,content +1335,3122766,"genie.py",6134,0,"",python,selection_command +1336,3125627,"genie.py",6133,0,"",python,selection_command +1337,3126198,"genie.py",6134,0,"",python,selection_command +1338,3127101,"genie.py",6133,0,"",python,selection_command +1339,3127354,"genie.py",6132,0,"",python,selection_command +1340,3128052,"genie.py",6133,0,"",python,selection_command +1341,3128662,"genie.py",6132,0,"",python,selection_command +1342,3129447,"genie.py",6133,0,"",python,selection_command +1343,3130035,"genie.py",6134,0,"",python,selection_command +1344,3136191,"genie.py",7005,0,"",python,selection_mouse +1345,3137610,"genie.py",7004,0,"",python,selection_mouse +1346,3144744,"genie.py",6650,0,"",python,selection_mouse +1347,3145635,"genie.py",6683,0,"",python,selection_mouse +1348,3146276,"genie.py",6650,0,"",python,selection_mouse +1349,3147894,"genie.py",6135,0,"",python,selection_mouse +1350,3148510,"genie.py",6130,5,"at[]\n",python,selection_mouse +1351,3148510,"genie.py",6129,6,".at[]\n",python,selection_mouse +1352,3148511,"genie.py",6114,21,"token_idxs_full.at[]\n",python,selection_mouse +1353,3148840,"genie.py",6127,0,"",python,selection_mouse +1354,3148841,"genie.py",6114,15,"token_idxs_full",python,selection_mouse +1355,3148999,"genie.py",6102,33," token_idxs_full.at[]\n",python,selection_mouse +1356,3149373,"genie.py",6102,33,"",python,content +1357,3150689,"genie.py",6618,0,"",python,selection_mouse +1358,3151728,"genie.py",6026,0,"",python,selection_mouse +1359,3153092,"genie.py",6102,0,"",python,selection_mouse +1360,3154305,"genie.py",6072,0,"",python,selection_mouse +1361,3155804,"genie.py",6071,1,"",python,content +1362,3155913,"genie.py",6070,1,"",python,content +1363,3155977,"genie.py",6070,0,"t",python,content +1364,3155977,"genie.py",6071,0,"",python,selection_keyboard +1365,3167870,"genie.py",6281,0,"",python,selection_mouse +1366,3168574,"genie.py",6209,0,"",python,selection_mouse +1367,3168756,"genie.py",6200,10,"next_token",python,selection_mouse +1368,3169941,"genie.py",6342,0,"",python,selection_mouse +1369,3170076,"genie.py",6342,3,"rng",python,selection_mouse +1370,3170282,"genie.py",6342,5,"rng, ",python,selection_mouse +1371,3170283,"genie.py",6342,13,"rng, step_rng",python,selection_mouse +1372,3170336,"genie.py",6342,15,"rng, step_rng =",python,selection_mouse +1373,3170359,"genie.py",6342,19,"rng, step_rng = jax",python,selection_mouse +1374,3170425,"genie.py",6342,26,"rng, step_rng = jax.random",python,selection_mouse +1375,3170536,"genie.py",6342,27,"rng, step_rng = jax.random.",python,selection_mouse +1376,3170537,"genie.py",6342,32,"rng, step_rng = jax.random.split",python,selection_mouse +1377,3170610,"genie.py",6342,33,"rng, step_rng = jax.random.split(",python,selection_mouse +1378,3170656,"genie.py",6342,36,"rng, step_rng = jax.random.split(rng",python,selection_mouse +1379,3170725,"genie.py",6342,37,"rng, step_rng = jax.random.split(rng)",python,selection_mouse +1380,3170992,"genie.py",6379,0,"",python,selection_mouse +1381,3171557,"genie.py",6378,1,")",python,selection_mouse +1382,3171557,"genie.py",6375,4,"rng)",python,selection_mouse +1383,3171558,"genie.py",6369,10,"split(rng)",python,selection_mouse +1384,3171610,"genie.py",6368,11,".split(rng)",python,selection_mouse +1385,3171610,"genie.py",6362,17,"random.split(rng)",python,selection_mouse +1386,3171651,"genie.py",6358,21,"jax.random.split(rng)",python,selection_mouse +1387,3171675,"genie.py",6357,22," jax.random.split(rng)",python,selection_mouse +1388,3171703,"genie.py",6355,24," = jax.random.split(rng)",python,selection_mouse +1389,3171727,"genie.py",6347,32,"step_rng = jax.random.split(rng)",python,selection_mouse +1390,3171845,"genie.py",6346,33," step_rng = jax.random.split(rng)",python,selection_mouse +1391,3171845,"genie.py",6345,34,", step_rng = jax.random.split(rng)",python,selection_mouse +1392,3171917,"genie.py",6342,37,"rng, step_rng = jax.random.split(rng)",python,selection_mouse +1393,3172252,"genie.py",6344,0,"",python,selection_mouse +1394,3172575,"genie.py",6428,0,"",python,selection_mouse +1395,3173516,"genie.py",6427,0,"",python,selection_command +1396,3174297,"genie.py",6125,0,"",python,selection_mouse +1397,3174829,"genie.py",5949,0,"",python,selection_mouse +1398,3175482,"genie.py",6032,0,"",python,selection_mouse +1399,3176613,"genie.py",6004,96," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)",python,selection_command +1400,3176806,"genie.py",6004,97," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n",python,selection_command +1401,3176952,"genie.py",6004,145," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch",python,selection_command +1402,3177452,"genie.py",6004,177," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:",python,selection_command +1403,3177549,"genie.py",6004,257," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)",python,selection_command +1404,3177549,"genie.py",6004,277," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:",python,selection_command +1405,3177613,"genie.py",6004,315," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:",python,selection_command +1406,3177702,"genie.py",6004,375," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)",python,selection_command +1407,3177702,"genie.py",6004,399," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:",python,selection_command +1408,3177770,"genie.py",6004,454," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)",python,selection_command +1409,3177771,"genie.py",6004,509," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)\n # next_token = jax.random.categorical(",python,selection_command +1410,3177771,"genie.py",6004,582," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)\n # next_token = jax.random.categorical(\n # step_rng, next_token_logits / temperature, axis=-1",python,selection_command +1411,3177814,"genie.py",6004,612," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)\n # next_token = jax.random.categorical(\n # step_rng, next_token_logits / temperature, axis=-1\n # ) # (B, N)",python,selection_command +1412,3177814,"genie.py",6004,613," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)\n # next_token = jax.random.categorical(\n # step_rng, next_token_logits / temperature, axis=-1\n # ) # (B, N)\n",python,selection_command +1413,3177868,"genie.py",6004,675," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)\n # next_token = jax.random.categorical(\n # step_rng, next_token_logits / temperature, axis=-1\n # ) # (B, N)\n\n # # Insert the generated tokens into the sequence",python,selection_command +1414,3178038,"genie.py",6004,751," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)\n # next_token = jax.random.categorical(\n # step_rng, next_token_logits / temperature, axis=-1\n # ) # (B, N)\n\n # # Insert the generated tokens into the sequence\n # token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)",python,selection_command +1415,3178290,"genie.py",6004,752," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)\n # next_token = jax.random.categorical(\n # step_rng, next_token_logits / temperature, axis=-1\n # ) # (B, N)\n\n # # Insert the generated tokens into the sequence\n # token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n",python,selection_command +1416,3178435,"genie.py",6004,809," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)\n # next_token = jax.random.categorical(\n # step_rng, next_token_logits / temperature, axis=-1\n # ) # (B, N)\n\n # # Insert the generated tokens into the sequence\n # token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n\n # # --- Decode all tokens at once at the end ---",python,selection_command +1417,3178678,"genie.py",6004,857," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)\n # next_token = jax.random.categorical(\n # step_rng, next_token_logits / temperature, axis=-1\n # ) # (B, N)\n\n # # Insert the generated tokens into the sequence\n # token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n\n # # --- Decode all tokens at once at the end ---\n # final_frames = self.tokenizer.decode(",python,selection_command +1418,3178757,"genie.py",6004,924," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)\n # next_token = jax.random.categorical(\n # step_rng, next_token_logits / temperature, axis=-1\n # ) # (B, N)\n\n # # Insert the generated tokens into the sequence\n # token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n\n # # --- Decode all tokens at once at the end ---\n # final_frames = self.tokenizer.decode(\n # token_idxs_full, video_hw=batch[""videos""].shape[2:4]",python,selection_command +1419,3178880,"genie.py",6004,936," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)\n # next_token = jax.random.categorical(\n # step_rng, next_token_logits / temperature, axis=-1\n # ) # (B, N)\n\n # # Insert the generated tokens into the sequence\n # token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n\n # # --- Decode all tokens at once at the end ---\n # final_frames = self.tokenizer.decode(\n # token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n # )",python,selection_command +1420,3179047,"genie.py",6004,966," # next_token_logits = dyna_outputs[""token_logits""][:, t, :, :] # (B, N, vocab_size)\n\n # # Sample or argmax for each patch\n # if sample_argmax:\n # next_token = jnp.argmax(next_token_logits, axis=-1) # (B, N)\n # else:\n # if rng is not None:\n # rng, step_rng = jax.random.split(rng)\n # else:\n # step_rng = jax.random.PRNGKey(0)\n # next_token = jax.random.categorical(\n # step_rng, next_token_logits / temperature, axis=-1\n # ) # (B, N)\n\n # # Insert the generated tokens into the sequence\n # token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\n\n # # --- Decode all tokens at once at the end ---\n # final_frames = self.tokenizer.decode(\n # token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n # )\n # return final_frames",python,selection_command +1421,3179369,"genie.py",6016,0,"",python,selection_command +1422,3179852,"genie.py",6949,1,"",python,content +1423,3179853,"genie.py",6937,1,"",python,content +1424,3179853,"genie.py",6874,1,"",python,content +1425,3179853,"genie.py",6822,1,"",python,content +1426,3179853,"genie.py",6765,1,"",python,content +1427,3179853,"genie.py",6692,1,"",python,content +1428,3179853,"genie.py",6630,1,"",python,content +1429,3179853,"genie.py",6603,1,"",python,content +1430,3179853,"genie.py",6534,1,"",python,content +1431,3179853,"genie.py",6475,1,"",python,content +1432,3179853,"genie.py",6424,1,"",python,content +1433,3179853,"genie.py",6396,1,"",python,content +1434,3179853,"genie.py",6340,1,"",python,content +1435,3179854,"genie.py",6298,1,"",python,content +1436,3179854,"genie.py",6274,1,"",python,content +1437,3179854,"genie.py",6198,1,"",python,content +1438,3179854,"genie.py",6162,1,"",python,content +1439,3179854,"genie.py",6114,1,"",python,content +1440,3179854,"genie.py",6016,1,"",python,content +1441,3179955,"genie.py",6931,1,"",python,content +1442,3179955,"genie.py",6920,1,"",python,content +1443,3179955,"genie.py",6858,1,"",python,content +1444,3179955,"genie.py",6807,1,"",python,content +1445,3179956,"genie.py",6751,1,"",python,content +1446,3179956,"genie.py",6679,1,"",python,content +1447,3179956,"genie.py",6618,1,"",python,content +1448,3179956,"genie.py",6592,1,"",python,content +1449,3179956,"genie.py",6524,1,"",python,content +1450,3179956,"genie.py",6466,1,"",python,content +1451,3179956,"genie.py",6416,1,"",python,content +1452,3179956,"genie.py",6389,1,"",python,content +1453,3179956,"genie.py",6334,1,"",python,content +1454,3179956,"genie.py",6293,1,"",python,content +1455,3179956,"genie.py",6270,1,"",python,content +1456,3179956,"genie.py",6195,1,"",python,content +1457,3179956,"genie.py",6160,1,"",python,content +1458,3179956,"genie.py",6113,1,"",python,content +1459,3179956,"genie.py",6016,1,"",python,content +1460,3180119,"genie.py",6015,0,"",python,selection_command +1461,3180475,"genie.py",5932,0,"",python,selection_command +1462,3180863,"genie.py",6015,0,"",python,selection_command +1463,3181082,"genie.py",5932,0,"",python,selection_command +1464,3181347,"genie.py",5864,0,"",python,selection_command +1465,3184439,"genie.py",5932,0,"",python,selection_command +1466,3184651,"genie.py",6015,0,"",python,selection_command +1467,3184860,"genie.py",6099,0,"",python,selection_command +1468,3185044,"genie.py",6111,0,"",python,selection_command +1469,3185217,"genie.py",6157,0,"",python,selection_command +1470,3186121,"genie.py",6187,0,"",python,selection_command +1471,3186397,"genie.py",6265,0,"",python,selection_command +1472,3186566,"genie.py",6283,0,"",python,selection_command +1473,3186718,"genie.py",6319,0,"",python,selection_command +1474,3187348,"genie.py",6283,0,"",python,selection_command +1475,3190726,"genie.py",6031,0,"",python,selection_mouse +1476,3192040,"genie.py",6293,0,"",python,selection_mouse +1477,3201927,"genie.py",9133,0,"",python,selection_mouse +1478,3202068,"genie.py",9132,3,"rng",python,selection_mouse +1479,3213706,"genie.py",9586,0,"",python,selection_mouse +1480,3213815,"genie.py",9580,8,"step_rng",python,selection_mouse +1481,3215964,"genie.py",9165,0,"",python,selection_mouse +1482,3216096,"genie.py",9165,3,"rng",python,selection_mouse +1483,3223794,"genie.py",10008,0,"",python,selection_mouse +1484,3223901,"genie.py",10006,5,"batch",python,selection_mouse +1485,3224154,"genie.py",10006,7,"batch[""",python,selection_mouse +1486,3224155,"genie.py",10006,10,"batch[""rng",python,selection_mouse +1487,3224279,"genie.py",10006,11,"batch[""rng""",python,selection_mouse +1488,3224433,"genie.py",10006,12,"batch[""rng""]",python,selection_mouse +1489,3224848,"genie.py",10018,0,"",python,selection_mouse +1490,3227541,"train_dynamics.py",0,0,"",python,tab +1491,3228471,"genie.py",0,0,"",python,tab +1492,3231441,"genie.py",5386,0,"",python,selection_mouse +1493,3231562,"genie.py",5384,3,"rng",python,selection_mouse +1494,3233324,"genie.py",5397,0,"",python,selection_mouse +1495,3234146,"genie.py",5396,0,"",python,selection_command +1496,3234233,"genie.py",5395,0,"",python,selection_command +1497,3234489,"genie.py",5395,1,"",python,content +1498,3235409,"genie.py",5395,1,"",python,content +1499,3235595,"genie.py",5395,1,"",python,content +1500,3235724,"genie.py",5395,1,"",python,content +1501,3235856,"genie.py",5395,1,"",python,content +1502,3236667,"genie.py",5395,0,"[]",python,content +1503,3236668,"genie.py",5396,0,"",python,selection_keyboard +1504,3238283,"genie.py",5408,1,"",python,content +1505,3238579,"genie.py",5404,4,"",python,content +1506,3238954,"genie.py",5403,1,"",python,content +1507,3239082,"genie.py",5402,1,"",python,content +1508,3240120,"genie.py",5402,0,"]",python,content +1509,3240121,"genie.py",5403,0,"",python,selection_keyboard +1510,3241658,"genie.py",5396,0,"",python,selection_command +1511,3241879,"genie.py",5397,0,"",python,selection_command +1512,3242106,"genie.py",5396,1,"",python,content +1513,3242326,"genie.py",5395,0,"",python,selection_command +1514,3242872,"genie.py",5422,0,"",python,selection_command +1515,3243046,"genie.py",5458,0,"",python,selection_command +1516,3243185,"genie.py",5486,0,"",python,selection_command +1517,3243336,"genie.py",5535,0,"",python,selection_command +1518,3243443,"genie.py",5580,0,"",python,selection_command +1519,3243689,"genie.py",5601,0,"",python,selection_command +1520,3243825,"genie.py",5681,0,"",python,selection_command +1521,3243990,"genie.py",5757,0,"",python,selection_command +1522,3244203,"genie.py",5827,0,"",python,selection_command +1523,3244354,"genie.py",5862,0,"",python,selection_command +1524,3244494,"genie.py",5930,0,"",python,selection_command +1525,3244665,"genie.py",6013,0,"",python,selection_command +1526,3244885,"genie.py",6089,0,"",python,selection_command +1527,3245059,"genie.py",6109,0,"",python,selection_command +1528,3245235,"genie.py",6155,0,"",python,selection_command +1529,3245416,"genie.py",6185,0,"",python,selection_command +1530,3245625,"genie.py",6260,0,"",python,selection_command +1531,3245834,"genie.py",6281,0,"",python,selection_command +1532,3245995,"genie.py",6317,0,"",python,selection_command +1533,3246151,"genie.py",6375,0,"",python,selection_command +1534,3246344,"genie.py",6397,0,"",python,selection_command +1535,3246640,"genie.py",6375,0,"",python,selection_command +1536,3246797,"genie.py",6317,0,"",python,selection_command +1537,3246957,"genie.py",6281,0,"",python,selection_command +1538,3248387,"genie.py",6317,0,"",python,selection_command +1539,3248560,"genie.py",6375,0,"",python,selection_command +1540,3249390,"genie.py",6356,75,"",python,content +1541,3249429,"genie.py",6372,0,"",python,selection_command +1542,3249563,"genie.py",6317,0,"",python,selection_command +1543,3249761,"genie.py",6281,0,"",python,selection_command +1544,3250113,"genie.py",6262,36,"",python,content +1545,3250127,"genie.py",6282,0,"",python,selection_command +1546,3250618,"genie.py",6278,4,"",python,content +1547,3250779,"genie.py",6277,0,"",python,selection_command +1548,3259332,"genie.py",6331,0,"",python,selection_command +1549,3259688,"genie.py",6384,0,"",python,selection_command +1550,3259829,"genie.py",6455,0,"",python,selection_command +1551,3259988,"genie.py",6468,0,"",python,selection_command +1552,3260231,"genie.py",6455,0,"",python,selection_command +1553,3260379,"genie.py",6384,0,"",python,selection_command +1554,3260548,"genie.py",6331,0,"",python,selection_command +1555,3260863,"genie.py",6384,0,"",python,selection_command +1556,3261040,"genie.py",6455,0,"",python,selection_command +1557,3261219,"genie.py",6468,0,"",python,selection_command +1558,3261390,"genie.py",6455,0,"",python,selection_command +1559,3261575,"genie.py",6384,0,"",python,selection_command +1560,3261722,"genie.py",6331,0,"",python,selection_command +1561,3262014,"genie.py",6384,0,"",python,selection_command +1562,3262241,"genie.py",6455,0,"",python,selection_command +1563,3262458,"genie.py",6384,0,"",python,selection_command +1564,3262647,"genie.py",6455,0,"",python,selection_command +1565,3262805,"genie.py",6384,0,"",python,selection_command +1566,3262902,"genie.py",6455,0,"",python,selection_command +1567,3263040,"genie.py",6384,0,"",python,selection_command +1568,3263071,"genie.py",6455,0,"",python,selection_command +1569,3263206,"genie.py",6384,0,"",python,selection_command +1570,3263238,"genie.py",6455,0,"",python,selection_command +1571,3263409,"genie.py",6384,0,"",python,selection_command +1572,3263608,"genie.py",6331,0,"",python,selection_command +1573,3263842,"genie.py",6277,0,"",python,selection_command +1574,3264013,"genie.py",6331,0,"",python,selection_command +1575,3264215,"genie.py",6384,0,"",python,selection_command +1576,3264394,"genie.py",6455,0,"",python,selection_command +1577,3264530,"genie.py",6468,0,"",python,selection_command +1578,3264688,"genie.py",6484,0,"",python,selection_command +1579,3264862,"genie.py",6544,0,"",python,selection_command +1580,3264988,"genie.py",6603,0,"",python,selection_command +1581,3265159,"genie.py",6619,0,"",python,selection_command +1582,3265678,"genie.py",6603,0,"",python,selection_command +1583,3266046,"genie.py",6544,0,"",python,selection_command +1584,3266355,"genie.py",6484,0,"",python,selection_command +1585,3266890,"genie.py",6679,0,"",python,selection_mouse +1586,3268159,"genie.py",6674,0,"",python,selection_mouse +1587,3268315,"genie.py",6667,12,"final_frames",python,selection_mouse +1588,3268882,"genie.py",6720,0,"",python,selection_mouse +1589,3269016,"genie.py",6717,15,"token_idxs_full",python,selection_mouse +1590,3269542,"genie.py",6779,0,"",python,selection_mouse +1591,3269592,"genie.py",6778,0,"",python,selection_command +1592,3269672,"genie.py",6779,0,"",python,selection_mouse +1593,3269713,"genie.py",6778,0,"",python,selection_command +1594,3269951,"genie.py",6778,1,")",python,selection_mouse +1595,3269952,"genie.py",6713,65," token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n ",python,selection_mouse +1596,3269952,"genie.py",6711,67," token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n ",python,selection_mouse +1597,3269952,"genie.py",6664,114," final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n ",python,selection_mouse +1598,3269953,"genie.py",6663,115," final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n ",python,selection_mouse +1599,3269982,"genie.py",6779,0,"",python,selection_command +1600,3269982,"genie.py",6662,117," final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )",python,selection_mouse +1601,3270028,"genie.py",6607,172," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )",python,selection_mouse +1602,3270289,"genie.py",6606,173," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )",python,selection_mouse +1603,3270439,"genie.py",6605,174," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )",python,selection_mouse +1604,3270861,"genie.py",6605,0,"",python,selection_mouse +1605,3270862,"genie.py",6604,8," ",python,selection_mouse +1606,3271010,"genie.py",6604,55," # --- Decode all tokens at once at the end ---\n",python,selection_mouse +1607,3271206,"genie.py",6604,101," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n",python,selection_mouse +1608,3271261,"genie.py",6604,166," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n",python,selection_mouse +1609,3271379,"genie.py",6604,176," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )\n",python,selection_mouse +1610,3271512,"genie.py",6604,204," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )\n return final_frames\n",python,selection_mouse +1611,3272019,"genie.py",6807,0,"",python,selection_mouse +1612,3272047,"genie.py",6806,0,"",python,selection_command +1613,3272159,"genie.py",6795,12,"final_frames",python,selection_mouse +1614,3272203,"genie.py",6796,11,"inal_frames",python,selection_command +1615,3272361,"genie.py",6779,17,"\n return f",python,selection_mouse +1616,3272444,"genie.py",6717,79,"token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )\n return f",python,selection_mouse +1617,3272526,"genie.py",6667,129,"final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )\n return f",python,selection_mouse +1618,3272632,"genie.py",6610,186," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )\n return f",python,selection_mouse +1619,3272633,"genie.py",6609,187," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )\n return f",python,selection_mouse +1620,3272673,"genie.py",6608,188," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )\n return f",python,selection_mouse +1621,3272686,"genie.py",6607,189," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )\n return f",python,selection_mouse +1622,3272732,"genie.py",6606,190," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )\n return f",python,selection_mouse +1623,3272762,"genie.py",6605,191," # --- Decode all tokens at once at the end ---\n final_frames = self.tokenizer.decode(\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\n )\n return f",python,selection_mouse +1624,3273132,"genie.py",6605,0,"",python,selection_mouse +1625,3275273,"genie.py",5835,0,"",python,selection_mouse +1626,3275808,"genie.py",5808,35,"",python,content +1627,3275877,"genie.py",5820,0,"",python,selection_command +1628,3276834,"genie.py",5888,0,"",python,selection_command +1629,3277340,"genie.py",5971,0,"",python,selection_command +1630,3277369,"genie.py",6054,0,"",python,selection_command +1631,3277412,"genie.py",6067,0,"",python,selection_command +1632,3277421,"genie.py",6113,0,"",python,selection_command +1633,3277467,"genie.py",6143,0,"",python,selection_command +1634,3277490,"genie.py",6221,0,"",python,selection_command +1635,3277523,"genie.py",6239,0,"",python,selection_command +1636,3277559,"genie.py",6293,0,"",python,selection_command +1637,3277622,"genie.py",6346,0,"",python,selection_command +1638,3277661,"genie.py",6417,0,"",python,selection_command +1639,3277662,"genie.py",6433,0,"",python,selection_command +1640,3277813,"genie.py",6446,0,"",python,selection_command +1641,3278182,"genie.py",6506,0,"",python,selection_command +1642,3278354,"genie.py",6567,0,"\n jax.debug.breakpoint()",python,content +1643,3278398,"genie.py",6580,0,"",python,selection_command +1644,3281630,"TERMINAL",0,0,"^DERROR:2025-07-17 10:38:44,341:jax._src.debugging:98: jax.debug.callback failed\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 96, in debug_callback_impl\r\n callback(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 336, in _flat_callback\r\n callback(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/core.py"", line 220, in _breakpoint_callback\r\n debugger(frames, thread_id, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 167, in run_debugger\r\n CliDebugger(frames, thread_id, **kwargs).run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 160, in run\r\n self.cmdloop()\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 138, in cmdloop\r\n stop = self.onecmd(line)\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 217, in onecmd\r\n return func(arg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 146, in do_quit\r\n sys.exit(0)\r\nSystemExit: 0\r\nERROR:jax._src.debugging:jax.debug.callback failed\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 96, in debug_callback_impl\r\n callback(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 336, in _flat_callback\r\n callback(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/core.py"", line 220, in _breakpoint_callback\r\n debugger(frames, thread_id, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 167, in run_debugger\r\n CliDebugger(frames, thread_id, **kwargs).run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 160, in run\r\n self.cmdloop()\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 138, in cmdloop\r\n stop = self.onecmd(line)\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 217, in onecmd\r\n return func(arg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 146, in do_quit\r\n sys.exit(0)\r\nSystemExit: 0\r\nE0717 10:38:44.346692 940210 pjrt_stream_executor_client.cc:2916] Execution of replica 0 failed: INTERNAL: CpuCallback error calling callback: Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 178, in \r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 141, in _autoreg_sample\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/traceback_util.py"", line 182, in reraise_with_filtered_traceback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 292, in cache_miss\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 153, in _python_pjit_helper\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 1877, in _pjit_call_impl_python\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/profiler.py"", line 354, in wrapper\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/interpreters/pxla.py"", line 1297, in __call__\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/callback.py"", line 782, in _wrapped_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 202, in _callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 99, in debug_callback_impl\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 336, in _flat_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/core.py"", line 220, in _breakpoint_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 167, in run_debugger\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 162, in run\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 145, in cmdloop\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 217, in onecmd\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 146, in do_quit\r\nSystemExit: 0\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 178, in \r\n vid = _autoreg_sample(rng, video_batch, action_batch)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 141, in _autoreg_sample\r\n generated_vid = sampling_fn(\r\njaxlib._jax.XlaRuntimeError: INTERNAL: CpuCallback error calling callback: Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 178, in \r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 141, in _autoreg_sample\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/traceback_util.py"", line 182, in reraise_with_filtered_traceback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 292, in cache_miss\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 153, in _python_pjit_helper\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 1877, in _pjit_call_impl_python\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/profiler.py"", line 354, in wrapper\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/interpreters/pxla.py"", line 1297, in __call__\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/callback.py"", line 782, in _wrapped_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 202, in _callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 99, in debug_callback_impl\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 336, in _flat_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/core.py"", line 220, in _breakpoint_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 167, in run_debugger\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 162, in run\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 145, in cmdloop\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 217, in onecmd\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 146, in do_quit\r\nSystemExit: 0\r\n",,terminal_output +1645,3283376,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +1646,3283989,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +1647,3284600,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +1648,3285768,"train_dynamics.py",0,0,"",python,tab +1649,3286300,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +1650,3287578,"genie.py",0,0,"",python,tab +1651,3288441,"TERMINAL",0,0,"2025-07-17 10:38:51.221949: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1652,3288903,"genie.py",6203,0,"",python,selection_mouse +1653,3289511,"genie.py",6054,0,"",python,selection_mouse +1654,3291016,"genie.py",6550,0,"",python,selection_mouse +1655,3291788,"genie.py",5854,0,"",python,selection_mouse +1656,3293323,"genie.py",5624,0,"",python,selection_mouse +1657,3293819,"genie.py",5705,0,"",python,selection_mouse +1658,3294769,"genie.py",5737,0,"\n ",python,content +1659,3295968,"genie.py",5738,12,"",python,content +1660,3297572,"genie.py",5738,0," ",python,content +1661,3297775,"genie.py",5742,0," ",python,content +1662,3298007,"genie.py",5746,0," ",python,content +1663,3298323,"genie.py",5750,0,"p",python,content +1664,3298323,"genie.py",5751,0,"",python,selection_keyboard +1665,3298494,"genie.py",5751,0,"r",python,content +1666,3298495,"genie.py",5752,0,"",python,selection_keyboard +1667,3298598,"genie.py",5752,0,"i",python,content +1668,3298599,"genie.py",5753,0,"",python,selection_keyboard +1669,3298647,"genie.py",5753,0,"n",python,content +1670,3298648,"genie.py",5754,0,"",python,selection_keyboard +1671,3299046,"genie.py",5753,1,"",python,content +1672,3299182,"genie.py",5752,1,"",python,content +1673,3299322,"genie.py",5751,1,"",python,content +1674,3299448,"genie.py",5750,1,"",python,content +1675,3299833,"genie.py",5746,4,"",python,content +1676,3299943,"genie.py",5742,4,"",python,content +1677,3300309,"genie.py",5738,4,"",python,content +1678,3300717,"genie.py",5662,0,"",python,selection_command +1679,3300919,"genie.py",5582,0,"",python,selection_command +1680,3301304,"genie.py",5582,80,"",python,content +1681,3301394,"genie.py",5594,0,"",python,selection_command +1682,3301415,"genie.py",5658,0,"",python,selection_command +1683,3301774,"TERMINAL",0,0,"2025-07-17 10:39:04.546273: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1684,3302009,"genie.py",5594,0,"",python,selection_command +1685,3302563,"genie.py",5611,0,"token_idxs_full shape: {}"", token_idxs_full.shape)\n jax.debug.print(""",python,content +1686,3302586,"genie.py",5738,0,"",python,selection_command +1687,3303173,"genie.py",5674,0,"",python,selection_command +1688,3303340,"genie.py",5594,0,"",python,selection_command +1689,3303782,"genie.py",5674,0,"",python,selection_command +1690,3304049,"genie.py",5737,0,"\n jax.debug.print(""token_idxs_full shape: {}"", token_idxs_full.shape)",python,content +1691,3304080,"genie.py",5750,0,"",python,selection_command +1692,3304789,"genie.py",5818,0,"",python,selection_command +1693,3305283,"genie.py",5818,1,"",python,content +1694,3305292,"genie.py",5830,0,"",python,selection_command +1695,3305359,"genie.py",5750,0,"",python,selection_command +1696,3305611,"genie.py",5751,0,"",python,selection_command +1697,3306133,"genie.py",5752,0,"",python,selection_command +1698,3306139,"genie.py",5753,0,"",python,selection_command +1699,3306173,"genie.py",5754,0,"",python,selection_command +1700,3306213,"genie.py",5755,0,"",python,selection_command +1701,3306241,"genie.py",5756,0,"",python,selection_command +1702,3306283,"genie.py",5757,0,"",python,selection_command +1703,3306291,"genie.py",5758,0,"",python,selection_command +1704,3306354,"genie.py",5759,0,"",python,selection_command +1705,3306358,"genie.py",5760,0,"",python,selection_command +1706,3306431,"genie.py",5761,0,"",python,selection_command +1707,3306435,"genie.py",5762,0,"",python,selection_command +1708,3306442,"genie.py",5763,0,"",python,selection_command +1709,3306509,"genie.py",5764,0,"",python,selection_command +1710,3306520,"genie.py",5765,0,"",python,selection_command +1711,3306529,"genie.py",5766,0,"",python,selection_command +1712,3306573,"genie.py",5767,0,"",python,selection_command +1713,3306609,"genie.py",5768,0,"",python,selection_command +1714,3306625,"genie.py",5769,0,"",python,selection_command +1715,3306664,"genie.py",5770,0,"",python,selection_command +1716,3306676,"genie.py",5771,0,"",python,selection_command +1717,3306742,"genie.py",5772,0,"",python,selection_command +1718,3306751,"genie.py",5773,0,"",python,selection_command +1719,3306795,"genie.py",5774,0,"",python,selection_command +1720,3306813,"genie.py",5775,0,"",python,selection_command +1721,3306844,"genie.py",5776,0,"",python,selection_command +1722,3306893,"genie.py",5777,0,"",python,selection_command +1723,3306916,"genie.py",5778,0,"",python,selection_command +1724,3306941,"genie.py",5779,0,"",python,selection_command +1725,3306986,"genie.py",5780,0,"",python,selection_command +1726,3307471,"genie.py",5779,0,"",python,selection_command +1727,3307687,"genie.py",5778,0,"",python,selection_command +1728,3307838,"genie.py",5779,0,"",python,selection_command +1729,3308009,"genie.py",5780,0,"",python,selection_command +1730,3308155,"genie.py",5781,0,"",python,selection_command +1731,3308286,"genie.py",5782,0,"",python,selection_command +1732,3308434,"genie.py",5783,0,"",python,selection_command +1733,3308779,"genie.py",5783,5,"",python,content +1734,3310014,"genie.py",5783,0,"a",python,content +1735,3310015,"genie.py",5784,0,"",python,selection_keyboard +1736,3310254,"genie.py",5784,0,"t",python,content +1737,3310255,"genie.py",5785,0,"",python,selection_keyboard +1738,3310345,"genie.py",5785,0," ",python,content +1739,3310345,"genie.py",5786,0,"",python,selection_keyboard +1740,3311556,"genie.py",5785,1,"",python,content +1741,3311710,"genie.py",5784,1,"",python,content +1742,3311815,"genie.py",5783,1,"",python,content +1743,3314556,"genie.py",5783,0,"0",python,content +1744,3314557,"genie.py",5784,0,"",python,selection_keyboard +1745,3316575,"genie.py",5807,5,"",python,content +1746,3317005,"genie.py",5806,1,"",python,content +1747,3317133,"TERMINAL",0,0,"2025-07-17 10:39:19.838341: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1748,3317840,"genie.py",5806,0,"[]",python,content +1749,3317841,"genie.py",5807,0,"",python,selection_keyboard +1750,3318932,"genie.py",5807,0,":",python,content +1751,3318933,"genie.py",5808,0,"",python,selection_keyboard +1752,3319166,"genie.py",5808,0,",",python,content +1753,3319166,"genie.py",5809,0,"",python,selection_keyboard +1754,3320152,"genie.py",5808,1,"",python,content +1755,3320306,"genie.py",5807,1,"",python,content +1756,3320650,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 34000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/034000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 36000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/036000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 35000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/035000/metrics/metrics not found.\r\n",,terminal_output +1757,3320846,"genie.py",5807,0,"0",python,content +1758,3320847,"genie.py",5808,0,"",python,selection_keyboard +1759,3321273,"genie.py",5808,0,",",python,content +1760,3321273,"genie.py",5809,0,"",python,selection_keyboard +1761,3321531,"genie.py",5809,0,":",python,content +1762,3321532,"genie.py",5810,0,"",python,selection_keyboard +1763,3322151,"genie.py",5810,0,",",python,content +1764,3322152,"genie.py",5811,0,"",python,selection_keyboard +1765,3322333,"genie.py",5811,0,"0",python,content +1766,3322334,"genie.py",5812,0,"",python,selection_keyboard +1767,3323505,"genie.py",5913,0,"",python,selection_mouse +1768,3389756,"TERMINAL",0,0,"2025-07-17 10:40:32.537655: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 10:40:32.538168: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 10:40:32.538843: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1769,3467570,"TERMINAL",0,0,"token_idxs shape: (Array(1, dtype=int32), Array(1, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\n",,terminal_output +1770,3467934,"TERMINAL",0,0,"Entering jdb:\r\n(jdb) ",,terminal_output +1771,3483301,"TERMINAL",0,0,"l",,terminal_output +1772,3483507,"TERMINAL",0,0,"\r\n> /hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/genie.py(183)\r\n rng, step_rng = jax.random.split(rng)\r\n next_token = jax.random.categorical(\r\n step_rng, next_token_logits / temperature, axis=-1\r\n ) # (B, N)\r\n \r\n # Insert the generated tokens into the sequence\r\n-> token_idxs_full = token_idxs_full.at[:, t, :].set(next_token)\r\n jax.debug.breakpoint()\r\n \r\n # --- Decode all tokens at once at the end ---\r\n final_frames = self.tokenizer.decode(\r\n token_idxs_full, video_hw=batch[""videos""].shape[2:4]\r\n(jdb) ",,terminal_output +1773,3488520,"TERMINAL",0,0,"token_idxs_full",,terminal_output +1774,3489654,"TERMINAL",0,0,"[?25l[[?25h",,terminal_output +1775,3491811,"TERMINAL",0,0,"[?25l0[?25h",,terminal_output +1776,3492131,"TERMINAL",0,0,"[?25l,[?25h",,terminal_output +1777,3493027,"TERMINAL",0,0,"[?25l:[?25h",,terminal_output +1778,3493276,"TERMINAL",0,0,"[?25l,[?25h",,terminal_output +1779,3493554,"TERMINAL",0,0,"[?25l0[?25h",,terminal_output +1780,3494164,"TERMINAL",0,0,"[?25l][?25h",,terminal_output +1781,3494499,"TERMINAL",0,0,"\r\nArray([148, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0], dtype=int32)\r\n(jdb) ",,terminal_output +1782,3498324,"TERMINAL",0,0,"n",,terminal_output +1783,3498445,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1784,3498663,"TERMINAL",0,0,"[?25lxt[?25h",,terminal_output +1785,3499105,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +1786,3499399,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +1787,3499583,"TERMINAL",0,0,"[?25lo[?25h[?25lk[?25h",,terminal_output +1788,3499724,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1789,3499777,"TERMINAL",0,0,"[?25ln[?25h",,terminal_output +1790,3500347,"TERMINAL",0,0,"\r\nArray([[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 545, 255, 255, 255, 309, 0, 255, 255, 255, 545, 18,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 375, 228, 0, 948, 26,\r\n 26, 439, 488, 948, 26, 261, 439, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 273, 853, 164, 338, 1, 1, 968, 610, 338,\r\n 68, 1, 1018, 0, 0, 0, 0, 0, 0, 0, 0,\r\n 0, 0, 0, 0, 0, 0, 0]], dtype=int32)\r\n(jdb) ",,terminal_output +1791,3506492,"TERMINAL",0,0,"\rnext_token",,terminal_output +1792,3507414,"TERMINAL",0,0,"\r_idxs_full[0,:,0]",,terminal_output +1793,3511201,"TERMINAL",0,0,"\r[5@next_",,terminal_output +1794,3511392,"TERMINAL",0,0,"\r",,terminal_output +1795,3511977,"TERMINAL",0,0,"\rnext_token",,terminal_output +1796,3512831,"TERMINAL",0,0,".",,terminal_output +1797,3513105,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1798,3513194,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +1799,3513299,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +1800,3513412,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +1801,3513485,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1802,3513693,"TERMINAL",0,0,"\r\n(1, 920)\r\n(jdb) ",,terminal_output +1803,3516274,"TERMINAL",0,0,"\rnext_token.shape",,terminal_output +1804,3516503,"TERMINAL",0,0,"\r",,terminal_output +1805,3517100,"TERMINAL",0,0,"\r_idxs_full[0,:,0]",,terminal_output +1806,3517672,"TERMINAL",0,0,"[?25l]\r[?25h",,terminal_output +1807,3517864,"TERMINAL",0,0,"[?25l0\r[?25h",,terminal_output +1808,3518017,"TERMINAL",0,0,"[?25l:,\r[?25h",,terminal_output +1809,3518155,"TERMINAL",0,0,"[?25l:\r[?25h",,terminal_output +1810,3518254,"TERMINAL",0,0,"\r",,terminal_output +1811,3518394,"TERMINAL",0,0,"[?25l0\r[?25h",,terminal_output +1812,3518503,"TERMINAL",0,0,"[?25l[\r[?25h",,terminal_output +1813,3519380,"TERMINAL",0,0,".",,terminal_output +1814,3519582,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +1815,3519722,"TERMINAL",0,0,"[?25lha[?25h",,terminal_output +1816,3519961,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1817,3520285,"TERMINAL",0,0,"[?25le\r[?25h",,terminal_output +1818,3520549,"TERMINAL",0,0,"[?25lp[?25h",,terminal_output +1819,3520642,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +1820,3520790,"TERMINAL",0,0,"\r\n(1, 16, 920)\r\n(jdb) ",,terminal_output +1821,3568352,"TERMINAL",0,0,"^DERROR:2025-07-17 10:43:31,081:jax._src.debugging:98: jax.debug.callback failed\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 96, in debug_callback_impl\r\n callback(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 336, in _flat_callback\r\n callback(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/core.py"", line 220, in _breakpoint_callback\r\n debugger(frames, thread_id, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 167, in run_debugger\r\n CliDebugger(frames, thread_id, **kwargs).run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 160, in run\r\n self.cmdloop()\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 138, in cmdloop\r\n stop = self.onecmd(line)\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 217, in onecmd\r\n return func(arg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 146, in do_quit\r\n sys.exit(0)\r\nSystemExit: 0\r\nERROR:jax._src.debugging:jax.debug.callback failed\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 96, in debug_callback_impl\r\n callback(*args)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 336, in _flat_callback\r\n callback(*args, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/core.py"", line 220, in _breakpoint_callback\r\n debugger(frames, thread_id, **kwargs)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 167, in run_debugger\r\n CliDebugger(frames, thread_id, **kwargs).run()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 160, in run\r\n self.cmdloop()\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 138, in cmdloop\r\n stop = self.onecmd(line)\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 217, in onecmd\r\n return func(arg)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 146, in do_quit\r\n sys.exit(0)\r\nSystemExit: 0\r\nE0717 10:43:31.087060 943237 pjrt_stream_executor_client.cc:2916] Execution of replica 0 failed: INTERNAL: CpuCallback error calling callback: Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 178, in \r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 141, in _autoreg_sample\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/traceback_util.py"", line 182, in reraise_with_filtered_traceback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 292, in cache_miss\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 153, in _python_pjit_helper\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 1877, in _pjit_call_impl_python\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/profiler.py"", line 354, in wrapper\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/interpreters/pxla.py"", line 1297, in __call__\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/callback.py"", line 782, in _wrapped_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 202, in _callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 99, in debug_callback_impl\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 336, in _flat_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/core.py"", line 220, in _breakpoint_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 167, in run_debugger\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 162, in run\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 145, in cmdloop\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 217, in onecmd\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 146, in do_quit\r\nSystemExit: 0\r\njax.errors.SimplifiedTraceback: For simplicity, JAX has removed its internal frames from the traceback of the following exception. Set JAX_TRACEBACK_FILTERING=off to include these.\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 178, in \r\n vid = _autoreg_sample(rng, video_batch, action_batch)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 141, in _autoreg_sample\r\n generated_vid = sampling_fn(\r\njaxlib._jax.XlaRuntimeError: INTERNAL: CpuCallback error calling callback: Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 178, in \r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 141, in _autoreg_sample\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/traceback_util.py"", line 182, in reraise_with_filtered_traceback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 292, in cache_miss\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 153, in _python_pjit_helper\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/pjit.py"", line 1877, in _pjit_call_impl_python\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/profiler.py"", line 354, in wrapper\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/interpreters/pxla.py"", line 1297, in __call__\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/callback.py"", line 782, in _wrapped_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 202, in _callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 99, in debug_callback_impl\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugging.py"", line 336, in _flat_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/core.py"", line 220, in _breakpoint_callback\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 167, in run_debugger\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 162, in run\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 145, in cmdloop\r\n File ""/home/hk-project-p0023960/tum_cte0515/.local/share/uv/python/cpython-3.10.18-linux-x86_64-gnu/lib/python3.10/cmd.py"", line 217, in onecmd\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/jax/_src/debugger/cli_debugger.py"", line 146, in do_quit\r\nSystemExit: 0\r\n",,terminal_output +1822,3569628,"TERMINAL",0,0,"srun",,terminal_focus +1823,3570572,"train_dynamics.py",0,0,"",python,tab +1824,3570763,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +1825,3572276,"genie.py",0,0,"",python,tab +1826,3573091,"genie.py",6131,0,"",python,selection_mouse +1827,3574672,"genie.py",6132,0,"",python,selection_command +1828,3574830,"genie.py",6178,0,"",python,selection_command +1829,3574974,"genie.py",6208,0,"",python,selection_command +1830,3575122,"genie.py",6286,0,"",python,selection_command +1831,3575267,"genie.py",6304,0,"",python,selection_command +1832,3575433,"genie.py",6358,0,"",python,selection_command +1833,3575568,"genie.py",6411,0,"",python,selection_command +1834,3575715,"genie.py",6482,0,"",python,selection_command +1835,3575886,"genie.py",6510,0,"",python,selection_command +1836,3576019,"genie.py",6511,0,"",python,selection_command +1837,3576178,"genie.py",6571,0,"",python,selection_command +1838,3576549,"genie.py",6645,0,"",python,selection_command +1839,3577488,"genie.py",6645,35,"",python,content +1840,3578333,"genie.py",6571,0,"",python,selection_command +1841,3578589,"genie.py",6645,0,"",python,selection_command +1842,3578883,"genie.py",6571,0,"",python,selection_command +1843,3579010,"genie.py",6645,0,"",python,selection_command +1844,3579190,"genie.py",6646,0,"",python,selection_command +1845,3579354,"genie.py",6645,0,"",python,selection_command +1846,3579429,"genie.py",6646,0,"",python,selection_command +1847,3579624,"genie.py",6645,0,"",python,selection_command +1848,3579799,"genie.py",6571,0,"",python,selection_command +1849,3579960,"genie.py",6511,0,"",python,selection_command +1850,3580095,"genie.py",6571,0,"",python,selection_command +1851,3580310,"genie.py",6645,0,"",python,selection_command +1852,3580449,"genie.py",6571,0,"",python,selection_command +1853,3580653,"genie.py",6511,0,"",python,selection_command +1854,3580827,"genie.py",6571,0,"",python,selection_command +1855,3581025,"genie.py",6511,0,"",python,selection_command +1856,3581193,"genie.py",6571,0,"",python,selection_command +1857,3581345,"genie.py",6511,0,"",python,selection_command +1858,3581467,"genie.py",6571,0,"",python,selection_command +1859,3581524,"genie.py",6511,0,"",python,selection_command +1860,3581600,"genie.py",6571,0,"",python,selection_command +1861,3581736,"genie.py",6511,0,"",python,selection_command +1862,3581791,"genie.py",6571,0,"",python,selection_command +1863,3581933,"genie.py",6511,0,"",python,selection_command +1864,3581996,"genie.py",6571,0,"",python,selection_command +1865,3582147,"genie.py",6511,0,"",python,selection_command +1866,3582224,"genie.py",6571,0,"",python,selection_command +1867,3582376,"genie.py",6511,0,"",python,selection_command +1868,3582437,"genie.py",6571,0,"",python,selection_command +1869,3582598,"genie.py",6511,0,"",python,selection_command +1870,3582702,"genie.py",6571,0,"",python,selection_command +1871,3582847,"genie.py",6511,0,"",python,selection_command +1872,3582917,"genie.py",6571,0,"",python,selection_command +1873,3583043,"genie.py",6511,0,"",python,selection_command +1874,3583104,"genie.py",6571,0,"",python,selection_command +1875,3583232,"genie.py",6511,0,"",python,selection_command +1876,3583333,"genie.py",6571,0,"",python,selection_command +1877,3584822,"TERMINAL",0,0,"srun",,terminal_focus +1878,3586358,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +1879,3587085,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +1880,3588947,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +1881,3591082,"TERMINAL",0,0,"2025-07-17 10:43:53.811767: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1882,3605194,"TERMINAL",0,0,"2025-07-17 10:44:07.929642: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1883,3620171,"TERMINAL",0,0,"2025-07-17 10:44:22.962428: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1884,3623921,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 35000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/035000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 34000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/034000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 36000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/036000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\n",,terminal_output +1885,3692897,"TERMINAL",0,0,"2025-07-17 10:45:35.686456: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 10:45:35.686981: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 10:45:35.687655: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1886,3740927,"TERMINAL",0,0,"token_idxs shape: (Array(1, dtype=int32), Array(1, dtype=int32), Array(920, dtype=int32))\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\n",,terminal_output +1887,3741269,"TERMINAL",0,0,"token_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full shape: (Array(1, dtype=int32), Array(16, dtype=int32), Array(920, dtype=int32))\r\naction_tokens shape: (Array(1, dtype=int32), Array(15, dtype=int32), Array(1, dtype=int32), Array(32, dtype=int32))\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\n",,terminal_output +1888,3741638,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 181, in \r\n ssim = pix.ssim(gt[:, args.start_frame + 1 :], recon[:, args.start_frame + 1 :]).mean()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/dm_pix/_src/metrics.py"", line 221, in ssim\r\n chex.assert_type([a, b], float)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/chex/_src/asserts_internal.py"", line 279, in _chex_assert_fn\r\n host_assertion_fn(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/chex/_src/asserts_internal.py"", line 197, in _assert_on_host\r\n raise exception_type(error_msg)\r\nAssertionError: [Chex] Assertion assert_type failed: Error in type compatibility check: input 0 has type uint8 but expected .\r\n",,terminal_output +1889,3743101,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +1890,3797336,"genie.py",0,0,"",python,tab +1891,3798956,"genie.py",5616,0,"",python,selection_mouse +1892,3800990,"genie.py",5582,156,"",python,content +1893,3801058,"genie.py",5594,0,"",python,selection_command +1894,3802168,"genie.py",5693,0,"",python,selection_command +1895,3802665,"genie.py",5763,0,"",python,selection_command +1896,3802688,"genie.py",5831,0,"",python,selection_command +1897,3802853,"genie.py",5763,0,"",python,selection_command +1898,3803089,"genie.py",5693,0,"",python,selection_command +1899,3803276,"genie.py",5616,0,"",python,selection_command +1900,3804359,"genie.py",5581,0,"",python,selection_mouse +1901,3804364,"genie.py",5580,0,"",python,selection_command +1902,3805887,"genie.py",6437,0,"",python,selection_mouse +1903,3806034,"genie.py",6427,15,"token_idxs_full",python,selection_mouse +1904,3827191,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",0,0,"from dataclasses import dataclass\nfrom typing import Optional\nimport time\nimport os\n\nimport dm_pix as pix\nimport einops\nimport jax\nimport jax.numpy as jnp\nimport flax.linen as nn\nimport numpy as np\nfrom flax.training.train_state import TrainState\nimport grain\nimport orbax.checkpoint as ocp\nimport optax\nfrom PIL import Image, ImageDraw\nimport tyro\n\nfrom genie import Genie\nfrom utils.dataloader import get_dataloader\n\n\n@dataclass\nclass Args:\n # Experiment\n seed: int = 0\n seq_len: int = 16\n image_channels: int = 3\n image_height: int = 90\n image_width: int = 160\n data_dir: str = ""data/coinrun_episodes""\n checkpoint: str = """"\n checkpoint_step: Optional[int] = None\n # Sampling\n batch_size: int = 1\n maskgit_steps: int = 25\n temperature: float = 1.0\n sample_argmax: bool = True\n start_frame: int = 0\n # Tokenizer checkpoint\n tokenizer_dim: int = 512\n latent_patch_dim: int = 32\n num_patch_latents: int = 1024\n patch_size: int = 4\n tokenizer_num_blocks: int = 8\n tokenizer_num_heads: int = 8\n # LAM checkpoint\n lam_dim: int = 512\n latent_action_dim: int = 32\n num_latent_actions: int = 6\n lam_patch_size: int = 16\n lam_num_blocks: int = 8\n lam_num_heads: int = 8\n lam_co_train: bool = True\n # Dynamics checkpoint\n dyna_dim: int = 512\n dyna_num_blocks: int = 12\n dyna_num_heads: int = 8\n param_dtype: jnp.dtype = jnp.float32\n dtype: jnp.dtype = jnp.bfloat16\n\n\nargs = tyro.cli(Args)\nrng = jax.random.PRNGKey(args.seed)\n\n# --- Load Genie checkpoint ---\ngenie = Genie(\n # Tokenizer\n in_dim=args.image_channels,\n tokenizer_dim=args.tokenizer_dim,\n latent_patch_dim=args.latent_patch_dim,\n num_patch_latents=args.num_patch_latents,\n patch_size=args.patch_size,\n tokenizer_num_blocks=args.tokenizer_num_blocks,\n tokenizer_num_heads=args.tokenizer_num_heads,\n # LAM\n lam_dim=args.lam_dim,\n latent_action_dim=args.latent_action_dim,\n num_latent_actions=args.num_latent_actions,\n lam_patch_size=args.lam_patch_size,\n lam_num_blocks=args.lam_num_blocks,\n lam_num_heads=args.lam_num_heads,\n lam_co_train=args.lam_co_train,\n # Dynamics\n dyna_dim=args.dyna_dim,\n dyna_num_blocks=args.dyna_num_blocks,\n dyna_num_heads=args.dyna_num_heads,\n use_maskgit=False,\n param_dtype=args.param_dtype,\n dtype=args.dtype,\n)\nrng, _rng = jax.random.split(rng)\nimage_shape = (args.image_height, args.image_width, args.image_channels)\ndummy_inputs = dict(\n videos=jnp.zeros((args.batch_size, args.seq_len, *image_shape), dtype=jnp.float32),\n mask_rng=_rng,\n)\nrng, _rng = jax.random.split(rng)\nparams = genie.init(_rng, dummy_inputs)\n\ndummy_train_state = TrainState.create(\n apply_fn=genie.apply,\n params=params,\n tx=optax.adamw(\n optax.warmup_cosine_decay_schedule(\n 0, 0, 1, 2 # dummy values\n )\n ), \n)\nhandler_registry = ocp.handlers.DefaultCheckpointHandlerRegistry()\nhandler_registry.add('model_state', ocp.args.StandardRestore, ocp.handlers.StandardCheckpointHandler)\ncheckpoint_manager = ocp.CheckpointManager(\n args.checkpoint,\n options=ocp.CheckpointManagerOptions(step_format_fixed_length=6),\n handler_registry=handler_registry\n)\nabstract_train_state = jax.tree_util.tree_map(\n ocp.utils.to_shape_dtype_struct, dummy_train_state\n)\n\nrestored = checkpoint_manager.restore(\n args.checkpoint_step or checkpoint_manager.latest_step(),\n args=ocp.args.Composite(\n model_state=ocp.args.StandardRestore(abstract_train_state),\n ),\n)\nrestored_train_state = restored[""model_state""]\nparams = restored_train_state.params\n\n\ndef _sampling_wrapper(module, batch):\n # return module.sample_maskgit(batch, args.seq_len, args.maskgit_steps, args.temperature, args.sample_argmax)\n return module.sample_causal(batch, args.seq_len, args.temperature, args.sample_argmax)\n\n# --- Define autoregressive sampling loop ---\ndef _autoreg_sample(rng, video_batch, action_batch):\n vid = video_batch[:, : args.start_frame + 1]\n sampling_fn = jax.jit(nn.apply(_sampling_wrapper, genie)) \n rng, _rng = jax.random.split(rng)\n batch = dict(videos=vid, latent_actions=action_batch, rng=_rng)\n generated_vid = sampling_fn(\n params,\n batch\n )\n return generated_vid\n\ndef _get_dataloader_iterator():\n array_record_files = [\n os.path.join(args.data_dir, x)\n for x in os.listdir(args.data_dir)\n if x.endswith("".array_record"")\n ]\n grain_dataloader = get_dataloader(\n array_record_files,\n args.seq_len,\n # NOTE: We deliberately pass the global batch size\n # The dataloader shards the dataset across all processes\n args.batch_size,\n *image_shape,\n num_workers=0,\n prefetch_buffer_size=1,\n seed=args.seed,\n )\n initial_state = grain_dataloader._create_initial_state()\n grain_iterator = grain.DataLoaderIterator(grain_dataloader, initial_state)\n return grain_iterator\n\n# --- Get video + latent actions ---\ngrain_iterator = _get_dataloader_iterator()\nvideo_batch = next(grain_iterator)\n\n# Get latent actions for all videos in the batch\nbatch = dict(videos=video_batch)\naction_batch = genie.apply(params, batch, False, method=Genie.vq_encode)\naction_batch = action_batch.reshape(video_batch.shape[0], args.seq_len - 1, 1)\n\n# --- Sample + evaluate video ---\nvid = _autoreg_sample(rng, video_batch, action_batch)\ngt = video_batch[:, : vid.shape[1]].clip(0, 1).reshape(-1, *video_batch.shape[2:])\nrecon = vid.clip(0, 1).reshape(-1, *vid.shape[2:])\nssim = pix.ssim(gt[:, args.start_frame + 1 :], recon[:, args.start_frame + 1 :]).mean()\nprint(f""SSIM: {ssim}"")\n\n# --- Construct video ---\ntrue_videos = (video_batch * 255).astype(np.uint8)\npred_videos = (vid * 255).astype(np.uint8)\nvideo_comparison = np.zeros((2, *vid.shape), dtype=np.uint8)\nvideo_comparison[0] = true_videos[:, :args.seq_len]\nvideo_comparison[1] = pred_videos\nframes = einops.rearrange(video_comparison, ""n b t h w c -> t (b h) (n w) c"")\n\n# --- Save video --- \nimgs = [Image.fromarray(img) for img in frames]\n# Write actions on each frame, on each row (i.e., for each video in the batch, on the GT row)\nfor t, img in enumerate(imgs[1:]):\n d = ImageDraw.Draw(img)\n for row in range(action_batch.shape[0]):\n action = action_batch[row, t, 0]\n y_offset = row * video_batch.shape[2] + 2\n d.text((2, y_offset), f""{action}"", fill=255)\nimgs[0].save(\n f""generation_{time.time()}.gif"",\n save_all=True,\n append_images=imgs[1:],\n duration=250,\n loop=0,\n)\n",python,tab +1905,3830872,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5525,0,"",python,selection_mouse +1906,3832424,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5577,0,"",python,selection_mouse +1907,3832556,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5576,2,"gt",python,selection_mouse +1908,3836257,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5611,0,"",python,selection_mouse +1909,3836410,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5607,5,"recon",python,selection_mouse +1910,3838294,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5512,0,"",python,selection_mouse +1911,3838443,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5509,5,"recon",python,selection_mouse +1912,3840049,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5437,0,"",python,selection_mouse +1913,3841039,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5385,0,"",python,selection_mouse +1914,3841179,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5378,15,"_autoreg_sample",python,selection_mouse +1915,3842234,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5374,0,"",python,selection_mouse +1916,3842379,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5372,3,"vid",python,selection_mouse +1917,3848011,"train_dynamics.py",0,0,"",python,tab +1918,3849801,"genie.py",0,0,"",python,tab +1919,3857305,"models/dynamics.py",0,0,"",python,tab +1920,3868310,"models/dynamics.py",1920,0,"",python,selection_mouse +1921,3868439,"models/dynamics.py",1905,22,"DynamicsAutoregressive",python,selection_mouse +1922,3872379,"train_dynamics.py",0,0,"",python,tab +1923,3887854,"train_dynamics.py",11808,0,"",python,selection_mouse +1924,3888193,"train_dynamics.py",11803,10,"train_step",python,selection_mouse +1925,3890261,"train_dynamics.py",11808,0,"",python,selection_mouse +1926,3892931,"train_dynamics.py",4482,0,"",python,selection_mouse +1927,3901281,"train_dynamics.py",1813,0,"",python,selection_mouse +1928,3901426,"train_dynamics.py",1808,11,"param_dtype",python,selection_mouse +1929,3902102,"train_dynamics.py",1852,0,"",python,selection_mouse +1930,3902263,"train_dynamics.py",1849,5,"dtype",python,selection_mouse +1931,3926554,"train_dynamics.py",2684,0,"",python,selection_mouse +1932,3926711,"train_dynamics.py",2681,7,"outputs",python,selection_mouse +1933,3928087,"train_dynamics.py",2704,0,"",python,selection_mouse +1934,3928328,"train_dynamics.py",2704,1,".",python,selection_mouse +1935,3928329,"train_dynamics.py",2704,5,".asty",python,selection_mouse +1936,3928329,"train_dynamics.py",2704,8,".astype(",python,selection_mouse +1937,3928331,"train_dynamics.py",2704,11,".astype(jnp",python,selection_mouse +1938,3928331,"train_dynamics.py",2704,13,".astype(jnp.f",python,selection_mouse +1939,3928331,"train_dynamics.py",2704,15,".astype(jnp.flo",python,selection_mouse +1940,3928359,"train_dynamics.py",2704,17,".astype(jnp.float",python,selection_mouse +1941,3928389,"train_dynamics.py",2704,19,".astype(jnp.float32",python,selection_mouse +1942,3928415,"train_dynamics.py",2704,20,".astype(jnp.float32)",python,selection_mouse +1943,3937888,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",0,0,"",python,tab +1944,3938858,"genie.py",0,0,"",python,tab +1945,3941891,"genie.py",5925,0,"",python,selection_mouse +1946,3943375,"genie.py",5952,0,"",python,selection_mouse +1947,3944645,"genie.py",6489,0,"",python,selection_mouse +1948,3946437,"genie.py",5952,0,"",python,selection_mouse +1949,3947537,"genie.py",5952,0,".astype(jnp.float32)",python,content +1950,3956793,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +1951,3957371,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +1952,3958850,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +1953,3960937,"TERMINAL",0,0,"2025-07-17 10:50:03.731602: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1954,3974717,"TERMINAL",0,0,"2025-07-17 10:50:17.512362: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1955,3989592,"TERMINAL",0,0,"2025-07-17 10:50:32.384191: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1956,3993451,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 34000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/034000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 35000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/035000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 36000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/036000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\n",,terminal_output +1957,4062236,"TERMINAL",0,0,"2025-07-17 10:51:45.025648: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 10:51:45.026186: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 10:51:45.026866: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +1958,4111100,"TERMINAL",0,0,"token_idxs shape: (Array(1, dtype=int32), Array(1, dtype=int32), Array(920, dtype=int32))\r\n",,terminal_output +1959,4111440,"TERMINAL",0,0,"token_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\n",,terminal_output +1960,4111916,"TERMINAL",0,0,"Traceback (most recent call last):\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py"", line 181, in \r\n ssim = pix.ssim(gt[:, args.start_frame + 1 :], recon[:, args.start_frame + 1 :]).mean()\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/dm_pix/_src/metrics.py"", line 221, in ssim\r\n chex.assert_type([a, b], float)\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/chex/_src/asserts_internal.py"", line 279, in _chex_assert_fn\r\n host_assertion_fn(\r\n File ""/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/chex/_src/asserts_internal.py"", line 197, in _assert_on_host\r\n raise exception_type(error_msg)\r\nAssertionError: [Chex] Assertion assert_type failed: Error in type compatibility check: input 0 has type uint8 but expected .\r\n",,terminal_output +1961,4113580,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +1962,4147504,"genie.py",0,0,"",python,tab +1963,4157436,"train_dynamics.py",0,0,"",python,tab +1964,4158700,"genie.py",0,0,"",python,tab +1965,4160102,"genie.py",5953,0,"",python,selection_mouse +1966,4161150,"genie.py",5906,0,"",python,selection_mouse +1967,4161330,"genie.py",5892,17,"next_token_logits",python,selection_mouse +1968,4161850,"genie.py",5924,0,"",python,selection_mouse +1969,4162574,"genie.py",5906,0,"",python,selection_mouse +1970,4162721,"genie.py",5892,17,"next_token_logits",python,selection_mouse +1971,4163195,"genie.py",5995,0,"",python,selection_mouse +1972,4165735,"genie.py",6651,0,"",python,selection_mouse +1973,4165869,"genie.py",6649,5,"batch",python,selection_mouse +1974,4166534,"genie.py",6664,0,"",python,selection_mouse +1975,4166650,"genie.py",6662,3,"""].",python,selection_mouse +1976,4167183,"genie.py",6659,0,"",python,selection_mouse +1977,4167344,"genie.py",6656,6,"videos",python,selection_mouse +1978,4169356,"genie.py",6598,0,"",python,selection_mouse +1979,4169522,"genie.py",6593,9,"tokenizer",python,selection_mouse +1980,4170124,"genie.py",6608,0,"",python,selection_mouse +1981,4170273,"genie.py",6603,6,"decode",python,selection_mouse +1982,4170783,"genie.py",6685,0,"",python,selection_mouse +1983,4171465,"genie.py",6659,0,"",python,selection_mouse +1984,4171627,"genie.py",6656,6,"videos",python,selection_mouse +1985,4177865,"train_dynamics.py",0,0,"",python,tab +1986,4179308,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",0,0,"",python,tab +1987,4181141,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5448,0,"",python,selection_mouse +1988,4181840,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5520,0,"",python,selection_mouse +1989,4187327,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",2366,0,"",python,selection_mouse +1990,4187488,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",2363,5,"dtype",python,selection_mouse +1991,4191391,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",6564,0,"",python,selection_mouse +1992,4193621,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",1397,0,"",python,selection_command +1993,4193994,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",1408,0,"",python,selection_command +1994,4194253,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",1432,0,"",python,selection_command +1995,4194667,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",1443,0,"",python,selection_command +1996,4195005,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",2324,0,"",python,selection_command +1997,4195884,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",2341,0,"",python,selection_command +1998,4197985,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",2352,0,"",python,selection_command +1999,4198354,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",2363,0,"",python,selection_command +2000,4198707,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",2568,0,"",python,selection_command +2001,4206448,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",3304,0,"",python,selection_command +2002,4209071,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5837,0,"",python,selection_command +2003,4232641,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5427,0,"",python,selection_mouse +2004,4233210,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5435,0,"",python,selection_mouse +2005,4233360,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5431,11,"video_batch",python,selection_mouse +2006,4250103,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5439,0,"",python,selection_mouse +2007,4250104,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5431,11,"video_batch",python,selection_mouse +2008,4255935,"train_dynamics.py",0,0,"",python,tab +2009,4257478,"train_dynamics.py",2799,0,"",python,selection_mouse +2010,4257538,"train_dynamics.py",2798,0,"",python,selection_command +2011,4259516,"train_dynamics.py",2409,0,"",python,selection_mouse +2012,4259678,"train_dynamics.py",2404,6,"videos",python,selection_mouse +2013,4260651,"train_dynamics.py",2441,0,"",python,selection_mouse +2014,4260786,"train_dynamics.py",2439,4,"args",python,selection_mouse +2015,4263417,"train_dynamics.py",2446,0,"",python,selection_mouse +2016,4263593,"train_dynamics.py",2444,5,"dtype",python,selection_mouse +2017,4266674,"train_dynamics.py",2431,0,"",python,selection_mouse +2018,4266896,"train_dynamics.py",2431,1,".",python,selection_mouse +2019,4266896,"train_dynamics.py",2431,3,".as",python,selection_mouse +2020,4266897,"train_dynamics.py",2431,57,".astype(args.dtype) / 255.0\n outputs = state.apply_fn(",python,selection_mouse +2021,4267083,"train_dynamics.py",2431,12,".astype(args",python,selection_mouse +2022,4267146,"train_dynamics.py",2431,13,".astype(args.",python,selection_mouse +2023,4267309,"train_dynamics.py",2431,14,".astype(args.d",python,selection_mouse +2024,4267310,"train_dynamics.py",2431,15,".astype(args.dt",python,selection_mouse +2025,4267310,"train_dynamics.py",2431,16,".astype(args.dty",python,selection_mouse +2026,4267310,"train_dynamics.py",2431,17,".astype(args.dtyp",python,selection_mouse +2027,4267345,"train_dynamics.py",2431,18,".astype(args.dtype",python,selection_mouse +2028,4267450,"train_dynamics.py",2431,19,".astype(args.dtype)",python,selection_mouse +2029,4273793,"train_dynamics.py",2458,0,"",python,selection_mouse +2030,4273802,"train_dynamics.py",2457,0,"",python,selection_command +2031,4273993,"train_dynamics.py",2457,1,"0",python,selection_mouse +2032,4274033,"train_dynamics.py",2458,0,"",python,selection_command +2033,4274047,"train_dynamics.py",2455,3,"5.0",python,selection_mouse +2034,4274047,"train_dynamics.py",2452,6," 255.0",python,selection_mouse +2035,4274085,"train_dynamics.py",2448,10,"e) / 255.0",python,selection_mouse +2036,4274085,"train_dynamics.py",2458,30,"\n outputs = state.apply_fn(",python,selection_mouse +2037,4274342,"train_dynamics.py",2438,20,"(args.dtype) / 255.0",python,selection_mouse +2038,4274395,"train_dynamics.py",2439,19,"args.dtype) / 255.0",python,selection_mouse +2039,4274489,"train_dynamics.py",2438,20,"(args.dtype) / 255.0",python,selection_mouse +2040,4274513,"train_dynamics.py",2437,21,"e(args.dtype) / 255.0",python,selection_mouse +2041,4274561,"train_dynamics.py",2436,22,"pe(args.dtype) / 255.0",python,selection_mouse +2042,4274561,"train_dynamics.py",2435,23,"ype(args.dtype) / 255.0",python,selection_mouse +2043,4274624,"train_dynamics.py",2434,24,"type(args.dtype) / 255.0",python,selection_mouse +2044,4274669,"train_dynamics.py",2433,25,"stype(args.dtype) / 255.0",python,selection_mouse +2045,4274718,"train_dynamics.py",2432,26,"astype(args.dtype) / 255.0",python,selection_mouse +2046,4274981,"train_dynamics.py",2431,27,".astype(args.dtype) / 255.0",python,selection_mouse +2047,4281966,"genie.py",0,0,"",python,tab +2048,4286333,"genie.py",4674,0,"",python,selection_mouse +2049,4288298,"genie.py",4673,0,"",python,selection_command +2050,4288753,"genie.py",4674,0,"\n ",python,content +2051,4289207,"genie.py",4683,0,"b",python,content +2052,4289208,"genie.py",4684,0,"",python,selection_keyboard +2053,4289354,"genie.py",4684,0,"a",python,content +2054,4289355,"genie.py",4685,0,"",python,selection_keyboard +2055,4289548,"genie.py",4685,0,"t",python,content +2056,4289549,"genie.py",4686,0,"",python,selection_keyboard +2057,4289661,"genie.py",4686,0,"c",python,content +2058,4289662,"genie.py",4687,0,"",python,selection_keyboard +2059,4289696,"genie.py",4687,0,"h",python,content +2060,4289697,"genie.py",4688,0,"",python,selection_keyboard +2061,4290944,"genie.py",4688,0,"[]",python,content +2062,4290945,"genie.py",4689,0,"",python,selection_keyboard +2063,4291281,"genie.py",4689,0,"""""",python,content +2064,4291282,"genie.py",4690,0,"",python,selection_keyboard +2065,4291560,"genie.py",4690,0,"v",python,content +2066,4291561,"genie.py",4691,0,"",python,selection_keyboard +2067,4291652,"genie.py",4691,0,"i",python,content +2068,4291653,"genie.py",4692,0,"",python,selection_keyboard +2069,4291778,"genie.py",4692,0,"d",python,content +2070,4291779,"genie.py",4693,0,"",python,selection_keyboard +2071,4291909,"genie.py",4693,0,"e",python,content +2072,4291910,"genie.py",4694,0,"",python,selection_keyboard +2073,4291994,"genie.py",4694,0,"o",python,content +2074,4291995,"genie.py",4695,0,"",python,selection_keyboard +2075,4292155,"genie.py",4695,0,"s",python,content +2076,4292156,"genie.py",4696,0,"",python,selection_keyboard +2077,4292516,"genie.py",4697,0,"",python,selection_command +2078,4292772,"genie.py",4698,0,"",python,selection_command +2079,4293110,"genie.py",4698,0," ",python,content +2080,4293110,"genie.py",4699,0,"",python,selection_keyboard +2081,4293537,"genie.py",4699,0,"=",python,content +2082,4293538,"genie.py",4700,0,"",python,selection_keyboard +2083,4293607,"genie.py",4700,0," ",python,content +2084,4293608,"genie.py",4701,0,"",python,selection_keyboard +2085,4294210,"genie.py",4701,0,".astype(args.dtype) / 255.0",python,content +2086,4296338,"genie.py",4698,0,"",python,selection_mouse +2087,4296574,"genie.py",4697,1,"]",python,selection_mouse +2088,4296575,"genie.py",4695,3,"s""]",python,selection_mouse +2089,4296575,"genie.py",4693,5,"eos""]",python,selection_mouse +2090,4296575,"genie.py",4692,6,"deos""]",python,selection_mouse +2091,4296575,"genie.py",4690,8,"videos""]",python,selection_mouse +2092,4296576,"genie.py",4689,9,"""videos""]",python,selection_mouse +2093,4296576,"genie.py",4688,10,"[""videos""]",python,selection_mouse +2094,4296624,"genie.py",4687,11,"h[""videos""]",python,selection_mouse +2095,4296669,"genie.py",4698,43," = .astype(args.dtype) / 255.0\n toke",python,selection_mouse +2096,4296694,"genie.py",4698,42," = .astype(args.dtype) / 255.0\n tok",python,selection_mouse +2097,4296729,"genie.py",4698,41," = .astype(args.dtype) / 255.0\n to",python,selection_mouse +2098,4296781,"genie.py",4698,40," = .astype(args.dtype) / 255.0\n t",python,selection_mouse +2099,4296834,"genie.py",4698,39," = .astype(args.dtype) / 255.0\n ",python,selection_mouse +2100,4297093,"genie.py",4683,15,"batch[""videos""]",python,selection_mouse +2101,4299955,"genie.py",4701,0,"",python,selection_mouse +2102,4300490,"genie.py",4701,0,"batch[""videos""]",python,content +2103,4303337,"genie.py",4726,0,"",python,selection_mouse +2104,4303479,"genie.py",4724,4,"args",python,selection_mouse +2105,4303693,"genie.py",4724,5,"args.",python,selection_mouse +2106,4303694,"genie.py",4724,10,"args.dtype",python,selection_mouse +2107,4304107,"genie.py",4732,0,"",python,selection_mouse +2108,4305439,"genie.py",4701,15,"",python,content +2109,4305638,"genie.py",4701,27,"",python,content +2110,4307070,"genie.py",4701,0,".astype(args.dtype) / 255.0",python,content +2111,4307211,"genie.py",4701,0,"batch[""videos""]",python,content +2112,4307823,"genie.py",4742,0,"",python,selection_command +2113,4308967,"genie.py",4675,69,"",python,content +2114,4309029,"genie.py",4683,0,"",python,selection_command +2115,4310597,"train_dynamics.py",0,0,"",python,tab +2116,4312171,"train_dynamics.py",2424,0,"",python,selection_mouse +2117,4312825,"train_dynamics.py",2408,0,"",python,selection_mouse +2118,4313846,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",0,0,"",python,tab +2119,4316539,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5089,0,"",python,selection_mouse +2120,4317181,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5124,0,"",python,selection_mouse +2121,4317954,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5176,0,"",python,selection_mouse +2122,4318836,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5083,0,"",python,selection_mouse +2123,4319439,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5102,0,"",python,selection_mouse +2124,4320263,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5076,0,"",python,selection_mouse +2125,4320432,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5067,11,"video_batch",python,selection_mouse +2126,4321117,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5087,0,"",python,selection_mouse +2127,4321945,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5102,0,"",python,selection_mouse +2128,4323001,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5102,0,"\n batch[""videos""] = batch[""videos""].astype(args.dtype) / 255.0",python,content +2129,4323055,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5111,0,"",python,selection_command +2130,4324725,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5107,4,"",python,content +2131,4325147,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5103,4,"",python,content +2132,4326012,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5103,5,"",python,content +2133,4326150,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5103,2,"",python,content +2134,4326398,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5103,6,"",python,content +2135,4326724,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5103,2,"",python,content +2136,4327904,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5103,0,"v",python,content +2137,4327905,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5104,0,"",python,selection_keyboard +2138,4328083,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5104,0,"i",python,content +2139,4328084,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5105,0,"",python,selection_keyboard +2140,4328225,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5105,0,"d",python,content +2141,4328226,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5106,0,"",python,selection_keyboard +2142,4329102,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5103,3,"video_batch",python,content +2143,4329886,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5117,0,"",python,selection_command +2144,4330199,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5117,5,"",python,content +2145,4330374,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5117,2,"",python,content +2146,4330636,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5117,6,"",python,content +2147,4330807,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5117,3,"",python,content +2148,4332844,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5117,0,"v",python,content +2149,4332845,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5118,0,"",python,selection_keyboard +2150,4333514,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5118,0,"i",python,content +2151,4333515,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5119,0,"",python,selection_keyboard +2152,4333773,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5117,2,"video_batch",python,content +2153,4334123,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5128,0,".",python,content +2154,4334124,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5129,0,"",python,selection_keyboard +2155,4334787,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5128,0,"",python,selection_command +2156,4338267,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5750,0,"",python,selection_mouse +2157,4338286,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5749,0,"",python,selection_command +2158,4341970,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +2159,4342161,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +2160,4344467,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +2161,4346534,"TERMINAL",0,0,"2025-07-17 10:56:29.324296: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2162,4360415,"TERMINAL",0,0,"2025-07-17 10:56:43.205786: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2163,4375516,"TERMINAL",0,0,"2025-07-17 10:56:58.276862: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2164,4379270,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 35000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/035000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 36000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/036000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 37000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/037000/metrics/metrics not found.\r\n",,terminal_output +2165,4448498,"TERMINAL",0,0,"2025-07-17 10:58:11.287114: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 10:58:11.287629: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 10:58:11.288296: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2166,4497287,"TERMINAL",0,0,"token_idxs shape: (Array(1, dtype=int32), Array(1, dtype=int32), Array(920, dtype=int32))\r\n",,terminal_output +2167,4497621,"TERMINAL",0,0,"token_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\n",,terminal_output +2168,4500706,"TERMINAL",0,0,"SSIM: 0.3542162775993347\r\n",,terminal_output +2169,4503177,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +2170,4612925,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",0,0,"",python,tab +2171,4622392,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5724,0,"",python,selection_mouse +2172,4623927,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5750,0,"",python,selection_mouse +2173,4623947,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5749,0,"",python,selection_command +2174,4624709,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5723,0,"",python,selection_mouse +2175,4624712,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5722,0,"",python,selection_command +2176,4625531,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5724,0,"",python,selection_mouse +2177,4626439,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5723,0,"",python,selection_mouse +2178,4626456,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5722,0,"",python,selection_command +2179,4627191,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5750,0,"",python,selection_mouse +2180,4627194,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5749,0,"",python,selection_command +2181,4628028,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5750,0,"",python,selection_mouse +2182,4628047,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5749,0,"",python,selection_command +2183,4628654,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5724,0,"",python,selection_mouse +2184,4673326,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5551,0,"",python,selection_mouse +2185,4673920,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5694,0,"",python,selection_mouse +2186,4674545,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5541,0,"",python,selection_mouse +2187,4674996,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5610,0,"",python,selection_mouse +2188,4675569,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5724,0,"",python,selection_mouse +2189,4676128,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5786,0,"",python,selection_mouse +2190,4676759,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5893,0,"",python,selection_mouse +2191,4677366,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",6092,0,"",python,selection_mouse +2192,4677372,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",6091,0,"",python,selection_command +2193,4677998,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5424,0,"",python,selection_mouse +2194,4677999,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",5423,0,"",python,selection_command +2195,4680326,"train_dynamics.py",0,0,"",python,tab +2196,4688316,"train_dynamics.py",3313,0,"",python,selection_mouse +2197,4694254,"genie.py",0,0,"",python,tab +2198,4696877,"genie.py",4900,0,"",python,selection_mouse +2199,4698333,"genie.py",4839,0,"",python,selection_mouse +2200,4698989,"genie.py",4882,0,"",python,selection_mouse +2201,4699131,"genie.py",4877,10,"token_idxs",python,selection_mouse +2202,4704360,"genie.py",6488,0,"",python,selection_mouse +2203,4707702,"genie.py",5945,0,"",python,selection_mouse +2204,4709987,"genie.py",5945,0,"-",python,content +2205,4709989,"genie.py",5946,0,"",python,selection_keyboard +2206,4710066,"genie.py",5946,0,"1",python,content +2207,4710067,"genie.py",5947,0,"",python,selection_keyboard +2208,4710531,"genie.py",5946,0,"",python,selection_command +2209,4711995,"genie.py",5792,0,"",python,selection_mouse +2210,4712542,"genie.py",5854,0,"",python,selection_mouse +2211,4720165,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +2212,4720629,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +2213,4722328,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +2214,4724464,"TERMINAL",0,0,"2025-07-17 11:02:47.254967: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2215,4738194,"TERMINAL",0,0,"2025-07-17 11:03:00.986005: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2216,4753257,"TERMINAL",0,0,"2025-07-17 11:03:16.028164: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2217,4756845,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 35000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/035000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 37000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/037000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 36000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/036000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\n",,terminal_output +2218,4825399,"TERMINAL",0,0,"2025-07-17 11:04:28.188739: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 11:04:28.189216: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 11:04:28.189880: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2219,4873038,"TERMINAL",0,0,"token_idxs shape: (Array(1, dtype=int32), Array(1, dtype=int32), Array(920, dtype=int32))\r\n",,terminal_output +2220,4873376,"TERMINAL",0,0,"token_idxs_full 0: [148 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 148 0 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 148 148 0 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 148 148 148 0 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 148 148 148 148 0 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 148 148 148 148 148 0 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 148 148 148 148 148 575 0 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 148 148 148 148 148 575 575 0 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 148 148 148 148 148 575 575 575 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 148 148 148 148 148 575 575 575 575 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 148 148 148 148 148 575 575 575 575 575 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 148 148 148 148 148 575 575 575 575 575 575 0 0 0 0]\r\ntoken_idxs_full 0: [148 148 148 148 148 148 575 575 575 575 575 575 575 0 0 0]\r\ntoken_idxs_full 0: [148 148 148 148 148 148 575 575 575 575 575 575 575 575 0 0]\r\ntoken_idxs_full 0: [148 148 148 148 148 148 575 575 575 575 575 575 575 575 575 0]\r\n",,terminal_output +2221,4876247,"TERMINAL",0,0,"SSIM: 0.6116350889205933\r\n",,terminal_output +2222,4878157,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +2223,5058880,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked",,terminal_output +2224,5060481,"TERMINAL",0,0," ",,terminal_output +2225,5060663,"TERMINAL",0,0,"[?25l-[?25h",,terminal_output +2226,5060776,"TERMINAL",0,0,"[?25l-[?25h",,terminal_output +2227,5061227,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +2228,5061648,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +2229,5061702,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +2230,5061880,"TERMINAL",0,0,"[?25lr[?25h",,terminal_output +2231,5061936,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +2232,5062523,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +2233,5062711,"TERMINAL",0,0,"[?25lf[?25h",,terminal_output +2234,5062820,"TERMINAL",0,0,"[?25lr[?25h",,terminal_output +2235,5062989,"TERMINAL",0,0,"[?25la[?25h",,terminal_output +2236,5063510,"TERMINAL",0,0,"[?25lm[?25h",,terminal_output +2237,5063621,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +2238,5064253,"TERMINAL",0,0,"[?25l=[?25h",,terminal_output +2239,5064306,"TERMINAL",0,0,"[?25l1[?25h",,terminal_output +2240,5064496,"TERMINAL",0,0,"[?25l0[?25h",,terminal_output +2241,5066267,"TERMINAL",0,0,"[?25l8[?25h",,terminal_output +2242,5067301,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +2243,5069131,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +2244,5071175,"TERMINAL",0,0,"2025-07-17 11:08:33.954623: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2245,5085028,"TERMINAL",0,0,"2025-07-17 11:08:47.821467: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2246,5099868,"TERMINAL",0,0,"2025-07-17 11:09:02.659807: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2247,5103395,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 37000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/037000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 35000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/035000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 36000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/036000/metrics/metrics not found.\r\n",,terminal_output +2248,5157741,"TERMINAL",0,0,"2025-07-17 11:10:00.531642: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 11:10:00.532209: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 11:10:00.532868: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2249,5191694,"TERMINAL",0,0,"token_idxs shape: (Array(1, dtype=int32), Array(9, dtype=int32), Array(920, dtype=int32))\r\n",,terminal_output +2250,5191856,"TERMINAL",0,0,"token_idxs_full 0: [148 842 707 148 148 148 148 148 148 0 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 842 707 148 148 148 148 148 148 148 0 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 842 707 148 148 148 148 148 148 148 148 0 0 0 0 0]\r\ntoken_idxs_full 0: [148 842 707 148 148 148 148 148 148 148 148 148 0 0 0 0]\r\ntoken_idxs_full 0: [148 842 707 148 148 148 148 148 148 148 148 148 575 0 0 0]\r\ntoken_idxs_full 0: [148 842 707 148 148 148 148 148 148 148 148 148 575 575 0 0]\r\ntoken_idxs_full 0: [148 842 707 148 148 148 148 148 148 148 148 148 575 575 575 0]\r\n",,terminal_output +2251,5194253,"TERMINAL",0,0,"SSIM: 0.8266993761062622\r\n",,terminal_output +2252,5196018,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +2253,5427432,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked --start_frame=8",,terminal_output +2254,5428265,"TERMINAL",0,0,"[?25l [?25h",,terminal_output +2255,5429014,"TERMINAL",0,0,"[?25l-[?25h",,terminal_output +2256,5429169,"TERMINAL",0,0,"[?25l-[?25h",,terminal_output +2257,5429452,"TERMINAL",0,0,"[?25lb[?25h",,terminal_output +2258,5429562,"TERMINAL",0,0,"[?25la \r[?25h",,terminal_output +2259,5429748,"TERMINAL",0,0,"[?25lt[?25h",,terminal_output +2260,5429945,"TERMINAL",0,0,"[?25lc[?25h",,terminal_output +2261,5430043,"TERMINAL",0,0,"[?25lh[?25h",,terminal_output +2262,5431049,"TERMINAL",0,0,"[?25l_[?25h",,terminal_output +2263,5432189,"TERMINAL",0,0,"[?25ls[?25h",,terminal_output +2264,5432259,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +2265,5432479,"TERMINAL",0,0,"[?25lz[?25h",,terminal_output +2266,5432616,"TERMINAL",0,0,"[?25le[?25h",,terminal_output +2267,5433056,"TERMINAL",0,0,"[?25l=[?25h",,terminal_output +2268,5433438,"TERMINAL",0,0,"[?25l1[?25h[?25l2[?25h",,terminal_output +2269,5436313,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",0,0,"",python,tab +2270,5437748,"train_dynamics.py",0,0,"",python,tab +2271,5442688,"genie.py",0,0,"",python,tab +2272,5444242,"genie.py",5604,0,"",python,selection_mouse +2273,5445811,"genie.py",5594,0,"",python,selection_command +2274,5446325,"genie.py",5594,0,"#",python,content +2275,5446327,"genie.py",5595,0,"",python,selection_keyboard +2276,5451299,"genie.py",5595,0," ",python,content +2277,5451300,"genie.py",5596,0,"",python,selection_keyboard +2278,5455621,"genie.py",4879,0,"",python,selection_mouse +2279,5457203,"genie.py",4878,0,"",python,selection_command +2280,5457767,"genie.py",4860,0,"",python,selection_command +2281,5458233,"genie.py",4860,0,"#",python,content +2282,5458234,"genie.py",4861,0,"",python,selection_keyboard +2283,5458341,"genie.py",4861,0," ",python,content +2284,5458342,"genie.py",4862,0,"",python,selection_keyboard +2285,5485638,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +2286,5487259,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +2287,5489356,"TERMINAL",0,0,"2025-07-17 11:15:32.139873: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2288,5518644,"TERMINAL",0,0,"2025-07-17 11:16:01.439387: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2289,5523863,"TERMINAL",0,0,"2025-07-17 11:16:06.644648: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2290,5527621,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 35000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/035000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 37000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/037000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 36000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/036000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\n",,terminal_output +2291,5596892,"TERMINAL",0,0,"2025-07-17 11:17:19.681381: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 11:17:19.681981: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 11:17:19.682684: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2292,5648354,"TERMINAL",0,0,"SSIM: 0.7451367378234863\r\n",,terminal_output +2293,5652477,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +2294,6121108,"TERMINAL",0,0,"srun",,terminal_focus +2295,6122357,"TERMINAL",0,0,"srun",,terminal_focus +2296,6123615,"TERMINAL",0,0,"python sample.py --checkpoint ""/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared//checkpoints/train_dyn_yolorun_new_arch/3352115"" --data_dir /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked --start_frame=8 --batch_size=12",,terminal_output +2297,6127986,"TERMINAL",0,0,"[?25l2[?25h",,terminal_output +2298,6128789,"TERMINAL",0,0,"[?25l[?25h",,terminal_output +2299,6128898,"TERMINAL",0,0,"",,terminal_output +2300,6129114,"TERMINAL",0,0,"",,terminal_output +2301,6130251,"TERMINAL",0,0,"[?25l8 --batch_size=12[?25h",,terminal_output +2302,6130775,"TERMINAL",0,0,"0 --batch_size=12",,terminal_output +2303,6131123,"TERMINAL",0,0,"[?25l  --batch_size=12[?25h",,terminal_output +2304,6131352,"TERMINAL",0,0,"[?25l- --batch_size=12[?25h",,terminal_output +2305,6131463,"TERMINAL",0,0,"[?25l- --batch_size=12[?25h",,terminal_output +2306,6131998,"TERMINAL",0,0,"[?25ls --batch_size=12[?25h",,terminal_output +2307,6132212,"TERMINAL",0,0,"[?25le -batch_size=12\r[?25h",,terminal_output +2308,6132451,"TERMINAL",0,0,"q --batch_size=12\r",,terminal_output +2309,6132869,"TERMINAL",0,0,"[?25l_ --batch_size=12\r[?25h",,terminal_output +2310,6133142,"TERMINAL",0,0,"[?25ll- --batch_size=12\r[?25h",,terminal_output +2311,6133232,"TERMINAL",0,0,"[?25l e --batch_size=12\r[?25h",,terminal_output +2312,6133289,"TERMINAL",0,0,"n --batch_size=12\r",,terminal_output +2313,6134512,"TERMINAL",0,0,"[?25l= --batch_size=12\r[?25h",,terminal_output +2314,6135919,"TERMINAL",0,0,"[?25l2 --batch_size=12\r[?25h",,terminal_output +2315,6139868,"TERMINAL",0,0,"\r\n[?2004l\r",,terminal_output +2316,6141639,"TERMINAL",0,0,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `param-dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/.venv/lib/python3.10/site-packages/tyro/_parsers.py:347: UserWarning: The field `dtype` is annotated with type ``, but the default value `` has type ``. We'll try to handle this gracefully, but it may cause unexpected behavior.\r\n warnings.warn(message)\r\n",,terminal_output +2317,6143810,"TERMINAL",0,0,"2025-07-17 11:26:26.605472: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2318,6157952,"TERMINAL",0,0,"2025-07-17 11:26:40.747443: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2319,6173230,"TERMINAL",0,0,"2025-07-17 11:26:56.023415: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2320,6176861,"TERMINAL",0,0,"WARNING:absl:Missing metrics for step 37000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/037000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 38000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/038000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 36000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/036000/metrics/metrics not found.\r\nWARNING:absl:Missing metrics for step 20000\r\nERROR:absl:File /hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/train_dyn_yolorun_new_arch/3352115/020000/metrics/metrics not found.\r\n",,terminal_output +2321,6222642,"TERMINAL",0,0,"2025-07-17 11:27:45.419339: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 11:27:45.419779: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n2025-07-17 11:27:45.420312: W external/xla/xla/service/gpu/autotuning/dot_search_space.cc:200] All configs were filtered out because none of them sufficiently match the hints. Maybe the hints set does not contain a good representative set of valid configs?Working around this by using the full hints set instead.\r\n",,terminal_output +2322,6246631,"TERMINAL",0,0,"SSIM: 0.7589486837387085\r\n",,terminal_output +2323,6248779,"TERMINAL",0,0,"]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +2324,6717374,"TERMINAL",0,0,"bash",,terminal_focus +2325,6718970,"TERMINAL",0,0,"srun",,terminal_focus +2326,6720569,"TERMINAL",0,0,"srun",,terminal_focus +2327,6724524,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",0,0,"",python,tab +2328,6725193,"train_dynamics.py",0,0,"",python,tab +2329,6726339,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",0,0,"",python,tab +2330,6730719,"train_dynamics.py",0,0,"",python,tab +2331,6734980,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +2332,6735112,"TERMINAL",0,0,"[?25ld[?25h",,terminal_output +2333,6735172,"TERMINAL",0,0,"[?25ll[?25h",,terminal_output +2334,6735390,"TERMINAL",0,0,"[?25li[?25h[?25ln[?25h",,terminal_output +2335,6735462,"TERMINAL",0,0,"[?25lg[?25h",,terminal_output +2336,6735666,"TERMINAL",0,0,"\r\n[?2004l\r[?1049h(B[?7hEvery 1.0s: sinfo_t_idlehkn0804.localdomain: Thu Jul 17 11:36:18 2025Partition dev_cpuonly: 10 nodes idle\rPartition cpuonly:\t 3 nodes idle\rPartition dev_accelerated:\t 2 nodes idle\rPartition accelerated: 52 nodes idle\rPartition dev_accelerated-h100 :\t 0 nodes idle\rPartition accelerated-h100:\t 0 nodes idle\rPartition large:\t 8 nodes idle",,terminal_output +2337,6736665,"TERMINAL",0,0,"9\t ",,terminal_output +2338,6737328,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0804:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0804 jafar]$ ",,terminal_output +2339,6739009,"train_dynamics.py",0,0,"",python,tab +2340,6904802,"train_dynamics.py",3319,0,"",python,selection_mouse +2341,6904809,"train_dynamics.py",3318,0,"",python,selection_command +2342,6943573,"train_dynamics.py",3319,0,"",python,selection_mouse +2343,6943574,"train_dynamics.py",3318,0,"",python,selection_command +2344,6973827,"train_dynamics.py",3241,0,"",python,selection_mouse +2345,6973828,"train_dynamics.py",3240,0,"",python,selection_command +2346,7024179,"train_dynamics.py",0,0,"",python,tab +2347,7025760,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",0,0,"",python,tab +2348,7027184,"genie.py",0,0,"",python,tab +2349,7030532,"models/dynamics.py",0,0,"",python,tab +2350,7031276,"genie.py",0,0,"",python,tab +2351,7032148,"models/dynamics.py",0,0,"",python,tab +2352,7033739,"genie.py",0,0,"",python,tab +2353,7034733,"/hkfs/home/project/hk-project-p0023960/tum_cte0515/Projects/jafar/sample.py",0,0,"",python,tab +2354,7036102,"train_dynamics.py",0,0,"",python,tab +2355,7869514,"TERMINAL",0,0,"srun",,terminal_focus +2356,7882538,"TERMINAL",0,0,"s",,terminal_output +2357,7882636,"TERMINAL",0,0,"[?25lm[?25h",,terminal_output +2358,7882719,"TERMINAL",0,0,"[?25li[?25h",,terminal_output +2359,7883079,"TERMINAL",0,0,"[?25l[?2004l\r[?25h[?1049h(B[?7hEvery 1.0s: nvidia-smihkn0605.localdomain: Thu Jul 17 11:55:25 2025Thu Jul 17 11:55:25 2025\r+-----------------------------------------------------------------------------------------+\r| NVIDIA-SMI 570.133.20Driver Version: 570.133.20 CUDA Version: 12.8 |\r|-----------------------------------------+------------------------+----------------------+\r| GPU NamePersistence-M | Bus-IdDisp.A | Volatile Uncorr. ECC |\r| Fan Temp PerfPwr:Usage/Cap |Memory-Usage | GPU-Util Compute M. |\r|||MIG M. |\r|=========================================+========================+======================|\r| 0 NVIDIA A100-SXM4-40GBOn | 00000000:31:00.0 Off |0 |\r| N/A 45C P056W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r| 1 NVIDIA A100-SXM4-40GBOn | 00000000:4B:00.0 Off |0 |\r| N/A 44C P062W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r| 2 NVIDIA A100-SXM4-40GBOn | 00000000:CA:00.0 Off |0 |\r| N/A 44C P052W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r| 3 NVIDIA A100-SXM4-40GBOn | 00000000:E3:00.0 Off |0 |\r| N/A 44C P053W / 300W |\t 27MiB / 40960MiB |\t 0%\t Default |\r|||Disabled |\r+-----------------------------------------+------------------------+----------------------+\r+-----------------------------------------------------------------------------------------+\r| Processes:|\r| GPU GI CIPID Type Process nameGPU Memory |\r|ID IDUsage\t |\r|=========================================================================================|",,terminal_output +2360,7883940,"TERMINAL",0,0,"[?1049l\r[?1l>]0;tum_cte0515@hkn0605:~/Projects/jafar[?2004h(jafar) [tum_cte0515@hkn0605 jafar]$ ",,terminal_output +2361,7923970,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun.sbatch",0,0,"",shellscript,tab +2362,7925223,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_yolorun copy.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun_new_arch\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n# source .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-new-arch-run-$slurm_job_id \\n --tags dynamics \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n ",shellscript,tab +2363,7934465,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=2\n#SBATCH --ntasks-per-node=4\n#SBATCH --time=48:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:4\n#SBATCH --output=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --error=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/logs/logs_mihir/%x_%j.log\n#SBATCH --job-name=train_dyn_yolorun_new_arch\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\n# source .venv/bin/activate\n\narray_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data_new/open_ai_minecraft_arrayrecords_chunked\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name/$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/big-runs/tokenizer-lr-scaling/train_tokenizer_lr_sweep_1e-4\n\nenv | grep SLURM\n\nsrun python train_dynamics.py \\n --save_ckpt \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=96 \\n --max_lr=1e-4 \\n --log_image_interval=1000 \\n --log \\n --log_checkpoint_interval=1000 \\n --name=dynamics-new-arch-run-$slurm_job_id \\n --tags dynamics \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --data_dir $array_records_dir\n ",shellscript,tab +2364,7947341,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1084,0,"",shellscript,selection_mouse +2365,7948239,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1083,0,"",shellscript,selection_command +2366,7949099,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1083,1,"3",shellscript,content +2367,7950652,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1130,0,"",shellscript,selection_mouse +2368,7950653,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1129,0,"",shellscript,selection_command +2369,7951417,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1105,0,"",shellscript,selection_mouse +2370,7964650,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",0,0,"#!/usr/bin/env bash\n\n#SBATCH --nodes=1\n#SBATCH --ntasks-per-node=1\n#SBATCH --time=10:00:00\n#SBATCH --partition=accelerated\n#SBATCH --cpus-per-task=5\n#SBATCH --gres=gpu:1\n#SBATCH --output=logs/logs_training/%x_%j.log\n#SBATCH --error=logs/logs_training/%x_%j.log\n#SBATCH --mail-user=mihir.mahajan2002@gmail.com\n#SBATCH --job-name=train_dynamics_minecraft_overfit_sample_1.5M\n#SBATCH --mem=50G\n#SBATCH --mail-type=ALL\n\n# Log the sbatch script\ncat $0\n\nmodule unload mpi/openmpi/5.0\nmodule unload devel/cuda/12.4\nsource .venv/bin/activate\n\ntf_records_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/data/knoms_tfrecords_500_shards\nws_dir='/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/'\n\njob_name=$SLURM_JOB_NAME\nslurm_job_id=$SLURM_JOB_ID\n\nCHECKPOINT_DIR=$ws_dir/checkpoints/$job_name_$slurm_job_id\nmkdir -p $CHECKPOINT_DIR\n\nenv | grep SLURM\n\necho Running dynamics model overfit run tiny model A, ~1.5M params. Slurm id: $slurm_job_id\n\ntokenizer_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290391/tokenizer_1750845012_50000/\nlam_ckpt_dir=/hkfs/work/workspace/scratch/tum_ind3695-jafa_ws_shared/checkpoints/3290392/lam_1750845133_180000/\n\nsrun python train_dynamics.py \\n --ckpt_dir $CHECKPOINT_DIR \\n --batch_size=1 \\n --min_lr=5e-7 \\n --max_lr=5e-6 \\n --warmup_steps=125 \\n --log_image_interval=100 \\n --log \\n --name=dynamics-model-size-scaling-1.5M-$slurm_job_id \\n --tags dynamics model-size-scaling 1.5M tiny A \\n --log_checkpoint_interval=500 \\n --entity instant-uv \\n --project jafar \\n --tokenizer_checkpoint=$tokenizer_ckpt_dir \\n --lam_checkpoint=$lam_ckpt_dir \\n --data_dir $tf_records_dir \\n --dyna_dim=128 \\n --dyna_num_blocks=2 \\n --dyna_num_heads=4",shellscript,tab +2371,7966392,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1769,0,"",shellscript,selection_mouse +2372,7966393,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1768,0,"",shellscript,selection_command +2373,7966905,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1769,0,"",shellscript,selection_mouse +2374,7966910,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1768,0,"",shellscript,selection_command +2375,7967041,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1768,1,"4",shellscript,selection_mouse +2376,7967042,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1769,0,"",shellscript,selection_command +2377,7967081,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1768,1,"4",shellscript,selection_mouse +2378,7967097,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1764,5,"ads=4",shellscript,selection_mouse +2379,7967128,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1761,8,"_heads=4",shellscript,selection_mouse +2380,7967129,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1758,11,"num_heads=4",shellscript,selection_mouse +2381,7967175,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1754,15,"yna_num_heads=4",shellscript,selection_mouse +2382,7967176,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1726,43,"-dyna_num_blocks=2 \\n --dyna_num_heads=4",shellscript,selection_mouse +2383,7967215,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1723,46," --dyna_num_blocks=2 \\n --dyna_num_heads=4",shellscript,selection_mouse +2384,7967216,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1721,48," --dyna_num_blocks=2 \\n --dyna_num_heads=4",shellscript,selection_mouse +2385,7967231,"slurm/jobs/mihir/horeka/modelsize_scaling/dynamics/A_train_dyn_1.5M.sbatch",1700,69," --dyna_dim=128 \\n --dyna_num_blocks=2 \\n --dyna_num_heads=4",shellscript,selection_mouse +2386,7972891,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",0,0,"",shellscript,tab +2387,7974101,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1369,0,"",shellscript,selection_mouse +2388,7974102,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1368,0,"",shellscript,selection_command +2389,7975333,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1369,0,"\n ",shellscript,content +2390,7975580,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1374,0," --dyna_dim=128 \\n --dyna_num_blocks=2 \\n --dyna_num_heads=4",shellscript,content +2391,7976985,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1378,0,"",shellscript,selection_mouse +2392,7977612,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1374,4,"",shellscript,content +2393,7978395,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1369,0,"",shellscript,selection_mouse +2394,7979349,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1369,0," ",shellscript,content +2395,7979350,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1370,0,"",shellscript,selection_keyboard +2396,7979558,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1370,0,"\",shellscript,content +2397,7979559,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1371,0,"",shellscript,selection_keyboard +2398,7980090,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1370,0,"",shellscript,selection_command +2399,7998280,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1443,0,"",shellscript,selection_mouse +2400,7998282,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1442,0,"",shellscript,selection_command +2401,8012086,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1130,0,"",shellscript,selection_mouse +2402,8012088,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1129,0,"",shellscript,selection_command +2403,8012562,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1112,0,"",shellscript,selection_mouse +2404,8013618,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1011,0,"",shellscript,selection_mouse +2405,8013619,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1010,0,"",shellscript,selection_command +2406,8020439,"train_dynamics.py",0,0,"",python,tab +2407,8024189,"train_dynamics.py",644,0,"",python,selection_mouse +2408,8024327,"train_dynamics.py",638,9,"num_steps",python,selection_mouse +2409,8027508,"train_dynamics.py",1127,0,"",python,selection_mouse +2410,8027643,"train_dynamics.py",1122,11,"lr_schedule",python,selection_mouse +2411,8028660,"train_dynamics.py",1144,0,"",python,selection_mouse +2412,8028790,"train_dynamics.py",1143,3,"wsd",python,selection_mouse +2413,8029779,"train_dynamics.py",1169,0,"",python,selection_mouse +2414,8030356,"train_dynamics.py",1174,0,"",python,selection_mouse +2415,8055512,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",0,0,"",shellscript,tab +2416,8056377,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1418,0,"",shellscript,selection_mouse +2417,8056379,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1417,0,"",shellscript,selection_command +2418,8056802,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1441,0,"",shellscript,selection_mouse +2419,8056802,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1440,0,"",shellscript,selection_command +2420,8058657,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1066,0,"",shellscript,selection_mouse +2421,8058658,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1065,0,"",shellscript,selection_command +2422,8059448,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1066,0,"\n ",shellscript,content +2423,8059851,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1071,0,"-",shellscript,content +2424,8059852,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1072,0,"",shellscript,selection_keyboard +2425,8059965,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1072,0,"-",shellscript,content +2426,8059966,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1073,0,"",shellscript,selection_keyboard +2427,8060251,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1073,0,"i",shellscript,content +2428,8060252,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1074,0,"",shellscript,selection_keyboard +2429,8060351,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1074,0,"n",shellscript,content +2430,8060352,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1075,0,"",shellscript,selection_keyboard +2431,8060494,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1075,0,"i",shellscript,content +2432,8060495,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1076,0,"",shellscript,selection_keyboard +2433,8060874,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1076,0,"t",shellscript,content +2434,8060875,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1077,0,"",shellscript,selection_keyboard +2435,8061057,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1077,0,"_",shellscript,content +2436,8061058,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1078,0,"",shellscript,selection_keyboard +2437,8061352,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1078,0,"l",shellscript,content +2438,8061353,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1079,0,"",shellscript,selection_keyboard +2439,8061505,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1079,0,"r",shellscript,content +2440,8061506,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1080,0,"",shellscript,selection_keyboard +2441,8062780,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1080,0,"=",shellscript,content +2442,8062781,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1081,0,"",shellscript,selection_keyboard +2443,8063593,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1081,0,"1",shellscript,content +2444,8063594,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1082,0,"",shellscript,selection_keyboard +2445,8064678,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1082,0,"3",shellscript,content +2446,8064679,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1083,0,"",shellscript,selection_keyboard +2447,8065048,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1083,0,"-",shellscript,content +2448,8065049,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1084,0,"",shellscript,selection_keyboard +2449,8066437,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1083,1,"",shellscript,content +2450,8066556,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1082,1,"",shellscript,content +2451,8066735,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1082,0,"e",shellscript,content +2452,8066736,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1083,0,"",shellscript,selection_keyboard +2453,8066982,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1083,0,"-",shellscript,content +2454,8066983,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1084,0,"",shellscript,selection_keyboard +2455,8067132,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1084,0,"3",shellscript,content +2456,8067132,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1085,0,"",shellscript,selection_keyboard +2457,8068162,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1085,0," ",shellscript,content +2458,8068163,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1086,0,"",shellscript,selection_keyboard +2459,8069356,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1086,0,"\",shellscript,content +2460,8069357,"slurm/jobs/mihir/horeka/yolo-runs/train_dynamics_new_arch_speedrun.sbatch",1087,0,"",shellscript,selection_keyboard +2461,8102405,"TERMINAL",0,0,"bash",,terminal_focus +2462,8107894,"TERMINAL",0,0,"fsacct_week",,terminal_command +2463,8107943,"TERMINAL",0,0,"]633;E;2025-07-17 11:59:10 fsacct_week;809dfc76-3a3b-47d1-9b49-bee31e96da6c]633;C",,terminal_output +2464,8107996,"TERMINAL",0,0," JobID JobName Partition All State Elapsed Timelimit \r\n--------------- ------------------------------ ---------------- --- ------------ ---------- ---------- \r\n 3331283 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331284 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331285 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331286 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331287 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331288 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3335345 train_dyn_yolorun accelerated 48 FAILED 00:00:27 01:00:00 \r\n 3335362 train_dyn_yolorun accelerated 48 TIMEOUT 01:00:18 01:00:00 \r\n 3335732 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:11 2-00:00:00 \r\n 3335733 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:23 2-00:00:00 \r\n 3335734 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:23 2-00:00:00 \r\n 3335735 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:21 2-00:00:00 \r\n 3335736 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:12 2-00:00:00 \r\n 3335737 train_tokenizer_lr_sweep_1e-4 accelerated 48 TIMEOUT 2-00:00:12 2-00:00:00 \r\n 3335738 train_tokenizer_lr_sweep_5e-5 accelerated 48 TIMEOUT 2-00:00:11 2-00:00:00 \r\n 3335744 train_dynamics_modelsize_scal+ accelerated 48 TIMEOUT 2-00:00:05 2-00:00:00 \r\n 3335745 train_dynamics_modelsize_scal+ accelerated 144 TIMEOUT 2-00:00:12 2-00:00:00 \r\n 3335746 train_dynamics_modelsize_scal+ accelerated 288 TIMEOUT 2-00:00:17 2-00:00:00 \r\n 3335747 train_dynamics_modelsize_scal+ accelerated 384 FAILED 00:00:58 2-00:00:00 \r\n 3335748 train_dynamics_modelsize_scal+ accelerated 768 FAILED 00:00:54 2-00:00:00 \r\n 3338176 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338177 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338178 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338179 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338180 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338240 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3341079 train_tokenizer_lr_sweep_1e-4 accelerated 48 TIMEOUT 2-00:00:11 2-00:00:00 \r\n 3341080 train_tokenizer_lr_sweep_5e-5 accelerated 48 TIMEOUT 2-00:00:11 2-00:00:00 \r\n 3345116 train_dynamics_modelsize_scal+ accelerated 48 TIMEOUT 2-00:00:17 2-00:00:00 \r\n 3348397 train_dynamics_lr_schedule_co+ accelerated 48 COMPLETED 1-03:21:05 2-00:00:00 \r\n 3348399 train_dynamics_lr_schedule_cos accelerated 48 COMPLETED 1-03:28:39 2-00:00:00 \r\n 3348400 train_dynamics_lr_schedule_wsd accelerated 48 COMPLETED 1-03:19:17 2-00:00:00 \r\n 3352115 train_dyn_yolorun_new_arch accelerated 48 RUNNING 16:22:05 2-00:00:00 \r\n]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;0",,terminal_output +2465,8118645,"TERMINAL",0,0,"alias",,terminal_command +2466,8118677,"TERMINAL",0,0,"]633;E;2025-07-17 11:59:21 alias;809dfc76-3a3b-47d1-9b49-bee31e96da6c]633;Calias egrep='egrep --color=auto'\r\nalias fgrep='fgrep --color=auto'\r\nalias fqueue='watch -n 1 ""squeue -o \""%.10i %.16P %.30j %.8u %.8T %.10M %.9l %.6D %R\""""'\r\nalias fsacct_week='sacct --format=""JobID%15,JobName%30,Partition%16,AllocCPUS%3,State%12,Elapsed%10,Timelimit%10"" --starttime $(date -d ""last week"" +%Y-%m-%d) | grep -vE ""*.batch|*.extern|*.inter|bash|python|CANCELLED|echo""'\r\nalias grep='grep --color=auto'\r\nalias idle='sinfo_t_idle'\r\nalias idling='watch -n1 sinfo_t_idle'\r\nalias l.='ls -d .* --color=auto'\r\nalias ll='ls -l --color=auto'\r\nalias ls='ls --color=auto'\r\nalias mc='. /usr/libexec/mc/mc-wrapper.sh'\r\nalias queue='watch -n1 squeue --me'\r\nalias runner='cd /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/'\r\nalias salloc_cpu='salloc --time=01:00:00 --partition=dev_cpuonly --nodes=1 --cpus-per-task=128'\r\nalias salloc_node='salloc --time=01:00:00 --partition=dev_accelerated --nodes=1 --gres=gpu:1 --cpus-per-task=8'\r\nalias smi='watch -n1 nvidia-smi'\r\nalias sync-runner='sh /home/hk-project-p0023960/tum_cte0515/sync_runner.sh /home/hk-project-p0023960/tum_cte0515/Projects/jafar /home/hk-project-p0023960/tum_cte0515/Projects/jafar_jobs/'\r\nalias xzegrep='xzegrep --color=auto'\r\nalias xzfgrep='xzfgrep --color=auto'\r\nalias xzgrep='xzgrep --color=auto'\r\nalias zegrep='zegrep --color=auto'\r\nalias zfgrep='zfgrep --color=auto'\r\nalias zgrep='zgrep --color=auto'\r\n]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;0",,terminal_output +2467,8147178,"TERMINAL",0,0,"sacct --format=""JobID%15,JobName%30,Partition%16,AllocCPUS%3,State%12,Elapsed%10,Timelimit%10"" --starttime $(date -d ""last week"" +%Y-%m-%d) ",,terminal_command +2468,8147270,"TERMINAL",0,0,"]633;E;2025-07-17 11:59:49 sacct --format=""JobID%15,JobName%30,Partition%16,AllocCPUS%3,State%12,Elapsed%10,Timelimit%10"" --starttime $(date -d ""last week"" +%Y-%m-%d) ;809dfc76-3a3b-47d1-9b49-bee31e96da6c]633;C JobID JobName Partition All State Elapsed Timelimit \r\n--------------- ------------------------------ ---------------- --- ------------ ---------- ---------- \r\n 3331282 train_tokenizer_batch_size_sc+ accelerated 48 TIMEOUT 1-12:00:22 1-12:00:00 \r\n 3331282.batch batch 24 CANCELLED 1-12:00:23 \r\n 3331282.extern extern 48 COMPLETED 1-12:00:51 \r\n 3331282.0 python 40 CANCELLED 1-12:00:28 \r\n 3331283 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331283.batch batch 24 CANCELLED 1-16:00:16 \r\n 3331283.extern extern 48 COMPLETED 1-16:00:44 \r\n 3331283.0 python 40 CANCELLED 1-16:00:23 \r\n 3331284 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331284.batch batch 24 CANCELLED 1-16:00:16 \r\n 3331284.extern extern 48 COMPLETED 1-16:00:44 \r\n 3331284.0 python 40 CANCELLED 1-16:00:22 \r\n 3331285 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331285.batch batch 24 CANCELLED 1-16:00:16 \r\n 3331285.extern extern 48 COMPLETED 1-16:00:44 \r\n 3331285.0 python 40 CANCELLED 1-16:00:22 \r\n 3331286 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331286.batch batch 24 CANCELLED 1-16:00:16 \r\n 3331286.extern extern 48 COMPLETED 1-16:00:44 \r\n 3331286.0 python 40 CANCELLED 1-16:00:22 \r\n 3331287 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331287.batch batch 24 CANCELLED 1-16:00:16 \r\n 3331287.extern extern 48 COMPLETED 1-16:00:44 \r\n 3331287.0 python 40 CANCELLED 1-16:00:22 \r\n 3331288 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331288.batch batch 24 CANCELLED 1-16:00:16 \r\n 3331288.extern extern 48 COMPLETED 1-16:00:44 \r\n 3331288.0 python 40 CANCELLED 1-16:00:23 \r\n 3333584 interactive accelerated 30 CANCELLED b+ 18:44:30 1-00:00:00 \r\n3333584.intera+ interactive 30 CANCELLED 18:45:00 \r\n 3333584.extern extern 30 COMPLETED 18:45:00 \r\n 3333584.0 bash 30 COMPLETED 00:29:25 \r\n 3333584.1 python 30 FAILED 00:00:36 \r\n 3333586 interactive accelerated 48 CANCELLED b+ 12:08:26 1-00:00:00 \r\n3333586.intera+ interactive 24 CANCELLED 12:08:55 \r\n 3333586.extern extern 48 COMPLETED 12:08:55 \r\n 3333586.0 bash 48 COMPLETED 00:05:02 \r\n 3333586.1 python 48 FAILED 00:00:54 \r\n 3333586.2 python 48 FAILED 00:00:26 \r\n 3333586.3 python 48 FAILED 00:00:24 \r\n 3333586.4 bash 48 FAILED 00:02:14 \r\n 3334542 interactive accelerated 0 CANCELLED b+ 00:00:00 12:00:00 \r\n 3334543 interactive accelerated 48 CANCELLED b+ 05:56:31 12:00:00 \r\n3334543.intera+ interactive 24 CANCELLED 05:57:01 \r\n 3334543.extern extern 48 COMPLETED 05:57:01 \r\n 3334543.0 python 40 FAILED 00:01:32 \r\n 3334543.1 python 40 CANCELLED b+ 00:02:54 \r\n 3334543.2 python 40 CANCELLED b+ 00:05:04 \r\n 3334543.3 python 40 FAILED 00:01:35 \r\n 3334543.4 python 40 FAILED 00:01:53 \r\n 3334543.5 python 40 FAILED 00:01:56 \r\n 3334543.6 python 40 FAILED 00:00:37 \r\n 3334543.7 python 40 COMPLETED 00:05:03 \r\n 3334543.8 python 40 COMPLETED 00:05:08 \r\n 3334543.9 python 40 COMPLETED 00:04:48 \r\n 3334543.10 python 40 COMPLETED 00:05:04 \r\n 3334543.11 python 40 COMPLETED 00:05:07 \r\n 3334543.12 bash 48 FAILED 00:06:28 \r\n 3334543.13 python 48 FAILED 00:00:31 \r\n 3334543.14 python 48 FAILED 00:00:21 \r\n 3334543.15 bash 48 FAILED 00:56:08 \r\n 3335250 interactive accelerated 0 CANCELLED b+ 00:00:00 12:00:00 \r\n 3335323 interactive dev_accelerated 0 CANCELLED b+ 00:00:00 01:00:00 \r\n 3335324 interactive accelerated 0 CANCELLED b+ 00:00:00 05:00:00 \r\n 3335325 interactive dev_accelerated 0 CANCELLED b+ 00:00:00 01:00:00 \r\n 3335335 interactive accelerated 48 CANCELLED b+ 03:32:02 10:00:00 \r\n3335335.intera+ interactive 24 FAILED 03:32:04 \r\n 3335335.extern extern 48 COMPLETED 03:32:05 \r\n 3335335.0 python 40 CANCELLED b+ 00:55:09 \r\n 3335335.1 python 40 CANCELLED b+ 00:06:42 \r\n 3335335.2 python 40 CANCELLED b+ 00:14:24 \r\n 3335345 train_dyn_yolorun accelerated 48 FAILED 00:00:27 01:00:00 \r\n 3335345.batch batch 24 FAILED 00:00:27 \r\n 3335345.extern extern 48 COMPLETED 00:00:27 \r\n 3335345.0 python 40 CANCELLED 00:00:00 \r\n 3335362 train_dyn_yolorun accelerated 48 TIMEOUT 01:00:18 01:00:00 \r\n 3335362.batch batch 24 CANCELLED 01:00:19 \r\n 3335362.extern extern 48 COMPLETED 01:00:47 \r\n 3335362.0 python 40 CANCELLED 01:00:26 \r\n 3335606 train_tokenizer_lr_sweep_1e-4 accelerated 48 CANCELLED b+ 00:14:11 2-00:00:00 \r\n 3335606.batch batch 24 CANCELLED 00:14:12 \r\n 3335606.extern extern 48 COMPLETED 00:14:41 \r\n 3335606.0 python 40 CANCELLED 00:14:19 \r\n 3335607 train_tokenizer_lr_sweep_5e-5 accelerated 48 CANCELLED b+ 00:04:48 2-00:00:00 \r\n 3335607.batch batch 24 CANCELLED 00:04:49 \r\n 3335607.extern extern 48 COMPLETED 00:05:18 \r\n 3335607.0 python 40 CANCELLED 00:04:57 \r\n 3335610 train_tokenizer_batch_size_sc+ accelerated 24 FAILED 00:03:25 2-00:00:00 \r\n 3335610.batch batch 24 FAILED 00:03:25 \r\n 3335610.extern extern 24 COMPLETED 00:03:25 \r\n 3335610.0 python 20 FAILED 00:02:58 \r\n 3335611 train_tokenizer_batch_size_sc+ accelerated 48 FAILED 00:02:57 2-00:00:00 \r\n 3335611.batch batch 24 FAILED 00:02:57 \r\n 3335611.extern extern 48 COMPLETED 00:02:57 \r\n 3335611.0 python 40 FAILED 00:02:30 \r\n 3335612 train_tokenizer_batch_size_sc+ accelerated 96 CANCELLED b+ 00:02:48 2-00:00:00 \r\n 3335612.batch batch 24 CANCELLED 00:02:49 \r\n 3335612.extern extern 96 COMPLETED 00:03:09 \r\n 3335612.0 python 80 FAILED 00:02:34 \r\n 3335613 train_tokenizer_batch_size_sc+ accelerated 0 CANCELLED b+ 00:00:00 2-00:00:00 \r\n 3335614 train_tokenizer_batch_size_sc+ accelerated 0 CANCELLED b+ 00:00:00 2-00:00:00 \r\n 3335660 interactive accelerated 0 CANCELLED b+ 00:00:00 10:00:00 \r\n 3335661 interactive accelerated 48 COMPLETED 00:53:32 10:00:00 \r\n3335661.intera+ interactive 24 COMPLETED 00:53:32 \r\n 3335661.extern extern 48 COMPLETED 00:53:32 \r\n 3335661.0 python 40 FAILED 00:02:51 \r\n 3335661.1 python 40 FAILED 00:02:28 \r\n 3335661.2 python 40 CANCELLED b+ 00:00:02 \r\n 3335661.3 python 40 CANCELLED b+ 00:06:00 \r\n 3335732 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:11 2-00:00:00 \r\n 3335732.batch batch 24 CANCELLED 2-00:00:12 \r\n 3335732.extern extern 48 COMPLETED 2-00:00:40 \r\n 3335732.0 python 40 CANCELLED 2-00:00:19 \r\n 3335733 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:23 2-00:00:00 \r\n 3335733.batch batch 24 CANCELLED 2-00:00:24 \r\n 3335733.extern extern 48 COMPLETED 2-00:00:52 \r\n 3335733.0 python 40 CANCELLED 2-00:00:27 \r\n 3335734 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:23 2-00:00:00 \r\n 3335734.batch batch 24 CANCELLED 2-00:00:24 \r\n 3335734.extern extern 48 COMPLETED 2-00:00:52 \r\n 3335734.0 python 40 CANCELLED 2-00:00:31 \r\n 3335735 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:21 2-00:00:00 \r\n 3335735.batch batch 24 CANCELLED 2-00:00:22 \r\n 3335735.extern extern 48 COMPLETED 2-00:00:50 \r\n 3335735.0 python 40 CANCELLED 2-00:00:29 \r\n 3335736 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:12 2-00:00:00 \r\n 3335736.batch batch 24 CANCELLED 2-00:00:13 \r\n 3335736.extern extern 48 COMPLETED 2-00:00:41 \r\n 3335736.0 python 40 CANCELLED 2-00:00:18 \r\n 3335737 train_tokenizer_lr_sweep_1e-4 accelerated 48 TIMEOUT 2-00:00:12 2-00:00:00 \r\n 3335737.batch batch 24 CANCELLED 2-00:00:13 \r\n 3335737.extern extern 48 COMPLETED 2-00:00:41 \r\n 3335737.0 python 40 CANCELLED 2-00:00:16 \r\n 3335738 train_tokenizer_lr_sweep_5e-5 accelerated 48 TIMEOUT 2-00:00:11 2-00:00:00 \r\n 3335738.batch batch 24 CANCELLED 2-00:00:11 \r\n 3335738.extern extern 48 COMPLETED 2-00:00:40 \r\n 3335738.0 python 40 CANCELLED 2-00:00:19 \r\n 3335739 train_tokenizer_batch_size_sc+ accelerated 384 TIMEOUT 2-00:00:14 2-00:00:00 \r\n 3335739.batch batch 24 CANCELLED 2-00:00:15 \r\n 3335739.extern extern 384 COMPLETED 2-00:00:43 \r\n 3335739.0 python 320 CANCELLED 2-00:00:20 \r\n 3335740 train_tokenizer_batch_size_sc+ accelerated 24 TIMEOUT 2-00:00:10 2-00:00:00 \r\n 3335740.batch batch 24 CANCELLED 2-00:00:10 \r\n 3335740.extern extern 24 COMPLETED 2-00:00:39 \r\n 3335740.0 python 20 CANCELLED 2-00:00:17 \r\n 3335741 train_tokenizer_batch_size_sc+ accelerated 48 TIMEOUT 2-00:00:05 2-00:00:00 \r\n 3335741.batch batch 24 CANCELLED 2-00:00:06 \r\n 3335741.extern extern 48 COMPLETED 2-00:00:34 \r\n 3335741.0 python 40 CANCELLED 2-00:00:09 \r\n 3335742 train_tokenizer_batch_size_sc+ accelerated 96 TIMEOUT 2-00:00:02 2-00:00:00 \r\n 3335742.batch batch 24 CANCELLED 2-00:00:03 \r\n 3335742.extern extern 96 COMPLETED 2-00:00:31 \r\n 3335742.0 python 80 CANCELLED 2-00:00:09 \r\n 3335743 train_tokenizer_batch_size_sc+ accelerated 192 TIMEOUT 2-00:00:28 2-00:00:00 \r\n 3335743.batch batch 24 CANCELLED 2-00:00:29 \r\n 3335743.extern extern 192 COMPLETED 2-00:00:57 \r\n 3335743.0 python 160 CANCELLED 2-00:00:35 \r\n 3335744 train_dynamics_modelsize_scal+ accelerated 48 TIMEOUT 2-00:00:05 2-00:00:00 \r\n 3335744.batch batch 24 CANCELLED 2-00:00:06 \r\n 3335744.extern extern 48 COMPLETED 2-00:00:34 \r\n 3335744.0 python 40 CANCELLED 2-00:00:12 \r\n 3335745 train_dynamics_modelsize_scal+ accelerated 144 TIMEOUT 2-00:00:12 2-00:00:00 \r\n 3335745.batch batch 24 CANCELLED 2-00:00:13 \r\n 3335745.extern extern 144 COMPLETED 2-00:00:41 \r\n 3335745.0 python 120 CANCELLED 2-00:00:19 \r\n 3335746 train_dynamics_modelsize_scal+ accelerated 288 TIMEOUT 2-00:00:17 2-00:00:00 \r\n 3335746.batch batch 24 CANCELLED 2-00:00:18 \r\n 3335746.extern extern 288 COMPLETED 2-00:00:46 \r\n 3335746.0 python 240 CANCELLED 2-00:00:32 \r\n 3335747 train_dynamics_modelsize_scal+ accelerated 384 FAILED 00:00:58 2-00:00:00 \r\n 3335747.batch batch 24 FAILED 00:00:58 \r\n 3335747.extern extern 384 COMPLETED 00:00:58 \r\n 3335747.0 python 320 FAILED 00:00:26 \r\n 3335748 train_dynamics_modelsize_scal+ accelerated 768 FAILED 00:00:54 2-00:00:00 \r\n 3335748.batch batch 24 FAILED 00:00:54 \r\n 3335748.extern extern 768 COMPLETED 00:00:54 \r\n 3335748.0 python 640 FAILED 00:00:24 \r\n 3337618 train_tokenizer_lr_sweep_1e-4+ accelerated 192 CANCELLED b+ 12:08:41 2-00:00:00 \r\n 3337618.batch batch 24 CANCELLED 12:08:42 \r\n 3337618.extern extern 192 COMPLETED 12:09:11 \r\n 3337618.0 python 160 CANCELLED 12:08:47 \r\n 3337619 train_tokenizer_lr_sweep_5e-5+ accelerated 192 CANCELLED b+ 12:08:44 2-00:00:00 \r\n 3337619.batch batch 24 CANCELLED 12:08:45 \r\n 3337619.extern extern 192 COMPLETED 12:09:14 \r\n 3337619.0 python 160 CANCELLED 12:08:52 \r\n 3338125 train_dynamics_modelsize_scal+ accelerated 0 CANCELLED b+ 00:00:00 2-00:00:00 \r\n 3338176 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338177 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338178 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338179 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338180 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338181 train_dynamics_modelsize_scal+ accelerated 0 CANCELLED b+ 00:00:00 2-00:00:00 \r\n 3338239 train_dynamics_modelsize_scal+ accelerated 0 CANCELLED b+ 00:00:00 2-00:00:00 \r\n 3338240 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338241 train_dynamics_modelsize_scal+ accelerated 576 CANCELLED b+ 1-22:44:52 2-00:00:00 \r\n 3338241.batch batch 24 CANCELLED 1-22:44:53 \r\n 3338241.extern extern 576 COMPLETED 1-22:45:22 \r\n 3338241.0 python 480 CANCELLED 1-22:44:58 \r\n 3338537 train_dynamics_modelsize_scal+ accelerated 288 CANCELLED b+ 1-22:10:24 2-00:00:00 \r\n 3338537.batch batch 24 CANCELLED 1-22:10:25 \r\n 3338537.extern extern 288 COMPLETED 1-22:10:53 \r\n 3338537.0 python 240 CANCELLED 1-22:10:30 \r\n 3340985 dynamics_cotraining_action_sp+ accelerated 48 CANCELLED b+ 1-05:42:39 2-00:00:00 \r\n 3340985.batch batch 24 CANCELLED 1-05:42:40 \r\n 3340985.extern extern 48 COMPLETED 1-05:43:09 \r\n 3340985.0 python 40 CANCELLED 1-05:42:46 \r\n 3340986 dynamics_cotraining_action_sp+ accelerated 48 CANCELLED b+ 1-05:36:50 2-00:00:00 \r\n 3340986.batch batch 24 CANCELLED 1-05:36:51 \r\n 3340986.extern extern 48 COMPLETED 1-05:37:20 \r\n 3340986.0 python 40 CANCELLED 1-05:36:57 \r\n 3340987 dynamics_cotraining_action_sp+ accelerated 48 CANCELLED b+ 1-03:44:55 2-00:00:00 \r\n 3340987.batch batch 24 CANCELLED 1-03:44:56 \r\n 3340987.extern extern 48 COMPLETED 1-03:45:25 \r\n 3340987.0 python 40 CANCELLED 1-03:45:02 \r\n 3340988 dynamics_cotraining_action_sp+ accelerated 48 CANCELLED b+ 1-03:45:04 2-00:00:00 \r\n 3340988.batch batch 24 CANCELLED 1-03:45:05 \r\n 3340988.extern extern 48 COMPLETED 1-03:45:34 \r\n 3340988.0 python 40 CANCELLED 1-03:45:11 \r\n 3340989 dynamics_cotraining_action_sp+ accelerated 48 CANCELLED b+ 1-03:45:19 2-00:00:00 \r\n 3340989.batch batch 24 CANCELLED 1-03:45:20 \r\n 3340989.extern extern 48 COMPLETED 1-03:45:49 \r\n 3340989.0 python 40 CANCELLED 1-03:45:26 \r\n 3341079 train_tokenizer_lr_sweep_1e-4 accelerated 48 TIMEOUT 2-00:00:11 2-00:00:00 \r\n 3341079.batch batch 24 CANCELLED 2-00:00:12 \r\n 3341079.extern extern 48 COMPLETED 2-00:00:40 \r\n 3341079.0 python 40 CANCELLED 2-00:00:18 \r\n 3341080 train_tokenizer_lr_sweep_5e-5 accelerated 48 TIMEOUT 2-00:00:11 2-00:00:00 \r\n 3341080.batch batch 24 CANCELLED 2-00:00:12 \r\n 3341080.extern extern 48 COMPLETED 2-00:00:40 \r\n 3341080.0 python 40 CANCELLED 2-00:00:18 \r\n 3341118 train_tokenizer_batch_size_sc+ accelerated 24 CANCELLED b+ 1-03:43:20 2-00:00:00 \r\n 3341118.batch batch 24 CANCELLED 1-03:43:21 \r\n 3341118.extern extern 24 COMPLETED 1-03:43:49 \r\n 3341118.0 python 20 CANCELLED 1-03:43:27 \r\n 3341119 train_tokenizer_batch_size_sc+ accelerated 48 CANCELLED b+ 1-01:19:46 2-00:00:00 \r\n 3341119.batch batch 24 CANCELLED 1-01:19:47 \r\n 3341119.extern extern 48 COMPLETED 1-01:20:15 \r\n 3341119.0 python 40 CANCELLED 1-01:19:53 \r\n 3341273 train_tokenizer_batch_size_sc+ accelerated 384 CANCELLED b+ 15:55:30 2-00:00:00 \r\n 3341273.batch batch 24 CANCELLED 15:55:31 \r\n 3341273.extern extern 384 COMPLETED 15:56:00 \r\n 3341273.0 python 320 CANCELLED 15:55:36 \r\n 3341274 train_tokenizer_batch_size_sc+ accelerated 96 CANCELLED b+ 22:21:26 2-00:00:00 \r\n 3341274.batch batch 24 CANCELLED 22:21:27 \r\n 3341274.extern extern 96 COMPLETED 22:21:56 \r\n 3341274.0 python 80 CANCELLED 22:21:33 \r\n 3341406 train_tokenizer_batch_size_sc+ accelerated 192 CANCELLED b+ 15:13:15 2-00:00:00 \r\n 3341406.batch batch 24 CANCELLED 15:13:16 \r\n 3341406.extern extern 192 COMPLETED 15:13:45 \r\n 3341406.0 python 160 CANCELLED 15:13:21 \r\n 3341476 train_dynamics_modelsize_scal+ accelerated 48 CANCELLED b+ 14:33:58 2-00:00:00 \r\n 3341476.batch batch 24 CANCELLED 14:33:59 \r\n 3341476.extern extern 48 COMPLETED 14:34:28 \r\n 3341476.0 python 40 CANCELLED 14:34:05 \r\n 3341477 train_dynamics_modelsize_scal+ accelerated 144 CANCELLED b+ 13:37:43 2-00:00:00 \r\n 3341477.batch batch 24 CANCELLED 13:37:44 \r\n 3341477.extern extern 144 COMPLETED 13:38:13 \r\n 3341477.0 python 120 CANCELLED 13:37:50 \r\n 3341478 train_dynamics_modelsize_scal+ accelerated 288 CANCELLED b+ 05:57:14 2-00:00:00 \r\n 3341478.batch batch 24 CANCELLED 05:57:15 \r\n 3341478.extern extern 288 COMPLETED 05:57:44 \r\n 3341478.0 python 240 CANCELLED 05:57:21 \r\n 3343465 interactive accelerated 0 CANCELLED b+ 00:00:00 10:00:00 \r\n 3343466 interactive accelerated 0 CANCELLED b+ 00:00:00 10:00:00 \r\n 3343467 interactive dev_accelerated 12 COMPLETED 00:36:30 01:00:00 \r\n3343467.intera+ interactive 12 COMPLETED 00:36:30 \r\n 3343467.extern extern 12 COMPLETED 00:36:30 \r\n 3344012 interactive dev_accelerated 6 TIMEOUT 01:00:08 01:00:00 \r\n3344012.intera+ interactive 6 CANCELLED 01:00:37 \r\n 3344012.extern extern 6 COMPLETED 01:00:37 \r\n 3344246 interactive accelerated 48 COMPLETED 08:34:29 10:00:00 \r\n3344246.intera+ interactive 24 CANCELLED 08:34:59 \r\n 3344246.extern extern 48 COMPLETED 08:34:59 \r\n 3344246.0 python 40 FAILED 00:03:15 \r\n 3344246.1 python 40 FAILED 00:02:25 \r\n 3344246.2 python 40 CANCELLED b+ 00:19:08 \r\n 3344246.3 python 40 FAILED 00:02:54 \r\n 3344246.4 python 40 FAILED 00:02:27 \r\n 3344246.5 python 40 FAILED 00:03:30 \r\n 3344246.6 python 40 CANCELLED 04:40:24 \r\n 3344502 interactive accelerated 0 CANCELLED b+ 00:00:00 10:00:00 \r\n 3344503 interactive accelerated 0 CANCELLED b+ 00:00:00 01:00:00 \r\n 3344505 interactive dev_accelerated 6 FAILED 00:10:16 01:00:00 \r\n3344505.intera+ interactive 6 FAILED 00:10:16 \r\n 3344505.extern extern 6 COMPLETED 00:10:16 \r\n 3344574 interactive dev_accelerated 20 TIMEOUT 01:00:04 01:00:00 \r\n3344574.intera+ interactive 20 FAILED 01:00:08 \r\n 3344574.extern extern 20 COMPLETED 01:00:08 \r\n 3344887 interactive dev_accelerated 20 TIMEOUT 01:00:28 01:00:00 \r\n3344887.intera+ interactive 20 CANCELLED 01:00:57 \r\n 3344887.extern extern 20 COMPLETED 01:00:57 \r\n 3345079 train_dynamics_modelsize_scal+ accelerated 0 CANCELLED b+ 00:00:00 2-00:00:00 \r\n 3345116 train_dynamics_modelsize_scal+ accelerated 48 TIMEOUT 2-00:00:17 2-00:00:00 \r\n 3345116.batch batch 24 CANCELLED 2-00:00:18 \r\n 3345116.extern extern 48 COMPLETED 2-00:00:46 \r\n 3345116.0 python 40 CANCELLED 2-00:00:25 \r\n 3345348 interactive dev_accelerated 0 CANCELLED b+ 00:00:00 01:00:00 \r\n 3345349 interactive dev_accelerated 0 CANCELLED b+ 00:00:00 01:00:00 \r\n 3345355 interactive accelerated 20 TIMEOUT 01:00:24 01:00:00 \r\n3345355.intera+ interactive 20 CANCELLED 01:00:53 \r\n 3345355.extern extern 20 COMPLETED 01:00:53 \r\n 3347272 interactive accelerated 48 FAILED 00:22:03 10:00:00 \r\n3347272.intera+ interactive 24 FAILED 00:22:03 \r\n 3347272.extern extern 48 COMPLETED 00:22:03 \r\n 3347272.0 python 40 CANCELLED b+ 00:13:45 \r\n 3347273 interactive accelerated 6 COMPLETED 02:16:38 10:00:00 \r\n3347273.intera+ interactive 6 CANCELLED 02:17:08 \r\n 3347273.extern extern 6 COMPLETED 02:17:08 \r\n 3347289 interactive accelerated 48 TIMEOUT 10:00:21 10:00:00 \r\n3347289.intera+ interactive 24 CANCELLED 10:00:50 \r\n 3347289.extern extern 48 COMPLETED 10:00:50 \r\n 3347289.0 python 40 FAILED 00:00:15 \r\n 3347289.1 python 40 COMPLETED 00:09:27 \r\n 3347289.2 python 40 CANCELLED b+ 00:00:27 \r\n 3347289.3 python 40 FAILED 00:00:03 \r\n 3347289.4 python 40 FAILED 00:00:02 \r\n 3347289.5 python 40 COMPLETED 00:09:57 \r\n 3347289.6 python 40 FAILED 00:00:14 \r\n 3347289.7 python 40 CANCELLED b+ 00:06:03 \r\n 3347289.8 python 40 CANCELLED b+ 00:06:22 \r\n 3347289.9 python 40 CANCELLED b+ 00:06:56 \r\n 3347289.10 python 40 FAILED 00:01:10 \r\n 3347289.11 python 40 CANCELLED b+ 00:01:20 \r\n 3347289.12 python 40 CANCELLED b+ 00:11:14 \r\n 3347289.13 python 40 CANCELLED b+ 00:07:04 \r\n 3347289.14 python 40 CANCELLED b+ 00:09:02 \r\n 3347289.15 python 40 CANCELLED b+ 00:25:32 \r\n 3347518 interactive accelerated 6 COMPLETED 09:04:25 10:00:00 \r\n3347518.intera+ interactive 6 CANCELLED b+ 09:04:25 \r\n 3347518.extern extern 6 COMPLETED 09:04:25 \r\n 3348397 train_dynamics_lr_schedule_co+ accelerated 48 COMPLETED 1-03:21:05 2-00:00:00 \r\n 3348397.batch batch 24 COMPLETED 1-03:21:05 \r\n 3348397.extern extern 48 COMPLETED 1-03:21:05 \r\n 3348397.0 python 40 COMPLETED 1-03:20:36 \r\n 3348399 train_dynamics_lr_schedule_cos accelerated 48 COMPLETED 1-03:28:39 2-00:00:00 \r\n 3348399.batch batch 24 COMPLETED 1-03:28:39 \r\n 3348399.extern extern 48 COMPLETED 1-03:28:39 \r\n 3348399.0 python 40 COMPLETED 1-03:28:10 \r\n 3348400 train_dynamics_lr_schedule_wsd accelerated 48 COMPLETED 1-03:19:17 2-00:00:00 \r\n 3348400.batch batch 24 COMPLETED 1-03:19:17 \r\n 3348400.extern extern 48 COMPLETED 1-03:19:17 \r\n 3348400.0 python 40 COMPLETED 1-03:18:48 \r\n 3348592 train_dyn_yolorun accelerated 48 CANCELLED b+ 19:29:04 1-00:00:00 \r\n 3348592.batch batch 24 CANCELLED 19:29:05 \r\n 3348592.extern extern 48 COMPLETED 19:29:34 \r\n 3348592.0 python 40 CANCELLED 19:29:13 \r\n 3349982 interactive accelerated 48 CANCELLED b+ 00:10:32 10:00:00 \r\n3349982.intera+ interactive 24 CANCELLED 00:11:02 \r\n 3349982.extern extern 48 COMPLETED 00:11:02 \r\n 3350109 interactive accelerated 0 CANCELLED b+ 00:00:00 10:00:00 \r\n 3350110 interactive dev_accelerated 0 CANCELLED b+ 00:00:00 01:00:00 \r\n 3350111 interactive accelerated 48 CANCELLED b+ 00:09:37 10:00:00 \r\n3350111.intera+ interactive 24 CANCELLED 00:10:06 \r\n 3350111.extern extern 48 COMPLETED 00:10:06 \r\n 3350245 interactive accelerated 48 CANCELLED b+ 00:17:59 10:00:00 \r\n3350245.intera+ interactive 24 CANCELLED 00:18:28 \r\n 3350245.extern extern 48 COMPLETED 00:18:28 \r\n 3350302 interactive accelerated 48 CANCELLED b+ 00:40:22 10:00:00 \r\n3350302.intera+ interactive 24 CANCELLED 00:40:52 \r\n 3350302.extern extern 48 COMPLETED 00:40:52 \r\n 3350302.0 python 40 FAILED 00:00:27 \r\n 3350302.1 python 40 FAILED 00:05:27 \r\n 3350302.2 python 40 FAILED 00:03:41 \r\n 3350302.3 python 40 CANCELLED b+ 00:03:28 \r\n 3350302.4 python 40 CANCELLED b+ 00:04:30 \r\n 3350302.5 python 40 FAILED 00:01:21 \r\n 3350302.6 python 40 FAILED 00:00:27 \r\n 3350302.7 python 40 CANCELLED 00:04:03 \r\n 3350418 interactive accelerated 48 COMPLETED 08:32:29 10:00:00 \r\n3350418.intera+ interactive 24 CANCELLED 08:32:59 \r\n 3350418.extern extern 48 COMPLETED 08:32:59 \r\n 3350418.0 python 40 CANCELLED b+ 00:01:30 \r\n 3350418.1 python 40 FAILED 00:02:41 \r\n 3350418.2 python 40 CANCELLED b+ 01:19:51 \r\n 3350418.3 python 40 FAILED 00:00:16 \r\n 3350418.4 python 40 FAILED 00:00:03 \r\n 3350418.5 python 40 FAILED 00:00:07 \r\n 3350418.6 python 40 FAILED 00:00:03 \r\n 3350418.7 python 40 FAILED 00:02:57 \r\n 3350418.8 python 40 CANCELLED b+ 00:08:48 \r\n 3350418.9 python 40 CANCELLED b+ 00:10:49 \r\n 3350418.10 python 40 CANCELLED b+ 00:05:51 \r\n 3350418.11 python 40 CANCELLED b+ 00:06:24 \r\n 3351743 train_dyn_yolorun_new_arch accelerated 48 CANCELLED b+ 01:58:23 2-00:00:00 \r\n 3351743.batch batch 24 CANCELLED 01:58:24 \r\n 3351743.extern extern 48 COMPLETED 01:58:53 \r\n 3351743.0 python 40 CANCELLED 01:58:29 \r\n 3352103 train_dyn_yolorun_new_arch accelerated 48 CANCELLED b+ 00:09:58 2-00:00:00 \r\n 3352103.batch batch 24 CANCELLED 00:09:59 \r\n 3352103.extern extern 48 COMPLETED 00:10:28 \r\n 3352103.0 python 40 CANCELLED 00:10:03 \r\n 3352115 train_dyn_yolorun_new_arch accelerated 48 RUNNING 16:22:45 2-00:00:00 \r\n 3352115.batch batch 24 RUNNING 16:22:45 \r\n 3352115.extern extern 48 RUNNING 16:22:45 \r\n 3352115.0 python 40 RUNNING 16:22:17 \r\n 3352588 train_dyn_yolorun_new_arch accelerated 0 CANCELLED b+ 00:00:00 2-00:00:00 \r\n 3352994 interactive accelerated 48 RUNNING 02:14:49 10:00:00 \r\n3352994.intera+ interactive 24 RUNNING 02:14:49 \r\n 3352994.extern extern 48 RUNNING 02:14:49 \r\n 3352996 interactive accelerated 6 RUNNING 02:14:32 10:00:00 \r\n3352996.intera+ interactive 6 RUNNING 02:14:32 \r\n 3352996.extern extern 6 RUNNING 02:14:32 \r\n",,terminal_output +2469,8216919,"TERMINAL",0,0,"sacct --format=""JobID%15,JobName%30,Partition%16,AllocCPUS%3,State%12,Elapsed%10,Timelimit%10"" --starttime $(date -d ""last week"" +%Y-%m-%d) | grep -vE ""*.batch|*.extern|*.inter|bash|python|CANCELLED|echo""",,terminal_command +2470,8216994,"TERMINAL",0,0,"]633;E;2025-07-17 12:00:59 sacct --format=""JobID%15,JobName%30,Partition%16,AllocCPUS%3,State%12,Elapsed%10,Timelimit%10"" --starttime $(date -d ""last week"" +%Y-%m-%d) | grep -vE ""*.batch|*.extern|*.inter|bash|python|CANCELLED|echo"";809dfc76-3a3b-47d1-9b49-bee31e96da6c]633;C JobID JobName Partition All State Elapsed Timelimit \r\n--------------- ------------------------------ ---------------- --- ------------ ---------- ---------- \r\n 3331283 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331284 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331285 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331286 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331287 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3331288 train_lam_action_space_scalin+ accelerated 48 TIMEOUT 1-16:00:15 1-16:00:00 \r\n 3335345 train_dyn_yolorun accelerated 48 FAILED 00:00:27 01:00:00 \r\n 3335362 train_dyn_yolorun accelerated 48 TIMEOUT 01:00:18 01:00:00 \r\n 3335732 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:11 2-00:00:00 \r\n 3335733 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:23 2-00:00:00 \r\n 3335734 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:23 2-00:00:00 \r\n 3335735 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:21 2-00:00:00 \r\n 3335736 dynamics_cotraining_action_sp+ accelerated 48 TIMEOUT 2-00:00:12 2-00:00:00 \r\n 3335737 train_tokenizer_lr_sweep_1e-4 accelerated 48 TIMEOUT 2-00:00:12 2-00:00:00 \r\n 3335738 train_tokenizer_lr_sweep_5e-5 accelerated 48 TIMEOUT 2-00:00:11 2-00:00:00 \r\n 3335744 train_dynamics_modelsize_scal+ accelerated 48 TIMEOUT 2-00:00:05 2-00:00:00 \r\n 3335745 train_dynamics_modelsize_scal+ accelerated 144 TIMEOUT 2-00:00:12 2-00:00:00 \r\n 3335746 train_dynamics_modelsize_scal+ accelerated 288 TIMEOUT 2-00:00:17 2-00:00:00 \r\n 3335747 train_dynamics_modelsize_scal+ accelerated 384 FAILED 00:00:58 2-00:00:00 \r\n 3335748 train_dynamics_modelsize_scal+ accelerated 768 FAILED 00:00:54 2-00:00:00 \r\n 3338176 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338177 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338178 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338179 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338180 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3338240 train_dynamics_modelsize_scal+ accelerated 0 FAILED 00:00:00 2-00:00:00 \r\n 3341079 train_tokenizer_lr_sweep_1e-4 accelerated 48 TIMEOUT 2-00:00:11 2-00:00:00 \r\n 3341080 train_tokenizer_lr_sweep_5e-5 accelerated 48 TIMEOUT 2-00:00:11 2-00:00:00 \r\n 3345116 train_dynamics_modelsize_scal+ accelerated 48 TIMEOUT 2-00:00:17 2-00:00:00 \r\n 3348397 train_dynamics_lr_schedule_co+ accelerated 48 COMPLETED 1-03:21:05 2-00:00:00 \r\n 3348399 train_dynamics_lr_schedule_cos accelerated 48 COMPLETED 1-03:28:39 2-00:00:00 \r\n 3348400 train_dynamics_lr_schedule_wsd accelerated 48 COMPLETED 1-03:19:17 2-00:00:00 \r\n 3352115 train_dyn_yolorun_new_arch accelerated 48 RUNNING 16:23:54 2-00:00:00 \r\n]0;tum_cte0515@hkn1990:~/Projects/jafar]633;D;0",,terminal_output +2471,8227788,"TERMINAL",0,0,"sacct --help",,terminal_command +2472,8227832,"TERMINAL",0,0,"]633;E;2025-07-17 12:01:10 sacct --help;809dfc76-3a3b-47d1-9b49-bee31e96da6c]633;Csacct [